Government 3.0 - Next Generation Government Technology Infrastructure and Services
Government 3.0 - Next Generation Government Technology Infrastructure and Services
Adegboyega Ojo
Jeremy Millard Editors
Government 3.0 –
Next Generation
Government
Technology
Infrastructure
and Services
Roadmaps, Enabling Technologies &
Challenges
Public Administration and Information
Technology
Volume 32
Series editor
Christopher G. Reddick, San Antonio, TX, USA
More information about this series at http://www.springer.com/series/10796
Adegboyega Ojo • Jeremy Millard
Editors
Policymakers and academics largely recognise the need for a fresh vision for public
sector innovation and the use of technology in government given the challenging
and turbulent contexts in which most public administrations operate. In response,
recent studies have sought to better understand the forces that will shape the future
evolution of the PA environment. For instance, in a report on Future Trends in
European Public Administration and Management, some megatrends that are
already shaping the future of PAs were identified (Pollitt 2014). These changes
include demographic change, climate change, economic trajectories, technological
developments, public trust in government and changes in the political
environment.
Historically, technological change has had a significant effect on the locus of
administrative activity, the costs involved, the nature of administrative tasks, the
skill sets needed by officials, rules and regulations and the types of interactions citi-
zens have with their public authorities (Pollitt 2014). In 2007, Frissen et al. (2007)
identified some disruptive technologies with strong potentials to transform govern-
ment functions, including mobile devices; intelligent agents (and robotics); sensors;
language processing technologies; semantic technologies; serious games; RFID and
biometrics; ICT infrastructures such as WiFi, WiMAX and broadband; Web 2.0
technologies (social software); and grid infrastructure. While mobile devices and
Web 2.0 and ICT connectivity technologies such as WiFi have had a transforma-
tional effect, some of these technologies are yet to have any major impact in the
government space. Unfortunately, we are yet to fully understand the reasons for this
very slow adoption of these technologies.
A recent study by the European Commission on ‘Powering European Public
Sector Innovation: Towards a New Architecture’ (EC-DG Research and Innovation
2013) has also identified new technology paradigms considered as enablers of inno-
vation and core to the delivery of public services or the design of public policy.
These technologies include the following:
• Social – social networking offers new ways to deliver public services and to
enable citizens to participate.
vii
viii Preface
• Analytics – big data and predictive analytics offer new service opportunities for
citizens and businesses.
• Mobile – the advent of the smartphone enables citizens to access public services
from anywhere at any time.
• Cloud – cloud-based solutions, both public and private, can transform interoper-
ability and service provision.
• Open and big data paradigm – new public services, transparency/democracy,
economic growth potential.
• Sensors and Internet of Things – harnessing an enormous amount of data gener-
ated from everything around with an Internet address for better decision-making
and problem-solving.
Among these new technological paradigms, open and big data stands out regard-
ing attention by policymakers. It is widely believed that big data will enable hitherto
slow-moving public services to move much faster and to treat citizens on an indi-
vidual rather than a categorised basis.
However, despite these exciting possibilities, many questions remain unresolved:
Can these new technologies deliver the radical innovation needed for the ‘entrepre-
neurial’ and ‘directing’ (Pollitt 2014) state? How should governments (at different
levels) reconfigure their relationships with citizens, the social sector and businesses
to effectively leverage these technologies to deliver public outcomes effectively? To
what extent can open data enable greater transparency that can increase social capi-
tal and public trust in the government? How can the public sector effectively tap into
the ‘data tsunami’ already engulfing us due to the explosion of social media and the
introduction of new low-cost data gathering tools that effectively make every citizen
with a smartphone a data source? What are the new data gathering trends most
likely to impact public services (Millard 2013)? What kind of capabilities must the
government develop to leverage these technologies? Finally, what are the negative
consequences (such as an exacerbation of the digital divide or threats to citizen
privacy) that the adoption of these technologies may present and what strategies are
available to mitigate undesirable effects?
This book attempts to answer some of these questions. Specifically, this book
will shed some light on the question about the next steps of e-government initiatives
and public sector innovation. This next generation public sector innovation is what
we have labelled ‘Government 3.0’. Technology policymakers should benefit from
the visions created by the various roadmaps in the first three chapters of the book
which describe some of the common strategies of the European Union member
states in the areas of open data and services, open processes and the use of digital
technologies in policymaking. The book also discusses in its fourth chapter some of
the issues associated with existing models for tracking progress in e-government
development and highlights how some of these shortcomings could be addressed.
Examples of emerging innovations in the areas of process engineering and open
innovation in the government domain based on linked open data are described in
chapters “Techniques for Reuse in Business Process Modeling in Public
Administration, Capability Development in Open Data Driven Organizations, Water
Preface ix
References
EC-DG Research and Innovation (2013) Powering European public sector innova-
tion: towards a new architecture, Brussels, pp 1–64
Frissen AV, Millard J, Huijboom N (2007) The future of eGovernment – an explora-
tion of ICT-driven models of eGovernment for the EU in 2020, vol 2020. pp 1–15
Millard J (2013) ICT-enabled public sector innovation : trends and prospects. In
ICEGOV’13, October 22–25, 2013, Seoul, Republic of Korea. ACM International
Conference Proceedings Series, ACM Press, pp 77–86
Pollitt C (2014) Future trends in European public administration and management:
an outside-in perspective, pp 1–45. Retrieved from www.cocops.eu.
Acknowledgment
The publication of this book would not have been possible without the contributions
and support of many people. We use this medium to express our profound apprecia-
tion to all these people who helped and supported us from the conception of the
book to its publication. First on the list of these people is the PAIT book series edi-
tor, Chris Reddick, whose guidance was key in getting the project off the ground.
Next, we thank all the authors for sharing their work through our book. Very special
thanks to all the reviewers who gave their time, effort and constructive comments
that enhanced the overall quality of the chapters in the book. We particularly recog-
nise the dedication and commitments of Ed Curry, Peter Winstanley, Michael
Hogan, Fatemeh Ahmadi Zeleti, Niall O’Brolchain and Lukasz Porwol as review-
ers. Last but not least, we are grateful for the support of Samuel Adebayo (the book
project manager) and the Springer staff members for guidance in the final steps of
the publication process.
xi
Contents
xiii
xiv Contents
Index������������������������������������������������������������������������������������������������������������������ 355
Contributors
Samuel Adebayo Insight Centre for Data Analytics, National University of Ireland
Galway (NUIG), IDA Business Park, Lower Dangan, Galway, Ireland
Jerry Andriessen Wise & Munro Learning Res, Den Haag, The Netherlands
Michael Baker CNRS – Telecom ParisTech, Paris, France
Sami Bhiri ISIMM, University of Monastir, Monastir, Tunisia
Paolo Boscolo Comune di Prato, Prato, Italy
Eoghan Clifford College of Engineering & Informatics, National University of
Ireland Galway, Galway, Ireland
Amy Conroy Carleton University, Department of Law and Legal Studies, Ottawa,
ON, Canada
Edward Curry Insight Centre for Data Analytics, National University of Ireland
Galway, Galway, Ireland
Wassim Derguech Insight Centre for Data Analytics, National University of
Ireland, Galway, Galway, Ireland
Françoise Détienne CNRS – Telecom ParisTech, Paris, France
Jonathan Groff CNRS – Telecom ParisTech, Paris, France
Louise Hannon College of Engineering & Informatics, National University of
Ireland Galway, Galway, Ireland
Owen Harney National University of Ireland, Galway, Ireland
Souleiman Hasan Insight Centre for Data Analytics, National University of
Ireland Galway, Galway, Ireland
Umair ul Hassan Insight Centre for Data Analytics, National University of Ireland
Galway, Galway, Ireland
xv
xvi Contributors
Adegboyega Ojo is senior research fellow and head of the e-Government Unit at
Insight Centre for Data Analytics, National University of Ireland, Galway (NUIG).
His expertise is in the areas of open data strategies and infrastructures, data-intensive
public services, e-government services, e-participation infrastructure and gover-
nance of smart cities. He is currently principal investigator on a number of EU
Horizon 2020 projects in the area of open data platforms and co-creation of data-
driven public services. He has served as expert advisor to different UN organisa-
tions such as the e-Government Branch of the United Nations Department of
Economic and Social Affairs/Division for Public Administration and Development
Management on the use of open and big data for integrating Sustainable Development
Goals (SDG2030). Adegboyega also led the development of technical paper for the
Open Government Partnership on how to benchmark initiatives from participating
countries. While at the United Nations University as research fellow and academic
programme officer at the Centre for Electronic Governance in Macao SAR,
Adegboyega led the development of National e-Government and IT Strategies for
countries in Asia and Africa. He has served as programme and track chair and pro-
gramme committee member in well over 70 international conference editions in the
domains of e-government, smart cities and formal software engineering. He cur-
rently serves as editorial board member for Elsevier’s Government Information
Quarterly journal and International Journal of Public Administration in the Digital
Age. He was elected secretary of the Digital Government Society (DGS) in
December 2016. Adegboyega obtained his doctorate in computer science from the
University of Lagos, Nigeria, in 1998.
Jeremy Millard director of the non-profit consultancy Third Millennium
Governance, senior research fellow at Brunel University (London) and senior policy
advisor at Danish Technological Institute, has 40 years1 global experience working
with governments, development agencies and private and civil sectors in all parts of
the world. In the last 20 years, he has focused on how new technical and organisa-
tional innovations transform the government and the public sector. Recent assign-
ments for the European Commission include studies on administrative burden
xvii
xviii Editors’ Biography
reduction and on developing business models for ICT and ageing. He also recently
led an impact assessment of the European e-Government Action Plan and a large-
scale Europe-wide survey and analysis of participation and developed the
e-Government 2020 Vision Study on Future Directions of Public Service Delivery.
He has worked on the European e-Government annual benchmark since 2009 and
has assisted the EC in planning their research programme for ICT for governance
and policy modelling between 2008 and 2012 and then in designing inputs to the
Horizon 2020 research work programme (2014–2015) on ICT-enabled public sector
innovation. He is currently leading the EC1 Advisory Group providing inputs to the
Horizon 2020 work programme for 2016–2017, including public governance and
new economic models, and is a member of the EC1 Digital Entrepreneurship Forum.
Jeremy has also worked since 2008 as an expert for the UN on their successive
global e-government development surveys and has undertaken work adapted from
the UN approach in Oman, Georgia, China and Russia. He also provided inputs to
the UN1 debate on the post-2015 development agenda in relation to governance
issues and how these can be measured and to the World Bank on digital public sec-
tor services. He recently led the assignment for ReSPA on the Western Balkans
Comparative e-Government Study, as well as a survey for the OECD on back-office
developments in support of user-centred e-government strategies. In the Gulf, he
assists Oman develop their e-government services, prepared an e-government strat-
egy for the six nations of the Gulf Cooperation Council and has supported the devel-
opment of the Bahraini e-government strategy. Currently, he is undertaking a study
for ESCWA on integrated service delivery across the Arab Region. He also provides
e-government support services in India, Malaysia and Brunei.
European Strategies for e-Governance to 2020
and Beyond
Jeremy Millard
Abstract This chapter examines both academic and grey literature on the transi-
tions and developments in e-government towards notions of open government and
open governance. This is viewed through the prism of European level strategies, the
EU’s research and innovation programmes, as well as common strategies like the
European E-Government Action Plan agreed to by all EU Member States. The three
strands of the proposed European open governance setting, consisting of open data,
open service and open process, are examined, as is the conceptualization of govern-
ment as an open source service platform as well as a broader platform for collabora-
tion between all societal actors. The purpose is to support societal-wide innovation
for tackling pressing societal challenges where the role of ICT is seen more broadly
than has traditionally been the case, i.e. as a general purpose technology. In this
context, the chapter also examines emerging technologies likely to impact govern-
ment in the short as well as longer-term, such as big data, artificial intelligence,
drones and blockchain.
Introduction and Context
This chapter derives from both the academic and grey literature of e-government
and similar developments by examining the main conceptual paradigms which have
had real impact on how Information and Communication Technology (ICT) is used
by, and impacts, government over the last 20 years in Europe. Drawing on a review
by Millard (2015), the notion of electronic (e)-government, starting in the late
1990s, was explicitly linked to the ‘New Public Management’ philosophy which
emphasised inter alia how ICT could make the public sector much more efficient by
adopting private sector management disciplines which had already shown how to
J. Millard (*)
Danish Technological Institute, Taastrup, Denmark
e-mail: Jeremy.millard@3mg.org
1
The terms ‘public sector’ and ‘government’ are in practice used interchangeably in this paper as
in many others. The term ‘governance’ refers to public governance as “the role of governments,
working alongside other actors, in building, facilitating and overseeing political, social and eco-
nomic development. Irrespective of any intrinsic value it might have, public governance is there-
fore a crucial means to desired development outcomes.” (Bevir 2013)
2
There are many established definitions of ‘public value’, for example “public value refers to the
value created by government through services, laws, regulation and other actions” (Kelly et al.
2002). For the present purpose it can be also thought of as similar to the older notions of ‘public
goods’ and ‘good governance’.
3
https://www.whitehouse.gov/the_press_office/TransparencyandOpenGovernment/ (accessed 10
May 2015).
4
http://www.opengovpartnership.org (accessed 10 May 2015).
European Strategies for e-Governance to 2020 and Beyond 3
Some clear conclusions emerge from the development of o-government. The opera-
tions of the public sector, public policy and public services are seen as needing to
become more open and innovative as well as efficient and effective. Indeed it is
argued in this chapter that these attributes are complementary, especially if seen
over the medium to longer term, but also that the public sector cannot successfully
tackle the global challenges on its own. The chapter goes further and argues that an
understanding of open government within an open governance system cannot sim-
ply look at what is taking place inside the public sector, but must also examine
wider developments in society and the manner in which other societal actors are
changing their roles and ways of operating. Government, as an actor, needs to col-
laborate, and a powerful tool in this context is ICT. This is the basic message of this
chapter which examines a new approach to public sector innovation based around
the notion of ‘open governance systems’, and which attempts to unpick its main
components as we can presently see them (Millard 2013, 2015).
The global financial crisis of 2007–2008 tended to mask the fact that there are
longer term and deeper rooted global societal challenges which preceded it, many
of which have since become even more acute. These include climate change,
increasing inequalities within countries, poverty, corruption, energy and job short-
ages, health and education under pressure, rapidly changing demographics (ageing,
4 J. Millard
As sketched above, the discourse and most literature to date have focused on
responses to the crisis which envisage the public sector, enabled and perhaps driven
by ICT, as becoming transformed, for example through business process reengi-
neering, as well as shrinking in size and becoming ‘lean’ in order to “do more with
less”. As also noted, these trends are well documented by Janssen and Estevez
(2013). The present chapter argues that the next step, and certainly a complemen-
tary perspective, is a notion of open government which is itself embedded in broader
open governance systems encompassing all of society’s actors. In this context, the
public sector needs to adapt its roles and relationships with these others actors. But,
according to Millard (2015) these adaptations do not insist that the public sector
necessarily reduces in size or becomes ‘lean’, although indeed that may happen in
some manifestations of the open governance system. Downsizing the public sector
is not a given nor is it always efficacious, but where it happens it is a politico-
economic response to specific situations and may not always be relevant, although
of course it can be so. Assuming that a smaller leaner government is always the
answer to every challenge or context is a very fundamentalist approach.
The open governance system, just as in lean government, orchestrates networks
of actors to tackle society’s needs, but unlike in lean government, the public sector
European Strategies for e-Governance to 2020 and Beyond 5
does not thereby always need to become smaller. Instead it leverages and coordi-
nates unrealised and untapped assets and resources which otherwise lie dormant or
need catalysing and are thus in effect ‘wasted’. The public sector does this both
internally and across society, so it may need to remain the same size or in some
instances even grow larger depending on the context and the challenge. The public
sector might flexibly decrease or increase or otherwise transform in size, influence
and role in different sectors and localities at different times for different purposes in
a constant ‘dance’ with other actors to maximise public as well as private value
across the whole of society. Becoming a lean government is just one option along
this continuum, even though the driving features of lean as efficiency and productiv-
ity always remain important. Instead, such features need to be seen as interlinked
between actors across the whole of society and not just confined to the government.
Thus, efficiencies and productivity improvements are conceptualised at the societal
level over at least the medium-term where trade-offs and interactions are present
between actors, not only at the individual actor level.
According to Millard (2015), this is a very important observation. A lean govern-
ment might indeed save some money in a narrow context over the short-term, but
this could lead to overall loss of public value and thus additional costs on society,
especially in the longer term, if other actors or actor configurations are not able to
produce the value needed in the context of a shrunken public sector. Examples
include environmental degradation, social and economic inequalities and in main-
stream services like health, care and education, and these would be false economies
indeed.5 As shown below, such a flexible response to address the global challenges
is now possible in the context of ICT, although of course political, governance and
other issues are also critical. This is not an argument against lean government which
may often be relevant, but an argument for flexibility in the context of open gover-
nance systems made possible for the first time by ICT.
The current, but admittedly still tentative, move from ‘l-government’ to
‘o-government’ is illustrated in Fig. 1, whilst emphasising that the four waves are
not mutually exclusive but instead complementary even though a clear progression
is envisaged. Open government (o-government) is the sine qua non for ICT-enabled
public sector innovation which is today one of the main policy agendas in Europe
and elsewhere (for example European Commission 2013a and 2013b, deriving from
European Commission 2010, and European Commission 2016, as well as the
OECD6).
5
A recent example is the Danish tax system which has for many years been driven by an NPM
approach leading to downsizing, outsourcing and seeing hastily developed IT systems as a pana-
cea. In 2016, it became clear that this has strongly contributed to losses amounting to billions of
Euros of tax revenue, both internationally and domestically. In August 2016, the tax minister
announced a reversal of these policies with massive re-investment in the tax system, the re-employ-
ment of dismissed tax personnel, and employment of thousands of new personnel, and in much
better IT. This is a clear example where political decisions leading to cutting and blind over-opti-
mistic faith in untested IT can lead to massive inefficiencies and losses.
6
See the OECD’s Observatory of Public Sector Innovation: https://www.oecd.org/governance/
observatory-public-sector-innovation/events/
6 J. Millard
Fig. 1 Four waves of e-government evolution (Source: Millard (2015), adapted from Janssen and
Estevez (2013))
Open government (o-government) is the sine qua non for ICT-enabled public
sector innovation which is today one of the main policy agendas in Europe and
elsewhere (for example European Commission 2013a and 2013b, deriving from
European Commission 2010, and European Commission 2016, as well as the
OECD7).
Although in a European Union context, the 28 Member States have full powers over
their own policies and strategies for the public sector and electronic government,
they have for many years participated in different types of mutually beneficial col-
laboration around the latter in particular. Since the early 2000s, one of the main
frameworks for this has been the regular five-yearly eGovernment Action Plans,
which, by the end of the eGovernment Action Plan 2011–2015, have assisted
Member States in putting many eGovernment enablers in place, both technical and
non-technical.
The rationale for the new 2016–2020 European eGovernment Action Plan
(European Commission 2016) is to promote efficient and effective digital public
services as important components of the EU’s Digital Single Market,8 and which
together enable cross-border public services. To achieve this, the underlying vision
is threefold:
• By 2020, public administrations and public institutions in the European Union
should be open, efficient and inclusive, providing borderless, personalised, user-
friendly, end-to-end digital public services to all citizens and businesses in the EU.
7
See the OECD’s Observatory of Public Sector Innovation: https://www.oecd.org/governance/
observatory-public-sector-innovation/events/
8
http://ec.europa.eu/priorities/digital-single-market_en
European Strategies for e-Governance to 2020 and Beyond 7
• Innovative approaches should be used to design and deliver better services in line
with the needs and demands of citizens and businesses.
• Public administrations should use the opportunities offered by the new digital
environment to facilitate their interactions with stakeholders.
The 2016–2020 Action Plan further stipulates that the following underlying prin-
ciples should be observed:
• Digital by default
• Once only principle
• Administrative burden reduction
• Inclusiveness and accessibility
• Openness and transparency
• Cross-border by default
• Interoperability by default
• Trustworthiness and Security
The policy framework for the Action Plan rests on the goal of opening up the
public sector between public administrations, across Member States and between
public administrations and other stakeholders. Three policy priorities make up the
framework of pillars:
• Pillar 1: Modernising public administration with ICT, using key digital enablers
• Pillar 2: Enabling cross-border mobility with interoperable digital public
services
• Pillar 3: Facilitating digital interaction between administrations and citizens/
businesses for high-quality public services, for example which are modular for
re-use, user-friendly and personalised, as well as for better policies based on
opening up.
The 2016–2020 Action Plan contains some new features compared with previous
plans.9 In order to remain relevant, up-to-date and to reflect as closely as possible an
evolving Europe, flexibility is being built-in to accommodate adjustments over the
next 5 years. The Action Plan is thus seen as a platform and catalyst where new
ideas, both for actions in the Action Plan itself as well as elsewhere, can be pro-
posed by Member States or other actors. A monitoring framework is being intro-
duced to track progress both on individual actions as well as overall using an
appropriate mix of indicators. In support of the dynamic nature of the Action Plan,
a stakeholder engagement plan is also being put in place, one aim of which is to
engage citizen and business interest groups through visits by the European
Commission to Member States. It relies on the use of multipliers, for example the
support of other Directorates General through inter-service collaboration and the
Regional and Structural Funds.
9
Parts of this text are derived from the author’s participation in an Expert Consultation Workshop
on eGovernment in the Horizon 2020 Work Programme for 2018–2020, held on 27 April 2016 in
Brussels.
8 J. Millard
https://ec.europa.eu/programmes/horizon2020/en/what-horizon-2020
10
Expert Consultation Workshop on eGovernment in the Horizon 2020 Work Programme for
11
As depicted in Fig. 2, the open government setting examines open data, open
service and open process, within an overarching open governance framework,
where each of these three components is open by default. It recognises that,
given that government cannot address societal problems on its own, it needs to
collaborate openly, transparently and participatively using ICT, both within and
across the public sector and with all legitimate external actors. We need greater
understanding of how shared services (across government and with non-govern-
ment actors) can be developed through co-creation, and rolled out in order to
improve take-up, personalisation and impact. Standards are required for this,
open by default, not only in technical terms such as semantic interoperability,
but also to support quality of service standards to ensure universality and cross-
border applicability where appropriate, for example through procurement, plan-
ning and decision-making. It is not immediately clear how these objectives can
be achieved and what specific roles the government should play as compared to
the other actors, particularly in the digital context. How to ensure that privacy
and security issues are adequately taken into account also needs careful
consideration.
Fig. 2 Open governance framework (“Open, Innovative and Collaborative Government: towards
a new action plan”, 1 July 2015: https://joinup.ec.europa.eu/sites/default/files/open_and_collab-
orative_government.pdf)
10 J. Millard
Open Data
Open data is seen as essential for facilitating collaboration, co-creation and policy
making, but the barrier is that for many users this is a blackbox requiring new capac-
ities, skills and incentives, so government needs to provide much more support and
many more incentives. Some countries are starting to make much of their data avail-
able publically as so-called open government data (OGD). To date there is still only
a limited number of governments which have substantially embarked down this
path, and even fewer local and regional governments where the benefits are likely to
be greater. In order to maximise the benefits of OGD, it normally needs to be suit-
ably aggregated so individual persons or organisations cannot be identified, and to
make this available in machine readable linked datasets which can also be searched,
analysed and mashed with other data. Standards for data, quality, licensing, struc-
turing, linking, searching, etc., need to be developed as well as standard tool mod-
ules for compiling, analysing and visualisation. Appropriate cloud and other systems
to provide the underlying infrastructure and services both across government and
between different actors are also necessary.
Apart from OGD made available by the public sector, citizens also collectively
generate an enormous amount of economically valuable data through interactions
with companies and government. Such data is a public sector asset, but the value
created does not always go to the benefit of the individual, particularly when third
parties (whether governments, businesses or civil organisations) collect and keep it
closed. Smart disclosure is a tool that helps provide people with greater access to the
information they need to make their own informed choices, for example in health
care, education, employment, etc. In comparison, traditional OGD focuses on trans-
parency, accountability and decreasing corruption in government.
The smart disclosure approach is a step on from this and starts from the premise that
people, when given access to data and useful decision tools built for example by govern-
ments, can use both their own personal data disclosed by them together with other
appropriate data. Smart disclosure could be a useful way forward so needs much greater
emphasis as it strives to enable the user to mash their own personal and private data
together with those of one or more service providers, including commercial services
from the private sector. This is starting to be an important feature in both the USA and
UK, for example in the utilities sector, such as energy, water and gas, as well as mobile
phone usage. In both countries, the government provides an appropriate regulatory
framework and works with the service providers (which can be other parts of govern-
ment) to make it as easy as possible for users to see their own consumption patterns, for
example via a personal dashboard, and thereby adjust future consumption. The aim is to
assist users in reducing waste or over-use and to take account of often highly complex
tariffs and service charges from typically multiple potential providers. Users need as
much support and advice as possible, but although most examples are still only pilots,
they seem to hold much potential for users to take more control of their service use. In
this context, however, there are serious issues around transparency in terms of who is
seeing and using whose data and whether or not the data owners can correct it? For
example, can technical solutions be developed which incorporate privacy by design?
European Strategies for e-Governance to 2020 and Beyond 11
Open Service
Open Process
opposite responses (Thaler and Sunstein 2008). Nudge theory focuses on changing
peoples’ behaviour without binding regulation or legislation, for example using the
insight that a very powerful influence on an individual’s behaviour is linking this to
what other people are doing through social networks and social norms in behavior
patterns.
Government as a Platform
by residents in the flight path12; Microsoft’s ‘health vault’ storing citizens’ health
records in the cloud13; ‘Fix-My-Street’ in the UK developed by the civil society
organisation MySociety not by government14; and the website ‘Patients know best’
which is a service provided by a social enterprise enabling patients to control their
own medical data when negotiating with public health authorities about their treat-
ment.15 An example from the ‘makers’ world uses digital technologies to open new
perspectives for locally manufactured and very cheap products for people who oth-
erwise have no chance of being helped. For example, in the health sector, using the
Internet to send algorithms for 3D printed prosthetic limbs designed for war victims
in developing countries for local production and use.16 These are examples where
ordinary citizens, civil organisations and many other actors have seen holes in what
government is doing and stepped in without always being invited to do so.
For the ‘government as a platform’ approach to succeed, Millard (2015) pro-
poses that at least four types of role and relationship changes are needed, and some
are already starting to be seen, as outlined below.
When government sets up collaboration platforms at many levels, its role changes
to become coordinator, facilitator and enabler, as well as regulator and arbiter for
the activities others undertake in delivering public value. Government’s role is to
ensure that public value is created by the most appropriate means in terms of what
works best in a given context and for given needs. As described earlier, this could
involve government having either a minor or major role in creating public value, but
even in the latter case government needs to be a facilitator and orchestrator to ensure
that it does.
The second new role for government is to provide tools, guidance and incentives for
collaboration. Although, the bottom-up, participatory co-creation of services can
lead to more effective and personalised experiences, doing so can increase the bur-
den placed on citizens and other actors to participate. The adoption of e-government
12
http://www.sensornet.nl/english
13
https://www.healthvault.com
14
https://www.fixmystreet.com
15
https://www.patientsknowbest.com
16
http://3dprintingindustry.com/2014/12/08/3d-life-print-3d-printing-prosthetics/
European Strategies for e-Governance to 2020 and Beyond 15
services often results in government outsourcing some of the work it has previously
done itself to the user. Co-created, or even fully user created, services take this step
much further. Developing more cost-effective and efficient public services should
mean more than assuming citizens will contribute time and other resources to create
their own services. To counter this, governments should provide structured guidance
within which service co-creation with service users can take place. ‘Guided’ sup-
port for co-creation should also be designed to reduce the burden on service users of
participating in this way, whilst optimising benefits for both public administrations
and citizens. In addition, governments should provide incentives by highlighting the
benefits service users can derive from the co-creation process, giving them more
power to make decisions about their services in adapting them to their own needs,
and supporting them with relevant data and other resources.
Third, government has an increasing role in managing the assets society has.
Especially in the context of Europe’s pressing global challenges, there is a need to
identify and deploy all society’s available assets and resources but which are often
under-used or not at all. These available assets, including government’s own, for
example, could encompass people’s time and expertise, finance, organisational
structures and competences, data, knowledge, content, networks, capacity, infra-
structures, service building blocks, things, places, buildings, spaces, vehicles, etc.
The role of government in using the power of ICT, particularly in collaboration with
other actors, is to identify, match, orchestrate, broker and coordinate assets which
can be shared and converted into public value impacts, instead of, if unused, going
to waste. Already many non-government actors are launching typically bottom-up
and small scale examples of ICT-based platforms that have such a role, for example
as part of the so-called sharing and collaborative economies, such as for example
the civil society organisation Shareable based in the USA17 (Gansky 2010).
Government has in many cases, however, greater power and scope to do this by link-
ing between actors as well as sharing its own assets internally, and this is both a
growing challenge as well as a huge opportunity. This would involve widening the
scope of ICT-based content management systems to become asset management
systems.
17
www.shareable.net
16 J. Millard
Fourth, as outlined above, seeing the public sector as a platform ensures that public
value is appropriately created and deployed. It is important to recognise, however,
that even when government collaborates with other actors in producing public
value, this does not necessarily imply that government becomes just one actor
amongst many, given that it still needs to fulfil roles that other actors normally can-
not. Such roles include being responsible for overall quality standards and mecha-
nisms for asset sharing, quality and legal frameworks, even in situations when these
are formally delegated to other actors. Accountability for services and performance,
and responsibility especially if things go wrong, is a critical issue. Other such roles
include data protection and security.
It is important to recognise that innovation and change in the public sector is not
the same as in the private sector. Government cannot pick and choose its customers
and government services cannot afford to ‘fail’ in the same way as in the private
sector. Because government is the only institution democratically accountable to
society as a whole, only it can ensure sustainable and balanced public value where
all parts of society derive benefit and where trade-offs are seen as proportionate and
fair. This shows how the overall sustainability of the governance system is impor-
tant. Governments provide longer term stability and continuity which other actors
are not able to do, and this is needed so that people and communities are able to live
stable lives, as well as so the market can have confidence that unpredictable gover-
nance changes will not upset their investment and innovation strategies. Governance
systems with short-term horizons encourage short-termism in business and an
unstable society. Instead of always the sole actor, the public sector is becoming one
player amongst many, albeit with unique responsibilities in new forms of open and
collaborative governance.
As described and exemplified above, open governance gives critical roles to the
whole range of non-government actors, and especially citizens. At the same time
that government is changing and needs to change much more, citizens are also
increasing their awareness and leverage on government but it is not yet clear whether
their future partnership with government will be a positive one. Although they need
strong support from a pro-active government as examined above, citizens should be
ready take more responsibility and become more constructively critical and produc-
tive in their own right, but this is in many ways the biggest challenge of all. Members
of the upcoming ‘net generation’ are already acting in this way in their private and
working lives enabled by digital technology, and are starting to demand that their
relationships and dealings with the public sector should take place on the same basis
European Strategies for e-Governance to 2020 and Beyond 17
(Tapscott 2009). The challenge is whether government can and will respond to these
demands, and this depends a lot on the adoption of appropriate policies, structures
and mindsets, as well as the education and incentives for citizens to support this.
Critically, it depends on government changing its roles in the ways described above.
Since President Obama launched the open government movement in early 2009
with his focus on transparency, participation and collaboration, making the USA the
first country to explicitly do so, there have been clear developments in how these
three pillars are perceived and are playing out in practice, particularly vis à vis citi-
zens. First, transparency has increasingly become the sine qua non of the successful
development of open governance systems but is also becoming better understood. It
is clear that total transparency is not the goal given that citizens, public employees
and politicians all have areas of legitimate privacy, the former in terms of the protec-
tion of their personal data and the latter two as they need confidential spaces for
dialogue and brainstorming as long as decisions themselves, as well as the evidence
and rationales for them, are transparent. Limits to transparency also need to be set
by legitimate interests, the potential for the misuse of information, slander, dis-
respect, etc., but the nature of such limits and their definitions need to be clear and
open to debate. However, robust transparency is clearly necessary as this is the basis
for accountability and for tackling corruption in government as well as in the rest of
society (European Commission 2014; OECD 2014).
Second, the understanding of participation in open governance is moving towards
a broader notion of engagement in open process. The latter sees citizens and other
actors being invited to engage in all legitimate aspects of public sector activities, not
just decision making which, in Europe at least, has tended to be the focus of
e-participation. In some ways therefore, participation perceived like this only
requires a re-active citizen, whilst engagement is more mixed and can—through
transparency and accountability—imply that citizens are more pro-active and take
into their own hands activities which traditionally have been purely public sector
responsibilities.
Third, collaboration is starting to be exemplified through co-creation and innova-
tion, as discussed above, and especially in the context of new forms of open, social
and inclusive innovation. The current governance and market systems are becoming
extremely good at ‘sweating’ assets on the supply side, so that both pubic ad private
producers become incentivized to squeeze their financial, human and other assets to
the maximum extent, and thereby increase their performance and productivity.
However, on the demand and consumption side, there is often massive asset waste,
resulting from the widespread practice of exclusive asset ownership. This has started
to be challenged in the last decade by a new sharing economy growing from a small
base, in which organisations, companies and individuals share with each other an
increasing range of their assets. These include skills, competences, time, spaces,
vehicles, tools, buildings, facilities of all types, organisational capacities and even
financial resources. Much of this sharing is enabled by ICT developments like
crowdsourcing and crowdfunding.
The sharing economy is starting to supplement exclusive asset ownership with
new forms of common, collective and collaborative ownership. The sharing
18 J. Millard
ovement started as mainly non-profit activities but is now spreading to the entre-
m
preneurial and profit sector with examples like ZipCar, Uber and Airbnb for renting
out shared cars and accommodation space respectively, and which have since grown
into global market leaders. In turn, this is threatening incumbent market and public
actors, current legal and regulatory systems as well as the frameworks of trust and
ethics we wish to maintain and build. In addition to the sharing of existing under-
used assets, a new important trend is their use for the collaborative creation, innova-
tion and production of new products, services and other assets. This collaborative
economy is already underway starting with ‘pro-sumers’ (individuals who are both
producers and consumers) mainly in the digital sector, but is now rapidly expanding
into the collaborative innovation of physical goods and services, as discussed above.
(See also Rifkin 2014).
An important underpinning of both the sharing and collaboration economy is the
trend towards co-creation, originally conceived as a business strategy for identify-
ing new forms of customer engagement, it is being increasingly applied in other
environments including in the public sector and by non-profits and citizen groups.
Co-creation is understood as the active flow and exchange of ideas, information,
components and products across society which allows for a better understanding of,
as well as participation, engagement and empowerment in, policy development, cre-
ating and improving services and tackling societal challenges. Co-creation encom-
passes co-innovation, co-configuration, and co-production of products, services and
content through modularisation and digitisation, the role of social entrepreneurs in
these new processes, and creating platforms for creative organisations, for example
around ‘standard toolboxes’ for niche needs or markets.
As noted earlier in this chapter, government is typically one of the largest single
users of ICT and other new technologies, but also is often the most hesitant. There
are arguably understandable explanations for this, but it is also clear that, sooner or
later, governments will wish or need to avail of new and emerging technologies.
This is not least in order to save resources and become more efficient, but also
because the demands on governments for new and better services of all types is
growing, including from the Internet generation.
However, it is also important to recognise that ICT has become a general purpose
technology (Perez 2009) underpinning most if not all technological innovation and
development. This means that examining ‘digital’ government purely in the tradi-
tional arrow sense, of back-office and process re-engineering and front-office online
services, no longer makes much sense. Many of the main emerging technologies
which are having, and are likely to have in the future, significant impacts on the way
governments are organised and operated, as well as on how governments are per-
ceived and used, are arising out of the so-called Fourth Industrial Revolution (World
Economic Forum 2016): “The First Industrial Revolution used water and steam
European Strategies for e-Governance to 2020 and Beyond 19
power to mechanize production. The Second used electric power to create mass
production. The Third used electronics and information technology to automate
production. Now a Fourth Industrial Revolution is building on the Third, the digital
revolution that has been occurring since the middle of the last century. It is charac-
terized by a fusion of technologies that is blurring the lines between the physical,
digital, and biological spheres”.
Many of these emerging technologies have potentially profound implications for
the way governance for both the near and longer-term future is configured and expe-
rienced, as outlined in the following.
The value and role of big data, and specifically big open linked data (BOLD), has,
as noted above, rapidly become as essential asset for developing and delivering both
commercial ad public services, as well as helping to determine and design public
policy. For example, for public sector resource planning and real time management
based on real time and archived data, for use by the police, hospitals, fire services,
the selection of politicians, staff recruitment by algorithm, etc. Big data is increas-
ingly derived, not only from archived information, but from real time sources
through the Internet of Things (IoT) as the network of physical objects and devices,
vehicles, buildings and other items that are embedded with electronics, software,
sensors, and network connectivity enabling them to collect and exchange data and
thereby also to interoperate. The IoT can optimise the use of physical objects, con-
structs and systems, such as buildings, electricity grids and utility systems, ensuring
efficient performance and reducing the carbon footprint through environmental
monitoring, disaster forecasting and management. IoT can enable the public sector
to better control and deploy its assets in real-time, such as vehicle fleets, buildings,
supplies and equipment, as well as for example manage and direct traffic flows and
other unfolding situations. In addition, geo-enabled service delivery and geo-related
information, for example on ownership, activities, functions and history, can be
used for tourism, cultural and business development.
Artificial Intelligence
For example, Benedikt Frey and Osborne (2013) estimate that 47% of US jobs will
be at risk from automation, whilst the WEF (2016) suggest that by 2025, “robots
could jeopardise between 40m and 75m jobs worldwide”. The WEF also estimates
that “65% of children entering school today will end up working in jobs that cur-
rently do not exist.” There is little doubt that this will dramatically alter the lives of
most people employed in the public sector.
Robotics
Robotics are automating much physical work across all sectors. In the public sector,
this includes, for example, routine maintenance, fabricating spare parts or special-
ised components for machines, as well as accessing difficult and dangerous environ-
ments (as in disasters, fires and floods). Robots are also starting to be deployed in
human-interface situations, such as in caring and supporting older, disabled or ill
people, although such use is proving controversial in some contexts and also raises
potential ethical issues. In the public sector context, robotics can thus have immense
impacts on care, health, elderly and frail people, cleaning and maintenance, as well
as component assembly (including components from digital fabrication).
Drones
Drones are unmanned aerial vehicles (UAVs) utilised to transport packages, food or
other goods, as well as to provide real-time surveillance of unfolding situations.
They can be used in the public sector to facilitate the delivery or collection of small
European Strategies for e-Governance to 2020 and Beyond 21
items, such as post, medical equipment and spare parts. Drones are highly flexible
and manoeuvrable vehicles that are indispensable for low-height monitoring of
natural disasters and dangerous situations, as well as for example in traffic and secu-
rity related incidents. Thus, drones have huge potential for postal services, surveil-
lance, climatic and environmental monitoring, the delivery of equipment and
supplies, etc.
Digital fabrication is the use of 3D printers, laser cutters and sinterers and other
equipment, to fabricate one-off or small production runs of unique, typically rela-
tively small objects using specifically designed algorithms. A variety of materials is
used, including metals, ceramics, plastics, glass, and increasingly organic matter
such as food and living tissue. This enables the public sector, for example, to drasti-
cally reduce its stock of equipment and components, given that these can be cheaply
fabricated only when required to highly precise and individual designs. Applications
in the health sector which are already significant include the decentralised fabrica-
tion of personalised prosthetic limbs as well as of dental replacements and implants,
and in the care sector of customised meals for people in hospitals or care homes who
have specific dietary needs. Further, and often more ethically controversial, implica-
tions include the development of genetically modified organisms (GMOs), espe-
cially in the context of rapidly advancing gene editing techniques such as CRISPR,
in sectors such as health, agriculture and food.
Blockchain Technology
Blockchain technology is a relative new, and still largely unknown, concept, par-
ticularly in the public sector, given that its main applications to date are in financial
technologies, for example as the basis of the ‘Bitcoin’. Blockchains are basically
decentralised databases that could be used, for example, for legitimation purposes,
registers, participatory decision-making, automatic taxation, social security, coun-
teracting fraud and corruption, fighting crime, etc. The impact of blockchain tech-
nology in particular on governance systems could thus be profound and lead to the
end of governance as we have known it for millennia to be replaced by, in effect, an
autonomous and independent system which everyone can contribute to and benefit
from, but which no one controls. There might be immense ‘democratic’ benefits
arising from such a scenario, but also dangers inherent in the fact that blockchains
are, in effect, an impenetrable black box.18
This brief analysis is partially based on the Wikipedia entry for blockchains (accessed 24–4-16)
18
and the Nesta blog of 24–3-16 “Why you should care about blockchains: the non-financial uses of
22 J. Millard
Conclusion and Reflection
References
Anderson C (2012) Makers: the new industrial revolution. Random House, New York
Bason C (2010) Leading public sector innovation: co-creating for a better society. Policy Press,
Bristol, UK
Benedikt Frey C, Osborne MA (2013) The future of employment: how susceptible are jobs to
computerization? Oxford University Press, Oxford
Bevir M (2013) A theory of governance. University of California Press
Chesbrough HW (2003) Open innovation: the new imperative for creating and profiting from tech-
nology. Harvard Business School Press, Boston
Cordella A, Bonina CM (2012) A public value perspec-tive for ICT enabled public sector reforms:
a theoretical reflection. Gov Inf Q 29:512–520
Dunleavy, P and Margetts H "New public management is dead: long live digital era governance",
J Public Adm Res Theory, July 2006
European Commission (2010) A European strategy for smart, sustainable and inclusive growth,
2010–2010. http://ec.europa.eu/eu2020/pdf/COMPLET%20EN%20BARROSO%20%20%20
007%20-%20Europe%202020%20-%20EN%20version.pdf
European Commission (2013a) A vision for public services. Prepared by DG CONNECT after
an expert workshop and open public consultation. http://ec.europa.eu/digital-agenda/en/news/
vision-public-services
European Commission (2013b) Powering European Public Sector Innovation: towards a new
architecture. Report of the Expert Group on Public Sector Innovation, Directorate-General for
Research and Innovation Innovation Union
European Commission (2014) EU anti-corruption report. Report from the Commission to the
Council and the European Parliament, Brussels, 3.2.2014 COM(2014) 38 final. http://ec.europa.
eu/dgs/home-affairs/e-library/documents/policies/organized-crime-and-human-trafficking/
corruption/docs/acr_2014_en.pdf
European Commission (2016) Communication on EU eGovernment Action Plan 2016–2020 –
accelerating the digital transformation of government. 19 April 2016: https://ec.europa.eu/
digital-single-market/en/news/communication-eu-egovernment-action-plan-2016-2020-acceler-
ating-digital-transformation
Gansky L (2010) The mesh – why the future of business is sharing. Penguin Group, New York
Gascó-Hernández (2014) Open government: opportunities and challenges for public governance.
Springer Science + Business Media, New York
Harrison TM, Pardo TA, Cook M (2012) Creating open government ecosystems: a research and
development agenda. Future Internet 4(4):900–928
Hood C (1991) A public management for all seasons. Public Adm 69(Spring):3–19
Janssen M, Estevez E (2013) Lean government and platform-based governance: doing more with
less. Gov Inf Q 30 Suppl 1:S1–S8
Klein N (2014) This changes everything. Penguin, Random House UK
Lee G, Kwak YH (2012) An open government maturity model for social media-based public
engagement. Gov Inf Q 29(4):492–503
McDermott P (2012) Building open government. Gov Inf Q 27(4):401–413
Millard J (2011) Are you being served? Transforming e-government through service personaliza-
tion. Int J Electron Govern Res 7(4)
Millard J (2013) ICT-enabled public sector innovation: trends and prospects. Published in the
proceedings of the 7th international conference on the Theory and Practice of Electronic
Governance (ICEGOV2012), Seoul, 22–25 Octo 2013, the ACM Press
Millard J (2015) Open governance systems: doing ore with less. Gov Inf Q. doi:10.1016/j.
giq.2015.08.003
Moore MH (1995) Creating public value: strategic management in government. Harvard University
Press, Cambridge, MA
European Strategies for e-Governance to 2020 and Beyond 25
Niehaves B (2007) Innovation processes in the public sector – new vistas for an interdisciplinary
perspective on e-government research? Electron Gov, vol. LCNS 4656. Springer, pp 23–34
OECD (2014) OECD Foreign Bribery Report: an analysis of the crime of Bribery of Foreign Public
Officials. http://www.oecd.org/corruption/oecd-foreign-bribery-report-9789264226616-en.
htm
Perez C (2009) Technological revolutions and techno-economic paradigms. TOC/TUT working
paper No. 20, 2009
Rifkin J (2014) The zero-marginal cost society: the internet of things, the collaborative commons
and the eclipse of capitalism. Palgrave Macmillan, New York
Stoker G (2006) Public value management: a new narrative for networked governance? Am Rev
Public Adm 3(1):41–57
Tapscott D (2009) Grown up digital: how the net generation is changing your world. McGraw-Hill,
New York
Tepsie (2014) Research Programme Synthesis Report: final reports for practitioners, research-
ers, and policymakers. A deliverable of the project: “The theoretical, empirical and policy
foundations for building social innovation in Europe” (TEPSIE), European Commission – 7th
Framework, Brussels: European Commission, DG Research. http://www.tepsie.eu. Accessed
30 Jan 2015
Thaler RH, Sunstein CR (2008) Nudge: improving decisions about health, wealth, and happiness.
Yale University Press
Van Veenstra AF, Janssen M (2012) Investigating out-comes of t-government using a public value
management approach. In Scholl HJ et al (eds) IFIP EGOV 2012, Springer LNCS 7443,
Kristiansand, pp 187–197
Weerakkody V, Dhillon G (2008) Moving from e-government to t-government: a study of process
re-engineering challenges in a UK local authority perspective. Int J Electron Gov Res 4(4):1–16
World Economic Forum (2016) The fourth industrial revolution: what it means and how to
respond. http://www.weforum.org/agenda/2016/01/the-fourth-industrial-revolution-what-it-
means-and-how-to-respond
Jeremy Millard is Director of the consultancy Third Millennium Governance, as well as having
senior research positions at the Danish Technological Institute in Denmark and Bradford University
in the UK. He has over forty years’ global experience on issues ranging from governance, ICT,
open and social innovation, participation, sustainable and socio-economic development, tackling
poverty and exclusion, the new economy, urbanization and nature-based solutions for growth, and
has published extensively in these and related fields. His many clients include governments, the
European Commission, United Nations, OECD and World Bank, as well as many non-profits and
companies around the world. Recent assignments in the area of e-government include on-going
support to both the United Nations regarding their biennial eGovernment Survey, and the European
Commission regarding e-government research and innovation, as well as a survey on back-office
developments in support of user-centred e-government strategies for the OECD. He has also
worked with the UN on ICT and governance issues for the 2030 sustainable development agenda,
and with ESCWA on integrated service delivery across the Arab Region.
Public Administration for the Next Generation
Peter Winstanley
Exposition
From the times of Hammurabi, and slightly later, Moses, people have used written
law and regulation to guide the focus and interactions within states and communi-
ties. Both the law of Hammurabi and the Ten Commandments given by God to
Moses were writings in stone and this indicates the degree of consideration given to
them and to some extent too the duration for which they were expected to be effec-
tive. They were also conveyed in words, and in the case of Hammurabi’s Code we
know that it was written in Akkadian, the local language of the people, so that it
might be read and understood by all – it wasn’t in some language only understood
by learned judges and civil servants. Another feature that is worthy of note in the
case of Hammurabi’s Code is the “if this, then that” approach. Specific penalties
were appropriate for specified behaviors and acts.
The world of laws and the regulation of society had its ups and downs over the
past 4000 years since these early Biblical times, and the process of developing
laws and regulations has moved from it being the efforts of one or a small number
of people involved in the determination, documentation and publication process to
a formidable coordinated effort that in many instances spans multiple organisa-
tions and, in the case of trade regulations, many continents. The development of a
professional civil service skilled in regulatory policy (Organisation for Economic
P. Winstanley (*)
Semantechs Consulting Limited, Glasgow, UK
e-mail: p.w@semantechs.co.uk
Co-operation and Development 2016a) and practice has moved on from small
numbers of scribes and sculptors to large teams covering a very wide range of
disciplines and capabilities. This development of a professional civil service hasn’t
been without its problems. In the United Kingdom, for example, the development
of a modern civil service that works with the legislature of the day to provide
expert and impartial assistance in the business of government was set on its mod-
ern course by the work in the mid-nineteenth century of Stafford H Northcote and
Charles Edward Trevelyan that derived lessons learned from the operation of the
British government and trading companies in colonial India to develop a set of
minimum educational standards, remuneration, and principles of behavior for a
professional and permanent civil service (Wikipedia 2016a). The point of the com-
petency is obvious, but sometimes lost on people who see the outputs from the
civil service in the area of public administration as being less that of developing a
smooth-running machine to being a burden to their ease of living and transacting
the business of commerce and, in some cases such as the regulations relating to the
public availability of medication, the business of life itself (Australian Government
Department of Health 2016). In more recent times the skills and competencies of
the civil services in most countries and pan-national organization (e.g. United
Nations, European Commission, etc.), together with the routes to expressing their
outputs in the area of secondary legislation and regulation have been modernising
to adopt digital technologies.
In this chapter I will take a look at the ways in which computing technologies
and the Internet are operating within some areas of public administration, and spec-
ulate as to the areas for future benefit from more widespread adoption of existing
tools and technologies. This is a personal and perhaps predominantly Euro-centric
approach rather than a systematic review, but its intention is to stimulate the reader
to further investigation and action. The field is extremely dynamic and any review
would quickly be getting out of date, but I hope that these views might give impetus
to further work in this area.
Development
Reading the Runes
The first part that I am addressing relates to the forces within which any government
operates where, as described in the early twentieth century by A.F. Bentley, it is the
attitudes and activities of groups within society that will determine the course of
government (Lemman 2008). Gauging and defining the interests and activities of
these groups is not only the core of the pollsters’ art in “reading the runes” prior to
any democratic election, but it is also part of the public administrator’s required
assessment when providing impartial advice to government on the likely impact of
particular proposed courses of action (Organisation for Economic Co-operation and
Development 2016b). Toolkits such as that used within the UK Government (UK
Government 2014) look to include some assessment of both monetized and non-
monetized costs/burdens of proposed regulation. In the UK the assessment of
Public Administration for the Next Generation 29
Policy Engineering
doing the thing right?”. This approach would be a step change from the usual usage
of scientific computing where scientific or engineering knowledge base is taken into
account in the couching of regulation, but neither in testing the drafting of regula-
tion nor in determining consistency within and between regulations – this being the
craft of parliamentary draftsmen and other legal experts. There have been several
approaches using natural language processing and statistical analyses of natural lan-
guage to extract meaning from written regulations/policies and converting this into
an RDF or UML (Brodie et al. 2006) model in an attempt to validate the consistency
of the regulation/policy and assess compliance, but these have the same underlying
problem that they are making a best estimate of meaning. This is exactly the same
uncertainty experienced with extracting meaning from social media streams. In
order to develop a knowledge engineering approach to the crafting of regulation
unconstrained natural language is too variable to be used with existing parsing and
extraction tools such as UIMA, to give a high degree of parsing accuracy, and,
unsurprisingly, having some constraint has been found to work much better (Brodie
et al. 2006). Within the domain of legal XML markup, there are moves in the UK to
use Akomo Ntoso in addition to the relatively more complex metamodel Crown
Legislation Markup Language (CLML) for the markup of legal documents. The
Akomo Ntoso model has been shown in comparative studies (McGibbney and
Kumar 2013) to be more suited to marking up the end representation of the legisla-
tion. A related markup within this area is LegalRuleML (OASIS 2016) which is a
specialization of RuleML (RuleML Inc 2016). This is less a presentation markup
and more an exchange format for machine-to-machine communication of informa-
tion (Paschke and Boley 2009). Although there are specific editors, such as LIME
(LIME – CIRSFID, University of Bologna 2016), for XML markup of general rules,
including policies and regulation, these editors have steep learning curves and the
inputs and outputs tend to be difficult for non-specialists, including ordinary citi-
zens, to understand (Beach et al. 2015). Communication of the reasoning underpin-
ning decisions is something citizens have a right to in many countries, as for example
in New Zealand (New Zealand Law Commission 2012). This points to the advan-
tage of having a multi-purpose expression format for rules, policies, regulations/
statutes etc. that can be understood by both people and machines and be serialized
if a form that permits error-free transmission across machine and human interfaces
is used. End-users of all types need system interfaces and rule bases that are easy to
interact with. These types of interactions might be as part of the development route
or feedback loop of policy development and refinement or the formal expression of
policy for the purpose of implementation through regulation, and this latter more
often than not in the modern world is mediated, at least in part, through software.
Formal Modelling
The work of Wimmer and colleagues in the “Open Collaboration for Policy
Modelling” (OCOPOMO) project (OCOPOMO Project 2012) recognizes that there
has to be a direct and discoverable link between the narrative texts contributing to
Public Administration for the Next Generation 31
the domain expertise introduced into the policymaking mix and the formal models
that policymaking will use. Provenance is important. But so too is the approach for
collecting stakeholder input. In the OCOPOMO approach narrative text is the raw
material of policy making and from this there is a process of expert interpretation to
develop a conceptual model based on the stakeholder textual inputs and other docu-
ments. Expert annotation and interpretation by the policy analyst is the key distilla-
tion process through which the input becomes crystallised as formal models that
permit, through imperative code in “Declarative Rule-based Agent Modelling
Software” (DRAMS) the running models to determine the effects of policies,
including discovery of emergent behaviours. But DRAMS rules look like computer
programme language rather than natural language (Lotzmann and Meyer 2011) and
so might be impenetrable to the citizen who provided the original narrative text. The
gap between the citizen with domain expertise or stakeholder position and the
implementing specialist is perhaps too large to bridge to ensure effective feedback
within the policy process.
Pictures of Policy
At the other extreme are approaches to bridge the communication between citizens
and policymakers using mainly visualisations. “Policy Compass” (Policy Compass
2016) is looking to Fuzzy Cognitive Maps and other visual widgets and tools as the
route to bridging the divide between the citizen and the policymaker. This presents
issues of ambiguous interpretation due to the non-standardisation of symbols (unlike
e.g. road signs), and the potential disenfranchising of the visually impaired. The
development of ‘personas’ within policy modelling (Bennet 2015) is popular within
some interested in trying to open up the policy-making process, but although it may
provide scope for an inventive workshop, the longevity of the message in the graphic
artefacts is questionable. The restricted semantics and semiotics of issue based infor-
mation systems (Wikipedia 2016b) (IBIS) such as Compendium (Compendium
Institute 2012) provide graphics that can be interpreted and the underlying arguments
replayed long after they were crafted. I think that the similar level of replay from
graphics such as in the ‘personas’ referred to above would be extremely challenging.
Given that across many cultures words are the preferred form communication, a
focus on words provides the specificity and longevity required to pass around and
debate about the direction and detail of policy and regulation. This is still the pre-
ferred way for citizens to respond to government proposals. The commentariat of the
US government are prolific though, and Tyrus Manuel describes (Manuel 2015) not
only the overwhelming volume of this feedback (e.g. 800,000 public comments to the
US consultation on Net Neutrality) but also the palpable relief to some civil servants
32 P. Winstanley
as they discover the benefits of natural language processing (NLP) in distilling the
core messages (and isolating the “weak signals”) from large volumes of information
like this. Manuel sees NLP as a part of the answer.
…We can also use NLP to gain a better understanding of what citizens are trying to tell us
on any given issue or in general. It allows for a clearer understanding of items that may need
to be addressed, from healthcare to consumer safety. NLPs can help us do a better job of not
just listening to the people, but answering them as well.
My view is that persistence with NLP will only shift the problem to elsewhere.
There is such a diversity of language that with NLP we don’t arrive at a shared rep-
resentation in a social and democratically consensual way, but are shoehorned into
consensus by algorithm and heuristic. One simple improvement to soliciting text
input from citizens is to augment it with some fixed sentiments. For example, the
website patientopinion.org.uk (Patient Opinion website 2016) gets users to input
anecdotes about the workings of the UK health service together with some marked
up facts about what was good and what could be improved. This small change is an
improvement on machine-determined sentiment, whilst allowing the contributor to
use free text. There is also scope for selecting entities from a controlled vocabulary.
An extension of this would be the proposition developed in the “Integrated Method
for Policy making using Argument modeling and Computer assisted Text analysis”
(IMPACT) EC FP7 project to use a controlled natural language for all of the textual
input (Integrated Method for Policy making using Argument modelling and
Computer assisted Text analysis 2012). There are many challenges in this approach,
including tracking the argument both across sentences in the contribution from one
individual, and also in the ping-pong of contrapuntal debate. In both these cases
incorporation of globally-unique identifiers for ‘things’ and ‘relationships’ – the
kernel of the Semantic Web – can provide this continuity. Illustrations of registries
such as those for legal entities in the GLEIF project (GLEIF – Global Legal Entity
Identifier Foundation 2016), Open Corporates (OpenCorporates 2016) and Open
Charities (OpenCharities 2016) combined with identifiers for concepts (ConceptNet
5 2012) and diverse predicates (Linked Open Vocabularies 2016) are providing the
Lego™ building blocks for a constrained but rich set of fixed points that can enrich
an existing controlled natural language (CNL) approach to describe and comment
on policy and statute. Simple “What You See Is What You Meant” (WYSIWYM)
interfaces (Power and Scott 1998) have given way in recent years to sophisticated
ontology editors such as Fluent Editor (Cognitum 2016) which uses the Ontorion
Controlled Natural Language (OCNL) to guide the creation of WYSIWYM docu-
ments including ontologies and rules bases (Seganti et al. 2016).
Bridging the Gap
done in a way that copes with computational aggregation and summarization so that
the 800,000 responses to a consultation are undertaken and the single thread of
argument counter to the 799,999 others is discovered and considered on its merits.
The computational capacity and transformational fidelity of controlled natural lan-
guage allows this, and much more besides. Some of these additional benefits are
being played out in work ongoing in the Dutch Finance and Customs Administration
(Belastingdienst) where controlled natural language is being used to provide the
rule bases that are parsed using ANTLR to an intermediary that can be compiled
directly into code. This ability, even in a restricted domain such as finance and cus-
toms, to develop software artefacts directly from a human readable set of rules
opens up a wide range of new possibilities in public administration. In short, busi-
ness rules are prepared in “RegelSpraak”, a Dutch CNL based on “RuleSpeak”
(RuleSpeak 2016) that is fully consistent with Semantics for Business Vocabulary
and Business Rules (SBVR) (Object Management Group 2015) and is easily
human-readable. As Chris Maple of MMG Insurance (another organization that has
adopted CNL rule language deep into its business processes) states, “A lot of really
smart people have done really good thinking in this area.” (Maple 2014). The busi-
ness rules couched in the CNL can then be parsed and compiled into a form suitable
for e.g. the DROOLS rule engine.
But just as NLP is not necessarily going to provide the magic wand that allows
administrators to get computers to read their input from citizens, so the full-on appli-
cation of CNL is not going to allow citizens to provide their input to the process of
public administration in part because it requires some prior learning, and many peo-
ple are not going to adapt to that. It also potentially constrains the concepts and
constructs that a citizen might be able to use to connect with the administration, and
that would be politically untenable as it is likely to be seen as coercive and restric-
tive. There is potential scope for the scenario where citizens could be using argu-
mentation and debating technologies (IBM 2016) to facilitate constructing their
response to an administration’s proposals or actions in the same way that the legal
profession has AI tools such as “Ross” (ROSS 2016) at its disposal, but the outputs
are still in natural language and so there is an additional ‘layer’ or aspect needed to
help people use computers to process more effectively and efficiently the points
made and to aggregate them accurately and integrate them into other knowledge
accurately and efficiently. Equally, this move to a wholly IT-mediated discourse
based on argumentation alone fails to develop the trend to more inclusive and par-
ticipative democracy that is gaining momentum in countries such as Scotland where
recent conversations within the country have recognized that there needs to be both
online and offline interactions between government and citizens and the creation of
safe spaces for dialogue to help infuse ideas from all sectors of society into the mix,
and to mitigate the political risks of “getting it wrong” and the fears that stoke risk
aversion in administrations (Stoddart 2014). Collaborative Government that includes
administrators going to where the conversations are taking place requires some
mechanism of channel discovery, and this again is bringing us to the challenge of
scale. How can government monitor the online dialogues to identify which to join?
We are back to analyzing Twitter streams and the complexities and hazards of either
getting administrators to do this job themselves or using some algorithmic approach.
34 P. Winstanley
Ideality?
Smaller populations such as Iceland have the capacity to sample opinion in a much
more authoritative and interactive way and this was realized in 2009 and subsequent
years during a period of constitutional reform (Bergsson and Blokker 2014). This
exercise to update the constitution involved taking a random sample of c.1000 peo-
ple and getting them to talk about issues and then 25 of this assembly were selected
by the voting population (Iceland Review Online 2010) to address the issues identi-
fied by the larger assembly and write the revision of the constitution (Wikipedia
2016c). The large group was split into 128 smaller groups and their ideas were
condensed into word clouds as a rapid means of determining the topics of interest
(Blokker 2012). Clearly this is a circumstance where CNL could not only advance
this Icelandic approach but also allow it to scale effectively to larger numbers of
participants and provide in computable form greater complexity of input than sim-
ple word clouds.
Recapitulation
References
Amazon Web Services (2015) Lambda architecture for batch and stream processing on AWS. Available
at https://d0.awsstatic.com/whitepapers/lambda-architecure-on-for-batch-aws.pdf
Australian Government Department of Health (2016) Regulation and red tape reduction. Available at
http://www.health.gov.au/internet/main/publishing.nsf/Content/regulation-and-red-tape-reduction
Beach T, Rezgui Y, Li H, Kasim T (2015) A rule-based semantic approach for automated regula-
tory compliance in the construction sector. Expert Syst Appl 42(12):5219–5231. doi:10.1016/j.
eswa.2015.02.029
Bennet S (2015) Using design and open policy making techniques to address complex problems.
In UK government “Future of Aging” blog. Available at https://futureofageing.blog.gov.
uk/2015/10/01/design-and-open-policy-making/
Bergsson BT, Blokker P (2014) The constitutional experiment in Iceland (September 4, 2013). In: Pocza
K (ed) Verfassunggebung in kon-solidierten Demokratien: Neubeginn oder Verfall eines Systems?
Nomos Verlag. ISBN 3848709996. Available at SSRN: http://ssrn.com/abstract=2320748
Public Administration for the Next Generation 35
Organisation for Economic Co-operation and Development (2016b) Regulatory impact analysis.
Available at http://www.oecd.org/gov/regulatory-policy/ria.htm
Paschke A, Boley H (2009) Rule markup languages and semantic web rule languages. In: Giurca
A, Gasevic D, Taveter K (eds) Handbook of research on emerging rule-based languages and
technologies: open solutions and approaches. Chapter: 1. IGI Global, pp 1–24
Patient Opinion website (2016) Available at https://www.patientopinion.org.uk/
Pedersen S, Baxter G, Burnett S, Goker A, Corney D, Martin C (2014) Backchannel chat: peaks
and troughs in a twitter response to three televised debates during the Scottish independence
referendum campaign 2014. Aberdeen Business School Working Paper Series 7(2). Available
at https://openair.rgu.ac.uk/handle/10059/1086
Policy Compass (2016) Project website available at https://project.policycompass.eu/the-project/
Power R, Scott D (1998) WYSIWYM: knowledge editing with natural language feedback.
Proceedings of the 9th international workshop on natural language generation (INLG 98),
Niagara-on-the-Lake
ROSS (2016) Website available at http://www.rossintelligence.com/
RuleML Inc (2016) http://wiki.ruleml.org/index.php/RuleML_Home
RuleSpeak (2016) Website available at http://www.rulespeak.com/en/
Seganti A, Kapłański P, Zarzycki P (2016) Collaborative editing of ontologies using fluent editor
and Ontorion. In: 12th international experiences and directions workshop on OWL, OWLED
2015, co-located with ISWC 2015, Bethlehem, 9–10 Oct 2015, pp 45–55
Stoddart A (2014) Collaborative government in Scotland. Available at http://www.gov.scot/
Resource/0045/00459780.pdf
Taieb D (2016) Real-time sentiment analysis of twitter hashtags with spark. IBM cloud data ser-
vices developer center. Available at https://developer.ibm.com/clouddataservices/2016/01/15/
real-time-sentiment-analysis-of-twitter-hashtags-with-spark/
U.S.C. § 553(c) (2006) ([The agency shall incorporate in the rules adopted a concise general state-
ment of their basis and purpose.). Quoted in Kevin M. Stack (2012) Interpreting regulations.
Mich Law Rev 111(3):355–422. available at http://repository.law.umich.edu/cgi/viewcontent.
cgi?article=1080&context=mlr
UK Government (2014) Regulatory impact assessments: guidance for govern-
ment departments. Available at https://www.gov.uk/government/collections/
impact-assessments-guidance-for-government-departments
UK Government Digital Service (2016) Digital People blog website. Available at https://digi-
talpeople.blog.gov.uk/
Wikipedia (2016a) Northcote–Trevelyan report. Available at https://en.wikipedia.org/wiki/
Northcote-Trevelyan_Report
Wikipedia (2016b) Issue-based information system. Available at https://en.wikipedia.org/wiki/
Issue-Based_Information_System
Wikipedia (2016c) Icelandic constitutional reform, 2010–13. Available at https://en.wikipedia.org/
wiki/Icelandic_constitutional_reform,_2010%E2%80%9313
Peter Winstanley is a data and standards specialist with experience covering more than 25 years
in both government and private sector operations. Peter is a member of the UK Government Data
Standards Panel and Glasgow University School of Computing Science Industrial Advisory Board
and is a Chartered Fellow of the British Computer Society. He has contributed to the development
of European Community and W3C standards and recommendations on semantic interoperability.
Peter works routinely with smart data structures such as XML and RDF to develop and deliver high
impact, low maintenance solutions.
The Citizen Scientist in the ePolicy Cycle
Introduction
The term science as well as the nature of conducting science evolved over time. Not
always has research revolved around the methodological approach as we know it,
and not always has it been driven by the measures of today. In this paper we start by
describing the nature of conducting science and how some scientific paradigms
changed over time. This is relevant for our analysis of citizen science in relation to
the ePolicy cycle, as changes like the focus on the openness paradigm combined
with available means for sharing and mass collaboration also changed how citizens
can participate in the research process.
The section Citizen Science focuses on how openness in the research process
combined with means for mass collaboration can empower citizens to enrich the
research arena. After describing these changes, we take a more detailed look at the
possible ways to engage citizens in this process. The section Enablers and Methods
of Citizen Engagement summarises some recent modes of citizen participation and
engagement, mostly in relation to ICTs and digitalization, and the participatory and
technological aspects of citizen engagement. We also briefly address the opposite of
those enablers in the form of hurdles to citizen science.
In the Big Data Enabled Policy Cycle we present the policy cycle as a theoretical
vehicle to structure public policy making in light of technological advance. Use
Cases for the Citizen Scientist in the Policy Cycle ties together intrinsic motivation
and external enablers in respect to the policy cycle. In Challenges, Issues and Future
Implications we discuss existing impediments to unleash the potential of Citizen
Science in policy making, ethical and cultural considerations as well as potential
implications of future research.
By combining insights from different disciplinary fields, we hope to point towards
the chance of engaging citizens on various stages of the policy cycle, in particular
with view to an increased culture of sharing and related possibilities for evidenced-
based and participatory policy making.
For the most part of history, science was not meant for everyone. In former times,
many people lacked the basic foundations of what was perceived to be a pre-requisite
for scientific work, namely mathematics, jurisprudence, medicine, theology, and
philosophy. The lingua scientia was dependent on epoch and geography and differed
many times from the theodiscus, the people’s language. Thus, only people capable
to communicate in the scientific language were able to participate in the discourse.
Aristoteles created the nomenclature of practical science containing, f.i. politics and
ethics, theoretical science, mathematics and theology and poietic science, including
medicine and poetry. Elaborations meant for wider consumption were called exo-
teric, whereas those works targeting the circle of like brethren esotheric.
Methodology and reproducible results did not play the crucial role as they do
today. Alchemy, occultism, and religion where all closely related disciplines and
influenced what would become modern science. None of these areas is known for a
deep methodological foundation and for good reasons: to believe rather than to
know was an integral part of a scientific approach back in these days.
In his seminal work Saggiatore, published 1623, Galileo Galilei argued to under-
stand nature requires understanding mathematics, otherwise the inner workings of
nature would remain unintelligible. He also dismissed both Alchemy and Astrology
as incapable to describe nature, a view Francis Bacon already shared 1597 in his
essays: To master nature requires to understand nature. Bacons notion of understand-
ing was freed from influential idols of its time like the Greek philosophers Platon
and Aristoteles and, with Bacon’s words, their illusions. However, even a generation
later, science was still deeply embedded into religion, occultism and alchemy. Isaac
The Citizen Scientist in the ePolicy Cycle 39
until he admitted in 2011 that he had been fabricating the data. Apart from these
more extreme examples, a scientific movement called reproducibility movement has
been formed, and the community pushes not only for publication and sharing of
data, but also for the possibility to reproduce results. While irreproducible evidence
does not mean that results are wrong, it could also refer to undetected variables.
In his highly disputed book Against Method, Paul Feyerabend claims that the
idea of a method that contains firm, unchanging, and absolutely binding principles
for conducting science meets considerable difficulty, when confronted with the
results of historical research. There is not a single rule, however plausible, and how-
ever firmly grounded in epistemology that is not violated at some time or another.
He claims that such violations are necessary for progress (Feyerabend and Hacking
2010). This and many more propositions discussed by Feyerabend bear lots of con-
troversy, as they are shaking on the still young pillars of what just became “tradi-
tional” science.
Neglecting the discussion onto which more attention should be laid upon – the
availability of scientific data or a sound methodological approach – there seems to
be agreement that scientific research should become tangible for many more people
than it is today. Furthermore, we observe a shift towards research impact, visible in
the increasing importance of quantitative research measures and automatized cita-
tion indexes, like Google Scholar for impact monitoring1 (Harzing and van der Wal
2008). With an increasing amount of people becoming part of the scientific com-
munity, a term, which constitutes no sharply-delineated area anyhow, new ways of
how to conduct research are emerging.
Open Science
How science emerged and was conducted changed significantly over the past centu-
ries and is still undergoing rapid shifts and changes today. In former times, scientific
activities were rather performed by the aristocratic society than by common people,
as the trustworthiness of the associated results was strongly interconnected with the
scientist being a “gentleman”. Yet the situation has changed more and more in
favour of repeatability and availability of data than relying purely on big names and
the reputation of huge organizations. While science has sought to include outside
expertise (Carpenter 2001), the view on the notion of the expert itself also under-
went a significant shift. Taleb notes that a great deal of important scientific discover-
ies with significant impact did not result from planning and foresight, but mostly
resulted from a trial and error approach and the unexpected (Taleb 2007).
With view to the inclusion of expertise in ideation systems, different approaches
to include outside knowledge or expertise have been classified, mostly focusing on
1
Also in the e-government or e-policy domain, cp. f.i. Scholl, H.-J. (2016), Profiling the Academic
Domain of Digital Democracy and Government, presentation at CeDEM16, conference for
e-democracy and open government, 18th May 2016, Krems, Austria.
The Citizen Scientist in the ePolicy Cycle 41
2
Enserink, M., In dramatic statement, European leaders call for “immediate” open access to all
scientific papers by 2020. Science, 27th May 2016, http://www.sciencemag.org/news/2016/05/
dramatic-statement-european-leaders-call-immediate-open-access-all-scientific-papers (accessed
15th July 2016).
42 J. Höchtl et al.
researchers in reflecting on their initial thoughts, while at the same time promot-
ing new ideas within the scientific community and beyond; even influence entire
research directions in the long run. These published ideas can be commented,
evaluated, or even challenged by other scientists or amateurs. Furthermore, the
initial starting point of a concept and its evolution over time can be traced more
easily this way, as the pre-published versions stay within the Internet even after
the final paper has been accepted and published by a publisher.
3 . Constructivistic perspective: Arising co-creational processes open up new ways
of publication development. This includes new and innovative business models
as well as associated user models. A prominent example for such an approach is
crowdsourcing in which the wisdom of the crowd is used solve problems in a fast
and flexible manner and citizens are required to support professional scientists’
work, but raising scientific issues or drawing upon problem-solving strategies
are still done by professional scientists (Dickel and Franzen 2016). Open plat-
forms with small groups of experts loosely moderated and support the discussion
and dialogue between involved parties. But not only problem-solving but also
data collection are part of these perspective.
4. Exploitative perspective: This perspective refers to real life applications and
application-orientated knowledge exploitation in cooperation with practitioners.
Citizen Science
Finke notes that the English term “citizen science” is related to a predominance of
the Anglo-Saxon countries in this research area. However, with view to the actual
content, he constitutes no big national or cultural differences (Finke 2014, p. 37):
everywhere people participate on the collective acquisition of knowledge and on
forms of knowledge transfer. While his claim that scientific engagement is not based
on profession, titles or control structures, but on interest, skills and activities can be
debated, it seems obvious that citizen science can only be realized on the basis of
such attitudes. For Finke, the term of the amateur or layman is significant for citizen
science. Rationality (German: “Laienrationalität”) enables citizen science in a con-
tinuously more complex world. Citizenship means to be engaged for something.
Citizen science according to Finke satirizes a too narrow understanding of a science
that is done only by professionals (Finke 2014, p. 40). Irwin defines the term:
“Citizen Science” evokes a science which assists the needs and concerns of citizens.
He further notes that the term also makes the point for a science that is developed
and enacted by citizens themselves. (Irwin 1995, p. xi). Feyerabend (1978) even
claims that the amateurs are the only citizens that can be trusted to criticize or moni-
tor science independently. Crucial in this regard is that the distinction between citi-
zens and scientists is blurred, and emphasis is put on the context of scientific work:
on everyday life and the lifeworlds of citizens. This claim corresponds well with
newer theories of citizenship and participation fostered by the affordances of every-
day life, hybrid media environments, e.g. the concept of mundane citizenship by
The Citizen Scientist in the ePolicy Cycle 43
3
The Extreme Citizen Science Group at UCL London is also working with marginalized commu-
nities in citizen science activities with the goal to enable wider participation by lay people.
44 J. Höchtl et al.
regard. The main motivational driver for participation in these projects is based on
individual interests regarding collaboration-related social aspects. The techno-
logical adoption rate has increased significantly, as online-based citizen science
resources such as blogs offer data publicly to be integrated in own projects.
Citizen science has been described as participatory science (Conrad and Hilchey
2010; Carr 2004). While the use of volunteers has always been an important com-
ponent, it has evolved into citizen science within the past two decades (Catlin-
Groves 2012). This can partly be explained by the use of ICTs fostering forms of
participatory science.
While some forms of citizen science refer to a more active form of engagement, a
good deal of participation in digital late modernity is based on more mundane, implicit,
opportunistic or more passive forms of engagement. As Bennett and Segerberg note
on the characteristics of contemporary networked societies, a different form of organ-
isational structure enabled by phenomena of connective individualism (Bennett and
Segerberg 2012) and expressive issue-engagement (Svensson 2014) emerged. This
has mostly been explained by a specific form of collective action, initiated by the per-
sonalisation of actions. With this different “logic of connective action” (Bennett and
Segerberg 2012) and the ubiquitous utilization of new media and technologies, the
structures of mobilisation and techniques for citizen engagement have transformed.
The argument has been put forward that communication technologies replace the need
for traditional communities of action, in other words: those technologies take over
what has traditionally been done by humans, making it easier for humans to organise
themselves and to reduce the cost of organization and sharing.
Research has emphasised the importance of civic engagement as the actual
strength of citizen science (Finke 2014). This can also be done in a more continuous
form. Also monitoring function thus does not have to come at the end of a process,
but can be executed permanently. In this context the potential of online media can
create a multitude of responses and reactions (Papacharissi 2009, p. 230).
While modern citizenship assumes an active role (cp. The term “DIY Citizenship”
by Ratto and Boler 2014), not all modes of participation in the digital networked
society have to be completely active. Research has also emphasised the importance
of less active form of participation, e.g. in the form of so called “lurkers” (Nonnecke
and Preece 2000), who are less active and remain silent, but nonetheless are an
important factor of online engagement. In an extreme form, citizens can provide
their data as sensors. Information can now be packed digitally and travel anywhere
in the world. On the basis of this speed of flow of information coupled with its “rela-
tive uncensorability” (McNair 2009, p. 223) and the collapse of time-space “distan-
tiation” (Giddens 1990) and the assumption that members of society have access to
and can afford to buy the hardware, the sharing of information has become a com-
monplace of cultural life, leading to a different form of communication with a lot of
The Citizen Scientist in the ePolicy Cycle 45
data remaining unused. This expanded information flow makes participants con-
stant producers of data, amounting to a globalized public sphere (McNair 2009).
Catlin-Groves distinguishes on the citizen landscape from volunteers, citizen sen-
sors and beyond (Catlin-Groves 2012). In this classification, virtual citizen science
refers to data mining in a passive framework (f.i. via social networking sites), which
can also have a more active form in the form of active participation. Furthermore,
citizen science can comprise “citizen sensing” as an active framework via mobile
submissions.4 Catlin-Groves notes a move “from standardised data collection meth-
ods to data mining available datasets”, well as the “blurring of the line between citi-
zen science and citizen sensors and the need to further explore online social networks
for data collection” (Catlin-Groves 2012). In the context of citizens providing data,
(Cooper et al. 2007) emphasise a distinction between “citizen science” and “partici-
patory action research”. Citizen science should ideally not use citizens on unequal
terms and treat them as scientists on equal terms and not foster a state of competition
(cp. Finke 2014).5 A framework for engaging expertise or knowledge has also been
proposed by Dickel and Franzen (2016), who categorize two dimensions in four
levels of expertise, which are comparable to science and relevant for policy makers
(Fig. 1).
These roles are not found in empirically pure form, but seek to conceptualise
inclusion efforts in citizen science. Apart from the differentiation along the needed
expertise, these roles distinguish whether the link to the expertise is characterised
by competition or cooperation. When characterized as competition, inclusion efforts
are expected to be rejected (Dickel and Franzen 2016; Finke 2014), and competition
4
It can be noted that these newer forms of citizen engagement re less standardized, but mostly
opportunistic or directed.
5
Data compilers should be able to utilize centralized data to produce scientific results in exactly the
same way as anyone else should be allowed.
46 J. Höchtl et al.
between amateur science and professional science is usually implicit. It can become
explicit f.i. when publications of amateur scientists are criticised by the academic
world or the other way round (Dickel and Franzen 2016).
Participation in the citizen science landscape can be based on more than intrinsic
motivation. The willingness to share can be based on civic engagement, the joy of
discovery, but also on more playful motives and play instinct (Finke 2014, p. 124).
Another enabler is the private knowledge motives of participants or self-selected
areas of interest, sometimes in the form of hobbies and the will to preserve and cre-
ate knowledge. Behavioural approaches to spatial data sharing have also empha-
sised the importance of the following contextual factors for the willingness to share:
attitude (f.i. strategic position or social outcomes), social pressure (f.i. of institutions,
moral norms or the market) and perceived control (f.i. technical or interpersonal
skills or finding sharing partners) (de Montalvo 2003).
While those motivational factors play a big enabling role it should also be noted
that limited access to technology or technophobia can play a role, and factors
explaining motivational access to technology can be of a social/cultural or a mental/
psychological kind (Van Dijk 2009). Many technologies do not have appeal for the
low-income or low-educated though, and if citizen science is to be appealing to such
people, computer anxiety or technophobia as major barriers to access has to be
taken into account, as these phenomena are not expected to disappear with the ubiq-
uity of networks in the digital age (Van Dijk 2009). However, technologies of com-
munities (Irwin 2001) make it easier for citizens to participate when they feel like.6
Another strategy in lowering the participation threshold is the integration of ele-
ments of gamification or game-related elements. Thiel (2016) undertook a meta-
analysis of the use of such elements in the field of digital participation. She concludes
that while gamification does not work similar in all domains, if situated carefully in
the relevant context, gamification could increase the level of participation in some
areas and under specific circumstances. However, several studies have already
proven that the strategy of adding game elements to influence users’ behaviour can
be successful: “The most common objective behind gamification is to increase the
usage of a system. Other scholars have shown that game elements can increase the
perception of effort, make tasks or services more enjoyable and control behaviour.”
(Thiel 2016, p. 7).7 Others have found that gamification had no effect in the context
of a citizen science application (Bowser et al. 2013): it was found that in an intrinsi-
cally motivated user group the game elements in a citizen science application were
almost incidental. This can be explained as citizens were intrinsically interested in
the non-game context and did not need an additional motivator. Thiel concludes that
only if game aspects are utilised correctly and contextualized, they can build a
highly motivational user experience (Thiel 2016, p. 8). However, the gamification
approach can be effective in terms of influencing or tapping into users’ motivation
6
Irwin explores the configuration of the scientific citizen within policy and consultation processes
and accesses the significance of such technologies for the practice of scientific citizenship.
7
Thiel also addresses that ethical considerations need to be considered.
The Citizen Scientist in the ePolicy Cycle 47
up to a certain level in order to create a first motivating environment (cp. f.i. on the
agenda setting level).
With view to digital infrastructures, methods of science-driven crowdsourcing
enabled by the digital are described by Dickel and Franzen (2016), in which a task
normally performed by members of an organization is outsourced. Forms of such
crowd science relevant in our context also comprise delegating online data collec-
tion and assessment to the public. That way, crowd science enables the implementa-
tion of large data-intensive projects, which could otherwise hardly be implemented
(Franzoni and Sauermann 2014). As Dickel and Franzen (2016) note, knowledge
production and the reception of knowledge are becoming increasingly socially
inclusive. This raises the question of how much more inclusive new institutions
should be and how confidence can be guaranteed if the cycle of experts is expanded.
They propose a typology of digitally-supported inclusion models, and on that basis
conclude that the line between certified experts and laypeople is blurring (Dickel
and Franzen 2016, p. 3).
The preceding section primarily dealt with intrinsic factors of motivating participa-
tion in citizen science, while this section focuses on extrinsic enablers, with a closer
look on big data related technology. We further ask what this could mean for sup-
porting and evaluating governance processes and policy.
It sometimes feels like our society is obsesses with numbers. Scientific theory
mostly sees this as a good thing – reproducibility requires prove on the basis of
facts, figures numbers. Deming, the inventor of modern quality management and
heavy influencer of the reconstruction of post-World War 2 Japan towards the
world economic powerhouse of the 1960s, 70s and 80s, coined the following
phrase: “In God we trust; all others must bring data”. Books on Amazon with titles
referring to data divination are selling well. What does this mean for the future
role of the citizen scientist and how does it affects our society? More precisely:
How will policy making be conducted in the future? Let’s start with some big
numbers first.
Our known universe consists of roughly 1080 atoms, a number impossible to
fathom. Written out it spells as one-hundred thousand quadrillion vigintillion. Yet
Peter Norvig, Director of Research at Google, tends to disagree and argues the
(small) number of atoms in the universe. In a blog post referring to Googles break-
through in beating a human being in the board game of Go, Norvig addresses com-
binational theory. For example, the number of combinations made possible by a
40-character passphrase, consisting of uppercase, lowercase, numbers and special
characters, already reaches the numbers of estimated atoms in our universe.
Comparatively, the board game of Go with a 19 by 19 field setup entails 10170 legal
positions. In other words, combinational theory, which is by nature multiplicative,
48 J. Höchtl et al.
dwarfs every number of our additive physical nature.8 Translated to the citizen sci-
ence domain, in 2015, 3.2 billion people had access to the internet and they are all
potentially connected (ITU 2015). This theoretically entails an incredible number of
possibilities to share and re-combine data and translate it into valuable knowledge
for individuals, business making (What will be the next product a customer buys?)
and government (Where is the best place to build a new hospital?)
Combinational theory is just one aspect of the transformational power of ICT
enabled by network-connected infrastructure. It is reminiscent of Metcalfe’s law,
which states that the value of a telecommunications network is proportional to the
square of the number of connected users of the system. In other words, every citizen
creating data theoretically exponentially increases the value of the network.
Around 2000, two remarkable events related to digitisation took place. First, the
amount of digital information surpassed the amount of analogue information.
Second, the speed of data and information creation significantly accelerated. Today,
a multitude of devices is available at comparatively low costs, enabling the mainte-
nance of networked connections and sensing a multitude of data points; be it RFID-
chips, the Internet of things, city sensors or connected sports gadgets. General
purpose computers with low power requirements like the Arduino9 or the Raspberry
Pi10 sell for around 50 € and enable their owners to conceive all sorts of integrated
gadgets like home automation devices, weather stations, and beer brewers11. However
the most widespread digitisation device in use is the smartphone. According to
Statista, in 2015 there were 1.8 billion smartphones in use worldwide,12 which are
connected to the Internet most of the time (Fig. 2).
How is this related to citizen science for twenty-first century policy making?
Another puzzle piece in our line of argumentation is that of intelligence. When
thinking about intelligence, what springs to our mind is human intelligence or secret
services. We do not know for sure if people’s intelligence changed much in the last
three hundred years – the time frame in which modern science formed. Certainly the
artefacts we create are increasingly impressive, but this may to a large extent be due
to collective intelligence and how we are able to pass knowledge through objects
rather than through genes. With view to citizen science something is of greater
importance: the ability of algorithms to cope with the plethora of data and informa-
8
http://norvig.com/atoms.html, retrieved 2016-07-12.
9
https://www.arduino.cc/
10
https://www.raspberrypi.org/
11
http://www.networkworld.com/article/2290609/computers/computers-153240-20-cool-things-
you-can-do-with-a-raspberry-pi.html, retrieved 2016-07-12.
12
http://www.statista.com/statistics/330695/number-of-smartphone-users-worldwide/ (data from
eMarketer), retrieved 2016-07-12.
The Citizen Scientist in the ePolicy Cycle 49
2007
ANALOG
18.86 billion gigabytes
2000
1986 1993
ANALOG
2.62 billion
ANALOG STORAGE DIGITAL
DIGITAL
0.02 billion
2007
DIGITAL
276.12 billion gigabytes
Fig. 2 As of 2000, more information is available in digital rather than analogue and the speed at
which data and information accrues tremendously increased (Hilbert and López 2011)
tion generated every day. While the combination of networked devices, exchanging
data and information can be the source for better decision-making, it’s the algo-
rithms that provide us with the means to actually do so.
Looking back at the combinatorial features we previously identified, the sheer
amount of data would be far too large to store, inspect and analyse by any computer
system using traditional algorithms. A new way of thinking about problem solving
emerged. Striving for optimal solutions in Big Data requires the usage of algorithms
which expose polynomial runtime behaviour. Dedicating more computational
power in terms of available computing cycles, network speed and storage capacity
becomes unfeasible and increasingly impossible. A practical solution outplays opti-
mal solutions which, due to their runtime complexity, may only be able to process a
fraction of the available data and thus lead to local optima. “Good-enough” algo-
rithms become necessary if the amount of available data gets too large to be handled
by traditional ICT systems (Mayer-Schönberger and Cukier 2013). Imagine an
international online retailer. Even such seemingly simple questions such as “How
many items of X have we sold today in region Y?” become impossible to answer,
given the amount of data accrued over time.
50 J. Höchtl et al.
Another crucial aspect of today’s ICT systems is the capability to speedily react
on external events. This requirement for speed may either be triggered from a single
sensor continuously transmitting data, a sensor network whose collectively gath-
ered data results in a continuous data stream, or diverse and heterogeneous data
sources combined, like sensor and social media. Instant access to analysis results is
paramount.
An illustrative example to this new sort of intelligence we would like to present
is the HyperLogLog-Algorithm (Heule et al. 2013). This algorithm on the one hand
can deal with enormous amounts of data, yet at the expense of being not 100%
accurate. However, this is made up by the ability to analyse many facets in the data-
base to potentially identify multi-perspective patterns. Additionally, this algorithm
operates stream oriented, i.e. directly on the data as it arrives at ICT systems. Instead
of requiring an additional analysis step, analysis data is available in real time. This
is the sort of intelligence we introduced before and which completes the triangle of
The Digital Virtuous Forces. It is also this breed to algorithms which prevents mis-
interpretations in data sets by an ill-chosen or arbitrarily chosen data sampling rate.
The importance of correct sampling is well known to statisticians and an integral
part of every 101 statistics course. The danger of taking adverse decisions based on
incorrect or skewed samples can be adverse to harmful, depending on the conse-
quences drawn from the data. If it’s a million dollar business behind, correct sam-
pling becomes paramount. Imagine an online retailer, collecting a vast amount of
behavioural data (the “user journey”) every day to improve the customer experience
and to early react on changes to interaction patterns. Taking no decisions at all can
be better at times instead of taking the wrong decision. That’s what has happened to
Internet giant Ebay in 2003. Back in 2003, Ebay collected a vast amount of web
interaction patterns but was only able to analyse parts of that precious data. Future
decisions were based on the reliance on correct or good sampling techniques.
Analysts knew that due to their inability to incorporate all the data into their deci-
sion and alert models, valuable data patterns will remain undiscovered and spurious
patterns arouse where there are actually none.13 Using algorithms, which can inspect
the data in its entirety yet at the cost at not arriving at absolutely exact results, was
favourable for Ebay.
By describing the changed characteristics of ICT systems we introduced an
important concept which we think will change the way government policy is made
at each and every level in the future: big data analytics. Big data may be defined as
the “cultural, technological and scholarly phenomenon” made up of the interplay of
algorithmic analysis of large datasets in order to identify patterns (Boyd and
Crawford 2012; Ulbricht 2016).
While the technological dimension is emphasised, it should also be noted that
big data also entails an important cultural dimension, in our context referring to the
Cliff Saran: How big data powers the eBay customer journey. Case study, Computer Weekly,
13
2014-04-29 (http://www.computerweekly.com/news/2240219736/Case-Study-How-big-data-
powers-the-eBay-customer-journey, retrieved 2016-12-11).
The Citizen Scientist in the ePolicy Cycle 51
The widely accepted model for the design of government policy making is the pol-
icy cycle. Originally described 1956 by US political science researcher Harold
Dwight Lasswell, the policy cycle provides a theoretical frame to explain govern-
ment policy making. Depending on the chosen abstraction level and granularity of
the step model, (a) Agenda Setting, (b) Policy Discussion, (c) Policy Formulation,
(d) Policy Acceptance, (e) Provision of Means, (f) Implementation and (g)
Evaluation can be distinguished. The cycle is a helpful instrument for all affected
stakeholders like politicians, public administration, NGOs, business entities, and
the public when organizing campaigns to respect regulations, or which supportive
or enabling ICT instruments can be considered. However, the policy cycle does not
come without criticism. First it should be understood as a heuristic which requires
tailoring to the actual needs. In practice, the sharply distinguished steps will overlap
or certain steps left out altogether (Prozesse—Der Policy-Cycle 2009, p. 110).
Everet et al. also identify an overemphasis on the process itself rather than quality
or performance (Everett 2003).
Arguably the biggest factor of influence to this approved model is technological
change. As we identified, the biggest amount of data today is digital, arrives at high
speed and is, due to its plentiful sources, of varying structure. Looking at the tradi-
tional policy cycle, the model is iterative, with evaluation happening at the last step.
This was justified at times when data was primarily analogue and information a
scarce good. However, Big Data methodologies provide the means to inspect mas-
sive quantities of data in or near real time, to discover new insights through mining
yet undiscovered patterns and to visualize complexities in such ways that action-
able results can be immediately derived from Kim et al. (2014). The most problem-
atic aspect of the traditional policy cycle is that evaluation happens as a separate
and detached process at the end of the policy making process, which wastes time
otherwise available for re-focusing of initiatives or dropping unsuccessful measures
altogether. It also does not account for the possibility of a continuous inclusion of
evaluation and simulation results to re-assess policies based on evidence (Höchtl
et al. 2015) (Fig. 3).
52 J. Höchtl et al.
Fig. 3 Left: The policy cycle as described by Nachmias and Felbinger, 1982 (Nachmias and
Felbinger 1982); Right: The big data enabled ePolicy cycle including continuous evaluation
With view to the key concept introduced by Höchtl et al. (2015) of continuous
evaluation happening all along the policy cycle, the crucial question is by whom and
how evaluation is executed? The administration itself can, will and already does
employ big data technologies to better detect tax evasion, forecast disasters based
on past damage records, or to address climate change and its effect on the availabil-
ity of food and water (Mather and Robinson 2016). The tighter integration of yet
dispersed data sources is expected to make data based evidence available quicker
with the aim to act or foresee large-scale, systemic changes. In the future, algo-
rithms will play an important role in helping policy makers to rectify changes to
agreed policies and to instantaneously act on change.
Despite algorithmic approaches, the human ingenuity still excels in detecting pat-
terns in seemingly unrelated data sets. Moreover, citizens increasingly own and
operate distributed computing and sensing devices, be it the smartphone or dedicated
small scale computers like Arduinos or Raspberries. Therefore the inclusion of citi-
zens into the policy evaluation phase in an organized, structured way including sci-
entific means could draw on citizens’ skills, creativity and curiosity for supporting
the evaluation of government policy making.
While the inclusion of citizens into government policy making is not new, the
ability of citizens to engage in evaluation and monitoring actions in a scientific way
is fostered by the availability of big data tools, methodologies and means. However,
in the same way as participation will not happen simply by providing the tools and
means, incentives and supportive measures will be required to promote citizen par-
ticipation in science. Depending on intrinsic motivations, personal skills, and inter-
ests, a different set of techniques can be employed to encourage citizens to engage
The Citizen Scientist in the ePolicy Cycle 53
in policy evaluation, which may vary from levels of passive participation (lurking),
active participation, participation without taking explicit notice (implicit participa-
tion) up to coordinated citizen science leagues. Participation enhancing methods
such as gamification approaches could also create a breed of citizen scientists with-
out them actually taking notice. The ethical implications of this possibility have to
be considered.
In this section we deduct three use cases of citizen scienceship in policy making,
summarise some evidence or enabling elements and analyse the required setup for
the successful application of these elements between government and citizens.
Augmented Reality and Gamification Assuming a local authority is undecided
whether it should invest in renovating a school or building a new park. There are no
legal obligations to prioritize one measure over the other, and even experts are unde-
cided. In a virtual reality environment the government city planners sketch a model
of the actual city. People from all around the world subsequently connect to this open
playfield and start to model their ideal city. Their activities will become immediately
visible to all the other participants of this virtual city. Additionally, every virtual city
planner can inspect the planning efforts of the others and what infrastructure he or
she has built. After every planning period an election takes place to vote for the chief
city planner.
The city has access to the process data of this virtual environment, containing
information about which infrastructure was built, which was demolished and how
the virtual residents are using their city. They can also see who was elected as chief
city planner and replay and analyse the measures taken by her or him. By overlaying
the design elements of the virtual city with the actual city by means of augmented
reality, the virtual artefacts become immediately tangible.
Enabling elements Assuming that a lot of people enjoy engaging in virtual environ-
ments, augmented reality methods for city planning can be successful. One example
of a city building simulation in the past is SimCity, which was a huge success even
when computers where not yet connected to the Internet. Today peoples’ interest in
creating an alternate or ideal world has not waned. Minecraft14 is one example of a
game which can be played in a massive multiplayer online mode to design virtual
worlds. In Civic Crafting in Urban Planning, Mather et al. discuss the potential of
using Minecraft for public consultations and argue that serious games in planning
can capture participants’ attention for a longer period of time, educate the public
about planning concepts and site-specific challenges (Mather and Robinson 2016).
https://minecraft.net/en/, available on PC, handhelds and gaming consoles and found its way into
14
Analysis In this use case scenario, the city planning council takes the role of a
facilitator by creating a model of the existing city. Additionally it sets the rules to
keep people engaged in participating in the virtual planning process, for example
by promoting participants to become planning directors, etc. through other players
vote. The citizens need not necessarily know that they are taking part in a serious
game and that their actions might have an influence in the real world. By choosing
a gamification approach, the citizen scientist uses his devices and means to partici-
pate, yet the incentives of participation can be “hidden”. Instead of scheduling
assignments, it is the quest and challenge of the virtual environment which will
attract the participants. By using virtual reality elements, the rules of the game can
be kept within reasonable constraints, reducing the risk that the citizen scientists
create infrastructure which in reality would be inconceivable. The application of
augmented reality and gamification to support policy making could be used in the
Agenda Setting step, where citizens’ wishes in the virtual world can be used to
prioritize actions in reality.
Ubiquitous computing devices Most smartphone apps fulfil a very specific user
need and most users accept trading usability in exchange for granting access to her
or his phones sensors (e.g. location) and even more so to contact details. The com-
bination of increased tools usability in conjunction with communicating the goals of
the authority could provide another use case. State services would need to provide
increased usability levels compared to the offline version or the browser version,
e.g. by being seamless integrated into more backend systems without requiring the
service user to log into multiple sites to collect information just to enter this infor-
mation onto another site. Users might then accept the fact that these apps access the
phones sensors to deliver data to the authorities, which could support a number of
goals, e.g. to reduce traffic jams, or to support early warning systems (rise of tem-
perature in certain regions) in exchange for increased usability. Depending on
whether the goal is communicated, users could become citizen sensors knowingly
or unknowingly.
Enabling elements The University of Vienna engaged in a joint venture with Samsung
to utilize the capacity of smartphones during charging. Cancer and Alzheimer research
is computationally intense and involves scanning protein sequences for patterns. Only
after the phone is fully charged, a roughly one megabyte large data package will be
downloaded by the app Power Sleep,15 which comes as an alarm clock. The App then
inspects and analyses the data package and sends results back to the medical research
units.
Analysis The capability to effectively distribute work to many participating nodes
in such a way that only little effort is wasted in the coordination of work, combined
with algorithms which can efficiently operate on a mere subset of the data, is an
achievement of big data research. The citizens’ role in the above scenario is that of
an active facilitator – he or she will most likely deliberately participate out of altru-
15
http://www.iflscience.com/technology/new-app-crunches-scientific-data-while-you-sleep/
The Citizen Scientist in the ePolicy Cycle 55
istic motives. In this role the citizen scientist is unable to influence the details, like
the used algorithms, of the performed analysis, which remains under the control of
the institution or organisation who is issuing the data for inspection. This is also true
for the research results: While the citizen scientist contributes resources, the bene-
fits are harvested elsewhere. The usage of citizen resources by the government is
best employed in the Provision of Means policy cycle step.
Co-creation Complementary to the voluntary offering of resources by citizen scien-
tists via smartphones in exchange for usability is the idea of planning, designing, and
implementing citizens’ devices or even infrastructures to sense social and/or environ-
mental phenomena, to collect and aggregate the associated data, and to stream them
to a central repository or to provide access to the device/installation via an open
API. Such an actively developed networking infrastructure goes beyond the concept
of pure data collection and enable participants to actively develop and enhance the
underlying scientific ICT infrastructure, transforming the associated projects into
living environments. Additionally, the gathered data as well as the research results
remain und the control of the.
Enabling elements A prominent example for such an user-implemented sensor net-
work infrastructure can be found in form of the Citizen Weather Observer Program
(CWOP),16 in which private individuals host weather stations that are either using
amateur radio or Internet connectivity to transmit collected data. The available sen-
sors range from humidity and temperature sensors, up to sensors for wind speed,
barometric pressure and rainfall. While a lot of vendor-sold setups for weather
observation exist, a huge group of individuals works with small computerized
boards such as the Arduino platform or Raspberry Pies, which provide a high level
of extensibility and interconnectivity with other devices and electronic components.
Furthermore, the open platforms enable users to freely program their setups in vari-
ous computer languages. This opens up a plethora of possibilities with view to ana-
lytical processes or visualizations.
Analysis Extending the idea to use citizens computing resources, co-creation by citi-
zens requires more intense and ongoing participation levels. Here, a crowd or com-
munity of citizen scientists needs to organize themselves, define the objectives, agree
on the tools and infrastructure, schedule tasks and governance structures to accom-
plish a goal. In the most likely case, the government will profit from the results, but
seek to secure methodological rigor and soundness of science projects’ outcome. The
government can support such efforts by legally endowing the opening up of govern-
ment data and APIs, through specialized research grants also targeting individuals,
by providing cloud computing infrastructure which can be used by the citizens like
EU’s FIWARE platform,17 or by providing crucial software components as open
source like NASA’s open source building blocks.18 Big data tools like platform as a
16
http://wxqa.com/, retrieved 18.07.2016.
17
https://www.fiware.org/
18
https://code.nasa.gov/#/
56 J. Höchtl et al.
Citizen science in combination with big data and evidence-informed decision mak-
ing raises some issues that should to be addressed at the beginning of projects and
throughout the course of scientific investigation (Resnik et al. 2015). In this con-
text, ethical, legal, social and project-related challenges can arise,19 not only as
technology is always situated in a political context (Feenberg 2010), and critical
data studies, while in its infancy, have addressed such issues. It seems that all
around the world, policy-makers have taken on a hype, and big data is often referred
to as the “new oil of the digital age” (European Commission 2012), while at the
same time criticised as support of techno-capitalism (Rieder and Simon 2016).
Going even further, there is an increasing tendency among citizenry to ignore facts
obtained by investigative and data driven journalism. The Trump election campaign
or the Brexit were two examples of phenomenon which we might increasingly
observe: Neglecting factual proof, irrespective of the efforts and clarity which has
been laid on data gathering, model crafting and visualisation making. People
believe in what they want to believe.20 This raises questions of which areas in policy
making do make sense to include the citizenry in data driven policy making and to
what extend large scale policy making will always remain driven by sentiments
rather than by facts, independent of how tangible and easy to understand these facts
will ever be presented. This situation is likely to be aggravated by recent advances
in non-deterministic and self-improving algorithms like Artificial Intelligence with
feedback loops or stacking of algorithms in deep learning arrangements. While the
results obtainable by these algorithms or algorithmic arrangements are stunning
and are an important aspect to master the complexity of e.g. autonomous vehicles,
they are hardly suited for automated decision making, affecting citizens life.
Transparency involves many areas such as the availability of data and information
for once - the ability to explain citizens why a decision has been made will rise in
its importance. The jurisdictions of Germany and Austria have already reacted and
grant citizens the right to access the algorithms which have been used to support
decision making. This, however, requires the used algorithms to be accessible in a
way so their inner working can be explained to the ordinary citizen.21
19
Metcalf and Crawford identified several cases of an “ethics divide” in the big data context and
address disputes about human-subjects research ethics in data science.
20
Down on the Data: facts are not the only truth in life. Greg Jericho, The Guardian, 2016-09-19
(https://www.theguardian.com/commentisfree/2016/sep/19/down-on-the-data-facts-are-not-the-
only-truth-in-life, retrieved 2016-12-11).
21
Data Protection Act Austria (Datenschutzgesetz, DSG), BGBl. I Nr. 165/1999, §49 (3).
The Citizen Scientist in the ePolicy Cycle 57
While the general consensus is that data analysis can lead to important insights,
significant power shifts and advantages and disadvantages for individuals, groups or
communities, can arise. Some voices, like cultural critic Slavoj Žižek, have empha-
sised that humans would not benefit from it, and leaders would probably make deci-
sions not based on data evidence, but still on their own ideological fantasies, claiming
that big data analytics would be like “showing Hegel’s logic to a cow”.22
Rieder and Simon (2016) argue that while the consequences of big data have been
a concern, the underlying culture of measurement and quantification has not, and
discussions have focused on modalities of change rather than forms of continuity,
framed in a narrative of novelty and disruption. Culturally, this can be explained by
an effort to reduce uncertainty in societies. The authors address the recent interest in
evidence-based policy making and more data-driven forms of governance and relate
big data to a distinct political culture based on public distrust and uncertainty.
However, more data does not necessarily equal better insights (Rieder and Simon
2016). With the demand for quantitative rigor increasing in societies, a culture of
quantification risks reducing the human element, and why the reasons for this shift
can be explained as a strategy to adapt to new external pressures, it can also be inter-
preted as a chance to de-politicize legislation (Rieder and Simon 2016). A framework
for addressing ethical challenges in citizen science has been provided by Resnik
et al. (2015). They propose that for promotion of ethical research, scientists should
develop guidelines and provide laymen with education and training on the conduct of
research.
Conrad and Hilchey (2010) identified three main areas for challenges regarding
the concept of citizen science. While these challenges are situated within their work
in the field of community-based monitoring, the authors see them as generic issues
regarding the concept of citizen science in general. The first area relates to the
aspect of the number of people involved as well as how to trigger their interest to
participate. This also interrelates to whether or not there exists an established and
well-curated network for communication and exchange, which furthermore is also
impacted by the provided funding, not only for the citizen science project itself but
also for related environmental, organizational, and infrastructural aspects.
The second area covers challenges in terms of data collection and associated
processes. In order to fulfil many analytical tasks, it is imperative that data are avail-
able on a continuous time basis. If the collected data is heavily fragmented, analyses
over time become very difficult. Furthermore, there have to be processes defined
which provide the necessary means of a guaranteed level of accuracy regarding
measurements. Mistakes or measurement errors in the early phase of the project can
negatively impact all other succeeding steps. Furthermore, data collected by indi-
viduals are always prone to a certain personal bias, and in a more general way,
modern data analysis software is often not understandable for the average citizen.
The third area is the actual use of the data collected within the actual policy/
scientific context, i.e., the adaption by policy-makers in their decision-making pro-
cess or the publication in a suitable journal. Due to the before-mentioned quality
22
https://www.youtube.com/watch?v=PBBzYG8szmc (accessed 15th July 2016).
58 J. Höchtl et al.
aspects, results are often disregarded as invalid or processes not compatible with the
expected level of scientific rigor.
Future citizen science projects have therefore to adapt their processes and overall
strategy to overcome these challenges, therefore Newman et al. (2012) foresees
future directions of citizen science strongly be based on concepts such as viral mar-
keting, e.g., using social media, interconnected databases, and the initiation of
cyberinfrastructures as flexible and scalable backbones. The development of research
questions will be predominantly via bottom-up approaches, bringing together prac-
tices of amateur research and open science and open source (Dickel and Franzen
2016), supported by intuitive visualization for displaying and navigation data, avail-
able in real-time. High quality data will be available 24/7 via globally distributed,
high-availability databases. In addition to accessibility, the newly designed cyberin-
frastructures offer high-performance, cloud-based computing for everyone, foster-
ing joint collaborations between quantitative and qualitative science fields such as
natural and social sciences. The dissemination process will improve due to peer-
assessments via social community platforms across the globe. At the same time, this
will lead to overall community-accepted key performance indicators, which can be
adapted to projects of various scales. The newly formed (virtual) citizen science
communities will bridge existing geographical gaps, to enable better and faster
exchange and adoption of gained knowledge. The motivation behind participating in
these communities will be based on gamification-driven processes, which reward
individuals not only with new technological insights but also with reputation within
the community, e.g., expressed via achievement badges or ranks.
If citizen science wants to address these challenges, it will be necessary to ask the
question how big data relates to power, and how we want to shape the big data soci-
ety. It is important to note that unethical use of big data can be controlled, and
unequal power balances can be recalibrated (Ulbricht 2016). Ulbricht mentions
granting wider access to data and data analysis as one way to challenge the privi-
leged position of data collectors and controllers, and also to provide data subjects
with participation rights and comprehensible formation. Open data initiatives and
increasing public transparency about datasets will be crucial in this context. However,
every project should address questions of possible power shifts that might arise, and
which unintended consequences they could cause. On the basis of wider knowledge,
it will be possible for policy makers to choose the appropriate protection measure-
ments against such threats (Ulbricht 2016). In this context, more empirical studies
about the consequences of such projects in the governance field will be necessary in
order to be able to make good use of the new instruments.
References
Bakardjieva M (2009) Subactivism: lifeworld and politics in the age of the internet. Inf Soc
25(2):91–104
Bennett WL, Segerberg A (2012) The logic of connective action. Digital media and the personal-
ization of contentious politics. Commun Inf Soc 15(5):739–768. http://dx.doi.org/10.1080/13
69118X.2012.670661
The Citizen Scientist in the ePolicy Cycle 59
Johann Höchtl is senior researcher at the Centre for E-Governance at Danube University Krems
and holds a doctoral degree in Business Informatics. The projects he is involved in include
EU-funded research projects and national grants in the domain of large-scale data governance,
social media application in administration, open data and inclusive approaches of ICT application
in public administration. He is former member of OASIS SET TC standardisation group, member
The Citizen Scientist in the ePolicy Cycle 61
of OKFN Austria and member of Cooperation Open Government Data Austria, where he is head-
ing data quality sub working group. He was advisor to the E-Georgia strategy for the public admin-
istration and advisor to Macedonia (FYROM) to raise interoperability capabilities among federal
ministries.
Judith Schossböck is a research fellow at the Centre for E-Governance at Danube University
Krems, Austria. Among her research interests are online participation and activism, open govern-
ment and open access, digital literacy, the sociopolitical effects of ICTs and occasionally cyber
utopia or dystopia. She was involved in a study on the internet skills of the 14 years old youth in
Austria, the development of a youth participation platform on which four European countries
cooperated and a project researching electronic identification in online participation. In 2016, she
won a scholarship from the Research Grants Council of Hong Kong (HK PhD Fellowship Award)
Peter Parycek is full professor for eGovernance and head of the department for eGovernance and
Administration at the Danube University Krems. He holds a doctoral degree in legal sciences
(University Salzburg) and a master’s degree in telematics management (Danube University Krems).
As head of the department he is responsible for coordinating the eGovernance research groupe,
development and coordination of academic programmes and for the management of national and
international cooperations with public and private partners. He conducts research at the intersection
of legal science, technology and political science. From 2006 until 2011, he was chairman of the
Austrian working group e-democracy and e-participation of the Federal Chancellery. From 2010
until 2011, he contributed to the conceptualization of an E-Government-Act for the Principality of
Liechtenstein. He has founded the International Conference for e-Democracy and Open Government
(CeDEM) and is co-initiator of the open access journal JeDEM
Governance Failure in Light of Government
3.0: Foundations for Building Next Generation
eGovernment Maturity Models
Morten Meyerhoff Nielsen
Abstract Demographic, economic and other challenges is putting the public sector
and service deliver under increasing pressure. ICT as an enabler of increased effi-
ciency, effectiveness and transformation has long been recognized as part of the
solution. National experiences show that the potential of ICT has not been fully
realized, especially not in relation to Government 3.0 (Gov3.0). Existing public
administration, information systems management and eGovernment literature and
individual studies all point to the role of governance and cross-organisational coop-
eration in successfully introducing eServices and citizens actual use of them.
With a specific focus on eGovernment and eGovernance maturity and stage mod-
els, the literature attempt to unearth the underlying reasons why countries with simi-
lar infrastructures and eService availability experience very different levels of
online interaction with the public sector, and in particular whether existing stage
models address governance and cooperation.
Unfortunately, the review highlight a number of gaps including: Focus on out-
comes and actual use is missing; most lack a real understanding of core government
service concepts; decision-making should not be considered an eGovernment matu-
rity level; front-office service provision and back-office integration is mixed-up;
none addresses governance directly; most models are merely restructure or adjust
existing ones, and none address Gov3.0 as such.
Introduction
With demographic, economic and even climatic changes, the public sector and ser-
vice delivery will to face change in the coming years. In this regard the potential of
Information Communication Technology (ICT) as an enabler of public sector
M. Meyerhoff Nielsen (*)
Tallinn University of Technology, Ragnar Nurkse Department of Innovation and Governance,
Akadeemia tee 3, 12618 Tallinn, Estonia
e-mail: morten.nielsen@ttu.ee; https://www.ttu.ee/nurkse
Background
ICT use in public administration is in the literature seen in two ways: As a tool to
rationalize existing process or as an instrument to rethink the public sector, re-
engineer processes and organisations (Cordella and Bonina 2012).
ICT as a tool to increase public sector performance and efficiency is closely
associated the New Public Management (NPM) literature (Cordella and Bonina
2012; Cordella 2007; Demmke 2006). NPM brings the private sector corporate way
of thinking to public administration, thus shifting the focus from effectiveness to
efficiency through a new management culture and a focus on measurable results,
often cost savings (Pollitt and Bouckaert 2011; Cordella and Bonina 2012; Self
2000).
The expectations of ICT enabled NPM reforms has nonetheless be questioned
due to the complexity of organisational change and the political ramifications
(Cordella and Bonina 2012; Peters and Pierre 1998; Iribarren et al. 2008). Authors
like Bannister highlight the ability of ICT to transform the public sector, creating a
Joined-up Government (JUG) where inter-governmental collaboration and coordi-
nation is supported by technology (Bannister 2001; de Bri and Bannister 2010).
In contrast to NPM, JUG (also known as collaborative public management or
Gov 2.0), aim to reintegrate the public sector often fragmented by NPM reform
(Cordella and Bonina 2012; Huijboom et al. 2009b; Christensen and Lægreid 2007;
O’Leary et al. 2006). Lips’ definition of Public Administration 2.0 (Lips 2012) goes
as far as dropping the “e” in order to accommodate the complex and dynamic none-
technical and contextual aspects of public administration reform.
What classical public administration literature seem to lack, is the merger NPM
and JUG, i.e. the role governance in the introduction of ICT in public administra-
tions in combination with measuring maturity levels and ICT take-up.
IS Management
r esearchers have explored less rigid alternatives. These in turn are modelled on the
operational realities of public sector organisations including vertical and horizontal
integration, centralized, federal, decentralized organizational forms of
government.
In contrast, IT governance contingency analysis unanimously agree that no uni-
versal best practice IT governance structure exist (Brown and Magill 1994; Brown
and Grant 2005). Research therefore explore the basic structural options available,
and attempt to unearth the logical and best options for different types of organisa-
tions. Similarly research focus on the contingencies which influence the adoption of
a particular IT governance model, the role of actors, organisational maturity, size,
structure, time frames, psychological climate, extra-organisational situations,
resources, rank and location of responsible executives and steering committees, risk
adversity, degree of centralisation etc. (Brown and Grant 2005).
For over 40 years, a recurring subtopic in this literature has been staged maturity
models: models that morphed into capability maturity models (CMM) for assessing
software development processes in the 1980s and, since 2002, the integration of
product and service development, management, and acquisition (Poeppelbuss et al.
2011; Röglinger et al. 2012). While IT governance models, such as the US Federal
Enterprise Architecture (Peters and Pierre 1998) and Chilean CMMI-inspired eGov-
ernment maturity model and toolkit (Iribarren et al. 2008), address political and
legal dimensions, most focus on business processes in single organisations, not the
cross-organisational, national, or international ones of PA and eGovernment
(Pöppelbuß and Röglinger 2011).What the IT governance literature lack, is the
political and legal dimensions found in the public administration and eGovernment
literature.
eGovernment and eGovernance
(2001), West (2004), Moon (2002), Heeks (2015), Andersen and Henriksen (2006),
Traunmüller and Wimmer (2003), Klievink and Janssen (2009) etc., have all argued
in favour of the usefulness of stage models to guide policymakers and to stimulate
the developments of capabilities needed by organisations to migrate from one stage
to another – albeit from different perspectives.
A gap in the stage models and eGovernment literature is a clear link between the
role governance and cooperation play in the successful implementation and subse-
quent use of ICT and eServices solutions. Similarly, most models merely focus on
supply and technology, and less on outcomes or results.
holistic approach, while Brown (Brown and Magill 1994) recommend an merger of
the classical IT governance streams of thinking.
To illustrate the importance of governance models and outcomes is the discrep-
ancy between Denmark and Japan online address changes (via the Internet). In
Demark close to 80% of address changes are made online, while this is a scant
0.0002% in Japan (Meyerhoff Nielsen and Igari 2012; Igari 2014). Statistical analy-
sis also fails to shed light on the underlying reasons why Danes use the Internet to
interact with public administration (85%) more often than their Dutch and Swedish
counterparts (79% and 78%, respectively) – although similar numbers of house-
holds in Denmark, the Netherlands, and Sweden pay for having access to the
Internet (all in the 90+ percentile), and why their citizens have similar patterns of
Internet use (also in the 90+ percentile) and private sector services such as online
banking (all, 82%) (EC 2014; Meyerhoff Nielsen 2014, 2016b; Eurostat 2016).
In light of these challenges, the OECD on 15 July 2014 adopted a number of
recommendations for public sector digitisation and eGovernment strategies (OECD
2014). The recommendations address the strategic direction of eGovernment,
implementation, governance, and cooperation models. The OECD’s recommenda-
tions are anchored in the realisation that, in order to successfully introduce ICT
infrastructure and online services for improved public-sector efficiency and effec-
tiveness, more than just a technological and supply-oriented approach is required
(OECD 2014; O’Leary et al. 2006).
These practical examples hint also at potential limitations in current research.
This chapter will therefore review the existing literature in an attempt to unearth the
underlying reasons why countries with similar infrastructures and eService avail-
ability experience very different levels of online interaction with the public sector,
and in particular whether existing stage models address governance and cooperation
(sometimes known as maturity models).
Based on the initial exploration of current literature (above), an appropriate theo-
retical framework to assess and map the degree to which governance and coopera-
tion models ensure the successful supply and use of online eServices, is found in the
eGovernment stage model literature, and therefore be the focus of this chapters lit-
erature review.
Methodology
To address the two potential gaps identified in current eGovernment and governance
research (in section “Background”), this chapter sets out two questions:
1. Does the literature address the degree to which, and in what way, governance and
cooperation models ensure success supply and use (i.e. demand) of online citi-
zen services?
70 M. Meyerhoff Nielsen
2. Does the literature identify the success factors and are they mapped and devel-
oped into a universal governance model for successful digitisation of public sec-
tor service delivery (i.e. supply) and eService take-up (i.e. demand) by citizens?
To address the two research questions, a literature review is carried out. The
focus of this review included the identification of existing models and their key dif-
ferences (i.e. can the identified models and theories be mapped). What does current
academic and practitioner debate focusing on? What is the current state-of-affairs?
What are the clusters of theory, models and critique? What is the real life applicabil-
ity of the theories and models?
The literature review follow a classical pattern for systematic information retrieval as
outlined by e.g. Roberts (1977) and the Walsh and Downe (2005) qualitative meta-
synthesis procedure. The seven-step Walsh and Downe model is adapted to include
‘berrypicking’ (Bates 1989). The adapted methodology consists of the following six
steps: Frame the exercise; Locate relevant studies; Decide what to include and a degree
of ‘berrypicking’; Appraise studies; Compare and contrast, and finally; Conclude.
Primary and secondary key word searches were used. Primary key words were:
eGovernment and stage, or model, or level, or tier, or development. Secondary key
works included: eGovernment and/or maturity, governance, cooperation models,
technology maturity, transformation, benchmarks, indexes. Other secondary key
words were: Use, take-up, benefits, impact, output, efficiency, effectiveness, return
of investment, eGovernment Readiness Index, eGovernment Benchmark.
To ensure that relevant literature and arguments were identified, Web of Science
managed by Thomson Reuters, Scopus managed by Elsevier and EGRL -
E-Government Reference Library (version 11.5) managed by the University of
Washington, Information School online libraries were selected based on their rele-
vance, scope and size to the literature review. Each of the reference libraries were
searched and cross-referenced to ensure as complete and up-to-date picture of the
academic discourse and the state-of-affairs as possible.
To ensure the quality of the literature review, the reference libraries was compli-
mented with online research for number of secondary sources including key topic
journals i.e.: GIQ – Government Information Quarterly by Elsevier, MIS Quarterly –
Management Information Systems Quarterly MIS Quarterly by the Management
Information Systems Research Center at the Carlson School of Management,
University of Minnesota, and Information Polity by published by IOS Press.
Governance Failure in Light of Government 3.0… 71
To frame and define the parameters of the literature review, a publication had to be
published in English, in the proceedings of an academic conference or in an aca-
demic journal (preferably GIQ, MISQ or Information Polity) or a recognized inter-
national body (mainly UN, EU or OECD), been subject to peer review (exception
possible if published by the UN, EU or OECD), a minimum seven pages (or approx.
3700 words) in length including references, after 1 January 1995.
Where appropriate a second stage of screening, or ‘berrypicking’ as outlined by
Bates (EC 2014), is applied. The robustness of the theoretic models identified, sec-
ondary sources and key words is of particular relevance in this regard.
Appraise Studies
developed in the field of IS management and computer science. This section there-
fore outlines the characteristics and focus of number of key eGovernment stage and
maturity models.
Layne and Lee refer to the experiences of eGovernment as chaotic and unmanage-
able, arguing for the division of development into distinguishable stages (Layne and
Lee 2001). To this effect eGovernment research has largely focused on stage, or
maturity, models.
Multiple stage models has been suggested by researchers, consultants, national
authorities and international organisations. In this context academics differentiate
between three types of stage-models (Fath-Allah et al. 2014; Persson and Goldkuhl
2005):
• Governmental models: Models developed by governments, consultants and aca-
demics to help authorities identify and improve their level of maturity (generally
using predefined models and toolkits).
• Holistic approach models: Models designed to assist authorities (generally pre-
defined models, toolkits and indicators) in project implementation and to deter-
mine if the project will be successful or not.
• Evolutionary eGovernment maturity models: Models which focus on sequential
evolutionary steps, for instance from immature to mature eGovernment with
improved quality (often from an academic perspective).
The primary focus of this review is on governmental and evolutionary stage
models, since the holistic maturity model approach focuses on project implementa-
tion and organisational capabilities, and particularly relevant in relation to IS man-
agement and CMM literature (Ross et al. 2006; Poeppelbuss et al. 2011; Persson
and Goldkuhl 2005).
Using the methodology outlined in section “Methodology”, 42 stage models are
identified. The following subsections clusters the various models based on their
respective characteristics.
ANAO – Australian National Audit Office’s (ANAO 1999) 1999, four-stage
maturity model was introduced to categorize and evaluate process to guide agencies
in their decision as to what services could and should provide. The model is national
in character and takes an abductive-deductive approach to eGovernment maturity.
The model is developed based on experiences in Australia. The levels of maturity
are: Publishing and information; Interaction; Transaction of secure information
(incl. login), and; Sharing information with other agencies (incl. business and
citizens).
Gartner Group (Baum and Di Maio 2000) published a four state model in 2000.
It is one of the earliest eGovernment maturity models not emerging out of a national
context. The Gartner model focus is on supply and technology with a degree of
Governance Failure in Light of Government 3.0… 73
Fig. 4 Layne and Lee model four-stage maturity model (Layne and Lee 2001)
oped based on observations in the USA and earlier models. The four-stages of matu-
rity are defined as (see Fig. 4): Catalogue i.e. online presence (i.e. websites with
static information and downloadable forms); Transactional incl. service and forms
(i.e. eServices); Vertical integration, that is local system integration, and; Horizontal
integration i.e. integration across function (i.e. life-events and personalisation).
United Nation’s (UNDESA 2014, 2008, 2010, 2012; Ronaghan 2002) is best
known for its biannual UN eGovernment Readiness Index. The model has been in
76 M. Meyerhoff Nielsen
use since 2001 when the first Index was first published. It covers pre-conditions
such as supply, technology and integration. The original focused has been on the
five-stages of maturity. The UN publishes the bi-annual eGovernment Readiness
Index, but has in the last few years refocused the models to include additional
aspects of engagement and transparency (e.g. the UN eParticipation Index). The
model is “international” in character and takes an abductive-deductive approach to
eGovernment maturity. The model consists of a biannual ranking of 193 countries.
The model has a pre-condition stage, which focus on at network preparedness,
access to PCs, the Internet and literacy and digital competences (i.e. TII Index). The
maturity levels are (see Fig. 5): Emerging presence such as basic websites with
static information; Enhanced presence e.g. emerging portals (i.e. a degree of vertical
and horizontal integration), interactivity, and customer services (i.e. eServices);
Interactive such as two-way interactivity (i.e. eServices and communication),
searchable intranet; Transactional i.e. eServices, and; Seamless incl. sully net-
worked government (i.e. horizontal and vertical integration).
Wescott’s (2001) 2001 model consist of six stages. It is a scientific model, with
an abductive approach to eGovenment maturity. It has been developed based on
Fig. 5 (a) The original four-stage UN model, 2002 (Ronaghan 2002). (b) The updated version of
the UN model, 2012 (UNDESA 2012)
Governance Failure in Light of Government 3.0… 77
observations in the Asia-Pacific. The maturity levels are: Setting up an email system
and internal network e.g. feature e-mail systems to improve information sharing,
coordination and feedback; Enabling inter-organisational and public access to infor-
mation e.g. information is department centric, shared between organisations and can
be accessed by the public over the Internet; Two-way communication such as basic
eServices and citizens can make suggestions using emails or ask questions in forums
and receive answers. Exchange of value e.g. eService features applications such as
tax assessments and license renewals. At this stage, the citizen can make secure pay-
ments online; Digital democracy incl. focus is on empowering the civil society (e.g.
increasing awareness of government corruption) and allowing citizens to vote and
express their opinions and feedback, and; Joined-up government incl. vertical and
horizontal integration allowing for citizens to execute services without knowing
which government agency is responsible for.
Chandler and Emanuel (2002) in 2002 proposed a four-stage model. It is a sci-
entific model, with an indicative approach to eGovernment maturity. The maturity
levels are: Information i.e. online information about government services and poli-
cies; Interaction such as basic level of interaction between governments and citizens
such as email systems; Transaction i.e. features eServices, and; Integration e.g. fea-
tures integrated services across various departments and agencies.
European Union (2012) has since 2002 used a eGovernment benchmark model
similar to the UN. The focus is on supply, technology and integration and initially
included five-stages of maturity. The European Commission publishes its eGovern-
ment benchmark yearly, but has since 2010 started including additional biannual
focus areas, and has moved from benchmarking services to life-events, user engage-
ment, access to personalized data and user-rating – through mystery shoppers and
surveys. The model is “international” in character and takes an abductive-deductive
approach to eGovernment maturity. The model is developed with inspiration from
the SAFAD model (Statskontoret 2000) and experiences in the EU+ member states.
The model forms the basis of the EU’s annual eGovernment Benchmarks and
Surveys. A pre-condition stage looking at PC and Internet accessibility as well as
digital literacy compliments its five stages (see Fig. 6): Emerging presence i.e. basic
websites with static information; Enhanced presence e.g. emerging portals (i.e. a
degree of vertical and horizontal integration), interactivity, and customer services
(i.e. eServices); Interactive, that is two-way interactivity (i.e. eServices and com-
munication), searchable intranet; Transactional i.e. eServices, and; Seamless such
as fully networked government (i.e. horizontal and vertical integration).
Hodgkingson (2002), in 2002, present a two phased, five-stage model, focusing
learning cycles and an s-shaped curve for learning (see Fig. 7). The model focus the
rate of technology diffusion in government, service impact and technical aspects
such as interoperability before data exchange and vertical and horizontal integration
is possible. It is inspired by diffusion of innovation (DOI) and innovation diffusion
theory (IDT), technology acceptance (TAM) and IS management models. The
stages are: Government online i.e. initiation of idea generation, analysis and pilot
implementation and contagion such as wider adoption of technology and benefits of
ICT, business needs, decentralization of strategy and resources; eGovernment i.e.
78 M. Meyerhoff Nielsen
Fig. 6 (a) The original four-stage EU model, 2002. (b) The updated five-stage EU model, 2009
(EC 2012)
control (i.e. re-focus on cost, efficiency and quality, re-centalisation of some strate-
gies and control), interoperability, and data management.
Moon’s (2002) 2002 five-stage model by M.J. Moon focus on supply, technology
and organisational integration and some aspects of participation in a democratic
sense. It is very similar to the 2001 model proposed by Hiller and Belanger (2001).
It is a scientific model, with an abductive approach to eGovernment maturity. It was
developed based on observations and data from 2000 US municipality websites.
The maturity levels are: Web presence i.e. technological leap-frogging, websites
with static information); Interaction such as simple interaction, e-mail and down-
loadable forms; Transaction i.e. eServices; Transformation/integration such as back
office automation and digitization of processes with aspects of vertical and horizon-
tal integration, and; Participation for transparency and release of data.
Governance Failure in Light of Government 3.0… 79
Fig. 7 Hodgkingson’s five-stage maturity model and learning curve (Hodgkinson 2002)
Accenture (Rohleder and Jupp 2003) in 2003 published a five-stage model. The
model is developed by a consultancy and takes a deductive approach to eGovern-
ment maturity. The model has been applied to Australia, Belgium, Brazil, Canada,
Denmark, Finland, France, Germany, Hong Kong, Ireland, Italy, Japan, Malaysia,
Mexico, the Netherlands, Norway, Portugal, Singapore, South Africa, Spain, the
UK and USA. Based on the model Accenture publish an annual eGovernment
Ranking of selected countries. The model consisting of the following maturity lev-
els: Online presence with information published online; Basic capability i.e. secu-
rity and certification is developed and the online presence is broadened; Service
availability with eServices increasingly available on portal(s) and features of cross
agency cooperation and services increasingly designed to meet customer needs;
Mature delivery with eServices clustered with clear ownership and authority – CIO
(Chief Information Officer) or central agency the involvement of customer in the
process of eGovernment and the services are marketed; Service transformation i.e.
improved customer service delivery is the objective and multi-channel integration is
common.
Koh and Prybutok (2003) in 2003 presented a three-element model (see Fig. 9).
The model is scientific and takes a inductive approach to eGovernment maturity.
The model focus on internal and external factors and three stakeholder groups i.e.
employees in public authorities, suppliers (i.e. IT vendors or IT departments) and
customers (i.e. citizens and businesses). Visualised as circles of there are overlaps
between the three elements thus providing a degree of granularity with a degree of
inspiration from the IS management and computer science literature. The elements
are: Informational i.e. online information; Transactional i.e. online transactions,
and; Operational i.e. operational, vertical and horizontal integration.
Reddick’s (2004) 2004, two-stage model, is one of the most simple maturity
models identified. It is a scientific model, with an abductive approach to eGovern-
ment maturity. The model is developed based on observations in the USA, The
Governance Failure in Light of Government 3.0… 81
Fig. 9 Koh and Prybotok’s three-stage and users of internet maturity model (Koh and Prybutok
2003)
maturity levels are: Cataloguing online information about the government and its
activities, and; Transactions incl. eServices and one-stop-shops.
Waseda (Obi 2014, 2012, 2015) first published the Waseda model and its annual
benchmarks for selected countries in 2004. The model is “international” in charac-
ter and takes an abductive-deductive approach to eGovernment maturity. The model
is used for an annual eGovernment ranking list for an increasing number of coun-
tries. The model differs somewhat from other stage models as it does not define
distinct levels of maturity. It covers managerial and organisational aspects also seen
in CMM/CMMI models and the IS management literature. The focus is on qualita-
tive and quantitative indicators including: Network preparedness and infrastructure;
Management optimization and efficiency; Online presence of information, services,
national portals and websites; Governance incl. cooperation and promotion; ePar-
ticipation and digital inclusion; Open government, and; Cyber security. The indica-
tors can be grouped into four-stages, that is: Networked preparedness and
infrastructure; Online services; Management optimization, and; eParticipation.
West (2004) first published the four-stages model in 2004. It is a scientific model,
with an abductive-deductive approach to eGovernment maturity. The model is
developed based on observations and data from 1813 and 1680 US municipality
websites model in 2000 and 2001. The West Index on US municipalities and a num-
ber of countries is based on the model. The maturity levels are: Bill-board i.e. web-
sites as billboards mainly used for posting information; Partial-service-delivery
with the ability to search for data via search engines with some eServices available;
Portals containing all information and eServices (i.e. a one-stop-shop), and;
Interactive democracy incl. ortals offers personal and proactive online service, uti-
lise push technology and feedback forms.
Windley’s (2002) 2002, four-stage model. It is a scientific model, which takes a
deductive approach to eGovernment maturity. It is developed based on observations
82 M. Meyerhoff Nielsen
from the US Utah.gov site and consists of the following maturity levels: Simple
website with static pages with downloadable forms; Online government featuring
interaction mechanisms such as e-mails, online forms, help and FAQs; Integrated
government incl. end-to-end eService transactions, but also emerging internal inte-
gration as information is shared between departments, and; Transformed govern-
ment: Customer centric eServices organized according to user needs and segmented
according to population groups and life events. Vertical and horizontal integration is
also a feature.
Davison et al. (2005) in 2005 presents a four-element model combining the
insights of the strategy and maturity alignment models form the IS management and
computer science literature (see Fig. 10). The model focus on internal and external
factors in both the government (i.e. the public sector in general) and eGovernment
domain (i.e. ICT within the public sector). Rather than looking at the supply-side
issues related to digitization of service delivery and ICT enabled reform Davison
et al. focus on the key elements enabling the successful use of ICT. The models
cover eGovernment services (i.e. information and transaction), processes (i.e. verti-
cal and horizontal integration) and transformation within the four elements of:
Government strategy with choices pertaining to positioning of government and
business strategies; Government infrastructure and processes incl. choices pertain-
ing to internal arrangements and configurations supporting authorities chosen
position including public sector culture; eGovernment strategy incl. choices per-
taining to IT scope, systemic capabilities and IT governance, and; eGovernment
infrastructure and processes e.g. internal arrangements and configurations deter-
mining data, applications and technology infrastructure used to deliver eGovern-
ment services.
Siau and Long’s (2005) 2005 five-stage maturity models focus on supply, tech-
nology and organisational integration and some aspects of participation in a demo-
cratic sense. It is a scientific model, with an inductive approach to eGovernment
maturity. It differs from the Moon (2002), Hiller and Belanger (2001) models by
including engagement and political decision making to the fifth stage in the form of
Fig. 10 Davison’s et al. four stage strategy and maturity model (Davison et al. 2005)
Governance Failure in Light of Government 3.0… 83
a “eDemocracy stage”. The maturity levels are: Web presence incl. technological
leap-frogging, websites with static information); Interaction e.g. simple interaction,
e-mail and downloadable forms; Transaction i.e. eServices; Transformation/integra-
tion such as back office automation and digitization of processes and aspects of
vertical and horizontal integration, and; eDemocracy incl. engagement, political
decision making, transparency, release of data.
Persson and Goldkuhl (2005) in 2005 evaluates a number of existing models and
propose a two-stage model from a computer science perspective. The maturity lev-
els are: Integration of services with a focus on public services, directed services,
concentrated services and portals, and; Integration in services incl. elements such as
the integration of services and agencies, transparency in processes of independent
processes, database access in information gathering, information or decision provi-
sion requirements and joint information services.
Andersen and Henriksen’s (2006) 2006 Public Sector Process Rebuilding Model
(PPR) builds on Layne and Lee four-stage maturity model (Layne and Lee 2001). It
is a scientific model, with an abductive-deductive approach to eGovernment matu-
rity. It is developed based on observations and data from 110 central government
sites in Denmark. The PPR model focus on supply, organizational integration, pro-
cesses and differs from other models by emphasising user-centricity rather than
technological aspects. Four-stages of maturity, of which the first two stages com-
prise the four-stages proposed in the Layne and Lee model (see Fig. 11): Cultivation
e.g. websites with static information, downloadable forms, vertical and horizontal
integration; Extension such as eServices, basic personalization and life-events and
a focus on data ownership; Maturity of eServices, none-Internet interphases,
General, widely applied
Phase I: cultivation
Exeption,
Fig. 11 Andersen and Henriksen PPR model (Andersen and Henriksen 2006)
84 M. Meyerhoff Nielsen
Document
Vitality Review
Compliance Communica-
tion
Fig. 12 (a) eGOV-MM’s three dimension and interrelated elements (NASCIO, N.A.o.S.C.I.O.
2006). (b) eGOV-MM’s domain level and key domain areas (NASCIO, N.A.o.S.C.I.O. 2006)
online taxes and eProcurement, and; Transformation citizen centric, i.e. consoli-
dated and shared administrative services at this stage are across various government
jurisdictions.
Almazan and Gil-Garcia’s (Almazan and Gil-Garcia 2008; Luna et al. 2013) six-
stage model published in 2008 (presence, information, interaction, transaction, inte-
gration and participation). It was updated in 2013 by merging the initial two stages
information based stages (i.e. presence and information) and adjusting the remain-
ing four stages. The 2013 model consists of five-stages and 172 indicators, which
aim to highlight the performance and efficiency of portals by including supply and
actual use in relation to the online population – thus, indirectly including pre-
conditions (Luna et al. 2013). It is a scientific model, with an abductive approach to
eGovernment maturity. It has been developed based on observations and data from
32 Mexican state portals and includes ranking (in 2013). The 2013 levels of matu-
rity are: Information Online information, static or updated; Interaction e.g. down-
loadable forms, communicate with the government via e-mail and forums;
Transaction such as secure eService transactions and payment options via portals;
Integration incl. one-stop-shops/portals, vertical and horizontal integration, and;
Political participation offering users voting and participation in opinion polls, sur-
veys and public forums.
Chan et al. (2008) in 2008 proposed a model focusing on supply, technology and
organizational integration. It adds none-governmental stakeholders to the mix. It is
a scientific model, with an abductive approach to eGovernment maturity. The model
is developed based on observations and data from regional government in China.
The five-stages of maturity are: Publish websites with static information; Interact
i.e. downloadable forms; Transact though eServices; Integrate though vertical and
horizontal integration of service providing agencies, and; Tri-party integration i.e.
integration of public, private and stakeholder organisations.
Iribarren et al. (2008) proposed an IT focused eGovernment Maturity Model
(eGov-MM) based on four domain levels, in 2008. It is a multi-dimensional model
and assessment tool in the form of a capability maturity framework to ensure con-
tinued measurement and control. It is a national model developed for the Chilean
government and borrows from experiences in the UK, US, Australia, Canada,
Sweden, South Korea and others. It distinguish between maturity and capabilities
and is inspired by the IS management’s US CMMI and EA models (NASCIO,
N.A.o.S.C.I.O. 2006), ISO/IEC 15504 in Europe and supports Wimmers holistic
view (Traunmüller and Wimmer 2003; Iribarren et al. 2008). The domain levels on
effectiveness, efficiency, confidentiality, integrity, availability, compliance, man-
ageability on one axis and IT resources like applications, data, infrastructure and
facilities on the other (see Fig. 13). The four domain levels are: eStrategy; IT gover-
nance; Process management; People and organisation capabilities.
Shahkooh et al. (2008) in 2008 proposed a five-stage model. It is a scientific
model, which takes an abductive approach to eGovernment maturity, proposing the
following maturity levels: Online presence i.e. online information; Interaction with
citizens interacting with governments through e-mail to officials and downloading
forms; Transaction though secure eService transactions like payments and tax fill-
86 M. Meyerhoff Nielsen
Fig. 13 (a) eGOV-MM’s three dimension and interrelated elements (Iribarren et al. 2008). (b)
eGOV-MM’s domain level and key domain areas (Iribarren et al. 2008)
economies around the world. Rather than mere levels of maturity it propose four-
stages of complexity depending on three binary variables: (1) whether digitisation
adds to internal work and structures of government without affecting them; (2)
transforms internal processes and structures; whether the transformation is internal
with, or without affecting the end-users; (3) whether the transformation is depend-
ing on a particular application context. The four levels of maturity are (see Fig. 16):
Digitisation or technology in government (i.e. precence); Transformation being
eGovernment (i.e. transaction and transformation); Engagement or eGovernance
(i.e. eParticipation/eDemocracy), and; Contextualisation i.e. policy-driven
eGovernance.
Heeks’ (2015) Manchester eGovernment Maturity Model from 2015, adapt the
Layne and Lee (2001) be less linear in it process, differentiate between the front-
and back-office and less “US-centric”. It is a scientific model, which takes an induc-
tive approach to eGovernment maturity. The model is developed based on
observations in developed and emerging economies around the world. The result is
two parallel dimensions consisting of three and four elements respectively, thus
forming a matrix (see Fig. 17). The stages are: Sophistication of digitised interaction
(i.e. front-office) incl. informed/one-way interaction, interact/two-way interaction
and transaction/complete service; Extent of process change (i.e. back-office) incl.
digitisation (simple automation), improvement (process integration), redesign (e.g.
proactive transaction) and transformation (fundamental change e.g. process
elimination).
organisations like the EU or UN, consultancy firms like Deloitte and CISCO or
academia. Table 1 present the number of models identified for each of these four
categories.
The first model was published by the Australian National Auditing Office in
1999 (ANAO 1999). The latest models are published by Heeks and Janowski in
2015 (Heeks 2015; Janowski 2015). Using the year of publication, the timeline (see
Fig. 18) highlight a number of developments.
The first models to emerge are from national authorities, international organiza-
tions and consultancies. National models from Australian ANAO (ANAO 1999) to
the UK equivalent were published in 1999–2001. International organisations fol-
lowed with the UN (UNDESA 2014) in 2001 and the EU (EC 2012) and World Bank
in 2002 (Toasaki 2003; Alhomod and Shafi 2012). The Deloitte (Deloitte and Touche
2001) through to the Accenture model (Rohleder and Jupp 2003) were published in
2000–2003. The first scientific models were published by (in alphabetical order) in
2001 by Hiller and Belanger (2001), Howard (2001), Layne and Lee (2001) and
Silcock (2001), followed by Wescott (2001), Chandler and Emanuel (2002), Moon
(2002) and Netchaeva (2002) in 2002. The most recent models includes Dias and
Gomas (2014) in 2014 and Janowski (2015) and Heeks (2015) in 2015.
Governance Failure in Light of Government 3.0… 91
2015
2014
2013
2012
2011
2010
2009
2008
2007
2006
2005
2004
2003
2002
2001
2000
1999
0 1 2 3 4 5 6 7
Academic National International Consultancy
What is also clear from the literature review is that stage models were of particular
interest in 2000–2004 when 23 of the 42 identified models were published (i.e.
54.8%) – including all models originating in international organisations and
consultancies.
As presented in Table 2, 22 (i.e. 52.4%) of the identified models are based on
practical experiences and case studies with 15 (i.e. 35.7%) being based largely on
observations of ICT use in a single country and at a single level of government
(e.g. local, regional, or central) public administration. Seven (i.e. 16.7%) models
are based on the experiences in multiple countries, i.e. Accenture, UN, EU,
Iribarren et al., Janokowski, Wescott and West. Three models (i.e. 7.1%), Windley,
Chan et al. and Almazan et al., are based on regional observations in a single
92 M. Meyerhoff Nielsen
country, whereas Moon, Reddick and Dias & Gomes (i.e. 7.1%) are based on case
studies in municipalities.
The most cited model is hard to asses as the original source of national, interna-
tional and consultant models are often not citied or referenced appropriately in the
literature, is neither publically available, not included in scientific databases, nor
available on sites such as research gate and Google scholar. Using Google scholar
(accessed on 15 April 2016) the most frequently cited models are all scientific:
Layne and Lee’s 2001 model (Layne and Lee 2001) with 2031 citations, Moon’s
2002 model (Moon 2002) with 1550 citations, Hiller and Belanger’s 2001 model
(Hiller and Belanger 2001) citied 952, and Andersen and Henriksen’s 2006 model
Governance Failure in Light of Government 3.0… 93
(Andersen and Henriksen 2006) model being cited 453 times. The most cited model
not published by academics are Gartner’s (Baum and Di Maio 2000) 2000 model
with 302 citations.
Fig. 19 (a) Identified stage models mapped in accordance with their different maturity levels. (b)
Identified stage models mapped in accordance with their different maturity levels
Governance Failure in Light of Government 3.0… 95
Fig. 19 (continued)
96 M. Meyerhoff Nielsen
The vast majority of the 42 models use different semantics and metaphors, simi-
larly many models and individual stages overlap (Meyerhoff Nielsen 2016a; b;
Alhomod and Shafi 2012; Lee 2010). This means that some general categories exist.
In fact, commonalities between national, international and consultancy models, are
also shared with many of the scientific models, and is visualized in Fig. 19 (Lee
2010).
As the various models are based on different perspectives and use different defini-
tion and metaphors, they can be difficult to understand and summarise. To alleviate
this difficulty, the 11 meta stages presented in Fig. 19 are distilled further using
Lee’s qualitative meta-synthesis framework (Lee 2010).
Using a detailed qualitative meta-synthesis procedure Lee use 12 stage-models
to develop a new semantic framework consisting of five general metaphors namely:
Presenting, Assimilating, Reforming, Morphing, eGovernance. The Lee’s five met-
aphors are defined and described in Table 3 below.
Table 3 Metaphors: their definitions, related stages, and themes (Lee 2010)
Metaphors Description Stages/concepts
Citizens and Operation and
services technology
Presenting Presenting information in the Information
information space
Assimilation Assimilates (or replicates) Interaction Integration
processes and service in the
information space with the ones
in the real world
Reforming Reform the processes and Transaction Streamlining
services in the real world to
match the information space
requirements, fitting for
efficiency
Morphing Change the shape and scope of Participation Transforming
processes and services in the
information space as well as the
ones in the real world, fitting for
effectiveness
eGovernance Processes and services in both Involvement Process
worlds are synchronously management
managed, reflecting citizen-
involved changes with
reconfigurable processes and
services
Governance Failure in Light of Government 3.0… 97
From the analysis of the 42 models, identified in the literature review, it becomes
clear that the 11 overarching stages identified represents six specific meta stage char-
acteristics. With respect to Lee’s framework, an initial ‘pre-condition’ stage is miss-
ing. A pre-condition stage is therefore added to Lee’s framework for the purpose of
this article (bringing the number of stages to six) (Meyerhoff Nielsen 2016a).
The 42 models and their respective overarching stages are, in Fig. 20, mapped in
accordance with the six meta characteristic described in Table 3. The models are
presented chronological and in alphabetical order within said year.
The stage and meta characteristic mapping in Fig. 20 highlight a number of interest-
ing aspects. Table 4 below summaries the number of models, which address each of
the six meta stages. The main differences in the models unearth relates to ICT
enabled morphing (i.e. transformation) of public administrations and eParticipation
and eDemocracy (i.e. user engagement and decision-making).
Preconditions
Models, like the UN (UNDESA 2014), EU (EC 2012), Waseda (Obi 2015) and
Iribarren et al. (Iribarren et al. 2008) which include preconditions generally focus
on the availability of key enabling factors such as digital literacy, Internet availabil-
ity and use, electronic identifiers (eID), availability of a basket of electronic ser-
vices, accessing public sector information, downloadable forms and transactional
eServices in aggregated terms. The aim is to enrich analysis and monitor the avail-
ability of key enablers. Unfortunately, none of the models addresses the actual use
of key enablers like eID’s.
While Lee’s framework include management and governance issues in the final
maturity level (Lee 2010), it may be argued – in line with the IT governance
literature (Brown and Magill 1994; Brown and Grant 2005) (see section
“Background”), recommendations by the OECD (OECD 2014) and authors like
Iribarren, NASCIO and Janowski (Iribarren et al. 2008; NASCIO, N.A.o.S.C.I.O.
2006; Janowski 2015) – that governance structures and cross-governmental coop-
eration are preconditions for successful ICT implementation and take-up. For
instance, is the eGovernment Strategy legally binding for one or all levels of govern-
ment, what mechanisms govern decision-making, legal changes and coordination
processes, benefit realization etc. While most would agree on the objective of IT
governance, the Waseda, NASCIO, Iribarren et al., Shareef et al. and Janowski mod-
els are the only one, which address it directly (e.g. governance, cooperation and
promotion structures, management optimization, policy driven eGovernment)
(Iribarren et al. 2008; Meyerhoff Nielsen 2016a; Obi 2015; NASCIO, N.A.o.S.C.I.O.
2006; Shareef et al. 2011; Janowski 2015; Lee 2010).
98 M. Meyerhoff Nielsen
Fig. 20 Identified stage models mapped in accordance with Lee’s qualitative meta-synthesis
framework (Adapted by author to incl. pre-conditions) (Lee 2010)
Governance Failure in Light of Government 3.0… 99
Emerging from a national context, the Australian ANAO and SAFAD models (see
Fig. 2) (Persson and Goldkuhl 2005) were introduced to categorize, evaluate pro-
cess and guide government organisations’ decisions on what services could and
should provide. Layne and Lee’s (2001) 2001 maturity model streamlines the devel-
opment stages online information and transactional services by merging different
aspects into two categories (see Fig. 4), that is: Catalogue of static information and
downloadable forms one websites and transactional aspects such as online service
and forms (i.e. eServices).
Dias and Gomes (2014) adjust the Layne and Lee (2001) and the EU benchmark-
ing models (EC 2012) in their 2014 evaluation of local eGovernment maturity in
100 M. Meyerhoff Nielsen
Portugal. The proposed model consists of three parallel dimensions each consisting
of four stages: (1) Information: Generic information (i.e. presence), downloadable
forms (i.e. interaction), search functionality (i.e. interaction), parameterize search
(i.e. interaction); (2) Service: Information (i.e. one-way), authentication of user (i.e.
two-way), eService transaction (i.e. two-way), authentication and eService transac-
tion (i.e. transaction); (3) Participation: features (i.e. two-way), authentication and
features (i.e. two-way), participative process (i.e. transaction), advanced participa-
tive process (i.e. transaction).
Iribarren et al. eGOV-MM model (see Fig. 13) (Iribarren et al. 2008) takes a
multi-dimensional approach including the front- and backoffice, policy, manage-
ment and oranisational capacities. Criticizing the Layne and Lee’s model (2001) for
being too linear and too ‘US-centric’ Heeks’ Manchester eGovernment Maturity
Model differentiate between the front- and back-office (Heeks 2015). The result is
two parallel dimensions which forms a matrix (see Fig. 17). One focus on the
sophistication of digitised interaction (i.e. one and two-way interaction plus transac-
tion) and the extent of process change (i.e. simple digitisation and automation,
improvement process integration, redesign/reform and transformation) which is
similar to Waseda (Obi 2015), IT governance and CMM/CMMI approach by
NASCIO, Iribarren and others (Iribarren et al. 2008; NASCIO, N.A.o.S.C.I.O.
2006).
While these adjustments to the presentation and publication of information
and eServices have evolved over time, none of the models includes actual use.
This is in sharp contrast to research in public administration reform – whether it
is a NPM efficiency or a JUG effectiveness approach (Bannister and Connolly
2011; Cordella and Bonina 2012; Bannister 2001; Meyerhoff Nielsen and Mika
2014). This is unfortunate as the value added of a project comes from its use, not
its existence.
Layne and Lee’s stage model breaks with the initial models, by including vertical
and horizontal integration as two distinct, and most advanced, levels of maturity to
their model (see Fig. 4) (Layne and Lee 2001). Both Deloitte (Deloitte and Touche
2001) and Gartner (Baum and Di Maio 2000) mirror this development.
Persson and Goldkuhl (2005) in 2005 evaluates a number of existing mod-
els and propose a two-stage model with a clear computer science perspective.
Based on Layne and Lee (2001), their focus is on the integration of services
(i.e. services, directed services, concentrated services and portals) and inte-
gration in services including horizontal and vertical integration of organisa-
tions, processes, the exchange and re-use of data - with the data focus being
similar to OGD Maturity Model by Kalambokis et al. (see Fig. 14) (Kalampokis
et al. 2011).
Governance Failure in Light of Government 3.0… 101
The review in section “Preconditions”. (see Table 4, Figs. 19 and 20) identified 36
(i.e. 85.7%) models which includes ICT enabled reform of public administration as
a maturity level. Of these only half (i.e. 18 models or 42.9% of all models) address
ICT enabled transformation (or morphing).
The Klievink and Janssen (2009) five-stage is of particular interest. The level of
customer orientation increases with every stage of the model, as does the level of
flexibility and includes: Stovepipes, integrated organisations, nationwide portals,
inter-organisational integration and customer-driven, joined-up government. The
Klievink and Janssen model clearly reflect joint-up government (i.e. integration)
and outcomes based thinking seen in public administration and eGovernment
literature.
Kim and Grant (2010) propose continuous improvement as a fifth and final matu-
rity level in their 2010 model. Featuring political activities and a focus on continu-
ous improvements and performance it sees ICT as a tool enabling public sector
innovation and reform – on par with the logic behind agile development in the IT
sector. Lee and Kwak’s (2012) takes a similar approach in their data based model
for collaboration and ubiquitous engagement. Although data and collaboration
forms the core of Lee and Kwak’s model, the development stages follow a ‘classi-
cal’ stage-model pattern, i.e. publication, assimilation, reform and transformation
and does therefore not cover Gov3.0. Janowski’s (Janowski 2015) model focus on
complexity of ICT enabled reform and move from a ‘classical’ model focus to a
fourth and final contextual stage.
The IT governance and CMM/CMMI models, like Davison, Iribarren et al.,
NACSIO and Waseda, provides a particular interesting multi-dimensional perspec-
tive and inclusion of both human, management and organisational capacities
(Iribarren et al. 2008; Davison et al. 2005; Obi 2015; NASCIO, N.A.o.S.C.I.O.
2006).
Considering the level of academic consensus of ICT as an enabler of public
sector reform and transformation, the limited attention paid to actual outputs
and results is surprising. Similarly, not of the models adequately address the
Gov3.0 concept.
Cooperation is indirectly addressed by all the models addressing reform and
transformation, but none look at the role governance play to ensure backoffice inte-
gration or the outcomes required to move from one stage to another. Here the IT
governance and CMM/CMMI models, like Davison, Iribarren et al., NACSIO and
Waseda, stands out with their multi-dimensional perspective and the inclusion of
both human, management and organisational capacities (Iribarren et al. 2008;
Davison et al. 2005; Obi 2015; NASCIO, N.A.o.S.C.I.O. 2006).
102 M. Meyerhoff Nielsen
The Hiller and Bélanger (2001) and Deloitte and Touche (2001) – and in 2003 the
World Bank (Toasaki 2003) with respect to legislative consultations – are the first to
add a dimension of engagement and co-creation (indirectly by none-governmental
stakeholders) and aspects of participation in a democratic sense. The focus is non-
the-less on supply, technology and organisational integration.
In contract active engagement, participation and democratic decision making are
aspects of the most advanced maturity levels proposed by authors like Moon (2002)
and Siau and Long (2005) while Chan et al. (2008) adds none-governmental stake-
holders to the mix of their five-stage model focusing on supply, technology and orga-
nizational integration. Similarly the UN eParticipation index was introduced in 2012
(UNDESA 2012) and EU benchmark has included aspects since 2013 (EC 2012).
Lee and Kwak’s (2012) five-stage model focus on engagement and data exchange
between authorities (i.e. horizontal and vertical integration), transparency by
increasing access to data, user-engagement and participation in decision making
(i.e. eParticipation and eDemocracy), and lastly on the total transformation of the
way public administration deliver services and make decisions (i.e. ubiquitous
engagement).
While increased levels of transparency in the government, political and demo-
cratic processes is laudable, the latter two does not necessarily constitute a maturity
level in their own right. Especially, when focusing on ICT use to improve the effi-
ciency, effectiveness, quality and value added of public sector service delivery.
While stage models like indexes and benchmarks are helpful in mapping the supply
and sophistication levels of eService offerings, they all have a technological focus.
The relevance of these different models is therefore limited in terms of governance,
cooperation and measuring the successful use of online offerings – and thus the
value added. In contrast to other stage models, Andersen and Henriksen (2006) fol-
low an activity- and user-centric approach to personalisation of online services in
their Public Sector Process Rebuilding (PPR) model (illustrated in Fig. 11).
Andersen and Henriksen extends the Layne and Lee’s model (see Fig. 4) (Layne and
Lee 2001) by making an online presence, horizontal and vertical integration the
foundation of their PPR model (Meyerhoff Nielsen 2015, 2016a; Alhomod and
Shafi 2012). Klievink and Janssen also address outcomes but anchor their model in
the joint-up government research stream (Klievink and Janssen 2009). The approach
is interesting as it also reflects ideas around whole-of-government approaches
(Frissen et al. 2007; Huijboom et al. 2009b; Traunmüller and Wimmer 2003; Millard
Governance Failure in Light of Government 3.0… 103
2010), JUG (Bannister and Connolly 2011; de Bri and Bannister 2010) and person-
alisation of online service delivery (Meyerhoff Nielsen and Igari 2012; Meyerhoff
Nielsen and Robert 2015).
The importance of outcomes is a key topic with the both the public administra-
tion reform (Bannister and Connolly 2011; Bannister 2001, 2007; de Bri and
Bannister 2010), IT-governance and computer science (Brown and Grant 2005) and
eGovernment literature (Cordella and Bonina 2012; Traunmüller and Wimmer
2003; Scholl 2009; Janowski 2015). Seven models are complimented with various
benchmarks, indexes and rankings (EC 2012; UNDESA 2014; West 2004; Rohleder
and Jupp 2003; Obi 2014; Almazan and Gil-Garcia 2008; Luna et al. 2013; Dias and
Gomes 2014) but several researchers have questions the value of their due to their
simplicity, their supply and technology focus (Lips 2012; Meyerhoff Nielsen 2015,
2016a; Bannister 2007; Heeks 2006, 2015; Rorissa et al. 2011). Andersen and
Henriksen are the first researchers, which have taken an outcomes based approach
but do not include take-up, qualitative or quantitative indicators. The Waseda (Obi
2015) model differs somewhat from other stage models as it does not define distinct
levels of maturity. The focus is on qualitative and quantitative indicators including
network preparedness and infrastructure, management optimisation and efficiency
etc. Unfortunately, it does not directly address the actual use of eServices, but rather
pre-conditions like internet and mobile subscriptions.
Conclusion
The review of the 42 stage-models identified, their respective maturity levels and
meta charateristics show that aspects of Gov3.0 aspects such as ICT enabled inte-
gration, transformation, sharing of data and increased participation a number of
weaknesses persists.
First, all models, with the exemption of the PPR (Andersen and Henriksen 2006),
Howard (2001) and Klievink and Janssen (2009) models, have a technology and
supply orientated, i.e. no focus on outcomes or actual use (Meyerhoff Nielsen
2016a, b; Alhomod and Shafi 2012; Lee 2010). This is unfortunate as the tangible
benefits of any ICT solution and eServices in particularly can only be realized
through the actual and effective use of supplied eServices by citizens (OECD 2014;
UNDESA 2014; Meyerhoff Nielsen 2016a; Meyerhoff Nielsen and Mika 2014;
Meyerhoff and Kelly 2011).
Second, most of the models have no real understanding of core government ser-
vice concepts. For instance individual service elements – that is information, trans-
action capability, personal data – are not separate maturity levels but rather elements
in a given service request and subsequent delivery. Similarly downloadable forms
are merely a type of static information and does not warrant a separate maturity
level (Meyerhoff Nielsen 2016a, b). This is particular surprising considering that 22
models (i.e. 52.4%) are partially based on observations, experiences and case stud-
ies in one or more countries (see Table 2).
104 M. Meyerhoff Nielsen
Acknowledgement This chapter is a result of the project “SmartEGOV: Harnessing EGOV for
Smart Governance (Foundations, methods, Tools)/NORTE-01-0145-FEDER-000037”, supported
by Norte Portugal Regional Operational Programme (NORTE 2020), under the PORTUGAL 2020
Partnership Agreement, through the European Regional Development Fund (EFDR). It work was
also supported in part by funding from Tallinn University of Technology, Project B42; OGI – Open
Government Intelligence project in the EU Horizon 2020 framework program, grant agreement
693849.
Governance Failure in Light of Government 3.0… 105
References
Alhomod SM, Shafi MM (2012) Best practices in e-government: a review of some innovative
models proposed in different countries. Int J Electri Comput Sci 12(2):1–6
Almazan RS, Gil-Garcia JR (2008) e-Government portals in Mexico. Electron Gov Concepts
Methodol Tools Appl 6:1726–1736
ANAO (1999) Electronic service delivery, including internet use by Commonwealth government
agencies. ANAO, Australian National Auditing Office, Canberra, p 87
Andersen KV, Henriksen HZ (2006) E-government maturity models: Extension of the Layne and
Lee model. Gov Inf Q 23(2):236–248
Bannister F (2001) Dismantling the silos: extracting new value from IT investments in public
administration. Inf Syst J 11(1):65–84
Bannister F (2007) The curse of the benchmark: an assessment of the validity and value of
e-government comparisons. Int Rev Adm Sci 73(2):171–188
Bannister F, Connolly R (2011) Transformation and public sector values, in tGov 11. Brunel
University, London
Bates MJ (1989) The design of browsing and berrypicking techniques for the online search inter-
face. Online Review 13(5):407–424
Baum C, Di Maio A. (2000) Gartner’s four phases of e-government model. In: Gartner Group
Brown AE, Grant GG (2005) Framing the frameworks: a review of IT governance research.
Commun Assoc Inf Syst 15(1):38
Brown CV, Magill SL (1994) Alignment of the IS functions with the enterprise: toward a model of
antecedents. MIS Q:371–403
Chan CM, Lau YM, Pan SL (2008) E-government implementation: a macro analysis of Singapore's
e-government initiatives. Gov Inf Q 25(2):239–255
Chandler S, Emanuels S (2002) Transformation not automation. In: Proceedings of 2nd European
conference on E-government. Management Center Europe, Brusseles
Charalabidis Y (2015) What is government 3.0? In: Charalabidis Y (ed) Governance and transfor-
mation. Yannis Charalabidis, Athens
Chen JYY, Mingins C (2011) A three-dimensional model for e-government development with
cases in China’s regional e-government practice and experience. In: ICMeCG, 2011 fifth
international conference on management of e-commerce and e-government. The Institute of
Electrical and Electronics Engineers Inc., Wuhan
Christensen T, Lægreid P (2007) The whole-of-government approach to public sector reform.
Public Adm Rev 67(6):1059–1066
Christine Leitner, J.-M.E., François Heinderyckx, Klaus Lenk, Morten Meyerhoff Nielsen, Roland
Traunmüller (2003) eGovernment in Europe: the state of affairs. p 66
Cisco IBSG (2007) e-Government Best Practices learning from success, avoiding the pitfalls.
Cisco IBSG
Cordella A (2007) E-government: towards the e-bureaucratic form? J Inf Technol 22(3):265–274
Cordella A, Bonina CM (2012) A public value perspective for ICT enabled public sector reforms:
a theoretical reflection. Gov Inf Q 29(4):512–520
Davison RM, Wagner C, Ma LC (2005) From government to e-government: a transition model. Inf
Technol People 18(3):280–299
de Bri F, Bannister F (2010) Whole-of-government: the continuing problem of eliminating silos.
Proceedings of the 10th European conference on eGovernment. National Centre for Taxation
Studies and University of Limerick, Ireland, pp 122–133
Deloitte and Touche (2001) The citizen as customer. In: CMS management. Deloitte and Touche,
p 58
Demmke C (2006) Governmental, organisational and individual performance. Performance myths,
performance “hype” and real performance. EIPAScope 2006(1):4–11
Dias GP, Gomes H (2014) Evolution of local e-government maturity in Portugal. In: Information
systems and technologies (CISTI), 2014 9th Iberian conference on. 2014. IEEE
106 M. Meyerhoff Nielsen
EC (2012) E.C., Public services online ‘Digital by default or by De-tour?’ Assessing user centric
eGovernment performance in Euorpe – eGovernment Benchmark 2012. European Commission,
Brussels
EC (2014) E.C., Delivering the European advantage? ‘How European governments can and
should benefit from innovative public services’. European Commission DG Communications
Networks, Content & Technology, Brussels
Edelmann N, Krimmer R, Parycek P (2008) Engaging youth through deliberative e-participation:
a case study. Int J Electron Gov 1(4):385–399
European Commission, D.R.a.I (2013) Powering European public sector innovation: towards a new
architecture. D.R.a. Innovation, Editor. European Commission, DG Research and Innovation,
Brussels, pp 1–64
Eurostat (2016) Information society household survey [cited 28 March 2016]; Available from:
http://ec.europa.eu/eurostat/web/information-society/data/database
Fath-Allah A et al (2014) eGovernment maturity models: a comparative study. Int J Software Eng
Appl 5(3):72–91
Frissen V et al (2007) The future of eGovernment: an exploration of ICT-driven models of eGov-
ernment for the EU in 2020. D. Osimo, D. Zinnbauer and A. Bianchi, Joint Research Centre
Gammon H (1954) The automatic handling of office paper work. Public Adm Rev 14(1):63–73
Heeks R (2005) Implementing and managing eGovernment: an international text. Sage, Los
Angeles
Heeks R (2006) Understanding and measuring eGovernment: international benchmarking studies.
UNDESA workshop “E-participation and e-government: understanding the present and creat-
ing the future”. Budapest, Hungary, pp 27–28
Heeks R (2015) A better eGovernment maturity model. In: iGovernment Briefing. Manchester,
University of Manchester
Heeks R, Bailur S (2007) Analyzing e-government research: Perspectives, philosophies, theories,
methods, and practice. Gov Inf Q 24(2):243–265
Hiller JS, Belanger F (2001) Privacy strategies for electronic government. E-government
200:162–198
Hodgkinson S (2002) Managing an e-government transformation program. Working Towards
Whole-of-Government Online Conference, Canberra
Howard M (2001) E-government across the globe: how will “e” change government? Gov Finan
Rev 17(4):6–9
Huijboom N, van der Broek T, Frissen V, Kool L, Kotterink B, Meyerhoff Nielsen M, Millard
J (2009a) Public services 2.0: key areas in the public sector impact of social computing. p 134
Huijboom N et al (2009b) Public Services 2.0: the impact of social computing on public services, in
Institute for Prospective Technological Studies, Joint Research Centre, European Commission.
Office for Official Publications of the European Communities, Luxembourg
Igari N (2014) How to successfully promote ICT usage: a comparative analysis of Denmark and
Japan. Telematics Inform 31(1):115–125
InfoDev, C.f.D.a.T (2002) The e-government handbook for developing countries. World Bank,
Washington, DC, pp 1–41
Iribarren M et al (2008) Capability maturity framework for eGovernment: a multi-dimensional
model and assessing tool. In: Electronic government. Springer, pp 136–147
Janowski T (2015) Digital government evolution: from transformation to contextualization. Gov
Inf Q 32(3):221–236
Janssen M, Chun SA, Gil-Garcia JR (2009) Building the next generation of digital government
infrastructures. Gov Inf Q 26(2):233–237
Janssen M, Charalabidis Y, Zuiderwijk A (2012) Benefits, adoption barriers and myths of open
data and open government. Inf Syst Manag 29(4):258–268
Jukić TT, Ljupčo N, Nameslaki A (2015) Investigation of e-government research field: what has
been done and how to proceed? NISPAcee J Public Admin Policy 23
Kalampokis E, Tambouris E, Tarabanis K (2011) Open government data: a stage model. In:
Electronic government. Springer, pp 235–246
Governance Failure in Light of Government 3.0… 107
Kim D-Y, Grant G (2010) E-government maturity model using the capability maturity model inte-
gration. J Syst Inf Technol 12(3):230–244
Klievink B, Janssen M (2009) Realizing joined-up government—dynamic capabilities and stage
models for transformation. Gov Inf Q 26(2):275–284
Klischewski R, Scholl HJ (2008) Information quality as capstone in negotiating e-government
integration, interoperation and information sharing. Electron Gov Int J 5(2):203–225
Koh CE, Prybutok VR (2003) The three ring model development of an instrument for measuring
dimensions of E-government functions. J Comput Inf Syst 43(3):34
Layne K, Lee J (2001) Developing fully functional E-government: a four stage model. Gov Inf Q
18(2):122–136
Lee J (2010) 10 year retrospect on stage models of e-Government: a qualitative meta-synthesis.
Gov Inf Q 27(3):220–230
Lee G, Kwak YH (2012) An open government maturity model for social media-based public
engagement. Gov Inf Q 29(4):492–503
Lips M (2012) E-government is dead: long live public administration 2.0. Inf Polity 17(3):239–250
Luna DE et al (2013) Improving the performance assessment of government web portals: a pro-
posal using data envelopment analysis (DEA). Inf Polity 18(2):169–187
Meyerhoff M, Kelly A (2011) Scandinavia 2.0: efficiency, cooperation and innovations to alleviate
the economic crisis. Eur J ePract 11:19–38
Meyerhoff Nielsen M (2014) Identifying eGovernment success factors: an analysis of selected
national governance models and their experiences in digitising service delivery. Proceedings
of the 2014 conference on Electronic Governance and Open Society: challenges in Eurasia,
2014, pp 19–25
Meyerhoff Nielsen M (2015) Supply and use of citizen eServices: an analysis of selected national
experiences in relation to existing governance and cooperation models. NISPAcee J Public
Admin Policy 23
Meyerhoff Nielsen M (2016a) The role of governance, cooperation, and eService use in current
eGovernment stage models. Hawaii
Meyerhoff Nielsen M (2016b) eGovernance and stage models: Analysis of identified models and
selected Eurasian experiences in digitizing citizen service delivery. Int J Electron Gov Res
x(x):2016
Meyerhoff Nielsen M, Igari N (2012) Speaking Danish in Japan. CeDEM 12 conference for
E-Democracy and Open Government 3–4 May 2012 Danube-University Krems, 2012, p 137
Meyerhoff Nielsen M, Mika Y (2014) An analysis of the Danish approach to eGovernment benefit
realisation. Internet Technologies and Society 2014 conference proceedings, 2014, pp 47–58
Meyerhoff Nielsen M, Robert K (2015) Reuse of data for personal and proactive service: an
opportunity not yet utilised. In: CeDEM 15 conference for e-democracy and open government
20–22 May 2015, Danube-University Krems, Austria. Krems an der Donau: Donau-Universität
Krems; eJournal of eDemocracy and Open Government
Millard J (2010) Government 1.5 – is the bottle half full or half empty? Eur J ePract (9):35–48
Millard J (2013) ICT-enabled public sector innovation: trends and prospects. In: Proceedings of the
7th international conference on theory and practice of electronic governance. ACM
Millard J, Luca C, Galasso G, Riedl R, Neuroni AC, Walser K, Sami Hamida A, Huijboom N,
Meyerhoff Nielsen M, Leitner C, Fehlmann RS (2007) European eGovernment 2005–2007:
Taking stock of good practice and progress towards implementation of the i2010 eGovernment
Action Plan. p 80
Millard J et al (2008) Social computing: trends in public services and policies. JRC-IPTS
Ministry of Interior Korea (2016) Government 3.0. Ministry of Interior Korea, Seoul
Moon MJ (2002) The evolution of e-government among municipalities: rhetoric or reality? Public
Adm Rev 62(4):424–433
NAO (2002) N.A.O., Government on the Web II. UK National Audit Office, London
NASCIO, N.A.o.S.C.I.O. (2006) Enterprise Architecture Maturity Model (EAMM), version 3.1.
National Association of State Chief Information Officers, Lexington
108 M. Meyerhoff Nielsen
Netchaeva I (2002) e-government and e-democracy a comparison of opportunities in the North and
South. Int Commun Gaz 64(5):467–477
O’Leary R, Gerard C, Bingham LB (2006) Introduction to the symposium on collaborative public
management. Public Adm Rev 66(s1):6–9
Obi T (2012) WASEDA – IAC Internationl e-Government Index. Waseda University and IAC
International Agency of CIO, Tokyo
Obi T (2014) WASEDA – IAC internationl e-government index. Waseda University and IAC
International Agency of CIO, Tokyo
Obi T (2015) WASEDA – IAC Internationl e-Government Index. Waseda University and IAC
International Agency of CIO, Tokyo
OECD (2014) Recommendation of the Council on Digital Government Strategies 15 July 2014 –
C(2014)88. OECD, Paris
Persson A, Goldkuhl, G (2005) Stage-models for public e-services-investigating conceptual foun-
dations. 2nd Scandinavian Workshop on e-Government, Copenhagen
Peters BG, Pierre J (1998) Governance without government? Rethinking public administration.
J Public Adm Res Theory 8(2):223–243
Poeppelbuss J et al (2011) Maturity models in information systems research: literature search and
analysis. Commun Assoc Inf Syst 29(27):505–532
Pollitt C (2014) Future trends in European public administration and management: an outside-in
perspective. COCOPS Coordination for Cohesion in the Public Sector of the Future
Pollitt C, Bouckaert G (2011) Public management reform: a comparative analysis-new public
management, governance, and the Neo-Weberian state. Oxford University Press, Oxford
Pöppelbuß J, Röglinger M (2011) What makes a useful maturity model? A framework of general
design principles for maturity models and its demonstration in business process management.
ECIS
Reddick CG (2004) A two-stage model of e-government growth: theories and empirical evidence
for US cities. Gov Inf Q 21(1):51–64
Roberts SE (1977) Theories and Models in Information Retrieval. J Doc 33(2):126–148
Röglinger M, Pöppelbuß J, Becker J (2012) Maturity models in business process management. Bus
Process Manag J 18(2):328–346
Rohleder SJ, Jupp V (2003) e-Government leadership: engaging the customer. Accenture,
Arlington, pp 1–94
Ronaghan SA (2002) Benchmarking e-government: a global perspective: assessing the progress of
the UN member states United Nations Division for Public Economics and Public Administration
Rorissa A, Demissie D, Pardo T (2011) Benchmarking e-Government: a comparison of frame-
works for computing e-Government index and ranking. Gov Inf Q 28(3):354–362
Ross JW, Weill P, Robertson D (2006) Enterprise architecture as strategy: creating a foundation for
business execution. Harvard Business Press, Boston
Scholl HJJ (2009) Profiling the EG research community and its core. In: Electronic government.
Springer, Berlin/Heidelberg, pp 1–12
Self P (2000) Rolling back the state. Economic dogma & political choice. St Martin’s Press,
New York
Shahkooh KA, Saghafi F, Abdollahi A (2008) A proposed model for e-Government maturity. In:
Information and communication technologies: from theory to applications, 2008. ICTTA 2008.
3rd international conference on. 2008. IEEE
Shareef MA et al (2011) e-Government Adoption Model (GAM): differing service maturity levels.
Gov Inf Q 28(1):17–35
Siau K, Long Y (2005) Synthesizing e-government stage models-a meta-synthesis based on meta-
ethnography approach. Ind Manag Data Syst 105(4):443–458
Silcock R (2001) What is e-government. Parliam Aff 54(1):88–101
Statskontoret (2000) 24-timmmarsmyndighet: Förslag til kriterier för statlige elektronisk förvalt-
ning i medborgarnas tjänst. Statskontoret, Stockholm, pp 1–80
Toasaki Y (2003) e-Government from a user’s perspective. World Bank, Taipei
Governance Failure in Light of Government 3.0… 109
Abstract As part of the Smart Cities movement, public administrations are con-
stantly in need to create new and innovative public services. Innovative services can
be derived from exiting best practices. Reuse is a key enabler for cost effective
customization of their processes for delivering effective and timely services. The
literature exhibits a wide variety of techniques that can be applied. This paper con-
ducts an analysis of major reuse-oriented process modeling techniques with respect
to available means of maintainability, user support, compression rate gained when
storing process models as well as traceability of modeling decisions. Furthermore,
we empirically evaluated the technique of configuration-based process modeling to
validate its applicability in modeling municipal processes.
Introduction
data to enhance the dynamic design of new and innovative public services (Chouikh
et al. 2016; Ojo et al. 2015).
While public administrations are constantly required to deliver improved and
efficient public services, they are facing various challenges such as cost reductions,
change management in organizational work concepts, political pressures, etc.
(Karow et al. 2008). Consequently, public administrations have to redesign their
processes and resource allocations to meet cost and time requirements. Reusing and
customizing existing proven practices is an important pillar for driving innovative
services in a cost-effective and rapid manner.
Reuse in process modeling has been proven to be effective using techniques that
vary from establishing a common repository of processes (Beeri et al. 2008a; Lu
and Sadiq 2007; Rosa et al. 2011; Vulcu et al. 2011) to creating reference process
models that can be tailored to each organization needs (Baran et al. 2013; Derguech
et al. 2010; Rosemann and van der Aalst 2007; Sadiq et al. 2001). However, choos-
ing the right technique to apply within an organization requires a proper analysis of
available tools for maintainability, user support and alignment with the organiza-
tions’ strategies regarding transparency and traceability (Karow et al. 2008).
The aim of this paper is to analyze major reuse-oriented process modeling tech-
niques with respect to a set of requirements that are identified in section
“Methodology”. The outcome of the analysis can serve as a guideline for choosing
which technique to apply in certain organizations. The analyzed approaches are
classified in two families: techniques using repositories of process models in section
“Business Process Models Repository” and techniques using reference process
models in section “Reference Business Process Modelling”. Before concluding the
chapter in section “Conclusion”, we conduct an evaluation of one of the techniques
that uses reference process models, in section “Configurable Models for
Municipalities”, to assess its applicability in modeling municipal processes.
Methodology
In this section, we define the methodology that we use in conducting the analysis of
the state of the art related to the topic of reuse in business process modelling. Our
analysis starts by classifying research contributions with respect to the categories
shown in Fig. 1. That is: sections “Business Process Models Repository” and
“Reference Business Process Modelling” respectively outline contributions in two
main categories of reuse-oriented business process modelling techniques: (i) using
Business Process Repositories and (ii) using Reference Business Process Models.
The first category is investigated in section “Business Process Models Repository”
by considering various implementations of business process repositories that permit
either to discover an entire business process model or to discover business process
building blocks that can be used later for composition.
The second category is investigated in section “Reference Business Process
Modelling” by considering three implementations of reference process models
Techniques for Reuse in Business Process Modeling in Public Administration 113
Business Process
Models Reuse
Reference BP
BP Repositories
Models
In the first part of the analysis, we study related work in the area of business pro-
cess model discovery. The discovery operation consists of querying a business
process repository in order to find a relevant business process model satisfying
particular needs. In this section, we investigate various implementations of pro-
cess repositories.
Description The Process Variant Repository (Lu and Sadiq 2006, 2007; Lu et al.
2009a) or PRV for short, defines a repository of both business process models and
associated “preferred work practices”. A preferred work practice is a process vari-
ant that is captured from the process execution logs and is suitable for a particular
situation. Each process model is stored with its historical information about the
execution instances in order to achieve new operational goals in similar situations.
For example, we can refer to the registration of a newborn child of parents from
either the local or a foreign country. Here the process will be the same with some
changes in the required documents.
PVR provides a support for querying business process models and their vari-
ants where a query is a partial or complete description of a process variant. On
the basis of similarity metrics, the authors measure the equivalence and sub-
sume relations between the process query and the stored processes using reduc-
tion techniques in graphs. The results are then ranked based on these similarity
values.
Analysis With respect to the identified requirements, the key points of analysis are
as follows:
–– Requirement 1: Compression Rate: not fulfilled. The PVR focus is on providing
a discovery mechanism while ignoring any challenges related to managing com-
mon process parts. Business process variants are stored individually without per-
forming any compression.
–– Requirement 2: Maintainability: not discussed.
–– Requirement 3: User Support: not fulfilled. Querying the repository requires
expert knowledge for creating queries.
–– Requirement 4: Traceability: partially fulfilled. In essence, the use of process
repositories guarantees traces of all process variants. However, traces of user
queries are not logged in this work.
Techniques for Reuse in Business Process Modeling in Public Administration 115
BP-Suite
System – IRIS1 is integrated with the semantic business process repository to reason
over the business processes described using ontologies.
Second, while the framework for querying business process models proposed by
Markovic et al. (2008) uses ontologies for describing business process models, Sakr
and Awad (Sakr and Awad 2010) use ontologies only in the query matching process
and tackle the problem of applying different terminologies when modelling pro-
cesses. The former (Markovic et al. 2008) uses Web Service Modeling Ontology
(WSMO) for describing functional and non-functional related properties and a pro-
cess algebra, pi-calculus, for the structural properties of a business process model.
They use Web Service Modeling Language (WSML) logical expressions as a query
language and ontological reasoning for query answering. Whereas the latter (Sakr
and Awad 2010) relies mainly on activity labels for describing functional properties
and uses BPMN-Q (Awad and Sakr 2012) for querying business process models
with an underlying classical database management system.
Last, the oryx (Decker et al. 2008) extension for semantically-enabled business pro-
cess discovery (Vulcu et al. 2011) proposes the use of ontologies for modelling and stor-
ing business process models. The authors propose an ontology for describing graph-based
and block-based business processes while capturing their functional (i.e., Input, Output,
Precondition and Effect) and non-functional properties at multiple levels of abstraction.
Analysis With respect to the identified requirements, the key points of analysis are
as follows:
–– Requirement 1: Compression Rate: not fulfilled. The reviewed solutions investi-
gate the use of ontologies for storing and querying business process models.
They use graphical querying mechanisms for supporting users to avoid learning
a complex querying language. However, none of them deals with how to effi-
ciently store process variants: compression is out of scope.
–– Requirement 2: Maintainability: fulfilled. Standard CRUD operations and ver-
sion management were investigated (Ma et al. 2007; Sakr and Awad 2010).
–– Requirement 3: User Support: partially fulfilled. Although an extensive work has
been put towards creating graphical query mechanisms, users still need to manually
define some difficult parameters such as Input, Output, Precondition and Effect.
–– Requirement 4: Traceability: fulfilled. In essence, the use of process repositories
guarantees traces of all process variants. Furthermore, version control adds
another traceability dimension for verifying the evolution of changes in the pro-
cess models.
APROMORE
http://www.iris-reasoner.org/
1
Techniques for Reuse in Business Process Modeling in Public Administration 117
Business Expert
Modeller
Techniques for Reuse in Business Process Modeling in Public Administration 119
Fig. 4 Using placeholders for managing business process variants (Adapted from Weber et al.
2009)
Techniques for Reuse in Business Process Modeling in Public Administration 121
Analysis With respect to the identified requirements, the key points of analysis are
as follows:
–– Requirement 1: Compression Rate: fulfilled. In essence, the use of a reference
model guarantees that duplicate process elements are merged, ensuring a high
compression rate.
–– Requirement 2: Maintainability: fulfilled. Maintainability, has been tackled from
a technical perspective. Indeed, the literature proposes various algorithms for
checking the satisfiability of the constraints (Lu et al. 2009b; Pesic et al. 2010)
used with the predefined set of activities.
–– Requirement 3: User Support: partially fulfilled. Even though it has been noticed
that there is a need to help users create sound and correct models (van der Aalst
et al. 2009), we could not find any contribution that creates and updates such
reference process models. However, during the customisation phase, users can
be assisted either with a set of activities and/or constraints as it has been high-
lighted by Sadiq et al. (2001).
–– Requirement 4: Traceability: not discussed.
Description In most cases, business process models tend to be very large and are
difficult to manage by end-users. Reducing the complexity of large models can be
achieved by representing them at different levels of detail. The general idea is to
reduce the complexity of business processes and reveal to the end-user a partial
model by applying abstraction techniques. This fosters the reuse of similar process
fragments as well as reducing inconsistency. In this context, some researchers tried
to manage reference process models at various levels of abstraction while explicitly
capturing variation points. The object of this section is to review the proposed
approaches that study such models, i.e., hierarchical reference process models.
Razavian and Khosravi (2008), propose a variability modelling method which is
specifically designed for the component and connector view of UML 2. The authors
introduce multiple mechanisms for modelling variation points depending on the
variable element (component, connector or interface). Variation points are presented
at various levels of abstraction by having optional or alternative architectural ele-
ments. An example is shown in Fig. 5 where the top level component “UI Manager”
can be further refined to one of the two associated variants: “JavaScript UI Manager”
and “HTML UI Manager”. Each element is annotated by specific stereotypes: the
variation point is marked by << alt vp >> and its lower level sub processes express
all details related to higher level activities and variabilities residing in them and they
are annotated by << variant >>.
Baran et al. (2013) investigated the use of hierarchical reference business process
models using BPMN. Such models are created in a two-step operation. First, the
proposed algorithm transforms the input BPMN models into two-level hierarchy.
122 W. Derguech et al.
<<component>> <<component>>
<<variant>> <<variant>>
JavaScript UI Mgr HTML UI Mgr
Customer
Registration*
Fig. 6 A hierarchical indexing structure for modelling one variation point of a process for register-
ing a customer to an insurance contract (Derguech and Bhiri 2010)
The authors use a very simple abstraction technique that takes as input a BPMN
model and the set of interlinked high-level and low-level tasks and delivers the cor-
responding hierarchical model. Second, the BPMN models are merged into a single
one that requires additional transformations to become well formed.
In previous works, we explored the use of hierarchical reference process models
(Derguech and Bhiri 2010; Derguech et al. 2010) by proposing the use of an index-
ing structure for representing process models at different levels of abstractions as
depicted in Fig. 6 We used the concept of abstract tasks for capturing variation
points, it is marked with a “*” at the end of the task label (see “Customer
Registration*” on Fig. 6). An abstract task can be refined/concreted by selecting one
of its concrete alternatives which are associated to it via dotted lines. In addition to
this customised notation, we proposed an algorithm for updating the reference
model by inserting a new node (either a task or sub-process). The work looked
promising, however, it has not been implemented or further investigated.
Analysis With respect to the identified requirements, the key points of analysis are
as follows:
–– Requirement 1: Compression Rate: fulfilled. In essence, the use of a reference
model in general, and hierarchical model in particular, guarantees that duplicate
process elements are merged, ensuring a high compression rate.
Techniques for Reuse in Business Process Modeling in Public Administration 123
Description A configurable business process model (Rosemann and van der Aalst
2007) is the result of merging process variants into a single model. This model can
be tailored to the analysts’ needs by enabling or disabling different branches of the
configurable model. Figure 7 depicts, in the left-hand side, two variants of the same
business process. These two variants reflect two common tasks (i.e., Task A and B),
however after this, each variant ends with a different task (i.e., C or D). This differ-
ence introduces the choice between the task C or D that represents a variability
depending on various indicators, e.g., cost, quality of service, user preference, etc.
The right-hand side of the Fig. 7 shows the configurable process model which is
a merger between the two process variants. The variation point is represented by a
configurable gateway: an inclusive split gateway marked with a thick red border.
Unlike a “normal” BPMN gateway, it does not represent a choice or a parallel split,
instead, it represents a design choice that needs to be made by an analyst to adapt
the configurable process model to a particular requirement. In this example, the
124 W. Derguech et al.
configurable gateway captures the fact that one needs to choose whether to select
one path (i.e., task C) or the other (i.e., task D), or possibly both.
In this case, the modelling phase consists of enabling or disabling different
branches of the configurable process model. This allows customization of the con-
figurable process model by choosing the right variant. However, the main weakness
of this solution is that it does not allow the business users to understand the relation-
ship each variant has with the business domain. There are two important challenges
for adopting these models: (1) automation support for creating a configurable model
and (2) assisting end-users during the configuration. Multiple contributions provid-
ing algorithms for automatically creating configurable process models either by
merging a set of input variants or mining process logs have been investigated (Assy
et al. 2015; Derguech and Bhiri 2011; La Rosa et al. 2013).
La Rosa recognises the need to make the configuration phase user-friendly and
proposes a questionnaire-driven configuration (La Rosa 2009; La Rosa et al. 2009).
The proposed approach is sketched in Fig. 8. A process modeller has to define the
configurable process model and meet with the domain expert in order to define
domain constraints (i.e., business capabilities) and their mapping to the model. The
configuration is then performed via an interactive questionnaire. The domain
expert’s answers are then mapped to the configurable model in order to “individual-
ize” it in a process model.
Analysis With respect to the identified requirements, the key points of analysis are
as follows:
Fig. 8 Questionnaire-driven approach for configurable business process modelling (La Rosa
2009)
Techniques for Reuse in Business Process Modeling in Public Administration 125
This section analysed reuse in the context of reference process modelling. The
results of the analysis are summarised in Table 2 that clearly shows the use of such
models fulfills already the first requirement of compression rate.
For maintainability, the approaches considered in this analysis exhibit a hetero-
geneous set of methods for maintainability and more specifically in automatically
creating such process models (Derguech and Bhiri 2011; Gerth et al. 2009, 2010;
Gerth and Luckey 2012; Kuster et al. 2008a; b; La Rosa et al. 2013).
For user support, customizing a reference model is difficult and has been
extensively discussed in the literature. Even though La Rosa (2009) proposes a
prevalent solution to this problem, it still suffer for a major shortcoming: the
need for an extensive manual matching between the model and the domain
constrains.
Traceability has been covered by covered by La Rosa et al. (2013) and Derguech
and Bhiri (2011) as it was considered as part of their requirements when merging
business process variants for creating configurable process models.
From our analysis, we found that configurable business process models are the
most mature contributions with respect to the considered requirements. The use of
configurable models in public administration is assessed in section “Configurable
Models for Municipalities”.
Tool Support
The designed business process merging algorithm (Derguech and Bhiri 2011) has been
implemented as an extension of EPCTools (Nicolas and Ekkart 2006). EPCTools is an
open source initiative toward a tool for Event Driven Process Chains (EPCs) that
Techniques for Reuse in Business Process Modeling in Public Administration 127
Fig. 9 Extended version of EPCTools that supports the creation of capability-annotated configu-
rable business process models
supports the tool independent EPC interchange format EPML (Mendling and Nuttgens
2006) implemented as an Eclipse Plug-in. As shown in Fig. 9, after opening one of the
two process models, the user has to click on the “Merging models” button (see 1 in
Fig. 9), then a new dialog window is open, the user selects the second process model
and clicks on ok, in this step the new configurable process is created. The user can
optionally decide to apply a reduction step by selecting the “Reduce” button (see 2 in
Fig. 9) to further reduce the generated model by applying reduction rule defined in
Derguech and Bhiri (2011). The tool support is a proof of concept that has been imple-
mented to carry out compression rate and execution time evaluations. Further evalua-
tions regarding the user interface and how the user interacts with this tool is part of our
future work. The user experience evaluation might be influenced by the modelling
environment and is out of the scope of the contribution of this research.
For organisations time is important and should not be spent on manual creation of
configurable models, this evaluation shows how quickly the merging algorithm
delivers configurable process models.
The evaluation of compression rate and execution time has been carried out as
follows:
1. A test collection of real-world municipal process models have been manually
created.
2. Each of the input models have been quantified in terms of the number of process
elements (i.e., events, functions and connectors).
3. Using the tool support, we have created configurable process models from the
input models.
4. Each resulting configurable process model has been quantified in terms of the
number of process elements.
5. Measure the compression rate by comparing the sizes of the input models and
the output configurable model.
6. Measure the execution time of the merging process.
Please note that the execution of the merging steps has not been interrupted with a
manual task. In this regard, all the model variants are merged at once (instead of
merging each pair one by one manually). Furthermore, the reduction step has been
carried out automatically after merging (no manual decision is needed regarding the
reduction step).
Test Collection The process variants that we used in the experiment are those that
have been used in a case study (Gottschalk et al. 2009) in which techniques for
managing configurable process models were extensively tested in a real-world sce-
nario. The process models used in this case study are four processes out of the five
most executed registration processes in the civil affairs department of Dutch munic-
ipalities (Gottschalk et al. 2009):
–– P1: Acknowledging an unborn child: This process is executed when a man wants
to register that he is the father of an unborn child in case he is not married to his
pregnant partner. Figure 10 shows an example of this process.
–– P2: Registering a newborn: This process describes the steps for registering a
newborn and get his birth certificate.
–– P3: Marriage: This process describes all the steps required before getting mar-
ried in a Dutch municipality.
–– P4: Decease: This process describes the steps required by relatives to burry the
deceased and get a death certificate.
The process variants considered in this evaluation are initially available in
Protos.2 Each process has five process variants. Consequently, a total of 5 × 4 = 20
process models were considered in this work (similar to the case study (Gottschalk
et al. 2009)). We have manually translated these models into EPC and used the
extended version of EPCTools (see section “Tool Support”) for merging them in
order to create configurable process models for each process.
Observations During the merging steps, two metrics were observed: process mod-
els sizes (before, and after the merging) and the execution time of the merging steps.
These metrics are shown in Table 3.
Table 3 shows the size of the input and output models (size in terms of number
of EPC nodes). The percentage value between parenthesis shows the compression
rate gained from the creation of the configurable process models. And the last
column shows the execution time in milliseconds needed for merging the input pro-
cess models.
Discussion The reduction approach can gain around 50% in terms of space for
storing several process variants. Besides the space gain, we can see that in a
few milliseconds a set of five process variants can be automatically merged
which would take much longer for a business analyst to perform the task
manually.
In general, compression rates are high because most of the process models
share various process elements. Indeed, all the used variants, are from various
Dutch municipalities that are initially defined from a high level reference
model (Gottschalk et al. 2009). Depending on the population and the available
resources of each municipality, few process tasks are either skipped or replaced
by other ones. This keeps most of the process functions sequentially aligned.
Consequently, the merged model observe a large number of common functions
and events.
130 W. Derguech et al.
Conclusion
As public administrations are constantly in need to create new and innovative public
services, they need to face challenges related to cost reductions, reorganizations and
political pressures. Reuse of their existing best practices is a key enabler for cost
effective customization of their processes. The literature exhibits a wide variety of
techniques that can be applied. However, in the absence of a guideline for choosing
what technique to apply, this task remains difficult.
This paper, helps overcoming this issue by providing an analysis of major reuse-
oriented process modeling techniques with respect to their maintainability, user
support, compression rate gained when storing the models as well as traceability of
modeling decisions.
The analysis shows that the use of configurable process models is a promising
technique that covers all the requirements. The technique has been evaluated in this
paper for assessing its applicability. However, the user support requirement has not
been validated in this paper and is kept as part of our future work.
Acknowledgment The research leading to these results has received funding from Science
Foundation Ireland (SFI) under Grant Number SFI/12/RC/2289. It is supported in part by the
European Commission’s Seventh Framework Programme from ICT grant agreement
WATERNOMICS no. 619660.
References
Assy N, Chan NN, Gaaloul W (2015) An automated approach for assisting the design of configu-
rable process models. IEEE T Services Computing 8(6):874–888
Awad A, Sakr S (2012) On efficient processing of bpmn-q queries. Comput Ind 63(9):867–881
Balan E, Milo T, Sterenzy T (2010) Bp-ex: a uniform query engine for business process execution
traces. In: Proceedings of the 13th international conference on extending database technology,
EDBT ‘10. ACM, New York, pp 713–716
Baran M, Kluza K, Nalepa GJ, Ligeza A (2013) A hierarchical approach for configuring business
processes. In: Ganzha M, Maciaszek LA, Paprzycki M (eds) FedCSIS. pp 915–921
Beeri C, Eyal A, Kamenkovich S, Milo T (2006) Querying business processes. In: Dayal U, Whang
KY, Lomet DB, Alonso G, Lohman GM, Kersten ML, Cha SK, Kim YK (eds) VLDB. ACM,
New York, pp 343–354
Techniques for Reuse in Business Process Modeling in Public Administration 131
Beeri C, Eyal A, Kamenkovich S, Milo T (2008a) Querying business processes with bp-ql. Inf
Syst 33(6):477–507
Beeri C, Eyal A, Milo T, Pilberg A (2008b) Bp-mon: query-based monitoring of bpel business
processes. SIGMOD Record 37(1):21–24
Bichler M, Hess T, Krcmar H, Lechner U, Matthes F, Picot A, Speitkamp B, Wolf P (eds) (2008)
Multikonferenz Wirtschaftsinformatik, MKWI 2008, Mu¨nchen, 26.2.2008–28.2.2008,
Proceedings. GITO-Verlag, Berlin
Chouikh A, Ojo A, Driss OB (2016) Exploring the affordances of social media platforms in support-
ing emerging public service paradigms. In: Bertot JC, Estevez E, Mellouli S (eds) Proceedings
of the 9th international conference on theory and practice of electronic governance, ICEGOV
2016, Montevideo, 1–3 Mar 2016. ACM, pp 177–186
Curry E, Dustdar S, Sheng QZ, Sheth AP (2016) Smart cities – enabling services and appli-
cations. J. Internet Services and Applications 7(1):6:1–6:3. http://dx.doi.org/10.1186/
s13174-016-0048-6
Decker G, Overdick H, Weske M (2008) Oryx – an open modeling platform for the bpm com-
munity. In: Dumas M, Reichert M, Shan MC (eds) Business process management, 6th interna-
tional conference, BPM 2008, Milan, 2–4 Sept 2008. Proceedings, Lecture Notes in Computer
Science, vol. 5240. Springer, pp 382–385
Derguech W, Bhiri S (2010) Reuse-oriented business process modelling based on a hierarchical
structure. In: zur Muehlen, M, Su J (eds) Business process management workshops. Lecture
Notes in Business Information Processing, vol. 66. Springer, pp 301–313
Derguech W, Bhiri S (2011) An automation support for creating configurable process models.
In: Bouguettaya A, Hauswirth M, Liu L (eds) Web information system engineering – WISE
2011 – 12th international conference, Sydney, 13–14 Oct 2011. Proceedings. Lecture Notes in
Computer Science, vol. 6997. Springer, pp 199–212
Derguech W, Vulcu G, Bhiri S (2010) An indexing structure for maintaining configurable process
models. In: Bider I, Halpin TA, Krogstie J, Nurcan S, Proper E, Schmidt R, Ukor R (eds)
BMMDS/EMMSAD. Lecture Notes in Business Information Processing, vol. 50. Springer,
pp 157–168
Dumas M, Reichert M, Shan MC (eds) (2008) Business process management, 6th international
conference, BPM 2008, Milan, 2–4 Sept 2008. Proceedings, Lecture Notes in Computer
Science, vol. 5240. Springer
Dumas M, Garcia-Banuelos L, Rosa ML, Uba R (2013) Fast detection of exact clones in business
process model repositories. Inf. Syst. 38(4):619–633
Gerth C, Luckey M (2012) Towards rich change management for business process models.
Softwaretechnik-Trends 32(4):32–34
Gerth C, Küster JM, Engels G (2009) Language-independent change management of process mod-
els. In: Schurr A, Selic B (eds) Model driven engineering languages and systems, 12th inter-
national conference, MOD-ELS 2009, Denver, 4–9 Oct 2009. Proceedings. Lecture Notes in
Computer Science, vol. 5795. Springer, pp 152–166
Gerth C, Luckey M, Küster JM, Engels G (2010) Detection of semantically equivalent fragments
for business process model change management. In: 2010 IEEE international conference on
services computing, SCC, Miami, 5–10 July 2010. IEEE Computer Society, pp 57–64
Gottschalk F, van der Aalst WMP, Jansen-Vullers MH (2008) Merging event-driven process
chains. In: Meersman R, Tari Z (eds) On the move to meaningful internet systems: OTM 2008,
OTM 2008 confederated international conferences, CoopIS, DOA, GADA, IS, and ODBASE
2008, Monterrey, 9–14 Nov 2008, Proceedings, Part I. Lecture Notes in Computer Science, vol.
5331. Springer, pp 418–426
Gottschalk F, Wagemakers TAC, Jansen-Vullers MH, van der Aalst WMP, Rosa ML (2009)
Configurable process models: experiences from a municipality case study. In: van Eck P,
Gordijn J, Wieringa R (eds) Advanced information systems engineering, 21st international con-
ference, CAiSE 2009, Amsterdam, 8–12 June 2009. Proceedings. Lecture Notes in Computer
Science, vol. 5565. Springer, pp 486–500
132 W. Derguech et al.
Pesic M, Bosnacki D, van der Aalst WMP (2010) Enacting declarative languages using ltl: avoid-
ing errors and improving performance. In: van de Pol J, Weber M (eds) SPIN. Lecture Notes in
Computer Science, vol. 6349. Springer, pp 146–161
Razavian M, Khosravi R (2008) Modeling variability in business process models using uml. In:
ITNG. IEEE Computer Society, pp 82–87
Rosemann M, van der Aalst WMP (2007) A configurable reference modelling language. Inf Syst
32(1):1–23
Sadiq SW, Sadiq W, Orlowska ME (2001) Pockets of flexibility in workflow specification. In:
Kunii HS, Jajodia S, Sølvberg A (eds) ER. Lecture Notes in Computer Science, vol. 2224.
Springer, pp 513–526
Sadiq SW, Orlowska ME, Sadiq W (2005) Specification and validation of process constraints for
flexible workflows. Inf Syst 30(5):349–378
Sakr S, Awad A (2010) A framework for querying graph-based business process models. In:
Proceedings of the 19th international conference on world wide web. WWW ‘10, ACM,
New York, pp 1297–1300
Schumm D, Dentsas D, Hahn M, Karastoyanova D, Leymann F, Sonntag M (2012) Web ser-
vice composition reuse through shared process fragment libraries. In: Brambilla M, Tokuda T,
Tolksdorf R (eds) ICWE. Lecture Notes in Computer Science, vol. 7387. Springer, pp 498–501
Sirbu A, Marconi A, Pistore M, Eberle H, Leymann F, Unger T (2011) Dynamic composition of
pervasive process fragments. In: ICWS. IEEE Computer Society, pp 73–80
Uba R, Dumas M, Garcıa-Banuelos L, Rosa ML (2011) Clone detection in repositories of business
process models. In: Rinderle-Ma S, Toumani F, Wolf K (eds) BPM. Lecture Notes in Computer
Science, vol. 6896. Springer, pp 248–264
van der Aalst WMP, Pesic M, Schonenberg H (2009) Declarative workflows: balancing between
flexibility and support. Computer Science – R&D 23(2):99–113
van der Aalst WMP, Dumas M, ter Hofstede AHM, Russell N, Verbeek HMWE, Wohed P Life
after bpel? In: Bravetti M, Kloul L, Zavattaro G (eds) EPEW/WS-FM, Lecture Notes in
Computer Science, vol 3670. Springer, New York, 2005, pp 35–50
Vulcu G, Bhiri S, Derguech W, Ibanez MJ (2011) Semantically-enabled Business process models
discovery. Int J Bus Process Integr Manag 5:257–272
Walmsley P (2007) XQuery. O’Reilly Media, Inc.
Weber B, Sadiq SW, Reichert M (2009) Beyond rigidity – dynamic process lifecycle support.
Computer Science – R&D 23(2):47–65
Zillner S, Becker T, Munne R, Hussain K, Rusitschka S, Lippell H, Curry E, Ojo AK (2016) Big
data-driven innovation in industrial sectors. In: Cavanillas JM, Curry E, Wahlster W (eds) New
horizons for a data-driven economy – a roadmap for usage and exploitation of big data in
Europe, pp 169–178. Springer. http://dx.doi.org/10.1007/978-3-319-21569-3
Dr. Edward Curry (www.edwardcurry.org) is Vice President of the Big Data Value Association
(www.BDVA.eu) a non-profit industry-led organisation with the objective of increasing the competi-
tiveness of European Companies with data-driven innovation. He is a research leader at the Insight
Centre for Data Analytics (www.insight-centre.org) and a funded investigator at LERO The Irish
Software Research Centre (www.lero.ie). Edward has worked extensively with industry and govern-
ment advising on the adoption patterns, practicalities, and benefits of new technologies. Edward has
published over 120 scientific articles in journals, books, and international conferences
134 W. Derguech et al.
Dr. Sami Bhiri is an associate professor in computer science at the ISIMM: Institut Suprieur
d’Informatique et Mathmatiques Monastir, at the University of Monastir, Tunisia. His research
interests include service computing and business process management. Before joining ISIMM, he
served as an associate professor at Telecom SudParis. Prior to that, he was the leader of the SOA
unit at DERI, and adjunct lecturer at the National University of Ireland, Galway. Before joining
DERI, he was a research and teaching assistant in the University of Nancy 1 and in the ECOO team
of the LORIA-INRIA research laboratory from where he holds an M.S. (2001) and a Ph.D. (2005)
Capability Development in Open Data-Driven
Organizations
Fatemeh Ahmadi Zeleti and Adegboyega Ojo
Abstract Open data (OD) is increasingly considered as a core resource for many
organizations in the emerging data economy. Open data-driven organizations
(ODDOs) like any other organizations must develop capabilities for generating
value from OD, agility, and competitiveness to survive. This chapter investigates the
salient factors for generating value from OD and agility in a dynamic data ecosys-
tem by developing an operationalization of the Resource-based View Theory (RBV)
and Dynamic Capability Theory (DCT) for ODDOs. As a first step towards deter-
mining the critical factors for developing value capabilities (VCs) and dynamic
capabilities (DCs) in these organizations, we analyzed the information gathered
from expert interviews on the saliency of the different aspects and stages of VCs and
DCs in developing VCs for a down-stream organization and the agility of an up-
stream organization or OD supplier in the data ecosystem. Both frameworks were
enhanced based on the feedbacks received from interviewees and as a result new
open data value capabilities are discovered. Our findings further suggest that critical
factors for DCs differ for organizations in the upstream and downstream sectors,
albeit some core elements are shared across sectors in data ecosystem.
Introduction
The increasing number of ODDOs and their centrality in the new data economy; calls
for scholarly works on the competitiveness and survivability of these entities. Studies
attempting to understand the required capabilities for value generation, agility and
competitiveness like (Zeleti and Ojo 2014) are emerging. Specifically Zuiderwijk
Theoretical Background
Many organizations today wonder what exactly organizational capability means and
why it is so important (Brits et al. 2007). While there are different definitions and con-
ceptualizations for the concept of organization capability in research literature, exten-
sive experience from practice clearly indicate that the concept represents “organization
capacity to successfully perform a unique organization activity period of overtime”.
Along this perspective, Brits (2006) defines capability as a “special type of a
resource whose function improves the productivity of other resources”. This implies
that resources can represent a cluster of elements that constitute a capability. In
addition, Townsend and Cairns (2003) argue that there is a considerable difference
between competency and capability. Competency, as it is more regularly defined
and theorized, is basically a term that covers observable current skills based on cur-
rent knowledge while capability is beyond competency. Capability is a more “holis-
tic, broad-based concept that includes the additional elements of values and
self-efficacy as core components and it describes how an individual or organization
applies their ability in a confident manner to problems in new and unfamiliar cir-
cumstances as well as in familiar situations” (Townsend and Cairns 2003). Townsend
and Cairns (2003) identified three fundamental organization capability attributes:
(1) ability (the current organization competence), (2) self-efficacy (belief in one’s
‘capability’ to perform satisfactorily) and (3) shared appropriate values (sharing
values across organization such as trust and valuing diversity).
In the study completed by Zeleti and Ojo (2014) and Bhatt and Grover (2005),
three types of organization capabilities are introduced based on the well-known
edicts of Resource-Based View and Dynamic Resource-Based Theory (Helfat and
Peteraf 2003). The Resource-Based View naturally evolved into studying how
intangible resources, such as intellectual assets, could be leveraged in order to
accelerate organizational learning and competitive advantage (Oliveira et al. 2002;
Bharadwaj 2000). Dynamic Resource-Based Theory simply facilitates the evolution
of these capabilities over time (Helfat and Peteraf 2003). The three capability types
include (1) Value capabilities, (2) Competitive capabilities and (3) Dynamic capa-
bilities (Bhatt and Grover 2005).
Value Capability this includes capabilities that are characterized by value, heterogene-
ity, and imperfect mobility. Value capabilities are necessary to produce value which was
promised by the organization. Value capabilities include all capabilities which assist an
organization to deliver the organization value to the customers. Value capabilities are not
source of competitive advantage as they only produce the promised necessary customer
value. For example, IT infrastructure falls into this type of capability. IT infrastructure
has been described as an important organization capability that can be an effective
source of value (Bhatt and Grover 2005; Bharadwaj 2000; Zeleti and Ojo 2014).
Dynamic Capabilities this includes capabilities involved in dynamic nature of com-
petitive environment. The concept of dynamic capability reflects the ability of the orga-
138 F.A. Zeleti and A. Ojo
nization to renew capabilities (integrate, build, and reconfigure internal and external
competences (Helfat and Peteraf 2003)) to address rapidly changing environments.
Dynamic capabilities capture the ability to search, explore, acquire, assimilate, and
apply knowledge about resources, opportunities, and how resources can be configured
to exploit opportunities (Bhatt and Grover 2005). Dynamic capability can also facilitate
branching of other capabilities as it is changing the organization capabilities. According
to Brits (2006), differential performance of organisations over time is because of their
capacity in the (1) accumulation, (2) deployment, (3) renewal, (4) reconfiguration of
resources in response to changes in the internal and external environment, (5) Attempts
to explain the process of how capabilities are created, (6) Emphasises the strategic
value of higher order resources because of its dynamic nature, and (7) Renewal of core
competencies and competitive advantage. For example, Research and Development
capability falls into this type of capability (Helfat and Peteraf 2003).
Competitive Capability this includes capabilities that foster the organization competi-
tive advantage and allow organizations stay competitive. These capabilities also impact
the future competitive capabilities because of the dynamic and long-term effect (Bhatt
and Grover 2005; Oliveira et al. 2002). For example, IT strategic choices fall into this
type of capability. IT strategic choices are source of competitive advantage because
they develop through years of experience by learning by doing (Oliveira et al. 2002).
IT experience allows the organization the ability to integrate IT strategy and organiza-
tion strategy, develop reliable and cost-effective systems for the organization and antic-
ipate organization needs sooner than the competitors (Bhatt and Grover 2005).
There are other types of capabilities such as competitive capabilities and dynamic
capabilities. In dividing capabilities, it is very important to distinguish the differ-
ence between capabilities that produce value and capabilities that are source of
competitive advantage and capabilities that are dynamic. Value capabilities are
effective and primary source of value while competitive and dynamic capabilities
are secondary which means value capability is prerequisite and is necessary for the
others to occur (Bhatt and Grover 2005; Zeleti and Ojo 2014).
Summary of the capability types and capabilities associated with each type is
illustrated in Table 1. Dynamic capability and competitive capability are briefly dis-
cussed in this section. The six types of value capabilities are described in section
“Value Capabilities: Theory Background”.
Value Capability
Individual/Competences Jaques and Stamp (1995) define it as the extent and com-
plexity of the context within which an individual can operate. For example, specific
employee skills required performing a specific task.
Business Process is a collection of related, structured activities or tasks that produce
a specific service or product for a particular customer or customers. For example,
Standardization and harmonization process, validation and visualization process
(Steiner et al. 1997; Symphony Technologies Pvt Ltd n.d.).
Organization refers to the way systems and people in the organization work
together to get things done. For example collaboration mechanisms, organization-
specific competencies (marketing, finance, etc.), employees motivation, efforts
towards organizational goal, adaptability, and flexibility, creativity and innovation
(Ambrosini and Bowman 2009).
IT Infrastructure IT infrastructure can provide an organization the ability to share
information across the organization (Bhatt and Grover 2005). Another word, IT
infrastructure is the technological foundation of equipment, computer, communica-
tions, data and basic systems used in common across an organization. It includes
software (ERP), internal and external network resources (servers and switches) and
services (software setup, help desk and computer administration) (Bhatt and Grover
2005; Bharadwaj 2000; Mithas et al. 2009; Gheysari et al. 2012; Xia and King 2002).
Technological Infrastructure technology is knowledge embedded in products and
processes on doing practical things, especially producing things or data. It includes
any sensor-based devices, sensing/sensor phones and smart grids (Gheysari et al.
2012; Arnold and Thuriaux 1997; Brazilian National Council for Scientific and
Technological Development (BNCSTD) 2011).
Management/Governance is about controlling things (people and resources) and
action of governing the organization. Different management skills or actions might
be required for different stages of the value chain. For example, staffing, training
programs, compensation, a quick response accepting additional data for advanced
features, technical management expertise and managing risks (Ambrosini and
Bowman 2009).
Dynamic Capability
Process Innovation is required to improve the processes for the production of new
product or output (Verworn and Herstatt 2002).
Knowledge Management encompasses identifying and mapping intellectual assets
within the organization, generating new knowledge for competitive advantage,
making vast amounts of corporate information accessible, sharing of best practices,
and technology (Tanriverdi 2005; Easterby-Smith and Prieto 2008).
140 F.A. Zeleti and A. Ojo
Manufacturing Performance are characterized by the set of practices in use for the
manufacturing system (Hallgren 2007).
Supply Chain Integration enables firms to share information with their chain to
create supply partners information-based approaches for superior demand planning,
for the staging and movement of physical products, and for streamlining volumi-
nous and complex financial work processes (Rai et al. 2006).
Competitive Capability
IT (Strategic Choices) are the main forces for competitive advantage. IT strategies
increase competitive pressure in the marketplace (Xia and King 2002).
Manufacturing Strategy are set of strategies organizations define for improvement
of manufacturing processes and performance (Hallgren 2007).
Business Operational capability of the whole system (organization) to operate
locally/globally (Cepeda and Vera 2007).
In addition, all the aforementioned capability types and their associated
capabilities have lifecycle which indicates the potential for development of a
capability over time. According to Helfat and Peteraf (2003), capability life-
cycle has three stages which are (1) Founding (capability is identified and
starts functioning), (2) Development (capability is developed gradually over
time) and (3) Maturity (capability meets its highest level of functionality and
impact) (Helfat and Peteraf 2003). Not all capabilities may reach the maturity
stage due to poor development.
Moreover, all aforesaid aspects of capabilities are found in both the Resource-Based
View and Dynamic Resource-Based View. That is why both are seen as essential
towards identifying, managing and in the end modelling organization capabilities.
In classifying capabilities, it is important to distinguish between those that have
value and those that can be a source of competitive advantage (Fig. 1). Value capa-
bility is necessary for the competitive advantage to occur but value capability alone
does not lead to competitiveness of an organization. Bhatt and Grover (2005) argue
that competitive capabilities are not only valuable but heterogeneously distributed
and difficult to transfer. Further, Bhatt and Grover (2005) argue that competitive
capability is a major source of competitive advantage of an organization. On the
same page, Mata et al. (2013) claim that an organization is said to have a competi-
tive advantage when it is deploying its DCs sufficiently. For example, an organiza-
tion has competitive advantage if it is formulating and implementing a strategy,
which is not simultaneously implemented by many other organizations and where
these other organizations face significant disadvantages in acquiring the resources
necessary to implement this strategy.
Although value and competitive capabilities are important, we also need to recog-
nize the dynamic nature of both. Organizations that are involved in such (dynamic)
activities have greater absorptive capacity and can build and renew value and com-
petitive capabilities, which could be a source of competitive advantage. For example,
knowledge management which is a dynamic concept involves accumulation, sharing,
and application of knowledge which could be heterogeneous across organizations,
and thereby also a source of competitive advantage (Bhatt and Grover 2005).
Value Chain
For a better understanding of the activities through which an organization creates and
develops value for shareholders, it is useful to separate the organization system into a
series of value-generating activities known as the value chain (Brits et al. 2007). Value
chain consists of stages of the process of creating value for stakeholders (Rayport and
Sviokla 1995). Value chain as described in Rayport and Sviokla (1995) is a model to
describe a series of value-adding activities and processes known as value production
connecting an organization’s supply side to its demand side. Moreover, value chain
offers organizations a means by which they can evaluate both existing and new strate-
gic opportunities to create customer value (Walters and Rainbird 2007).
Organizations should oversee a physical value chain but, they must also build
and exploit a virtual value chain. This is possible in three stages (1) Visibility (orga-
nization can see physical operations more effectively through information), (2)
Mirroring capabilities (organization substitute virtual activities for physical activi-
ties and they start building capabilities) and (3) Customer relationship (organiza-
tions can deliver value to customers in new ways based on the flow of information
in the virtual value chain). The three processes (Fig. 2) show that physical value
chain is linear and deliberates on physical activities while virtual value chain is
nonlinear and deliberates on information space or the flow of information in an
organization (Rayport and Sviokla 1995).
Mirroring Customer
Visibility
Capabilities Relationship
Given the sufficiency in research in the area of the value chain, three well-known
value chain frameworks are illustrated. Porter’s Value Chain which lies on the concept
of physical value chain of the organization (Bhatt and Emdad 2001; Porter 1985);
Rayport and Sviokla’s Value Chain which lies on the concept of virtual value chain of
the organization (Rayport and Sviokla 1995) and the Open Government Data Value
Chain which lies on the concept of Public Sector Information (Ubaldi 2013).
Below we present the physical and virtual value chain, the value chain that inte-
grates both physical and virtual, and the Open Government Data Value Chain. As
we employ the context of Open Government Data Value chain in this research – due
to the nature of our research – Open Government Data Value chain is extensively
explained in section “Synthesis of an Open Data Dynamic Capability Framework”.
Porter’s Conceptualization of Value Chain Porter’s value chain lies in the con-
cept of the physical value chain of the firm (University of Cambridge 2016) which
means that value chain is targeted toward manufacturing firms in which value of the
organization activities are mostly concerned with the physical flow of material
(Bhatt and Emdad 2005). Porter’s value chain consists of two sets of activities: pri-
mary and secondary activities.
Primary Activities this includes Inbound Logistics or Input (receiving, storing, and
disseminating inputs to the product), Process or Operation (transforming inputs into
the final product), Outbound Logistics or Output (collecting, storing, and physically
distributing the product to buyers), Marketing and Sales or Share (providing a
means by which buyers can purchase the product and inducing them to do so),
Service or Maintain (providing service to enhance or maintain the value of the prod-
uct) (Finne 1997).
Secondary/Support Activities this includes Procurement (the function of purchas-
ing inputs used in the organization value chain), Human Resources Management
(the recruiting, hiring, training, development, and compensation of all types of
personnel), Technology Development (know-how, procedures, or technology
embodied in process equipment) and Infrastructure (general management, planning,
finance, accounting, legal, government affairs, and quality management which sup-
port the entire chain and not individual activities) (Porter 1985; University of
Cambridge 2016; Julien 2012; W3C Brazil 2012).
Figure 3 shows Porter’s value chain.
Rayport and Sviokla’s Conceptualization of Value Chain Rayport and Sviokla
value chain lie on the concept of the virtual value chain of the firm which means
‘information’ play the key role in the chain. The virtual value chain is all about
utilizing information to enhance the value chain. Therefore, in the virtual value
chain, strategic decisions, and activities are built around information (Rayport and
Sviokla 1995; Bhatt and Emdad 2005).
According to Rayport and Sviokla (1995), a virtual value chain consists of five
stages; Gathering, Organizing, Selecting, Synthesizing, and Distributing. Figure 4
shows Rayport and Sviokla’s value chain.
Capability Development in Open Data-Driven Organizations 143
Primary Activities
Inbound Operations Outbound Marketing
Logistics Logistics & Sales Service
Support Activities
Firm Infrastructure
Human resources management
Technology development
Procurement
Fig. 4 Rayport and Sviokla virtual value chain (Rayport and Sviokla 1995)
Phase 1: Data Generation according to OECD working paper (Ubaldi 2013), data
generation phase covers all capabilities required for generating data. This phase
requires capabilities related to ‘generating data’, for example, technologies to col-
lect substantial amount of data.
Phase 2: Data Collection, Aggregation and Processing Raw data may not have
enough quality and meaning to be used, therefore; as it was reported in OECD
working paper (Ubaldi 2013); data need to be aggregated, linked, and or manipu-
lated in order to add value before being open and freely distributed. This phase
requires capabilities related to ‘data processing’ and ‘data storage and computing
facilities’, for example, data cleansing, mash-up, analysis, invalid or duplicate data
deletion, standardization. Moreover, data storage and computing facilities are
necessary to be pooled together for efficiency of data processing and aggregating,
for example, computing facilities.
Phase 3: Data Distribution and Delivery according to OECD working paper
(Ubaldi 2013), data processed need to be distributed to enable access and re-use.
Public sector entities and other organizations are obliged to define precise publish-
ing solutions, providing access to data and APIs and ultimately releasing data. This
phase requires capabilities related to ‘publishing solution’, ‘providing access to data
and APIs’ and ‘data release’, for example, publishing as linked data, data exposure
via APIs and proactively releasing data.
Phase 4: Final Data use data previously distributed need to be re-used by different
users to sustain public value creation (Ubaldi 2013). This phase requires capabilities
related to ‘data retrieval’ and ‘data usage’, for example, guidelines on how to use
data and supporting intermediaries.
The next section underlines what each aforementioned capability areas mean and
briefly describes capabilities associated with them.
Organization is seen as a tree, “mission and vision feed the tree, core competencies
serve as roots and processes produce the fruits in terms of services and products”
(Brits et al. 2007). According to Brits (2006), with the organization capability
framework insight, an organization should be analyzed to extract the critical organi-
zational information. This information includes (1) strategic artefacts such as vision,
mission, objectives and goals, (2) organization entities such as suppliers and cus-
tomers, (3) organization rules such as facts, derivations and definitions and (4) orga-
nization processes such as corporate, business units and operational. This information
will serve as the foundation to construct business capability framework.
According to Keller (2010), capability is the ability to perform actions and in the
context of the organization, capabilities:
• are the building blocks of the organization;
• represent stable organizational functions;
Capability Development in Open Data-Driven Organizations 145
• are unique and independent of each other, however, one’s result affect the
others;
• are abstracted from the organizational model;
• capture the organizational interests.
Keller (2010) further presents a top-level capability model that he claims to be
applicable to any industry and organization domain. Model is shown in Fig. 5.
While Keller’s capability model captures top-level capabilities, the model
described in Moller and Torronen (2003) presents a more detailed framework
(Fig. 6) describing the capabilities organizations require for value production.
As shown in Fig. 6, value production is obviously dependent on the capabilities
organization defines. Being able to produce core value is a necessary condition for
achieving innovation and competitive advantage. Figure 6 clearly highlighted that
different capabilities are the essence of value adding processes and value produc-
tion. For example, a supplier with broad knowledge of process improvement and its
respective capabilities could come up with more effective and efficient products and
services for the customers (Moller and Torronen 2003).
D. Logistic providers
C. Suppliers
5. Collaborate
3. Deliver 4. Plan and
product/ service manage the
business
Fig. 6 Capability base and value production (Moller and Torronen 2003)
146 F.A. Zeleti and A. Ojo
To be competitive and generating robust and thriving revenue streams, open data-
driven organizations tend to increase efficiency and effectiveness in respect to
value-adding processes related to generating data, processing data and re-using
data. To increase efficiency and effectiveness, organizations are required to identify
set of capabilities. List of open data value capability areas have been identified and
extracted from open data literature. Capability areas are ‘data generation’, ‘data
processing’, ‘data storage and computing facilities’, ‘data release’, ‘providing
access to data and APIs’, ‘publishing solution’, ‘data retrieval’ and ‘data usage’.
Data Generation This capability is associated with generating a new set of data
from existing information, text, other raw data or from any device or software col-
lecting data. This can include data generation from sensors or smart grids.
Data Processing This capability is associated with processing the generated or the
original data to meet its potential purpose of use. Examples in this vein include
utilizing processing software to mash-up of original data with other sources of
information, harmonization of data with a specific application and cataloging data
to suit the expected need and to the fruitful use of such data to enhance the organiza-
tion (Ferro and Osella 2013).
Data Storage and Computing Facilities This capability is associated with data stor-
age and back-ups such as storage capacity and computing facilities such as com-
puter hardware or software, computer networks and communications systems and
all networking and communications provision including connections to external
computers. It is essential for an organization to estimate data storage and computing
capacity appropriately to ensure data quality.
Data Release This capability is associated with the release of processed data to its
users to enable data reuse. Capabilities such as data structuring, classification, and
regular update. Data is considered as a good. Therefore, data should be released by
the data release rules and regulations of a particular organization (Zuiderwijk et al.
2015; HM Government 2013).
Providing Access to Data and APIs This capability is associated with availability
and accessibility of APIs to outside- organization users such as developers.
Capabilities such as API development, data exposure via GUI and APIs and testing
and bug fixing. There is still plenty more to do on making more data and APIs acces-
sible (HM Government 2013).
Publishing Solution This capability is associated with publishing data in compel-
ling formats which require methods and mechanisms. For example, publishing as
Linked Data is one publishing solution.
Data Retrieval This capability is associated with data query. This includes extract-
ing the requested data from data storage or datasets. This process requires sophisti-
cated querying and appropriate planning for data retrieval.
Capability Development in Open Data-Driven Organizations 147
Data Usage This capability is associated with enabling and supporting data users
such as data intermediaries and developers to be able to use data. Data re-use will
enrich the value of data.
Open data value capability areas and capabilities associated with each area are
presented in Table 2.
DCs have their antecedents in the RBV of the organization (Daniel and Wilson
2003). They are those specific physical (e.g., specialized equipment, geographic
location), human (e.g., expertise in chemistry), and organizational (e.g., superior
sales force) resources that can be used to implement value-creating strategies
(Wernerfelt 1984). They include the local abilities or competencies’ that are funda-
mental to the competitive advantage of an organization (Eisenhardt and Martin
2000; Alsos et al. 2008). Distinctive processes support the creation, protection, and
augmentation of organization-specific resources and competencies (Griffith and
Harvey 2013). If an organization possesses processes, resources, and competencies
but lacks DCs, it has a chance to make a competitive return for a short period, but
superior returns cannot be sustained. The possession and deployment of DCs pro-
vide the business enterprise with a chance to generate superior profitability over the
longer run. When organizations are dynamically competitive, management will be
active at sensing and seizing opportunities (Augier and Teece 2009).
According to Griffith and Harvey (2013), “DCs are derived from an organization
leveraging its internal and external resources which in turn enhance its power in its
global relationships, thereby enabling it to coordinate inter-organizational activities
and respond rapidly, in a flexible manner, to global competitors' strategies”.
Therefore, the organization has to be continuously alert and in a process of identify-
ing and exploiting new opportunities in order to transform its resources effectively
into new competitive advantages (Alsos et al. 2008).
A study by Teece and Pisano (1994) advances the argument that the capabilities of
an organization rest on three main constructs: organizational processes, position,
and the path/strategies available to it and these capabilities can provide competi-
tive advantage. Teece (2014) also identified the core building blocks of DCs under
the tripartite rubrics of processes, positions, and path/strategies. Below, we classify
the dynamic capability types (Table 1) according to the three main constructs.
1. Process
According to Teece and Pisano (1994), the organizational process is referred to
as the way things are done in the organization or what is called the organization’s
148 F.A. Zeleti and A. Ojo
Table 2 (continued)
OD value capability areas and capabilities
associated with each area References
Data exposure via APIs Ferro and Osella (2013) and Musings
(2012)
Freeing data Julien (2012)
API development Musings (2012)
Using APIs Musings (2012)
Testing and Bug Fixing Musings (2012)
Data change feed Musings (2012)
Publishing Solution
Publishing as Linked Data Julien (2012)
Sustainable Publishing Solution Ferro and Osella (2013)
Publishing in different format; machine-readable van den Broek et al. (2012), Ubaldi
data (2013), Julien (2012) and Musings (2012)
Publishing on the web as API to be queried or data Julien (2012)
dump to be downloaded as a whole
Development of software tools to visualize and Julien (2012)
create API services on the web
Data Retrieval
Sophisticated Querying Musings (2012)
Data Usage
Help and guideline on accessing, using and adding van den Broek et al. (2012), Julien (2012)
data, information or knowledge to the original data and Musings (2012)
source
Available data on the Web to the public and in Rojas et al. (2013)
formats that citizens can reuse
Support data intermediaries van den Broek et al. (2012)
A general search engine helping to locate data Julien (2012)
Dedicated service searching purely for datasets and Julien (2012)
providing useful categorization and tagging
breaking strategic logic of change (Eisenhardt and Martin 2000). Although often
neglected, jettisoned resource combinations that no longer provide competitive
advantage are also critical DCs as markets undergo change (Eisenhardt and Martin
2000). The organization’s processes and positions collectively encompass its capa-
bilities or competencies. According to Ambrosini et al. (2009) “DCs are built rather
than bought in the market”, and they include organizational processes or ‘routines’
that are employed to reconfigure or to combine the organization’s resources and
paths which are the choices open to the organization today and likely to be in the
future. The main four processes are reconfiguration – transformation and recombi-
nation of resources and resources; leveraging – extending a resource by deploying
it into a new domain; learning – allows tasks to be performed more effectively and
efficiently; and integration – ability of the organization to integrate and coordinate
its resources and resources). Similarly, Teece (2014) identifies the three classes of
processes that are relevant to DCs: integration, guided learning, and
reconfiguration/transformation.
In the same vein, Teece et al. (1997) identified technological, complementary
(technological or otherwise), financial, reputational, market structure, and institu-
tional resources. Teece and Pisano (1994) further claim the importance of external
integration and sourcing, integration of external activities and technologies, and
reconfiguration of resources on the competitiveness of the organization. In rapidly
changing environments, there is obviously value in the ability to sense the need to
reconfigure the organization’s processes, and to accomplish the necessary internal
and external transformation (Teece and Pisano 1994). The processes of organiza-
tional renewal are essential for the long-term survival and prosperity of the business
organization. Enterprises must also combine the exploration of new opportunities
with exploitation and renewal (Augier and Teece 2009). This requires constant sur-
veillance of markets and technologies and the willingness to adopt best practices. In
this regard, benchmarking is of considerable value as an organized process for
accomplishing such ends (Teece and Pisano 1994).
Knowledge Management The essence of the DCs approach is that competitive
success arises from the continuous development, alignment, and reconfiguration of
organization-specific resources (Griffith and Harvey 2013; Augier and Teece 2009).
This broader organizational capability is concerned with how organizations create
and/or access new knowledge (Augier and Teece 2007; Bhatt and Grover 2005;
Griffith and Harvey 2013), search, explore, acquire, assimilate, and apply knowl-
edge about resources (Vivas López 2005; Griffith and Harvey 2013), opportunities,
and how resources can be configured to exploit opportunities (Augier and Teece
2009), and how organization makes investment choices, and achieve necessary busi-
ness model and organizational transformation. This is referred to as the ‘intensity of
organizational learning’, which involves accumulation, sharing, and application of
knowledge (Bhatt and Grover 2005). Some scholars including Weerawardena et al.
(2007) suggest that whilst market-based learning enables the organization to learn
what the market needs, the organization must acquire knowledge from other sources
to develop leading edge innovative products and services that will fulfill organiza-
tion’s needs. This is called ‘acquisition’ and is one of the additional knowledge
Capability Development in Open Data-Driven Organizations 151
kets shape organizations however, these alone do not shape markets and provide
information manager needs to implement goals. The organization and managers also
require organization-level responses by competitors, suppliers, and customers
(Augier and Teece 2009; Griffith and Harvey 2013). The network relationship spe-
cifically with the suppliers plays a significant role in enhancing the supplier network,
sensing and seizing opportunities, knowledge creation, resource configuration and
integration and know-how exchange (Augier and Teece 2009; Teece 2007).
3. Path / Strategy
Path is refered to as the strategic alternatives available to the organization (Teece and
Pisano 1994). Authors of Griffith and Harvey (2013) highlight both internal (RBV) and
external (Market Based View) resources which provide the power basis necessary for
developing strategies. To better understand and achieve the necessary capabilities for
renewing organizational path or strategies, Managerial Strategic Functions or capabili-
ties are required. Managerial Strategic Functions is presented below.
Managerial Strategic Functions Once an organization is established, continuing
to succeed in an open competitive economy requires high management and employ-
ees skills with capacities to combine and integrate (Augier and Teece 2007, 2009;
Alsos et al. 2008). In particular, managers must think strategically and execute flaw-
lessly (Augier and Teece 2009) to access niche markets and for building market
positioning (Weerawardena et al. 2007) if they want to succeed (Augier and Teece
2009). They must also figure out how to harness competences (Daniel and Wilson
2003) and the skills of highly skilled employees who play a much more significant
role in creative success and performance of the organization. Survival of a organiza-
tion is not only about executing well but, about figuring out where to put resources,
realizing opportunities and then moving on when competition arises (Augier and
Teece 2009). Such capabilities, if built, constitute the DCs of a organization through
allowing managers to strategically combine, recombine, and reconfigure resources
and resources inside and outside of the organization’s boundaries in order to gener-
ate and exploit strategic internal and external organization-specific competences
(Augier and Teece 2009). Not many managers have the necessary skills, and fewer
still succeed in building them into their businesses (Augier and Teece 2009; Daniel
and Wilson 2003).
From the DC Theory literature, we identify three core constructs: Process, Position,
and Path or Strategy. Based on extensive literature review of the domain, we have
found sub-constructs and related dimensions to each sub-construct. In addition, pre-
vious studies show and investigate four types of DCs (Table 1). Here, we categorize
DCs into five types: Process Innovation, Knowledge Management, Manufacturing
Performance, Supply Chain Integration, and Strategic Managerial Function. In
Table 3, we establish relations between the three main DCs constructs, DCs types,
DCs sub-constructs and their respective dimensions.
Table 3 DCs constructs linked to the dynamic capability types
Three Constructs of DCs Dynamic Sub-Constructs
(Brazilian National Council for Capability Types (Tanriverdi 2005;
Scientific and Technological (Bhatt and Verworn and
Development (BNCSTD) 2011; Grover 2005) Herstatt 2002)
Verworn and Herstatt 2002) Dimensions
Process Process Integration Integrating and Adapting Resource; Combine the Exploration of new
Innovation Opportunities with Exploitation and Strategic Renewal
Leveraging Extending, Building and Releasing Resources by Deploying it into a new
Domain; Leveraging Critical Co-specialized Resources; Innovative Capability
Reconfiguration Transformation, Recombination, and Reconfiguration of Resources; Jettison
un-necessary Resources; Quick Response to Strategic Renewal; Innovative
Capability
Knowledge Learning and Coordinate Resources; Resource Alignment: Resource Orchestration;
Management Knowledge Mngt. Resource Continuous Development; Search, Explore, Configure, Acquire,
Share, Assimilate, Accumulation, Integrate, and Apply knowledge; Strength
Organizational Learning; Disseminate Information; Unlearning Routines;
Unlearning Knowledge-based Practices; Creativity and Idea Management;
Integrate Internally Generated Information
Capability Development in Open Data-Driven Organizations
Drawing on existing empirical findings (Wang and Ahmed 2007), we identify three
main stages for DCs: (1) Adaptive capability, (2) Absorptive capability and (3)
Innovative capability.
Adaptive Capabilities (Search/Variations/External Observation) DCs which
monitor the environment, to discover external knowledge (Büchel and Sorell 2012)
and new possibilities. Searching for new ideas in this manner can provide an insight
into how existing problems or new challenges may be managed and solved. To
reveal the potential in the environment may be said to be the core of all entrepre-
neurial and innovative activities. An organization must have the ability to appraise
the environment so as to constantly develop new ideas and business opportunities.
This adaptive ability to appraise markets and technologies, and the willingness to
adopt best practice, are therefore important (Alsos et al. 2008). In addition, adaptive
capabilities can also help to trigger and guide strategic renewal processes (Rouse
and Ziestma 2008).
Absorptive Capabilities (Selection/Evaluation/Acquisition) DCs which recog-
nize the value of new, external information, assimilate it, and apply it to commercial
ends (Wang and Ahmed 2007). Knowledge may be acquired through external
contacts and connections. However, the acquisition of new knowledge is very time
consuming and challenging as there may be considerable risks involved in investing
in new acquired knowledge (Alsos et al. 2008). Yet, the expectations of advantages
derived from new ideas may be achieved by analysis and debate concerning the
values and risks. The ideas are evaluated on the basis of previous experience, exper-
tise, and capabilities (Alsos et al. 2008). Stronger ability of learning from partners,
integrating external information and transforming it into organization-embedded
knowledge are the outcome of organizations with higher absorptive capability
(Wang and Ahmed 2007).
Innovative Capabilities (Routinisation/Implementation/Reconfiguration and
Renewal) DCs comprise product development routines, development and launch
of new profitable products and services, strategic decision-making, introduce, com-
bine or modify resources, and integrate new resources (Alsos et al. 2008) with inno-
vative behaviors and processes (Wang and Ahmed 2007). This includes implementing
newly approved initiatives to change within the organization and provides the
opportunity to reorganize the organization’s resources and the possibility to experi-
ment with new ideas (Alsos et al. 2008). In this stage, organization puts the ideas
from the selection phase into place in a competitive business platform. Thus, the
process ends in a form of utilization through the implementation of the ideas (Alsos
et al. 2008). Authors of (Wang and Ahmed 2007) suggest a range of possible inno-
vative alternatives, such as developing new products or services, developing new
methods of production, identifying new markets, discovering new sources of supply
and developing new organizational forms.
Capability Development in Open Data-Driven Organizations 155
In this section, we define what DCT constructs (Process, Position, and Path) means
in each the dynamic capability stages – Adaptive, Absorptive, and Innovative. In
Table 4, we specify nine conditions for agility of ODDCs following the descriptions
in section “Dynamic Capabilities: Theory Background”. In addition, corresponding
propositions are provided to succinctly capture these conditions.
Adaptive Position Capabilities (Sense and Search) ODDCs which monitor the
environment, to discover external knowledge (Büchel and Sorell 2012) and new
possibilities for positioning ODDOs in the OD industry. This capability includes:
Search for knowledge that can be acquired from OD ecosystem such as maturity of
Case Studies
To profoundly understand the open data value and dynamic capability framework in
practice, this research conducts two case studies (open-ended interview) on both
public and private ODDOs in Ireland.
This section describes findings from the use of the operationalization in Table 2 in
investigating the VCs of Xpreso; one of the ODDOs in the downstream sector of the
Irish OD Ecosystem. This case is used to validate our value capability framework oper-
ationalization and identification of critical factors based on the interviewee’s opinion.
158 F.A. Zeleti and A. Ojo
Equally important is to note that Xpreso verifies the significance of utilizing the
capability framework in open data-driven organizations more especially start-ups in
order to generate and capture the real value of data. With reference to this, the inter-
viewee adds:
The framework allowed us to, for the first time, fully examine the processes required to
produce and publish the datasets which we are considering, which helped to greatly clarify
the requirements of producing such data. It also allowed us to examine our organization
from in terms of the key business areas described in the framework, which we found to be
a very useful tool in its own right.
Consequently, Xpreso has never carried out any capability audit before engaging
in this research, and they find the capability framework a very useful tool to identify,
create, develop and manage open data capabilities.
Table 5 Xpreso’s capability framework
Data Data Collection, Aggregation and
Generation Processing Data Distribution and Delivery Final data Use
Data Storage Providing
Generating & Computing Access to Data Publishing Data
Data Data Processing Facilities Data Release & APIs Solution Data Retrieval Usage
Individual/ App- Capability to Understanding Have member Understanding Competency
competences development, efficiently process of Database of OSM of REST, with REST
Competency geographical data architecture community Understanding APIs and OSM
with GPS understanding of GIS, verify data of API design APIs,
technology, Understanding of integrity Knowledge of
Ability to OSM data-formats, Xpreso API
produce Ability to aggregate implementation
reliable API, Lat-Lng occurrences, specifics
Ability to add Understanding of
GPS recording Database Architecture
redundancy
Business Address data, Geocoding, Parcel Data is
Process Parcel data data are correlated to automatically
source, Drivers GPS data from uploaded to
Capability Development in Open Data-Driven Organizations
This section describes findings from the use of the operationalization in Table 4 in
investigating the available DCs and their relative importance at the Irish Marine
Institute; one of the major players in the upstream sector of the Irish OD Ecosystem.
This case is used to validate our dynamic capability framework operationalization
and identification of critical factors based on expert opinion.
The Marine Institute is an agency which operates under the aegis of the Department
of Agriculture, Food and Marine (DAFM), and the national agency, responsible for
undertaking marine research and development that critically informs policy, regula-
tory objectives, management and sustainable development strategies for Ireland’s
marine resources. The Marine Institute Act states that the Institute will have the
following general functions:
to undertake, to co-ordinate, to promote and to assist in marine researchand development
and to provide such services related to marine research and development that in the opinion
of the Institute will promote economic developmentand create employment and protect the
environment
During the interview with Marine Institute, a variety of ODDCs and identified:
Adaptive Position Capabilities Marine Institute’s DCs include: Searching for exist-
ing OD Products and services, technological opportunities, potential government
agencies, potential partners for collaboration, skills and expertise necessary, exist-
ing interest groups or agencies, future market.
Absorptive Position Capabilities Marine Institute’s DCs include: Marine data col-
lection; developing agreement with other government agencies for data services;
identifying new technologies, platforms and applications; Identifying skills and
expertise required.
Innovative Position Capabilities Marine Institute’s DCs include: Producing as
much marine data as possible; leading provider of data on marine environment and
Data Cataloging in Ireland; Feeds information into making decisions and support
growing resources of marine environment; Experts for technologies for provision of
162 F.A. Zeleti and A. Ojo
Adaptive Process Capabilities Marine Institute’s DCs include: Searching for exist-
ing knowledge from outside organization in order to identify potential processes,
OD standards and European directives, linked-data opportunities, data management
tools, set of requirements to develop prototypes, best practices around adding value
to data and processes, agencies and companies for resource exchange and integra-
tion, other potential project resources; aquaculture process opportunities; discover-
ing new online cataloging systems.
Absorptive Process Capabilities Marine Institute’s DCs include: Assess and evalu-
ate processes, platforms, and applications in order to define potential tools for add-
ing value to marine data; Adopting new online cataloging systems such as
GeoNetwork; assessing and identifying series of best practices (W3C best prac-
tices); Open license for Marine Institute; Developing a set of technical requirements
and specifications for developing the planned prototype; adopting appropriate data
standards such as ISO 19139 (Data Standardization) and ISO 19156 (Observation
and Measurements); adopting appropriate European Directives such as OD Standard
for Inspire Directive, Standards for Marine Strategy Framework Directive, and Data
Standards for Water Framework Directive.
Innovative Process Capabilities Marine Institute’s DCs include: Improve marine
data management processes; data cataloging; data management processes (using
generic marine related Data Models); software development and project manage-
ment processes; linked marine data; connecting to processes of agencies; develop
and enhance internal processes; utilizing defined specification to develop the proto-
type further; added value to data through new prototype and initiatives such as
Ireland’s Marine Atlas, Irish Spatial Data Exchange, Ireland’s Digital Ocean,
Ireland’s Marine Renewable Energy Portal and connecting to Ireland’s OD Portal;
utilizing evaluated tools such as ERDDAP to add value to marine data and enhance
data cataloging; Encourage and enhance Marine Institute Data License (existing for
11 years); developing new application that handles standards; developing new
application that deliver data to user.
Adaptive Path Capabilities Marine Institute’s DCs include: Searching for smart
strategies from potential and influential actors/players and experts in the industry;
searching for best practices around strategies in general and data strategies in spe-
cific; Seeking new and unique expertise; organize workshops for strategic decision
making; Searching for other organization’s advanced projects In order to identify
new areas and new knowledge.
Capability Development in Open Data-Driven Organizations 163
Absorptive Path Capabilities Marine Institute’s DCs include: Assessing and evalu-
ating identified actors/players and experts and connect with them; Developing new
expertise; Adopting best possible collaborative approach; identifying and assessing
series of best practices that could help organization’s expert groups to define strate-
gic areas and decisions; define strategic objectives or areas to tackle during the
workshop; assess and evaluate new knowledge gained from existing projects and
develop new objectives for organization.
Innovative Path Capabilities Marine Institute’s DCs include: Big and leading con-
tributor of environmental data in Ireland; Strengthening marine environment status;
Close and constant collaboration with companies and continuously providing them
with the data being produced; open and constant communication with expert groups
within organization; Working with and in parallel with big Irish players such as
Sustainable Energy Ireland and Department of Communication Energy and Natural
Resources; Provide high quality support services for marine food safety; use marine
data for service area collaborations; acting as one primary communication forum
between agencies; directive driven organization; standard driven organization; strength-
ening the organization brand; high level and educated employees; share capabilities
within organization and with other agencies; strengthening the collaborative environ-
ment; Organization’s expert groups to make smart decisions; adopting OD strategy best
practices; moving from 3 (CSV) star to 5 star (linked-data); adopting other organiza-
tion’s strategic best practices; follow and maintain citation strategy; encouraging orga-
nization’s employees to use best practices for their tasks; making sure all data are of
high quality and available online; easy and usable data; more datasets; other project
connectivity; always use powerful tools; always be ahead of other agencies; always
monitor market; Access resources from other agencies; to grow jobs in sector.
After carefully analysing the interview, in Table 6, we present a set of critical
factors for developing ODDCs.
Discussion
MA for upstream Process Discovering Data Management Tools Define and assess Tools, Adopt Tools, Processes, Platforms,
organizations and Processes, OD online cataloging Processes, Platforms, and and Applications to add value to the
systems, OD standards and related Applications to add value to Data
European Directives, and Linked- the Data Improve Data Cataloging Process
Data opportunities Define and assess appropriate and (Relational) Data Model
Discovering OD best practices European Directives Improve Data Management,
Discovering Data Licenses Define and assess series of Software Development, and Project
Discovering successful (Relational) best practices Management Processes
Data Models Define and assess OD license Alliance-based processes
Define and assess appropriate Heavily utilize well-known and
Data Standards Define and completed projects for adding more
assess appropriate value to the OD product or service
(Relational) Data Model and to fasten the customer-supplier
communication (Sufficient
delivery)
Active participation to the Local
and National Portals
F.A. Zeleti and A. Ojo
Applicable for Path/Strategy Strategic coordination/collaboration Define strategic objectives Strengthening the collaborative and
downstream (Strategic Alliances) through knowledge gained communicative environment –
organizations Seeking new and unique data/OD from existing projects, Internally and externally (Expert
Applicable for strategic solutions from potential and partners and OD products and groups, projects, and resources)
upstream influential actors/players, successful services Educated human resource –
organizations national and international level Define and assess best Domain specific education
projects, and experts in the industry possible strategic Continuously provide available and
Discovering best practices on data/ collaborative approach easy-to-use high quality OD
OD strategies Assess and evaluate identified products and services and support
Discovering differentiation strategies actors/players and experts in Differentiating OD products and
the field services
Identify new domain experts Maintain standard-driven
Define and assess series of organization status
OD strategic best practice Constantly magnify and strengthen
Define and assess unique organization’s brand
strategy for differentiation Use organization’s capabilities to
identify areas of collaborations
Share capabilities and resources
within organization and with other
organizations
Capability Development in Open Data-Driven Organizations
in this study. During our various attempts in involving other organizations, we dis-
covered that new and small establishments have difficulty understanding different
elements of the framework and that how framework can be used in assisting them in
identifying value capabilities while, the framework was fully understood and appre-
ciated by medium to larger organizations as they could relate to all the elements and
the integration. In parallel, our experience in engaging organizations specially new
and small establishments in this research process show that open data-driven orga-
nizations and start-ups also need to be eager and more engaged with the research
community in the domain if they want to defeat the challenges of the dynamic
market.
Previously in open data and organization literature, no open data value capability
framework exist. Therefore, comparison of the developed open data capability
framework with similar frameworks is not possible. Therefore, we have sought to
analyze this convergence as a form of alignment in which we expect open data value
chain to support directly open data value capability framework to shape open data-
driven organization value capabilities. The analysis of the value capability frame-
work and case study research show that value chain stages (data generation; data
collection, aggregation and processing; data distribution and delivery; and final
data use) extracted from open government data value chain by providing us the
baseline to cluster top-level capabilities have significant impact on shaping open
data value capabilities. As the framework can be explored in practice and by open
data-driven organizations, new open data value capabilities can emerge that can be
useful to other organizations in building their value capabilities (Table 7).
Moreover, we observe that open data-driven organizations out more emphasis on
‘generating data’, ‘data processing’ and ‘publishing solution’ capabilities. This
shows that the open data-driven organizations are more eager to develop capabilities
which result on generating data, processing data and publishing data. Other open
data capability areas receive less attention.
Despite existing critiques and ambiguities in literature on RBV, DCT and DCs, we
have found these three paradigms not as challenging as presented in the literature
and we have been able to join them very adequately and generate an easy-to-
understand dynamic capability framework taking into an account essentialities of
the two theories and the DCs of the organization. The framework can be utilized by
Capability Development in Open Data-Driven Organizations 167
all types of organizations specially ODDOs regardless of their size and maturity
levels. However, we encountered number of challenges in identifying the critical
aspects of the theories and the organizational DCs for addressing the objectives of
our study but, relying on our experience and knowledge in the domain, and by uti-
lizing the critical aspects identified, we developed a framework that can address
critiques and ambiguities exist in the literature. The framework is applicable for all
types of organizations and the application of the framework is fairly simple how-
ever, we have number of observations regarding how framework could be more
effectively used by organizations.
Positioning Upstream and leading organizations have already positioned them-
selves in the market. In this regard, defining organizational position in the market is
mostly applicable for downstream organizations, new entrants to the industry and
small organizations.
Process Development Upstream and leading organizations should constantly seek
and discover powerful data management tools, platforms and technologies and
reconfigure their existing value-added processes due to the fact that many down-
stream and small organizations rely on the OD products and services they provide.
In this regard, engaging in process configuration and reconfiguration is highly rec-
ommended and MA to upstream organizations.
Strategy Development Strategy must go hand in hand with processes. Strategy
needs to be consistent, coherent, and embrace innovation. While it is necessarily
shaped by the legacy of the past, it also shapes the path ahead. Strategy will deter-
mine which products to make, which customers to target, how to deploy the organi-
zation’s resources, what the optimal timing will be, and how to keep competitors at
bay. Downstream organizations must develop an effective strategy and renew
strategies every often in order to defeat the challenges in the dynamic market, while
upstream organizations set long-term strategies.
ODDOs regardless of being at upstream or downstream, need to recognize and
utilize the successful experiences or what is so called “best practices” of other orga-
nizations. It is very essential to develop a collaborative and communicative environ-
ment with other organizations and use organization’s capabilities to identify areas of
collaborations. In OD industry, being connected to a network of organizations plays
a significant role in the success of the organizations in the network as shared resources
and capabilities can lead to innovative OD products and services (Zhenbin Yang and
Kankanhalli 2013; Conradie and Choenni 2014; Wang and Lo 2015). ODDOs’ man-
agers and experts groups should constantly monitor the market for changes in order
to move toward the dynamic market and be able to be ahead of the competitors.
ODDOs must establish strategies to ensure desired participation rate and those orga-
nizational technological and human resources are used in a way that innovative OD
products and services are produced. In addition, the ODDOs should increase the
number of domain-specific experts as in OD industry, domain knowledge is very
vital as many OD products and services are domain specific. However, success hap-
pens in an environment with learning, sharing, and collaborating culture.
168 F.A. Zeleti and A. Ojo
Conclusion
A number of list of political, social and economic benefits have been associated
with the use and impact of OD. The economic aspect of OD has naturally generated
a lot of interest resulting in a number of OD business models. In this paper, we
developed an OD capability matrix as a tool to support design of OD business capa-
bilities. In addition, the OD Capability Matrix could help the implementation of OD
business models. Matrix can assist OD business managers to understand and
describe how capabilities should be utilize and extended throughout the OD value
chain of the business. As there is no OD capability framework in literature, our OD
Capability Matrix provides a significant starting point for OD businesses to plan and
develop the requisite capabilities to support their business models. With time, con-
crete experience from practice will be useful in refining the capability framework.
Regarding future work, our main interest is to develop the framework into a con-
crete tool (similar to the Business Model Canvas) to support OD practitioners. In
this regard, business modelers and managers are encouraged to utilize the Matrix.
From the research perspective, we intend to refine the capability matrix to reflect
maturity levels of OD capabilities. To represent the matrix as a modeling technique,
a potential future work would be to offer more specific guidelines for designing the
matrix. Moreover, we are considering the possibility of generation of OD capability
patterns from the OD capability matrix. Another potential future work would be to
study how capability driven development approach can support evolving OD busi-
nesses and facilitates adjustment of capabilities according to changing context.
References
Alsos GA, Borch OJ, Ljunggren E, Madsen EL (2008) Dynamic capabilities – conceptualization
and operationalization. In: The academy of management conference, pp 1–28
Ambastha A, Momaya K (2004) Competitiveness of firms: review of theory, frameworks and mod-
els. Singapore Manag Rev 26(1):45–61
Ambrosini V, Bowman C (2009) What are dynamic capabilities and are they a useful construct in
strategic management? Int J Manag Rev 11(1):29–49
Ambrosini V, Bowman C, Collier N (2009) Dynamic capabilities: an exploration of how firms
renew their resource base. Br J Manag 20(SUPP. 1):1–41
Arnold E, Thuriaux B (1997) Developing firms’ technological capabilities
Augier M, Teece DJ (2007) Dynamic capabilities and multinational enterprise: Penrosean insights
and omissions. Manag Int Rev 47(2):175–192
Augier M, Teece DJ (2009) Dynamic capabilities and the role of managers in business strategy and
economic performance. Organ Sci 20(2):410–421
Avital M, Bjorn-Andersen N (2012) The value of open government data: a strategic analysis
framework. In: SIG eGovernment pre-ICIS workshop, no. 2002
Bharadwaj AS (2000) A resource-based perspective on information technology capability and firm
performance: an imprical investigation. MIS Q 24(1):169–196
Bhatt GD, Emdad AF (2001) An analysis of the virtual value chain in electronic commerce.
J Enterp Inf Manag 14(1):78–84
Capability Development in Open Data-Driven Organizations 169
Bhatt GD, Emdad AF (2005) An analysis of the virtual value chain in electronic commerce. Logist
Inf Manag 14(1):78–85
Bhatt GD, Grover V (2005) Types of information technology capabilities and their role in competi-
tive advantage: an empirical study. J Manag Inf Syst 22(2):253–277
Brazilian National Council for Scientific and Technological Development (BNCSTD) (2011) The
relationship between technological capability and firm performance in an emerging economy,
Rio
Brits J-P (2006) Conceptual framework for modeling business capabilities. Tshwane University
of Technology
Brits J, Botha GHK, Herselman ME (2007) Conceptual framework for modeling business capa-
bilities. In: Proceedings of the 2007 informing science and IT education joint conference
conceptual
Büchel B, Sorell M (2012) Assessing your adaptive capability: Where do you ‘stand out’ within
your industry
Cepeda G, Vera D (2007) Dynamic capabilities and operational capabilities: a knowledge manage-
ment perspective. J Bus Res 60(5):426–437
Conradie P, Choenni S (2014) On the barriers for local government releasing open data. Gov Inf
Q 31:S10–S17
Daniel EM, Wilson HN (2003) The role of dynamic capabilities in e-business transformation. Eur
J Inf Syst 12(4):282–296
de Vries M (2012) 14 years of PSI policy and its impact. In Madrid conference, February
den Hertog P, van der Aa W, de Jong MW (2010) Capabilities for managing service innovation:
towards a conceptual framework. J Serv Manag 21(4):490–514
Easterby-Smith M, Prieto IM (2008) Dynamic capabilities and knowledge management: an inte-
grative role for learning? Br J Manag 19(3):235–249
Eisenhardt KM, Martin JA (2000) Dynamic capabilities: what are they? Strateg Manag
J 21(10–11):1105–1121
Ferro E, Osella M (2013) Eight business model archetypes for PSI re-use, London
Finne T (1997) Information security implemented in: the theory on stock market efficiency,
Markowitz’ s portfolio theory and porter’ s value chain. Comput Secur 16(May 1967):469–479
Gheysari H, Rasli A, Roghanian P, Jebur H (2012) The role of information technology infrastruc-
ture capability (ITIC) in management. IJFPSS 2(2):36–40
Griffith DA, Harvey MG (2013) A resource perspective of global dynamic capabilities. J Int Bus
Stud 32(3):597–606
Guidoin S The value of open data
Hallgren M (2007) Manufacturing strategy , capabilities and performance. Linköping University,
Linköping
Helfat CE, Peteraf MA (2003) The dynamic resource-based view: capability lifecycles. Strateg
Manag J 24(10):997–1010
HM Government (2013) Seizing the data opportunity: a strategy for UK data capability, UK
Jaques E, Stamp G (1995) Level and type of capability in relation to executive organization, Brunel
Julien N (2012), Business opportunities arising from open data policies. Imperial College London
Keller W (2010) Using capabilities in enterprise architecture management, Lochham
Mata FJ, Fuerst WL, Barney JB, Mata BFJ (2013) Information technology and sustainable com-
petitive advantage: a resource-based analysis. MIS Q 19(4):487–505
Mithas S, Ramasubbu N, Krishnan MS, Sambamurthy V (2009) Information technology infra-
structure capability and firm performance: an empirical analysis, Michigan
Moller KEK, Torronen P (2003) Business suppliers’ value creation potential: a capability-based
analysis. Ind Mark Manag 32:109–118
Musings J (2012) Open data business models. [Online]. Available: http://www.jenitennison.com/
blog/node/172
Oliveira P, Roth AV, Gilland W (2002) Achieving competitive capabilities in e-services. Technol
Forecast Soc Chang. 69:721–739
170 F.A. Zeleti and A. Ojo
Osterwalder A (2004) The business model ontology: a proposition in a design science approach,
UNIVERSITE DE LAUSANNE ECOLE
Pant S, Hsu C (1996) Business on the web: strategies and economics. In: Fifth international world
wide web conference
Porter ME (1985) Competitive advantage: creating and sustaining superior performance. Free
Press, New York
Rai A, Patnayakuni R, Seth N (2006) Firm performance impacts of digitally supply chain integra-
tion capabilities. MIS Q 30(2):225–246
Rayport JF, Sviokla JJ (1995) Exploiting the virtual value chain. McKinsey Q
Rojas LAR, Lovelle JMC, Tarazona GM, Montenegro CE (2013) Open data as a key factor for
developing expert systems: a perspective from Spain. Int J Interact Multimed Artif Intell 2(2):51
Rouse MJ, Ziestma C (2008) Responding to weak signals: the emergence of adaptive dyanmic
capabilities for strategic renewal. In: The international conference on organizational learning,
knowledge and capabilities, pp 1–23
Steiner S, Abraham B, Mackay J (1997) Understanding process capability indices, Waterloo
Symphony Technologies Pvt Ltd, Measuring Your Process Capability, India
Tanriverdi H (2005) Information technology relatedness , knowledge management capability , and
performance of multibusiness firms. MIS Q 29(2):311–334
Teece DJ (2007) Explicating dynamic capabilities: the nature and microfoundation of (sustainable)
enterprise performance. Strateg Manag J 28(August):1319–1350
Teece DJ (2014) A dynamic capabilities-based entrepreneurial theory of the multinational enter-
prise. J Int Bus Stud 45(1):8–37
Teece D, Pisano G (1994) The dynamic capabilities of firms: an introduction. Ind Corp Chang
3(3):537–556
Teece DJ, Pisano G, Slwen A (1997) Dynamic capabilities and strategic management. Strateg
Manag J 18(7):509–533
Townsend P, Cairns L (2003) Developing the global manager using a capability framework. Manag
Learn 34(3):313–327
Ubaldi B (2013) Open government data: towards empirical analysis of open government data
initiatives, 22
University of Cambridge (2016) Porter’s Value Chain. University of Cambridge. [Online].
Available: http://www.ifm.eng.cam.ac.uk/research/dstools/value-chain-/
Verworn B, Herstatt C (2002) The innovation process: an introduction to process models, 12
Vivas López S (2005) Competitive advantage and strategy formulation: the key role of dynamic
capabilities. Manag Decis 43(5):661–669
W3C Brazil (2012) Open government data value chain, Brussels
Walters D, Rainbird M (2007) Cooperative innovation: a value chain approach. J Enterp Inf Manag
20(5):595–607
Wang CL, Ahmed PK (2007) Dynamic capabilities: a review and research agenda. Int J Manag
Rev 9(1):31–51
Wang H, Lo J (2015) Adoption of open government data among government agencies. Gov Inf Q.
Weerawardena J, Mort GS, Liesch PW, Knight G (2007) Conceptualizing accelerated international-
ization in the born global firm: a dynamic capabilities perspective. J World Bus 42(3):294–306
Wernerfelt B (1984) A resource based view of the firm. Strateg Manag J 5(2):171–180
van den Broek T, Rijken M, van Oort S (2012) Towards open development data
Xia W, King WR (2002) Determinants of organizational IT infrastructure capabilities: an empiri-
cal study
Zeleti FA, Ojo A (2014) Capability matrix for open data. In: 15th IFIP working conference on
virtual enterprises.
Zeleti FA, Ojo A, Curry E (2014) Business models for the open data industry: characterization and
analysis of emerging models. In: 15th annual international conference on digital government
research
Capability Development in Open Data-Driven Organizations 171
Zhenbin Yang, Kankanhalli A (2013) Innovation in government services: the case of open data.
IFIP Adv Inf Commun Technol 402:644–651
Zuiderwijk A, Janssen M, Poulis K, van de Kaa G (2015) Open data for competitive advantage:
insights from open data use by companies. In: Proceedings of the 16th annual international
conference on digital government research (dg.o 2015), pp 79–88
Fatemeh Ahmadi Zeleti is a researcher at the e-Government unit at the Insight Centre for Data
Analytics @ National University of Ireland Galway (formerly Digital Enterprise Research Institute
(DERI) – a leading center in Semantic Web and Linked Open Data research). At the Insight Centre,
her research and development work addresses capabilities for open data-driven organizations and
how business (for-profit and nonprofit) entities can effectively leverage [Linked] Open
[Government] Data for optimizing business processes and product and service innovation. In 2016,
her research results have been selected as the best practice to address and achieve European PSI
Directive (2003/98/EC). She is the moderator of the World Bank Open Data Innovation Network
and an active member of the Swedish Research Network in e-Government and ImmigrationPolicy2.0.
Dr. Adegboyega Ojo is Senior Research Fellow at the Insight Centre for Data Analytics, National
University of Ireland Galway (NUIG). He leads the E-Government Group at Insight Centre, serves
as Adjunct Lecturer at the College of Engineering and Informatics. His current research interests
include data driven innovations in government, Open data policies and Infrastructures, data analyt-
ics and governance of smart cities. He is member of the Editorial Boards of the Government
Information Quarterly and International Journal of Public Administration in the Digital Age.
Water Analytics and Management
with Real-Time Linked Dataspaces
Abstract Due to predictions of water scarcity in the future, governments and public
administrations are increasingly looking for innovative solutions to improve water
governance and conservation. The problem is exasperated due to low levels of
awareness about water consumption among the general public. This calls for a
holistic approach to effectively manage resources during all stages of water usage.
Implementation of such an approach heavily relies on advanced analytics technolo-
gies that combine data from different sources to enable decision support and public
engagement. The next-generation of water information management systems must
overcome significant technical challenges including integration of heterogeneous
and real-time data, creation of analytical models for diverse users, and exploitation
of ubiquitous devices to disseminate actionable information. This chapter presents
a new approach for water analytics in public spaces that is built upon the fundamental
concepts of Linked Data technologies. The chapter also presents a concrete realization
of the Linked Data approach through the development of water analytics applications
for buildings in public educational institutions.
Introduction
One of the sustainable development goal set out by the United Nations, as part of its
agenda for 2030, is to ensure availability and sustainable management of water and
sanitation for all (General Assembly, United Nations 2015). Furthermore, recent
projections by the Organization for Economic Cooperation and Development esti-
mate that more than 40% of the world’s population will be living in areas under
severe water stress by 2050 (OECD 2012). This problem is expected to worsen due
to a high global demand for water from manufacturing, thermal electricity genera-
tion and domestic use. Commercial uses of water are depleting the world’s freshwa-
ter supply in both quantity and quality. A key factor contributing towards scarcity of
water is the historical belief that water is not a vital resource that needs to be man-
aged. Nonetheless, a recent study has highlighted the effects of water scarcity on
economic growth (Hertel and Liu 2016). The same study also recommends conserv-
ing water through increased efficiency in existing uses. This underlines a significant
opportunity for research and development of ICT tools to raise awareness, improve
management, and increase conservation of water (Pereira et al. 2003).
In order to manage water holistically, it is important to use decision support tools
that present meaningful and contextual information about usage, pricing, and avail-
ability of water in an intuitive and interactive way. Different users have different
information requirements to manage water, from home users managing their per-
sonal water usage, business users managing the water consumption of their com-
mercial activities, to municipalities managing regional distribution and consumption
at the city level. In order to develop water information services for such diverse
users, it is necessary to leverage knowledge from across a number of different
domains, including metering, collection and catchment management, environmen-
tal, water quality, energy usage, utility information, end-user feedback, occupancy
patterns, meteorological data, etc. However, many barriers exist to interoperability
across domains and there is little interaction between these islands of information.
The design of next-generation water information management systems poses sig-
nificant technical challenges in terms of information management, integration of
heterogeneous data, and real-time processing of dynamic data.
Linked Data technology leverages open protocols and W3C standards for shar-
ing structured data on the Web. In this chapter, we discuss the use of Linked Data as
an enabling technology for water data services. The objective of this approach is to
create an integrated well-connected Real-time Linked Dataspace (Franklin et al.
2005; Heath and Bizer 2011) of information relevant to managing water in public
spaces. Representing water usage data within the Linked Data format makes it open;
thus, allowing it to be easily combined with data from other relevant domain silos.
This chapter describes the fundamentals of the Linked Data approach for water data
services (Curry et al. 2014); in addition, it details a concrete implementation of this
approach for water analytics in public spaces. Section “Motivation” motivates the
need for contextual water information management. Section “Linked Data for Water
Information Integration” introduces the main concepts of the Linked Data approach.
Water Analytics and Management with Real-Time Linked Dataspaces 175
Section “Linked Real-time Dataspace for the Waternomics Project” details the
architecture developed for enabling this approach, in the context of Waternomics
project. Section “Water Management in Public Spaces” describes the pilots used for
testing and validation of proposed approach. Section “Realization of Waternomics
Platform” details the water management applications designed a university building
and a school. Section “Related Work” discusses related literature and section
“Summary” provides a brief summary of this chapter.
Motivation
Sustainability requires information on the use, flows, and destinies of energy, water,
and materials including waste, along with monetary information on environmental
costs, earnings, and savings. This type of information is essential if we are to under-
stand the causal relationships between the various actions that can be taken, and
their impact on sustainability. However, the problem is broad in scope, and the nec-
essary information may not be available, or difficult to collect. Within the context of
water management, improving the sustainability of water consumption, especially
through changing the way a household, organization, or city operates (Curry and
Donnellan 2012). This requires a number of practical steps that will include the
need for a systematic approach for information-gathering and analysis.
One of the key problems of modern water management systems is the lack of data
management and decision support tools that present meaningful and personalized
information about usage, pricing, and availability of water in an intuitive and inter-
active way to end-users. This introduces limitations in the efforts to manage water
as a resource, including:
• User Awareness: End-users do not have access to water information (i.e. avail-
ability, consumption, and pricing) at the moment water consumption decisions
are being taken.
• User Incentives: Due to billing, pricing, awareness, or metering aspects, end-
users may not have an incentive to change their behavior.
• Integrated Information Provision & Analysis: Decision makers do not have
access to information platforms to make organizational changes. Personalized
water information can only be created by combining publicly available water
data with private water usage data that is only available to water service
providers.
• Benchmarking: End-users do not know if their individual water consumption
pattern is high or low compared to others.
176 U. ul Hassan et al.
http://lod-cloud.net/
1
https://tools.ietf.org/html/rfc3986
2
Water Analytics and Management with Real-Time Linked Dataspaces 177
r ecommends the use of Hyper Text Transfer Protocol (HTTP3) for URIs, so that data
can be retrieved from names using standard protocols. The third principle promotes
the use of standard web formats, such as the Resource Description Framework
(RDF4) or the JavaScript Object Notation (JSON-LD5), for making data available
through URIs. The fourth principle encourages contextualization of data by provid-
ing links to other related URIs, thus creating a data network. Within the context of
water analytics, following these principles enables standardized access and supports
interoperability for applications that aim to exploit water information.
The goal of the Waternomics6 project is to provide personalized and actionable infor-
mation about water consumption and water availability to households, companies,
and cities in an intuitive and effective manner at a time-scale that is relevant for effec-
tive decision making (Curry et al. 2014). Access to such information will increase
end-user awareness and improve the quality of the decisions regarding water man-
agement and governance. Waternomics accomplishes this by combining water usage
related information from various sources and domains to offer water information
services to end-users. The Waternomics platform enables sharing of water informa-
tion services across different groups of users by providing a convergence layer on top
of existing water infrastructures with minimal disruption. The objective is to expose
the data within existing systems, but only linking the data when it needs to be shared.
Representing water usage data within the Linked Data format makes it open; thus,
allowing it to be easily combined with data from other relevant domains.
Architecture
The main components of the envisioned architecture, as illustrated in Fig. 1, are the
data sources of water usage on existing metering systems, a dataspace consisting of
Linked Data, a set of support services, and the resulting applications for water
management.
• Water Metering: At the bottom of the architecture are the operational and legacy
information systems. Adapters perform the “RDFization” process, which trans-
forms multiple formats and legacy data to lifts it to the dataspace. Linked Data
principles play a crucial part here since they enable interoperability a cross-linking
3
https://tools.ietf.org/html/rfc2616
4
https://www.w3.org/RDF/
5
http://json-ld.org/
6
http://www.waternomics.eu
178 U. ul Hassan et al.
support services exploit Linked Data technologies and provide additional tools for
aggregation, analysis, and improvement of basic data gathered through water meter-
ing. Furthermore, these services enrich the aggregated data for complex analytical
queries. The primary purpose of these services is to provide Application
Programmable Interfaces (APIs) over the dataspace that can be re-used by applica-
tion developers.
• Water Analytics Applications: At the top of the architecture are the water usage
and management applications that consume the resulting data and events from
the Real-time Linked Dataspace. These applications not only consume the infor-
mation from the dataspace but also generate user-friendly views over the under-
lying data.
The support services play a crucial role in realization and exploitation of the
Real-time Liked Dataspace. These services include but are not limited to:
• Search & Query Service: The query service concerns the technical aspects of
enabling access to the data in the Real-time Linked Dataspace through structured
queries or RESTful API calls. The query service also enables low latency data
analysis. The search service provides keyword-based lookup queries over under-
lying data sources and their descriptions. The objective of such a service is to
help developers and applications in a situation when their queries are not
well-defined.
• Entity Management Service: This service provides a catalog that serves as the
central registry of entities, datasets, and data sources. Within the catalog, all
water related datasets, entities, and other sources of information are declared
along with their descriptions. This includes a) the list of entities such as sensors
or locations that are important for understanding water data and b) open data
sources that are relevant to water management such as weather observation sta-
tions or forecast services. Besides the APIs and query endpoints provided by the
individual data sources, the catalog also provides queries services over the
descriptions of entities and datasets.
• Event Processing Service: The event processing service allows automatic
matching of events similar to users defined rules based on a semantic model for
water management. Thus, it simplifies the task of water sensor management. It
allows the system to go up early, while administrators can add more meta-data
for sensor management in a pay-as-you-go manner (Derguech et al. 2015; Hasan
and Curry 2014; Hasan et al. 2013b).
• Human Computation Service: The support services, as described above, are pri-
marily focused on providing management tools and programmable access to the
constituent information of the Real-time Linked Dataspace. These services are fur-
ther complemented with a human computation service that is concerned with the
collaborative aspect of data management (Ul Hassan et al. 2013, 2016). Essentially,
it allows small tasks for data management to be distributed among people who are
willing to participate in the dataspace management and improvement process (Ul
Hassan et al. 2012). The same service is further utilized for spatial tasks of data
management in public spaces (Ul Hassan and Curry 2016).
180 U. ul Hassan et al.
One of the distinguishing aspects of the Waternomics project is its wide variety of
end-users. Waternomics has four pilot sites to test and validate its research activi-
ties, data platform, and applications. The pilot sits represent use cases of water
management in public spaces, as summarized in Table 1.
Linate Airport
The Linate Airport pilot targets corporate users that are staff members of the air-
port including building managers, technicians, and engineers. These are adult users
that have an advanced level of education and skills to work in such environment.
Besides staff members of the airport, target users also include passengers that
range from a wide variety of casual to business travelers from different age groups
from kids to adults.
Linate Airport is deeply embedded in the urban belt of the city of Milan in
Italy. It has a total area of approximately 350 hectares. The airport clientele is
predominantly passengers on particular national and international particular
routes. In 2012, the Linate airport has operated for 6.3% of the passengers, and
2.2 % of the goods in transit through Italian airports. The airport has two runways
for landing and take-off. The first runway (2442 m long) is intended for commer-
cial aviation and the second runway (601 m long) is intended for general aviation.
The airport aprons, ramps, and parking stands allow for the simultaneous parking
of 46 aircraft. The passenger terminal extends over five levels with a total area of
about 75,000 m2 (of which about 33,000 m2 are open to the public). The terminal
is equipped with 71 check-in counters and 24 gates, five of which serve as a load-
ing bridge. Approximately 21% of the area open to the public is dedicated to com-
mercial activities (shops, restaurants, bars, car rentals, banking services, post
offices, branches of public services) and 7.5% to the services provided by airlines
(check-in, ticketing).
Given the complexity of an airport, a key aspect of this pilot site has been the
cooperation with the company that operates the airport. In particular, information on
commercial activities and information on key water consumers within the airport, as
well as the water and wastewater infrastructure, have been readily made available.
Municipality of Thermi
The pilot concerning the Municipality of Thermi in Greece targets domestic users.
Families are the primary users including children, young adults, and adults. The
Municipality of Thermi is situated in the eastern area of the prefecture of
Thessaloniki, at a distance of 15 km from the metropolitan center of Thessaloniki.
The Municipality of Thermi consists of 14 communities with Thermi being the
seat of the Municipality, covering an area of 38.34 hectares. The total population of
the Municipality of Thermi is 53,070 according to the census in 2011; however, the
actual population is now estimated at 70,000.
The main land use in the area is agriculture; however, land use is changing with
more land being dedicated to various types of buildings and infrastructure. Thermi
has a strong developmental relationship with an urban area located in close proxim-
ity: the Thessaloniki Urban Agglomeration (TUA). It is a rapidly growing and eco-
nomically viable zone, which is developed as a residential expansion of the TUA,
but also as a pole for the location of industrial plants, tertiary sector activities, and
highly specialized services, maintaining, at the same time, the characteristics of a
developed suburban agricultural economy. At the southeast part of the settlement,
there is a planned area of soft manufacturing activities. Finally, there are some large
land property areas, such as the military installations, the airport, the American
Farm School and the buildings of Aristotle University of Thessaloniki (AUTH),
which cover a significant amount of land in the area.
For the purpose of water management pilot, a selection of 10 households was
made. These households were selected so that they represent a wide variety of types
of houses and families in order to examine the effects of different types of domestic
users.
The pilot in National University of Ireland Galway (NUIG) targets staff members
(including managers, technicians, and researchers) and students (including under-
graduates and postgraduates). While staff members are interested in understanding
water usage behaviors and detecting saving opportunities, students are interested in
visualizing the building consumption and water consumption data in their projects
and research works. The age groups of this pilot site range from young adults to
adults. NUIG is one of the Ireland’s national universities, founded in 1845, it is
ranked among the top 2% of universities in the world. Located in the city of Galway
(population 70,000 approximately) on the west coast of Ireland, NUIG has more
than 17,000 students and 2500 staff. The Engineering Building at NUIG is a state of
the art educational facility designed to be a “living laboratory” where the building
itself is an interactive teaching tool. The Engineering Building opened in 2011; it is
the largest engineering building in Ireland and includes lecture halls, classrooms,
182 U. ul Hassan et al.
offices, laboratory facilities, a café, showers, and bathrooms. The building accom-
modates approximately 1100 students and 100 staff on four floors (in 14,000 m2 of
floor space). The majority of students are undergraduates aged 18–24 years.
Similar to the university, a secondary school in Galway has both staff members and
students as target users. The main difference is the age groups of users which range
from kids to adults on this site. Coláiste na Coiribe (CnaC) is an Irish language
secondary school with approximately 350 students and 25 teaching and administra-
tive staff. The existing school is housed at a small location in the city center. To
facilitate the demand for places at the school and to address space pressures, a new
school (7400 m2) was under constructed at a suburban location in Galway.
This new school building serves as a pilot for Waternomics. The new school
accommodates up to 720 students (aged 12–18) and includes classrooms, offices,
sports halls and associated toilet and shower facilities. As the school was identified
as a suitable pilot site at the early stages of construction it provided an opportunity
for the Waternomics project team to engage with the designers and contractors in
the deciding on the provision of water metering and water information infrastruc-
ture for the building. In addition, it provided a unique opportunity to monitor this
new building from the beginning of its occupation.
The new school building opened in October 2015, it facilitated engagement with
students at an early age regarding water consumption behavior. Furthermore, these
students tested and gave feedback to the project on how the platform functions in
communicating complex water-related data to a wider audience. The collaboration
between the school and the Waternomics project resulted in students actively pro-
viding inputs to the project (e.g. user interface design, applications etc.). The school
management faces key budgetary and conservation targets; to date reporting on
water and associated energy consumption has been very limited. The pilot informs
future design of similar buildings with a particular focus on water conservation
measures and rainwater harvesting systems.
Fig. 2 Lambda architecture realization of the Real-time Linked Dataspace for the Waternomics
project
implementation of the batch, speed, and serving layers. The support services in the
dataspace are mainly implemented through customization of following open source
software: Druid, Apache Spark, MySQL, Apache Kafka, and Apache Cassandra.
Figure 2 shows the data from a building management system (BMS) and water sen-
sors in the Galway pilot being processed in the dataspace. All data sources and entities
are defined in the catalog (WKAN) . The batch layer is implemented using Spark SQL
when historical data from BMS is fed into the indexer node of Druid. Real-time data
from sensors is fed into the Kafka message broker, which provides a high availability
integration point for speed layer data from the different pilots. Real-time data from
Kafka is processed through Spark Streaming to a real-time node of Druid. The com-
bined code from Spark Streaming and Spark SQL provides a standardized way of gen-
erating dimensional data that is served using the Druid cluster. The Druid nodes use
Cassandra as deep storage for historical data. The batch data is made available through
the historical node and streaming data is made available through the real-time node.
Periodically, the streaming data is pushed to the historical node as new data arrives. The
broker node of Druid seamlessly exposes batch data and real-time data, without the need
for writing queries for real-time and batch data separately.
The pilots in the Waternomics project aim to collect both real-time and historical
data for water management. For instance, the NUIG and CnaC pilots include fol-
lowing data sources for large buildings.
184 U. ul Hassan et al.
Fig. 3 Datasets and data sources in the WKAN catalog for the NUIG pilot
Water Analytics and Management with Real-Time Linked Dataspaces 185
Applications
The applications that may be built on top of Waternomics dataspace are diverse;
they include water awareness dashboards, decision support for the different targeted
users (i.e. domestic users, organizations, cities), and water availability/forecasting,
dynamic pricing, and water footprints.
• Water awareness: Low comprehension of water flows by users and over usage
is one of the biggest causes of water wastage. A lack of awareness on the amount
of water consumed leads to the lack of incentives to monitor and affect the
situation. Water awareness requires different information for household, com-
pany, and city level, and where different decisions are taken to manage water on
these levels. Therefore, water awareness dashboards need to be tailored to differ-
ent needs of different water usage levels. The data collected by smart water
meters is enriched with contextually Linked Data and processed in real-time;
hence, allowing for deeper data analysis and faster reactions.
• Water consumption: Hydro-meteorological forecasts predict natural demand and
supply of water and can be used to prepare and adjust water supply. Forecasting
systems can achieve different goals depending on the level of the system deploy-
ment. At the household level, forecasts include analysis of occupants’ behavior and
water consumption based on similar historical water usage. These forecasts can be
incorporated into dashboards and used as the drivers for water saving goal.
Forecasting models can further leverage Linked Open Data at the neighborhood or
city-level. At the company-level forecasts similar to those of the household level
are also augmented by models or simulations of the water needs of subsystems
within the organization. Linked Data can be used to perform benchmarking
between similar organizations to identify areas of potential water optimization.
• Water education: Understanding the impact of a product or service requires an
analysis of all potential water consumption associated with its entire lifecycle.
For instance, a water footprint of a product would provide a quantitative cradle-
to-grave analysis of the product/services global water costs (i.e., water used in
raw materials extraction, through materials processing, manufacture, distribu-
tion, use, repair and maintenance, electricity generation, and disposal or recy-
cling). Building a water footprint requires the gathering of water data from many
participants within the supply chain. Linked Open Data can be a key enabler for
the development of a global information ecosystem of water footprint inventory
data on products, services, and organizations.
In the following, we present a set of applications developed for the NUIG
Engineering Building and the CnaC School. Since both pilot sites have many com-
monalities, they share two main applications: the Public Display and the Manager
Dashboard. Table 2 presents an overview of the applications developed in terms of
their objective and target user groups.
186 U. ul Hassan et al.
Public Display
A key objective for both pilot sites in Galway was to increase water usage awareness
in public spaces. Towards this objective, setup of a kiosk with an interactive dashboard
can help attract people and engage them with discovering water usage details of their
building. A public dashboard is a web application that shows generic information of
the site’s water consumption as compared with social norms; in addition, it displays
information related to consumption per student, toilet flushes per day, etc.
Water Analytics and Management with Real-Time Linked Dataspaces 187
Fig. 4 Public Display showing water data from the NUIG Engineering Building
The web application developed for Galway pilots, as shown in Fig. 4, serves as a
communication medium to display the amount of water being consumed in various
parts of the building. This application shows volumetric values of water usage in
other dimensions such as cost, metaphors, and footprints. The image on left side of
Fig. 4 shows the amount of water in terms of the number of standard size of water
cooler bottles. This application visualizes water quantities in circles using colors to
indicate if the water usage is high or low. This application also uses social media to
inform users about the technology used within the Waternomics project, and its
updates. Users can interact further with the application to explore the water usage
data over a full month; furthermore, they can get more details about technologies
used within the Waternomics project.
The public dashboard application uses the Water Analytics Support Service for
querying the data from the dataspace. This service has been extended to serve as an
extension of the public dashboard and allow users to explore further the water data
by scanning QR codes near the sensors. As shown in Fig. 5, users are able to visual-
ize the entire month’s consumption. This extension aims to support students and
researchers who can retrieve the data from this service via its API to use in their
research projects.
Manager Dashboard
Managers in the NUIG Engineering Building and CnaC school are interested in
watching the consumption at different points of the water network. In both pilots,
dashboards can be considered rather as a family of applications targeting the
188 U. ul Hassan et al.
Fig. 5 Visualizing Water Analytics for the sensor 309 in February 2016
specific needs of managers than as a single application aiming to solve all prob-
lems for all users. One of the key elements used in the Manager dashboards are
historical graphs showing the consumption in various points or groups of inter-
ests (see Fig. 6). Goal setting and tracking is also an important aspect for manag-
ers in the Galway pilots so comparison graphs are an important part of their
dashboards.
Making drinking water available becomes a major concern in public spaces. This can
be guaranteed through a carefully selected location for drinking water fountains in
order to make sure that water is always flowing in the pipes. However, in spaces such
as a university building, drinking water fountains can remain unused during long holi-
days and weekends. Consequently, drinking water can reside in the pipes for long
periods. Building managers want to make sure that residual water is still safe to drink.
In this context, the water retention time observer application can assist managers
to guarantee that they receive timely notifications regarding water that has been
residing for a long period in drinking water pipes. This is done by allowing them to
setup a set of rules for tracking periods of inactivity in specific measurement points
and automatically send notifications through the system to selected user groups.
Figure 7 shows the list of active alarms detected by the application. The application
is well aligned with one of the objectives of Waternomics project, i.e. giving action-
able information to water users and managers.
Water Analytics and Management with Real-Time Linked Dataspaces 189
Both pilot sites in Galway also aim to improve water network management by
assisting staff in coordinating and making better-informed decisions. An addi-
tional aspect in this context is the ability to communicate messages to specific
user-groups related to their consumption in order to require actions or encour-
age behavior change. The observations control panel is an application that gives
an overview of the status of all notifications within a timeframe. It provides an
interface for managing notifications that can originate from any application that
uses data from the dataspace. Based on the activity logging on different notifica-
tions, the user can see how much time it takes from the time of creation of a
notification to the time of action or expiry. The application also allows to filter
notifications based on the group they were targeting, criticality level and the
source application. The control panel allows managers to not only show but
generate custom notifications themselves to facilitate this communication with
specific user groups.
Wearable Info-Centre
Managers at the Galway pilots are very mobile and they require instant notifications
of important aspects of the water consumption in their building. In this case, how-
ever, users are more technology friendly and expressed their willingness to use more
Water Analytics and Management with Real-Time Linked Dataspaces 191
Fig. 8 Wearable
info-centre, receiving
notification from the water
retention time observer
advanced mechanisms for receiving notifications through smart devices. The wear-
able info-center application was developed for mobile notifications.
The wearable info-centre is an application that the user installs on a smartwatch
to display notifications as they are received on the mobile phone. This way users
don’t have to check on their phones every time they receive notifications from the
platform. Instead, they can check their smartwatches which is less obtrusive while
communicating the information at any time. Figure 8 shows an example of using the
wearable info-centre. The application provides an interface for displaying notifica-
tions that can originate from any application which uses data from the real-time
dataspace. So, the application is indirectly uses all kinds of data provided by the
real-time dataspace.
One of the ideas explore during the user tests was the concept of allowing users to
track their own personal consumption patterns. The patterns are based on the appli-
cations that are activated and connected with specific micro-sessions of user con-
suming water such as preparing coffee, drinking water, washing hands etc. One of
the key outcomes in user tests was that mobile and wearable devices can offer a
great opportunity in personalized tracking but this is hard to do if it requires an
additional action to already existing routine (e.g. if it requires you to get your phone
out of your pocket and scan a QR code). So, in the goal-oriented accessing water
application we experimented with the idea of replacing an action in user’s routine
while in parallel providing some short pieces of information (Fig. 9).
This concept challenges the centuries old mechanism of operating a faucet, which
in fact is a valve of various designs. The new system transforms water usage into goal-
oriented activity such that accessing water is no longer just about enabling a valve. By
setting up a touch enabled sink display next to a faucet (without its original turning
knob), users were able to choose certain water activity such as “one cup of tea” or
192 U. ul Hassan et al.
Fig. 9 Goal-oriented
Accessing water application
“one bottle of water”. This message will be sent through wireless to a solenoid valve
connected to a water pipe or faucet that provides a certain amount of water. In this
manner, users were always aware of their water usage thus lowering the chance of
wastage. A social network system was also implemented into the system such that
users could report issues to each other or even to the building manager so that urgent
problems can be solved more rapidly to prevent waste of water in any case.
Related Work
Summary
This chapter motivates the need for efficient water information management in pub-
lic spaces and presents a Real-time Linked Dataspace approach for water data ser-
vices. A high-level architecture, for the Real-time Linked Dataspace, realizes this
approach in the context of the Waternomics project. The Waternomics project estab-
lished the utility of this approach with the help of four pilot sites that represent dif-
ferent scenarios of public spaces. This chapter describes a concrete instantiation of
the Real-time Linked Dataspace approach for two educational institutions, along
with applications supported by the water data services.
Acknowledgments The research leading to these results has received funding under the European
Commission’s Seventh Framework Programme from ICT grant agreement WATERNOMICS no.
619660. It is also supported in part by Science Foundation Ireland (SFI) under Grant Number
SFI/12/RC/2289.
References
Barua S, Ng AWM, Perera BJC (2011) Comparative evaluation of drought indexes: case study on
the Yarra River catchment in Australia. J Water Resour Plan Manag 137(2):215–226
Boken VK (2009) Improving a drought early warning model for an arid region using a soil-
moisture index. Appl Geogr 29(3):402–408
Cancelliere A, Mauro GD, Bonaccorso B, Rossi G (2007) Drought forecasting using the standard-
ized precipitation index. Water Resour Manag 21(5):801–819
Cavanillas JM, Curry E, Wahlster W (2016) New horizons for a data-driven economy: a roadmap
for usage and exploitation of big data in Europe. Springer, Cham
Curry E, Donnellan B (2012) Sustainable information systems and green metrics. In: Harnessing
green it: principles and practices. Wiley, Chichester, pp 167–198
Curry E, Degeler V, Clifford E, Coakley D, Costa A, van Andel SJ, van de Giesen N, Kouroupetroglou
C, Messervey T, Mink J, Smit S (2014) Linked water data for water information management.
In: 11th international conference on hydroinformatics (HIC)
194 U. ul Hassan et al.
Curry E, Dustdar S, Sheng QZ, Sheth A (2016) Smart cities enabling services and applications.
J Internet Serv Appl 7(1):6
Derguech W, Bhiri S, Hasan S, Curry E (2015) Using formal concept analysis for organizing and
discovering sensor capabilities. Comput J 58(3):356–367
Fang-Yuan Xu, Long Zhou, Yi Lin Wu, Yingnan Ma (2010) Standards, policies and case studies in
smart metering. In: IEEE PES general meeting, pp 1–5
Franklin M, Halevy A, Maier D (2005) From databases to dataspaces. ACM SIGMOD Rec
34(4):27–33. doi:10.1145/1107499.1107502
Froehlich J, Findlater L, Landay J (2010) The design of eco-feedback technology. In: Proceedings
of the 28th international conference on human factors in computing systems, pp 1999–2008
General Assembly, United Nations (2015) Transforming our world: the 2030 agenda for sustain-
able development. Tech. Rep. 1
Hasan S, Curry E (2014) Approximate semantic matching of events for the internet of things. ACM
Trans Internet Technol 14(1):1–23
Hasan S, Medland R, Foth M, Curry E. (2013a) Curbing resource consumption using team-based
feedback. In: Lecture notes in computer science (including subseries Lecture notes in artificial
intelligence and lecture notes in bioinformatics), vol 7822 LNCS, pp 75–86
Hasan S, O’Riain S, Curry E (2013b) Towards unified and native enrichment in event processing
systems. In: Proceedings of the 7th ACM international conference on distributed event-based
systems (DEBS ’13). ACM Press, New York, p. 171. doi:10.1145/2488222.2488347
Heath T, Bizer C (2011) Linked data: evolving the web into a global data space. Synthesis lectures
on the semantic web: theory and technology. Morgan & Claypool, San Rafael
Hertel TW, Liu J (2016) Implications of water scarcity for economic growth
Kappel K, Grechenig T (2009) “Show-me”: water consumption at a glance to promote water con-
servation in the shower. In: Persuasive 09 Proceedings of the 4th international conference on
persuasive technology, pp 1–6
Kogan FN (1995) Application of vegetation index and brightness temperature for drought detec-
tion. Adv Space Res 15(11):91–100
Kuznetsov S, Paulos E (2010) UpStream: motivating water conservation with low-cost water flow
sensing and persuasive displays. In: Proceedings of the 28th international conference on human
factors in computing systems – CHI ’10. ACM Press, New York, p 1851
Lepe-Salazar F, Yamabe T, Alexandrova T, Liu Y, Nakajima, T (2012) Family interaction for
responsible natural resource consumption. In: Proceedings of the 2012 ACM annual confer-
ence on human factors in computing systems (CHI ’12), pp 2105–2110
Liotta A, Geelen D, van Kempen G, van Hoogstraten F (2012) A survey on networks for smart
metering systems. Int J Pervasive Comput Commun 8(1):23–52
Makonin S, Pasquier P, Bartram L (2011): Elements of consumption: an abstract visualiza-
tion of household consumption. In: Lecture notes in computer science (including subseries
Lecture notes in artificial intelligence and Lecture notes in bioinformatics), vol. 6815 LNCS,
pp 194–198
Nam T, Aldama FA, Chourabi H, Mellouli S, Pardo TA, Gil-Garcia JR, Scholl HJ, Ojo A, Estevez
E, Zheng L (2011). Smart cities and service integration. In: Proceedings of the 12th Annual
International Digital Government Research Conference on Digital Government Innovation in
Challenging Times, dg.o ’11, New York, pp 333–334
OECD (2012) OECD environmental outlook to 2050. OECD environmental outlook. OECD
Publishing, Paris
Pereira AG, Rinaudo JD, Jeffrey P, Blasques J, Quintana SC, Courtois N, Funtowicz S, Petit V
(2003) ICT tools to support public participation in water resources governance & planning:
experiences from the design and testing of a multi-media platform. J Environ Assess Policy
Manag 05(03):395–420
Ul Hassan U, Curry E (2016) Efficient task assignment for spatial crowdsourcing: a combinatorial
fractional optimization approach with semi-bandit learning. Expert Syst Appl 58:36–56
Water Analytics and Management with Real-Time Linked Dataspaces 195
Ul Hassan U, O’Riain S, Curry E (2012) Towards expertise modelling for routing data cleaning
tasks within a community of knowledge workers. In: Proceedings of the 17th international
conference on information quality, Paris
Ul Hassan U, Bassora M, Vahid AH, O’Riain S, Curry E (2013) A collaborative approach for
metadata management for Internet of things: linking micro tasks with physical objects. In:
Proceedings of the 9th international conference on collaborative computing: networking, appli-
cations and worksharing. IEEE, pp 593–598
Ul Hassan U, Zaveri A, Marx E, Curry E, Lehmann J (2016) ACRyLIQ: Leveraging DBpedia for
Adaptive Crowdsourcing in Linked Data Quality Assessment. In: Blomqvist E, Ciancarini P,
Poggi F, Vitali F (eds) Knowledge engineering and knowledge management. Springer, Cham,
pp 681–696
Yang F, Tschetter E, Léauté X, Ray N, Merlino G, Ganguli D (2014) Druid: a real-time ana-
lytical data store. In: Proceedings of the 2014 ACM SIGMOD international conference on
Management of data – SIGMOD ’14. ACM Press, New York, pp 157–168
Louise Hannon is a Senior Research Associate in Civil Engineering at NUI Galway, she is a
Chartered Engineer with extensive experience in engineering infrastructure design and project
management. Louise has spent most of her career to date in civil engineering consultancy involved
in the design and management of highway infrastructure projects. Louise has been involved in civil
engineering research at NUI Galway research since 2013, completing a Research Masters in 2016
as well as managing and contributing to a number of EU and Irish research projects. Her research
interests include water and wastewater management, she has authored/contributed to a number of
papers in the area in particular related to Sustainable management of water resources, Fault
Detection and Diagnosis, Sustainable wastewater treatment, Effecting Water Use Behavioural
Change, Operational Testing and monitoring at De-centralized Wastewater Plants.
196 U. ul Hassan et al.
Eoghan Clifford is a lecturer in Civil Engineering, NUI Galway with 15 years’ experience in the
areas of water and wastewater engineering in the academic, research and private spheres. His key
research interests lie in the development of tools and technology that can improve outcomes in the
water and wastewater sectors. He also has a particular interest in areas such as policy development
and improving interaction between end users and technology that that lead to more sustainable
water and energy usage at a local, national and international scale. He has managed national, EU
and industry funded research projects with a total value of € 8 m in the last 7 year, has 100+ scien-
tific articles in peer reviewed international journals and conference proceedings. He has written
numerous reports for the Irish Environmental Protection Agency and has 3 patents in the wastewa-
ter technology sphere. Eoghan is a Chartered Engineer with a PhD in Environmental Engineering.
Christos Kouroupetroglou was awarded his PhD in 2010 and his thesis subject was “Semantically
enhanced web browsing interfaces”. During his PhD research he investigated the impact of
Semantic-Web based assistive interfaces in web browsing for blind people and worked as a pri-
mary researcher of SeEBrowser project. In parallel, he also taught as a scientific and laboratory
associate at the ATEI of Thessaloniki and in Mediterranean college of Thessaloniki in modules
related to Web Applications Development. His latest research efforts include the participation in 2
EU funded projects dealing with the use of ICT to raise awareness for water conservation issues
(http://www.waternomics.eu) and the use of robotics to combat loneliness and isolation of people
with dementia (http://www.mario-porject.eu). He is the author of “Enhancing the Human
Experience through Assistive Technologies and E-Accessibility” and has chaired a number of con-
ference sessions and online symposia.
Sander Smit is founder and owner of BM-Change, a Dutch consultancy firm specialised in stra-
tegic innovation. For the past twenty years he has been working as a Project Manager and Business
Consultant in product- and service-development projects in the domains Consumer Electronics,
Telecom & Media, Smart Mobility, Smart Energy Networks and Smart Water Systems. Since
2006, Sander has specialised himself in strategic innovation. He has been responsible for business
model design and exploitation activities in national- and international research projects. Sander has
published multiple papers on business model innovation and is co-creator of the international best-
selling book “Business Model Generation”.
Dr. Edward Curry (www.edwardcurry.org) is Vice President of the Big Data Value Association
(www.BDVA.eu) a non-profit industry-led organisation with the objective of increasing the com-
petitiveness of European Companies with data-driven innovation. He is a research leader at the
Insight Centre for Data Analytics (www.insight-centre.org) and a funded investigator at LERO The
Irish Software Research Centre (www.lero.ie). Edward has worked extensively with industry and
government advising on the adoption patterns, practicalities, and benefits of new technologies.
Edward has published over 120 scientific articles in journals, books, and international
conferences.
Fostering Citizens’ Participation
and Transparency with Social Tools
and Personalization
Introduction
each local context. When such interactions become more transparent, this may
mean greater accessibility, more sharing of ideas, better understanding on how the
other works and thinks, etcetera. Such processes can be studied at different levels of
abstraction, from that of individuals to that of democratic systems. It is our conten-
tion that social media can support transparency, especially when the interactions are
meaningful and deliberately focused on joint issues and backed up with knowledge
about ongoing developments and facts related to the issue at stake. Such knowledge
comes from the use of Open Data. Wide-spread access to the Internet has greatly
reduced the cost of collecting, distributing, and accessing government information.
But an important effect of the diffusion of networks in the population is the potential
of ICT, by promoting good governance, strengthening reform-oriented initiatives,
reducing potential for corrupt behaviours, enhancing relationships between govern-
ment employees and citizens, allowing for citizen tracking of activities, and by
monitoring and controlling behaviours of government employees, is able to effec-
tively reduce corruption (Bertot et al. 2010).
Open data are significantly seen, in general, as the main instrument to improve
transparency, at every level. In fact, transparency for local government can be
defined in different ways. The traditional view of transparency is that governments
provide information on their work, on the other hand, governments are requiring
transparency also from their dependents (such as non-profit-organizations and enti-
ties that they regulate in the private sector) (Michener and Bersch 2011). Hence,
transparency is an interactive concept.
The early attempts and most of the current open data efforts are somehow failing
to enable the transparency in its interactive aspects. Some of the main barriers often
cited by researchers is the complexity of the information that is provided and the
lack or inadequateness of tools that can help citizens in making sense out of the
highly specialized datasets that are provided by government. Of course, this is cru-
cial, as noticed by Michener and Bersch (2011), since the quality of transparency
does not only depend on how (and how much) information is made visible and
accessible, but also on how the information provided can be fruitfully used to accu-
rate inference.
Some researchers (Mishory 2013) have indeed noticed that transparency is not
an “object” but it is rather a “relationship”, and, in order to yield better outcomes
from transparency programs, it is of primary importance to design a relationship
toward greater trust between the “supplier” of open-data based transparency (gov-
ernment) to the “recipients” (i.e. the citizens). In this context, it is crucial to ensure
that citizens must be facilitated in their work toward comprehension of what data is
important for them.
Many have observed (Bonsón et al. 2012) that local governments in Europe are
often well behind their citizens in the use of the social dimension of the Internet,
since they usually prefer to use ICT technologies (and social networks and the
Internet) only as a one-way channel where they can communicate easily and at
low cost.
As noticed by Davies (2012), the task of opening up data to citizens is seen as
completed once data is correctly published, missing the important dimension of
Fostering Citizens’ Participation and Transparency with Social Tools and Personalization 199
Organization of the Chapter
A series of carefully designed workshops were conducted, one in each pilot site, for
the purpose of developing a comprehensive set of user needs, from the viewpoint of
key stakeholders. Each workshop brought together experts, academics, industry
specialists, open data practitioners, representatives of governments, open data
researchers, and potential users (including citizens, representatives of citizens and
social service institutes, and journalists) to brainstorm on open data platform adop-
tion challenges, solutions to the challenges and a set of needs and requirements
necessary for consideration in the design of the ROUTE-TO-PA platform. The
emphasis on citizen participation and collaborative design in the methodology seeks
to address the goals of improved government transparency and accountability for
decision-making. Each workshop began with a collective intelligence (CI) analysis
of barriers to accessing, understanding and using open data, followed by an analysis
of options that may overcome these barriers. Participants then worked to develop
scenario-based user needs, which involved profiling user needs in light of the barri-
ers and options and high level scenarios of open data usage.
The methodology used to gather user-level requirements is inspired by a scenario-
based design (SBD) approach (Rosson and Carroll 2002), but builds upon this
approach by adding a collective intelligence (Warfield 2006) and agile user story
development (Cohn 2004) approach. In the current application of CI, workshop
participants worked to develop scenario-based user needs, which involved profiling
user needs in light of the barriers and options and high level scenarios of open data
usage. This included a separate focus on (1) information needs, (2) social and col-
laborative interaction needs, and (3) understandability, usability and decision-
making needs. Idea writing was used for each cluster of needs. High level scenarios
including multiple users were used to prompt thinking in relation to user needs. All
the short user stories generated by participants were generated in the form:
As user type ________ , I want ______, so that I can _______
The wants (or needs) generated by participants across each pilot site were then
analysed and key categories of user needs identified. Reasons for specified user
needs were also analysed, and this analysis was used to advance our understanding
of the scenarios and prospective use case models. This work in turn has shaped the
test and evaluation framework (see Fig. 1).
The scenarios used addressed various contextual issues, relevant to each of the
workshop sites, and aligned with the primary case focus in each pilot site. For exam-
ple, the Dublin workshop focused on community networking and opportunity cre-
ation; the Groningen workshop focused on the use of Open Data in overcoming
issues associated with population decline; the Den Haag workshop focused on Open
Data in relation to employment and opportunity creation; the Prato workshop cen-
tered on local policy and budget issues; and finally, the workshop in Paris focused
on Open Data in relation to start-up companies and the digital economy.
As such, there was some variety in user needs generated, across all there catego-
ries of needs: information needs, social and collaborative needs, and understand-
ability, usability and decision-making needs (Hogan et al., submitted).
202 V. Scarano et al.
Fig. 1 Workflow for collective intelligence and link between collective intelligence and evaluation
framework and test specifications
As workshop participants in each pilot site were working with a variety of sce-
narios, the user information needs generated were numerous and diverse. The infor-
mation needs included, for example, demographic information needs; legal
information needs; health information needs; social and community information
needs; planning information needs; services, amenities and event information
needs; business and financial information needs; jobseeker information needs.
Essentially, the data and information that different pilot sites need depend on the
problems they are working to solve in their scenarios. The ROUTE-TO-PA team are
working to collate all available open data to make it available on the platform.
Participants then moved on to identify social and collaborative needs based on
the user stories, in order to provide input for the design of SPOD. Social and col-
laborative needs were commonly specified across pilot sites. Categories of needs
here included: dialogue and discussion spaces; moderation and maintenance of
these spaces; platform tool capabilities for interaction; varied forms of social media
interaction; personalisation of user spaces; and requesting and sharing information.
Broadly speaking, participants identified a variety of forms of interaction which
could be used over Open Data, and suggested a number of considerations and
affordances which would increase the impact and appeal of such social and collab-
orative platforms.
Participants also used the scenarios provided to design user stories around under-
standability, usability, and decision-making needs, which will inform the design of
the TET. The major categories of understandability, usability and decision-making
needs were common across sites. Categories of needs here included: Affordances
for the visualisation of complex information; data analysis and reporting tools;
decision-making support tools; guidance and usage support tools; affordances for
personalising platforms and/or data; and certification tools. Broadly speaking, par-
ticipants frequently cited the need for data visualisation tools, among others, which
would make data more easily understood, whether for personal or professional use.
Fostering Citizens’ Participation and Transparency with Social Tools and Personalization 203
Next to identifying user requirements for an open data platform from the bottom-up
via collective design based workshops, we also identified user requirements from a
top down perspective. After all, in order to design and implement a successful ICT
platform, “the context” that includes people and their relations (Kuutti 1999) needs
to be taken into account as well.
Information technologies should be able to support active users, while dealing
with the organizational and societal context (Kuutti 1999). Yet, often this context of
broader social forces and structures that influences the interaction between users
and information technology, is left unexamined (Engeström 2005).
Therefore, based on democracy, transparency and activity theory, the Societal
Activity model of Open Data use (Ruijer et al. 2016) was developed. The model
takes three democratic processes as a starting point for the design of open data plat-
forms: monitorial, deliberative and participatory democracy (Meijer 2012).
The Societal Activity Model of Open Data Use (Ruijer et al. 2016) enhances our
understanding of user requirements of open data in a societal context. It helps to find
the best fit between; on the one hand, the impetus for governmental organizations to
provide open data, to increase accountability and transparency, and on the other
hand the specific needs of citizen-users in particular domains.
The model was tested in five pilot sites, using interviews, analysis of official
documents (where available), and workshops or focus groups where open-data pro-
viders and users met and discussed The findings show that different societal pro-
cesses call for different roles of citizens and government and different user
requirements for the design of open data platforms, and, also, provided input for the
design of the ROUTE-TO-PA platform.
Our project will produce software by using open-source licensing model, and the
platform will be given to the community of PAs and developers that, after the end of
the project, will ensure further development and widespread, sustainable and scal-
able exploitation of the results achieved. The results of the project (both software
and guidelines) will allow PAs to follow the economic and budgetary pressures that
204 V. Scarano et al.
The SPOD platform architecture has multiple decoupled and modular components
that communicate together. The architecture is based on mainstream, open source
and modular technologies to guarantee interoperability with other external systems.
The overall architecture is distributed, as the load of different tasks is taken by dif-
ferent servers (components), both server-side and client-side (e.g., the client-side
visualization of data), thereby achieving the important non-functional requirement
of Performance Efficiency.
SPOD is a Social Platform for Open data, so its primary requirement is the
retrieval of data from Open data Providers. Therefore, SPOD interoperates with
Fostering Citizens’ Participation and Transparency with Social Tools and Personalization 205
TET, any CKAN based platform, UltraClarity and OpenDataSoft (and additional
interoperability with OASIS is planned).
In addition, SPOD can retrieve open data from other existing third party data
providers that use restful API. Hence, the interoperability with data provider plat-
forms is based on Web 2.0 mainstream technologies (fulfilling an important non-
functional requirement of Interoperability); in this way, SPOD can retrieve the open
data to use within the social discussions. For instance, the user can create visualiza-
tions from the data available in the open data provider and use them to support its
argumentation. SPOD can be configured to allow easy access to associated data
providers, so that their datasets can be shown easily used to build visualizations (see
Datalets below).
The platform administrator, using the administration pages, can add another data
provider and make it available to end-users. In addition, in order to maximize the
flexibility, during the creation of a data visualization, any user can copy and paste
the data URL from any other external open data Provider as well as post directly the
link or content on SPOD (Fig. 3).
The architecture has a ROUTE-TO-PA Authentication Server (RAS), which acts
as OpenID Authentication Provider and administration tool to manage users’
accounts. In accessing to the ROUTE-TO-PA platforms, users must seamless switch
between SPOD and TET features in a user-friendly way. Therefore, a user must
access to SPOD and TET, and any other feature federated system, with a unique
username and password credential. In order to provide this, the architecture has a
ROUTE-TO-PA Authentication Server (RAS) based on the OpenID protocol. Any
time a user logs in SPOD or TET, his/her browser redirects to the authentication
provider log in page. All the authentication server pages have a consistent Graphical
User Interface (GUI) with SPOD and TET, indeed they have been specifically
designed within the ROUTE-TO-PA project and it is based and compliant with the
material design. In this way, TET, SPOD and authentication look and feel is the
same, and the switching among their pages is seamless.
The overall architecture has a specific server for the authentication and the entire
platform has deployed with a dedicated authentication server. In order to support the
authentication through OpenID, SPOD team developer designed a new Oxwall
plug-in that supports OpenID. SPOD can be deployed without the activation of the
OpenID plug-in so it works without a dedicated authentication server and uses the
existing Oxwall registration, log in pages and user accounts management.
The introduction of a ROUTE-TO-PA Authentication Server allows the interop-
erability of other systems with the ROUTE-TO-PA platforms, following the non-
functional requirement of Interoperability. Both SPOD and TET can provide specific
services and data to external systems in form of Restful API. Therefore, not only
SPOD interoperates with existing data providers, but also itself can provides ser-
vices to other platforms. Any other external system can authenticate to RAS and
interact with SPOD through the restful API services. For instance, based on this
architecture, the platform can provide a social widget to embed within any web site
to easily share open data, add the content in the own private room or participate in a
discussion. In a federated architecture, after the authentication the federated system
can invoke a restful API service to perform an action on SPOD (e.g., post of content
on SPOD, etc.).
SPOD enables the social collaboration around open data; in particular, it aims to
support the collaboration around visualization of open data, allowing their creation,
sharing, change and comment. The SPOD architecture provides the visualizations
and their services through the DatalEt-Ecosystem Provider (briefly DEEP). DEEP
is a repository of visualization web-components to use within SPOD and within any
other web site or system that needs to visualise data.
The overall architecture is also useful to envision a “federation” of ROUTE-
TO-PA systems. Several federated SPODs and TETs or other institutional systems
can access to the ROUTE-TO-PA Authentication Server and share the same ID.
The SPOD architecture is modular and scalable: for instance, the DEEP compo-
nent can be replicated and distribute around the world to improve performances and
serve visualizations to end-users with high availability. Of course, multiple architec-
ture instances can be deployed in different places.
Our distributed architecture, in general, follows the non-functional requirement
of Replaceability, as each single component can be substituted with another specific
one with the same interface (like a different dataset provider, or another
Authentication server with OpenID).
Fostering Citizens’ Participation and Transparency with Social Tools and Personalization 207
SPOD on Oxwall
Oxwall is a free and Open Source Software (FOSS) social network engine that is
able to power customizable online social networks and community-enabled web-
sites. It provides all the basic functionalities of a social network, such as users’
friendship, posting text or media comments, handling (private) groups group or
event creation/joining. Oxwall architecture is based on plugins, few core plugins
handle user/platform/access management. Additional features can be provided by
plugins.
SPOD consists of several additional plugins for Oxwall that add to the “stan-
dard” social network (friends, status, etc.) the following functionalities:
• It is possible to attach to any social comment, status or answer a Datalet (i.e. a
re-usable Web Component that provides real-time visualization of datasets,
located on any compatible server (see the subsection below). In this way social
interactions become enriched with the actual datasets, providing discussions
with easy-to-use and easy-to-understand factual evidence. The process to build a
datalet is provided by a user-friendly, wizard-like component, that provides the
choice of the dataset (among the suggested ones from known providers or from
a new one added by providing the link to the RESTful call), the filtering capabili-
ties (choicing columns and rows) and the visualization (choice of the charts,
parameters, preview). It is a very important characteristics of Datalets that load
and show the actual dataset (1) directly from the source and (2) in real-time,
when the user is loading the page, i.e., ensuring authoritative datasets.
• Any user is given the possibility to access a Personal space, i.e. a place where he/
she can collect and annotate material that can be fruitfully re-used in social dis-
cussions later. The user can collect links to webpages, by providing URLs and a
datalet is showing the real page in a miniature (that can be also navigated),
Datalet for particular visualizations, and plain text notes. All the items can be
further annotated, and full text search capabilities allow easy management of the
material. The main purpose of the Personal space is to provide a space for
reflection as the citizen may need time and thoughts to build an argument to be
re-used in discussions.
• Discussions occur in Public Rooms grouped in an Agora. Each Public Room is a
traditional threaded chat on the left, with the possibility to add Datalets to the
discussion and the possibility to add an opinion (Neutral, Agree, Disagree) to the
comment. The right part of the screen is used to provide synthetic information
about the discussion, that also makes easier to navigate through long discussions.
A graph based representation of the discussions is shown, where nodes are the
comments and edges join the answer to a comment, with colors to represent
opinions and size of the node representing the number of answers below the
comment (see Fig. 6). Navigation is synchronized: clicking on a node on the tree
shows the corresponding comment on the threaded chat on the left. Other graphs
showing the datalets and the users can be also shown.
208 V. Scarano et al.
Datalets and DEEP
What Is TET? The TET comprises a set of tools designed to extend available
features on popular Open Data Platforms (ODP) to more adequately support trans-
parency related qualities desirable by different categories of ODP end-users. Starting
with the Comprehensive Knowledge Archive Network (CKAN) Platform (OKF
Fostering Citizens’ Participation and Transparency with Social Tools and Personalization 209
2014), the vision for TET is to extend major open ODPs platforms with features
enabling easier access to the relevant dataset, a better understanding of these datasets
and integration with social platforms for sharing and discussing datasets.
Technical Features TET is implemented as a set of plugins to extend the available
features on the well-known CKAN Open Data Platform. The Alpha version of TET
described in this report supports the following eight extensions: (1) Support for the
use of the WordPress Content Management System as a rich client for CKAN, (2)
Enhanced metadata schema to support provenance and alignment with latest W3C
guidelines for publishing data on the web, (3) Validation of metadata quality, (4)
Linking of related datasets, (5) Enhanced user profiles for personalization and rec-
ommendation, (6) Personalised search and dataset recommendation to users, (7)
Recommendation of similar dataset to users and (8) Extension of data analytics
function on CKAN to support pivot operations on datasets, (9) Interface with the
Social Platform for Open Data (SPOD). These features are explained in briefly
explained below.
1. Integration of Content Management System for Richer Client Experience –
CKAN Integration with content management system enables publishers to pub-
lish content related to datasets and publish updates related to portal in an easy
way. TET Wordpress plugin seamlessly integrates TET enabled CKAN to pro-
vide rich content publishing features to publishers and intuitive interface to
end-users.
2. Enhanced Metadata for Improved Context – More metadata fields are added to
dataset upload form to enable data publishers to specify richer metadata that
will help users in discovery and in getting better understanding of the datasets.
The metadata fields are guided by and comply with aligned with the “W3C
guidelines for publishing data on the web”. Additional fields supported
include: Basic details related to the dataset; Target audience of the dataset;
Theme/Category; Versioning; Provenance; Geospatial Coverage; and Temporal
Coverage.
3. Metadata Quality Check and Validation – Additional validations to dataset entry
form are added to prevent data entry errors and to ensure consistency. Quality
check indicators guide publishers about the quality of metadata being entered. The
features will also help end-users in assessing the metadata quality of dataset.
4. Relating Datasets – Dataset linking feature allow users to specify explicit links
between datasets, which can be exploited for recommendations and data integra-
tion purposes.
5. Enhanced User Profiles for Personalization and Recommendation – Default
CKAN user registration page is modified to allow more details related to the user
to be captured, the feature plays essential role in creating a personalized user
experience for the end user.
6. Personalization Search and Dataset Recommendation for users – enables users
to search for datasets based on their profile or based on the desired category.
210 V. Scarano et al.
Users can select appropriate profile from the list of profiles provided or could
select the category they are interested in from the list (see Fig. 7).
7. Recommendation for related datasets – enables recommendation of more datas-
ets based on user group and dataset category selected in the user profile in addi-
tion to other contextual information. The feature guides users to find potentially
useful and relevant datasets.
8. Extension of Available of Data Analytics Functions – CKAN platform lacks data
analysis capabilities essential for working with data. To overcome this limitation
as the first step, we added PivotTable feature which allows users to view, sum-
marize and visualize data.
9. Enhanced Interface with the SPOD – builds on the CKAN APIs to enable the
SPOD platform access datasets managed on the CKAN with the enhanced fea-
tures for visualisation, sharing and discussion.
The above features resulted from analysis of the information on the barriers to open
data use and needs of end-users gathered from the series of Collective Intelligence
Sessions hosted by pilot partners in respective partner countries. In addition,
transparency qualities including Accessibility, Usability, Understandability,
Informativeness and Auditability described in Cappelli et al. (2013) and other trans-
parency constructs such as those in Fung (2013) underpinned the development of
the above features.
In implementing the above features, the base CKAN platform were extended with a
number of additional components implemented as “plug-ins”. The architectural
decision to implement TET as plug-ins is to enable easy coupling and removal of
TET components and consequently minimal operational impact to the based open
data platforms. This architectural style allows us to experiment easily with alterna-
tive design and implementations of the feature (Wang and Avrunin 2008). The TET
components are grouped into three categories: Frontend, Data Platform and
Analytics components as shown in Fig. 4.
In addition, an additional element (or plugin) enables integration with the social
platform. The development of the above features was carried out through an Agile
Software Development process which enabled the development an early prototype
and subsequent short “develop-test-release” cycles to engage potential end-users of
TET platform.
Status and Exploitation The above features have been successfully implemented
as part of the Alpha release available at http://srvgal100.deri.ie:8081/. Some of the
functionalities have already been deployed as part of CKAN instance managed by
the Dublin City Council, Republic of Ireland.
Fostering Citizens’ Participation and Transparency with Social Tools and Personalization 211
The City of Prato (Italy) scenario case concerns city budget management. The goal
is for citizens to monitor the allocation of the city budget by municipality and to
propose expenditure priorities and suggestions. The city would like to increase trans-
parency on budget management and possibly collect some of citizens’ expenditure
suggestions. It is important to know that this process is intended to be an improved
version of existing democratic procedures of involving citizens in budget decisions.
The object or policy issue in Groningen (NL) focuses on population decline.
One of the areas with the highest population decline in the Netherlands is situated
in Groningen. The potential community of stakeholders is diverse consisting of citi-
zens, public organizations (schools, health care organizations etc.) and private com-
panies. Open data as an instrument can provide insight in the consequences of and
might be able to contribute to innovative and collaborative solutions for population
decline.
The Dublin, Ireland City Council is one of the founding partners of Dublinked.
Dublinked is an ideas and information sharing network which connects the Dublin
212 V. Scarano et al.
region’s four local authorities with universities, companies and entrepreneurs. The
scenario chosen for ROUTE-TO-PA evolves around capacity building, which
focuses on increasing citizen engagement in a deliberative process with their city or
more specifically their community. The scenario thus focuses on building commu-
nity awareness with the ultimate outcome to “make my city great”.
Issy- les- Moulineaux (France) is a city located near Paris. The city hosts many
IT companies and welcomes start-up companies in the field of new technologies.
The scenario focuses on the activity of two central groups: young entrepreneurs in
ICT domain who form a major part of the local economic base and public adminis-
trators from Paris Region who collaborate with Issy-les-Moulineaux in the global
open data policy.
The pilot in The Hague (NL) focuses on collaboration between public adminis-
trators and employers and can be characterized as a participatory process. Employers
and the City of The Hague have a longer history of collaboration and meeting, the
relationship between the Department and the local employers is quite good. The
specific scenario or policy issue suitable for exploiting Open Data will be jointly
developed, whereby the focus is on finding solutions for existing problems together.
This is called co-creation.
Concerning evaluation of developments and outcomes at the five user sites, in terms
of transparency, we have several sources of criteria. The identification of users’
needs (section “Requirements and User Involvement”), are translated into design
specifications for the tool, that is, the actions users are supposed to perform with the
tool leads to a set of evaluation criteria at the technology and user levels. The model-
ling activity (section “Societal Activity Model of Open Data User”) has provided
abstract models at the society and community-levels. On the basis of these models
we will be able to characterise and compare all cases with respect to their current
and future states at four different levels:
1. The technology functioning according to design specifications, tested by usabil-
ity studies and user consultation.
2. The individual user carrying out different actions: we shall develop a framework
based on the well-known and studied Technology Acceptance Model (TAM)
(more details in the next subsection).
3. The community, or small group working together to generate new ideas: the
main dimensions of evaluation on this level can be formulated in terms of the
characteristics of the participants, the structure of their interactions and their
content within the OECoP, studied from a developmental perspective (Engeström
1987). The developmental, or diachronic, analysis draws on synchronic analysis
of the OE-CoP (at a given point in time), and identifies relevant differences
across time-points, with a view to identifying the overall trajectory (Dreier 1999;
Ludvigsen et al. 2011) of the community, i.e. where it is heading.
Fostering Citizens’ Participation and Transparency with Social Tools and Personalization 213
4. The society, or the organisational context, adapting to the new possibilities for
creating transparency. We will investigate the impact of our project within the
organisation by observations, interviews and surveys.
Our evaluation activity also relates closely to the Models and Methods
developed:
1. The Societal Model (year 1) relates to the above in terms of analysis of the object
of activity (including community rules), and the expression of tensions. In these
cases, the level of content analysis is relevant, for identifying ‘what the partici-
pants are trying to achieve’ (i.e. object of activity) in given exchanges, and, on
the level of communicative functions, given that tensions will be correspond to
argumentative functions and, most likely, to the salient expression of strong
emotions.
2. The Community Model (year 2) provides the theoretical background and meth-
odological tools for addressing the main research question, concerning evolution
of SPOD-TET mediated collective activity towards a veritable epistemic
community.
3. The Social Representations Model (year 3) concerns the evolutions of the OE-
CoP participants’ representations (attitudes, appraisals) of the community, of
their perceived self-efficacy, of the degree of transparency and cooperativity of
the Public Administration, and of the SPOD-TET tools themselves. This will be
studied using interviews and questionnaires, but also on the basis of content
analysis (what are the attitudes expressed using SPOD?). This can draw on
appraisal theory (White 2002), i.e. the positive and negative attitudes and affects
expressed in language.
To evaluate the artifacts, we shall develop a framework based on the well-known and
studied Technology Acceptance Model (TAM). The model was published by Davis
(1989) and is the most widely accepted model for understanding the usage of
Information Systems (IS) and its acceptance. It suggests, that external variables (such
as system design and rich features) influence the perceived usefulness and perceived
ease of use. Over time the model evolved to TAM2 (Venkatesh and Davis 2000) and
was extended with additional external variables, relevant to IS utilised in the work-
place: the social influence variables (i.e. subjective norm) and the cognitive instru-
mental variables (i.e. result demonstrability). Original TAM is presented in Fig. 5.
The definitions of the additional variables in TAM2 are defined as follow:
1. Voluntariness – “the extent to which potential adopters perceive the adoption
decision to be non-mandatory”.
2. Subjective norm – “a person’s perception that most people who are important to
him think he should or should not perform the behavior in question”
214 V. Scarano et al.
The team has just released the first Alpha prototypes of SPOD and TET as the proj-
ect first year just finished (January 2016). SPOD and TET will be tested (in a first
round) in the five pilots for the year 2016, starting late September. In 2017 a new
version, beta, will be submitted to another round of testing from the pilots, leading
to the release of a highly tested, jointly designed and citizen-centered software.
Fostering Citizens’ Participation and Transparency with Social Tools and Personalization 215
Acknowledgments This project has received funding from the European Union’s Horizon 2020
research and innovation programme under grant agreement No 645860.
References
Bertot J, Jaeger P, Grimes JM (2010) Using ICTs to create a culture of transparency: E-government
and social media as openness and anti-corruption tools for societies. Government Information
Quarterly 27:264–271
Bonsón E, Torres L, Royo S, Flores F (2012) Local e-government 2.0: social media and corporate
transparency in municipalities. Government Information Quarterly 29(2):123–132
Cappelli C, Engiel P, Araujo RMD, Cesar J, Leite P (2013) Managing transparency guided by
a maturity model. In: 3rd global conference on transparency research, HEC, Paris, 24–26
October 2013, pp 1–17
Cohn M (2004) User stories applied for agile software development. Addison-Wesley, Boston
Colpaert P, Sarah J, Peter M, Mannens E, Van de Walle R (2013). The 5 stars of open data por-
tals. In: 7th international conference on methodologies, technologies and tools enabling
e-Government (MeTTeG), University of Vigo, Spain, pp 61–67
Davies T (2012) Supporting open data use through active engagement. In: Proceedings of W3C
workshop “using open data: policy modeling, citizen empowerment, data journalism”, 19–20
June 2012. http://www.w3.org/2012/06/pmod/pmod2012_submission_5.pdf
Davis FD (1989) Perceived usefulness, perceived ease of use, and user acceptance of information
technology. MIS Quarterly 13:313–339
Dreier O (1999) Personal trajectories of participation across contexts of social practice. Outlines:
Critical Social Studies 1:5–32
Engeström Y (1987) Learning by expanding: an activity theoretic approach to developmental
research. Orienta Konsu, Helsinki
Fostering Citizens’ Participation and Transparency with Social Tools and Personalization 217
Michael Baker is a Research Director (tenured Research Professor) in psychology and language
sciences, of the Centre National de la Recherche Scientifique (CNRS). He currently works in the
Social and Economics Science Department of Telecom ParisTech, the French National
Telecommunications Engineering School. His research aims to analyse and model the processes
of knowledge elaboration in dialogues produced in learning and work situations, focussing on
argumentation and the role of technological mediation. Michael Baker recently published the
following book, with Baruch Schwarz: “Dialogue, Argumentation and Education: History,
Theory and Practice” (Cambridge University Press, New York, 2017).
218 V. Scarano et al.
Françoise Détienne is research director at CNRS, Paris. Her research in cognitive ergonomics
focusses on collaboration, design, creativity and online communities as well as on the role of
socio-technical systems. She is associate editor of International Journal of Human Computer
Studies, and member of the editorial board of Interacting With Computers and CoDesign.
Jerry Andriessen is a senior scientist and research director of Wise & Munro learning Research,
The Hague, Netherlands. Wise & Munro participates in many innovative European projects, on
technology, open data, collaboration, primary education, and cybersecurity. These projects add
societal relevance to their scientific implications. Jerry has more than 25 years of experience in
uses of technology for support of learning and collaboration in various settings. His current interest
is in the qualitative interpretation of discourse and action, including the roles of emotions.
Mirjam Pardijs is an independent educational scientist. She is co-director of Wise & Munro
Learning Research, the Hague, Netherlands. She specialises in the design, coaching, and interpre-
tation of collaborative learning in educational and professional contexts. Her main interests are
narratives as a tool for learning, appropriation of technology, and the role of pedagogical support
for innovation in learning settings.
Dr. Adegboyega Ojo is Senior Research Fellow at the Insight Centre for Data Analytics, National
University of Ireland Galway (NUIG). He leads the E-Government Group at Insight Centre, serves
as Adjunct Lecturer at the College of Engineering and Informatics. His current research interests
include data driven innovations in government, Open data policies and Infrastructures, data analyt-
ics and governance of smart cities. He is member of the Editorial Boards of the Government
Information Quarterly and International Journal of Public Administration in the Digital Age.
Michael Hogan is a senior lecturer and researcher at NUI, Galway, Ireland. His research foci
include individual, social and technology factors contributing to adult learning, motivation, and
collaborative performance. He is a co-director of the Structured PhD in Perception, Cognition and
Action, co-director of the Structured PhD in Learning Sciences, and a co-leader of the Health and
Wellbeing theme at the Whitaker Institute for Innovation and Social Change, NUI, Galway.
Albert Meijer is a Professor of Public Innovation at the Utrecht School of Governance in the
Netherlands. His research focusses on e-government, government transparency, co-production in
the public sector and public innovation.
Erna Ruijer is a postdoctoral researcher at the Utrecht School of Governance in the Netherlands.
Her research focuses on open data, government transparency, open government and government
communication.
The 6-Values Open Data Business Model
Framework
Fatemeh Ahmadi Zeleti and Adegboyega Ojo
Abstract Business models for open data have emerged in response to the eco-
nomic opportunities presented by the increasing availability of open data. However,
scholarly efforts providing elaborations, rigorous analysis and comparison of open
data models are very limited. This could be partly attributed to the fact that most
discussions on open data business models are predominantly in the practice com-
munity. This shortcoming has resulted in a growing list of open data business mod-
els which, on closer examination, are not clearly delineated and lack clear value
orientation. This has made the understanding of value creation and exploitation
mechanisms in existing open data businesses difficult and challenging to transfer.
Following the Design Science Research (DSR) tradition, we describe a 6-Value
(6-V) business model framework built as a design artifact to facilitate the explica-
tion and detailed analysis of existing open data business models in practice. Based
on the results of the analysis, we identify business model patterns and emerging
core value disciplines for open data businesses. Our results not only help streamline
existing open data business models but helps in linking them to the overall business
strategy through value disciplines.
Introduction
Recently, attention of major stakeholders in the open data (OD) community, includ-
ing policymakers have shifted to the economic value of OD assets. OD constitute
an important resource around the world due to its potential to empower citizens,
businesses, change how government performs, and improve the delivery of public
Literature Review
Open Data
Nowadays, a surprising amount of data is generated and stored than at any other
time in history (van den Broek et al. 2012; Avital and Bjorn-Andersen 2012).
However, not all data can be published or made available to public for free. Some
data is commercially confidential; some are sensitive personal information, which
cannot be shared for reasons of privacy and security. However, where it is appropri-
ate to do so, and the right protections have been taken, such as removing personal
identifiers or aggregating data, sharing or linking data can bring both social and
economic benefits (HM Government 2013; Fensel 2013).
The 6-Values Open Data Business Model Framework 221
Business Models
Osterwalder and Pigneur (Osterwalder 2004; Osterwalder and Pigneur 2009) pres-
ents a business model canvas with nine building blocks. Model is presented in
Fig. 1. The model includes key partnership, key activities, key resources, value
222 F.A. Zeleti and A. Ojo
Shafer et al. (2005) based their framework on the four elements common to most
business models: Strategic choices; value creation; value network; and capture
value. Figure 2 presents the model.
Strategic choices: It defines strategies a business has to be able to develop to offer a
unique product to the customer. This is an element of the strategy formulation
process. Strategic choice adds value to a strategy.
Value network: It defines a network of suppliers and partners required to implement
the business model.
Create value: It describes value creation mechanisms from the different activities.
Capture value: It defines the process of recovering some or all of the value created
for the customer.
The 6-Values Open Data Business Model Framework 223
Strategic Choices
Value Network
Customer (Target Market,
Scope)
Value Proposition Supplier
Capabilities/Competencies Customer Information
Revenue/Pricing Customer Relationship
Competitors Information Flow
Output (Offering)
Strategy Capture Value
Differentiation Cost
Mission Create Value Financial Aspect
Resources/Assets
Various business models have been identified in the literature, mainly in the
practice community. These include: Howard (2014), Ferro and Osella (2013)
identified eight ODBMs: Premium, Freemium, Open Source, Infrastructural
224 F.A. Zeleti and A. Ojo
Conceptualization
Value Capture Value capture is the process of retaining some percentage of the
value provided in every transaction. This allows the business to use the output from
the value in return to rethink and redesign to support the value proposition.
Value Management top managers play a significant role in the whole process.
This includes mindset, organization, governance, stakeholders and shareholders.
Value Network all the business activities are done within the value network. This
includes customers, suppliers, information flow, product flow, service flow and part-
ner businesses.
Extending the 6-V business model framework presented in Fig. 4 and to better
understanding the model components, we present the 6-V model in table form that
provides second-level and third-level components. Each of the 6-V business model
main component includes sub-components (second-level components) in which
each sub-component consists of other sub-components (third-level components).
Fig. 4 (continued)
For example, value proposition can include an offer, channel, and value in which
offer can include product, services and information; channel can include a delivery
method and value can include better price or the value for money. Table 2 shows this
classification.
Model Elaboration
entire model to ensure the components are set appropriately to meet the objective/s.
The resulting information is presented in Table 3 and highlighted below.
Cost Avoidance offers sustainable publishing solution, cost avoidance, and improved
meaning of data and data integration as a value in return.
Sponsorship offers free and useful data to the public and provides availability of
data to public as a value in return.
Freemium offers free but limited data and high-quality data at some cost and pro-
vides limited availability of useful free data to public and perceived value of data
as a value in return.
Premium offers specific customer need and provides perceived value of data as a
value in return.
Dual-Licensing offers free data for non-commercial uses and high-quality data for
commercial use and provide limited availability of useful free data to public and
perceived data as a value in return.
Support and services offer high value-adding data services and provide perceived
value of data as a value in return.
Charging for changes offers free but limited data services and high-quality data at
some cost and provides limited availability of useful free data to public and per-
ceived value of data as a value in return.
Increasing quality through participation offers a higher quality of data and provides
higher data quality as a value in return.
Supporting primary business offers strategic support to the business objective and
provides improved in business results as a value in return.
Open source offers free data for non-corporate use and quality data for corporate
use and provides limited availability of useful free data to public and perceived
value of data as a value in return.
Infrastructural razor and blades offer incomplete data at a discount price while the
complementary parts cost higher. It provides perceived value of data as a value
in return.
The demand-oriented platform offers high quality and reliable data at some cost and
provides commoditization and democratization of data as a value in return.
Supply oriented platform offers efficient and scalable infrastructure and provides
perceived value of data as a value in return.
Free as branded advertising offers useful data for public and provides perceived
value of data as a value in return.
The white-label development offers useful data services and Apps and provides sav-
ing in development time and budget as a value in return.
The ultimate goal of understanding the business model variations in the digital
world is to be able to analyze them to address the real-world problems that the busi-
ness faces. It’s one thing to understand what business model mean for different
230
• Developers
Cost Avoidance • Sustainable • Publishing data as Linked Data • EU, • Improve the meaning • Sustainable
publishing • Data retrieval parliaments of data and data publishing practice
solution • Government integration • Proactive data
• Cost avoidance department release
• People
Sponsorship • Free data and • Publishing process • Sponsors • Availability of data to • Revenue from
useful for public • Clients public sponsors
Dual Licensing • Free data for • Publishing data • Developers • Limited availability of • Revenue from
non-commercial • Data maintenance • Clients useful free data to added value
use public services
• High-quality data • Perceived value of
for commercial data
use
231
(continued)
Table 3 (continued)
232
Value Proposition Value Adding Process Value Network Value in Return Value Capture
Support and • High value-adding • Guarantees on data availability • Mostly • Perceived value of • Revenue
Services data service • Prioritization on bug fixes (both in business clients data • Presence in the
data and its provision) for paying service market
customers
• Timely help for customers using the
data
• Services around data visualization
• Analysis and mashing with other data
Charging for • Free, but limited • Update data • Mostly • Limited availability of • Revenue from
Changes data services • Availability of different machine- business clients useful free data to added value
• High-quality data readable formats public services
at some cost • Unconstrained numbers of API calls • Perceived value of
• More sophisticated querying data
• Access to data dumps rather than
through an API (or vice versa)
• Provision of feeds of changes to the
data
• Enhancement of the data with
additional information
• Early access to data
Increasing • Availability of • Update data • Developers • Higher quality data • Revenue
Quality through higher quality data • Cleansed data • Lawyers • Client satisfaction
Participation • Feedback • Academics and
government
• Clients as an
active player
Supporting • Open data • Publishing data • Developers • Improved business • Revenue
Primary Business supporting • Providing APIs • Clients results • Customer
strategic business satisfaction
objective
F.A. Zeleti and A. Ojo
The 6-Values Open Data Business Model Framework 233
businesses, but it’s quite another for a business to be able to distinguish different
business models and understanding what business model suits their business. The
elaboration in Table 3 shows that the 15 reported business models are ODBMs as
they are well defining each component in the context of OD and therefore can be
utilized by OD businesses. We further seek commonalities in the 15 ODBMs, and
this will give us insight into what are the core ODBM patterns available and what
OD business value disciplines can best define the model a business wish to employ.
ODBMs patterns and value disciplines aid businesses especially innovative start-
ups to define the right business model. Business model patterns and value disci-
plines are described below.
The first part of our analysis identifies the major patterns of business models
focusing on value propositions – a central element of the business model which
are directly associated with customers and external entities. The centrality of
the value proposition in the design of business models is clearly reflected in our
6V business model conceptual model in section “Conceptualization”.
Specifically, we determined the business model patterns from the 15 ODBMs
(see Table 3) by examining the similarities between value propositions as well
as a careful comparison of what each model offers, tries to achieve and how. Our
analysis resulted in five major business model patterns including Freemium,
Premium, Cost Saving, Indirect Benefits and Parts of Tools categories. In
Table 4, we describe specific ODBMs comprising each pattern. We also identify
what ‘offer,’ ‘Channel for delivering value’, and ‘Price/Value for money’ mean
and can include in each pattern.
Fermium includes Fermium, Dual-Licensing, Charging for Changes, Open Source,
and Free as Branded Advertising models. All the models in this category offer lim-
ited data free of charge and apply fees for additional request for complete and higher
quality datasets.
Premium includes Sponsorship, Support and Services, Demand-Oriented Platform,
Supply-Oriented Platform, White-Label Development and Premium models. Data in
this category is not offered free of charge. However, data are offered in high quality
and complete form at some cost.
Cost Saving includes Increase Quality through Participation and Cost Avoidance
models. Models in this category do not entirely cover the cost, but reduce the cost
of opening and releasing data by engaging participants and publishing data as
Linked Data. Data user or re-user participants play a vital role in this category as by
active participation publishing data can happen at a lower cost.
Indirect Benefit includes Supporting Primary Business model. Opening up data in
this category is strategic and releasing open data naturally supports the primary goal
234 F.A. Zeleti and A. Ojo
of the business. Model in this category allows the business to develop its data and
data infrastructure by using the third-party infrastructures that are created because
the data is open and available.
The Parts of Tools includes Infrastructural Razor and Blades model. The business
strategy in this category is to offer the first set of data at a discount while offering
complementary or dependent data at a considerable higher price.
As can be seen from Table 4, most of the 15 ODBMs belong to Fermium and
Premium categories. Consequently, in the open data business community, more
emphasize is given to Fermium and Premium models than the other three categories.
Value Disciplines
vision and objectives. Value discipline helps a business to tailor value proposition to
exactly match the need. Therefore, before identifying the business model, defining
business value discipline is necessary.
Our approach to identifying the implicit value disciplines for ODBMs patterns is
based on the analysis of the model attributes such as value proposition and value in
return. Determination of the value disciplines enables analysis of the required capa-
bilities to enable attainment overall business objectives. A Delphi-like process
involving the three co-authors of the research was adopted in the analysis of Table 3,
resulting in four types of value disciplines for OD businesses. The identified value
disciplines converged on Usefulness, Process Improvement, Performance and
Customer Loyalty, which are explained below:
Usefulness, tailors, value proposition to support directly the needs of consumers in
one way or another. Business strategic focus, corporate vision, and business
objectives should be defined to meet usefulness of the offer. Usefulness is associ-
ated with the Freemium, Dual-Licensing, Charging for Changes, Open Source
and Free as Branded Advertising. These models all somewhat focus on the use-
fulness of the data offered to the clients as the business value disciplines.
Process Improvement, tailors value proposition to match to the needs of the cus-
tomer specifically for improving processes. Process improvement is associated
with Cost Avoidance model. Business oriented on Process Improvement, aim at
greater efficiency to reduce cost by optimizing its processes. OD published based
on this discipline targets improving business processes.
Performance, tailors, value proposition for better performance. Performance is
associated with Support Primary Business model. Businesses with this orienta-
tion aim to release data which support their primary business objectives.
Customer Loyalty, tailors, value proposition to target customer loyalty. This is asso-
ciated with Premium and Infrastructural Razor and Blades. Business with
Customer Loyalty value discipline should apply Premium or Infrastructural
Razor and Blades model to adjust their processes to meet the clients’ satisfaction
and build customer loyalty.
Table 5 shows that Usefulness value discipline is the most popular value disci-
pline in the open data industry followed by the Customer Loyalty.
Summary
Finally, we organize existing ODBMs regarding their inherent value disciplines and
their respective categories as shown in Table 5. For instance, an OD business which
aims to focus on customer loyalty can have two choices for their business model
which are Infrastructural Razor and Blades and Premium. Business can choose one
depending on the business model category they aim to target.
For OD businesses aiming at increasing performance as their value discipline
can have one choice for a business model which is Support Primary Business.
236 F.A. Zeleti and A. Ojo
Conclusion
disciplines; is missing and literature on ODBMs is also very limited to some num-
ber of websites and presentation files. Besides, regarding business models, various
scholars present generic business model differently.
Our research findings clearly answered to the problems above both at the research
and business levels. We also confess that the 6V business model conceptual frame-
work, core ODBMs patterns – Freemium, Premium, Cost Saving, Indirect Benefit
and Parts of Tools – and new OD business value disciplines – Usefulness, Process
Improvement, Performance and Customer Loyalty – contribute significantly to busi-
ness model and ODBMs literatures and assist not only start-ups and SMEs but also
big businesses to deliver full value to their stakeholders.
This study provides insight to governments and government authorities by pro-
viding knowledge of the importance of availability and accessibility of OD for inno-
vation and transparency. This allows more businesses and development of OD
products like APIs. For example, with a focus on realistic local solutions, initiatives
like CitySDK are working with pilot cities to create uniform APIs that have standard
approaches to how APIs expose local government data. Therefore, governments
have a new way of saving and making money by becoming a provider for the city.
By opening the data, governments allow city (businesses and developers) to create
products. Governments can also establish a partnership with private sectors to ben-
efit. Therefore, governments should seek to identify how publishing OD can be
done in a way that it provides value to general public and facilitates the development
of both free and commercial products.
References
Davies T, Perini F, Alonso JM (2013) Researching the emerging impacts of open data ODDC
conceptual framework
Deloitte (2012) Open data driving growth, ingenuity and innovation
Doligalski T (2010) Strategies of value proposition on the Internet. Perspect Innov Econ Bus
5(2):2–4
Epimorphics Ltd (2012) Linked open data business models. Epimorphics Ltd. [Online]. Available:
http://www.epimorphics.com/web/blogs/martin.
Fensel D (2013) Steps towards a data value chain, Salzburg
Ferro E, Osella M (2013) Eight business model archetypes for PSI re-use, London
Graeme Pietersz (2013) Razor-blade model. Graeme Pietersz. [Online]. Available: http://money-
terms.co.uk/razor-blade-model/
Hamel G (2000) Leading the revolution, New York
HM Government (2013) Seizing the data opportunity: a strategy for UK data capability, UK
Howard A (2014) Open data economy: eight business models for open data and insight from
Deloitte UK. O’Reilly Media, Inc.,. [Online]. Available: http://radar.oreilly.com/2013/01/
open-data-business-models-deloitte-insight.html.
Huber MH (2011) Designing your business model slideshare.net. [Online]. Available: http://www.
slideshare.net/hubesm/designing-your-business-model
IBM Business Consulting Services (2005) Component business models Making specialization real
IBM Institute for Business Value Government (2011) Opening up government: how to unleash the
power of information for new economic growth, New York
Janssen M, Zuiderwijk A (2014) Infomediary business models for connecting open data providers
and users. Soc Sci Comput Rev 32:694–711
Julien N (2012) Business opportunities arising from open data policies. Imperial College London
Krcmar, H, Böhm, M, Friesike S, Schildhauer T (2011) Innovation , society and business: Internet-
based business models and their implications. In: 1st Berlin symposium on Internet and society
Lambert S (2008) A conceptual framework for business model research. In: 21st Bled eConference
on eCollaboration: overcoming boundaries through multi-channel interaction, pp 277–289
Lambert SC, Davidson RA (2013) Applications of the business model in studies of enterprise suc-
cess, innovation and classification: an analysis of empirical research from 1996 to 2010. Eur
Manag J 31(6):668–681
Lüdeke-Freund F (2009) Business model concepts in corporate sustainability contexts for “busi-
ness models for sustainability”
Mai L, Zhang Z, (2011) The Freemium Business Model in Gävleborg’s Open Source Software
Industry – a Case Study on ArcMage AB
Manyika J, Chui M, Groves P, Farrell D, Van Kuiken S, Doshi EA (2013) Open data: unlocking
innovation and performance with liquid information
Morris M, Schindehutte M, Allen J (2005) The entrepreneur’s business model: toward a unified
perspective. J Bus Res 58(6):726–735
Musings J (2012) Open data business models. [Online]. Available: http://www.jenitennison.com/
blog/node/172
Osterwalder A (2004) The business model ontology: a proposition in a design science approach,
UNIVERSITE DE LAUSANNE ECOLE
Osterwalder A. Pigneur Y (2009) Business model generation
Plé L, Lecocq X, Angot J (2008) Customer-integrated business models: a theoretical framework.
Management 13:226–265
Relations K, Huijboom N, Van Den Broek T (2011) Open data: an international comparison of
strategies
Ren G-J, Glissmann S (2012) Identifying information assets for open data: the role of business
architecture and information quality. In: IEEE 14th international conference on commerce
enterprise computing, pp 94–100
Shafer SM, Smith HJ, Linder JC (2005) The power of business models. Bus Horiz 48(3):199–207
Teece DJ (2010) Business models, business strategy and innovation. Long Range Plann
43(2–3):172–194
The 6-Values Open Data Business Model Framework 239
The 451 Group (2008) Open source is not a business model, New York
Välimäki M (2003) Dual licensing in open source software industry. Syst d’Information Manag
8(1):63–75
van den Broek T, Rijken M, van Oort S (2012) Towards open development data
Vickery G (2011) Review of recent studies on PSI re-use and related market development, Paris
Zeleti FA, Ojo A, Curry E (2014) Business models for the open data industry: characterization and
analysis of emerging models. In: 15th annual international conference on digital government
research
Zott C, Amit R, Massa L (2010) The business model: theoretical roots, recent developments and
future research, Madrid
Fatemeh Ahmadi Zeleti is a researcher at the e-Government unit at the Insight Centre for Data
Analytics @ National University of Ireland Galway (formerly Digital Enterprise Research Institute
(DERI) – a leading center in Semantic Web and Linked Open Data research). At the Insight Centre,
her research and development work addresses capabilities for open data-driven organizations and
how business (for-profit and nonprofit) entities can effectively leverage [Linked] Open [Government]
Data for optimizing business processes and product and service innovation. In 2016, her research
results have been selected as the best practice to address and achieve European PSI Directive
(2003/98/EC). She is the moderator of the World Bank Open Data Innovation Network and an active
member of the Swedish Research Network in e-Government and ImmigrationPolicy2.0.
Dr. Adegboyega Ojo is Senior Research Fellow at the Insight Centre for Data Analytics,
National University of Ireland Galway (NUIG). He leads the E-Government Group at Insight
Centre, serves as Adjunct Lecturer at the College of Engineering and Informatics. His current
research interests include data driven innovations in government, Open data policies and
Infrastructures, data analytics and governance of smart cities. He is member of the Editorial Boards
of the Government Information Quarterly and International Journal of Public Administration in the
Digital Age.
Technology Innovations in Public Service
Delivery for Sustainable Development
Jeremy Millard
Abstract This chapter focuses on how ICT can be deployed to assist in the design
and delivery of innovative public services in support of sustainable development. In
many parts of the world, and especially in developing countries, basic public ser-
vices like education, health, basic infrastructures, as well as water and sanitation,
are often poor and patchy even when available. Such services are reflected in the
United Nations Sustainable Development Agenda for 2030, and also address chal-
lenges like poverty, food, housing and employment. All of these need innovative
public service delivery if targets are to be achieved by 2030. However, the provision
of such services is increasingly challenged by the diversity of social needs across
different locations and population segments. Mainstream, largely off-the-shelf, ICT
has tremendous potential today and in the near future to innovatively address these
needs and challenges, and there are already many valuable experiences both from
developed and developing countries about how this can make huge differences to
public service delivery. The chapter also addresses the governance and policy issues
that need to be addressed in this context.
Introduction
The use of new technologies by governments in many countries around the world
has led to widespread innovations and transformations across many aspects of the
public sector over the last 15 years. The most significant technological advance over
this period has been in information and communications technology (ICT) which
has dramatically impacted public services and their delivery, both via websites and
portals over the Internet, mobile and especially smart phones and social media, as
well as being available through kiosks located in places accessible to the public.
ICT-enabled public service delivery is having significant impact by generally being
more cost effective than traditional supply investments. They also give the service
user large benefits in terms of access, convenience through 24/7 availability, savings
J. Millard (*)
Danish Technological Institute, Taastrup, Denmark
e-mail: Jeremy.millard@3mg.org
in time and the cost of travel to physical premises, as well as the possibility of com-
pletely new types of public services.
The role of ICT in public service delivery needs to be seen in context. First, espe-
cially in developing countries, non-digital service delivery channels, such as tradi-
tional post, telephone call centers, over the counter face-to-face services in citizen
centers, as well as TV and radio, remain important. However, these can be signifi-
cantly improved by adding a digital channel, for example, by using satellite broad-
casting and multi-channel learning services through mobile Internet centers
connecting teachers, learners and communities. The back offices of service provid-
ers can also be digitized and joined-up to provide innovative solutions for enhancing
service delivery.
Secondly, many service components require direct human interaction in health,
care, education and building personal and trusting relationships through dialogue
and empathy, where ICT can be a valuable support tool for front-line staff. ICT is
being used innovatively to provide instant access to remote and hard-to-reach people
regardless of time or location over large areas and distances. ICT solutions have
been used to better handle and analyze large amounts of data in more standard, rou-
tine and rule-governed processes and transactions, thereby reducing overall transac-
tion costs and increasing process efficiency. It can, when this is rationally planned,
enable the re-deployment of staff and other resources away from routine government
processes into face-to-face engagement with users where this adds most value.
Third, ICT has emerged as a key tool for capacity building by streamlining admin-
istrative processes and providing opportunities for the learning and training of public
servants. ICT should thus be seen as a very powerful additional channel enabling new
types of innovative service delivery. It should be on the agenda of all public service
providers alongside existing traditional channels, and for delivering new services for
which previously there were neither adequate resources nor means of delivery.
A basic aim of innovative public service delivery is to ensure the universality of
basic services in order to “leave no one behind”1 as this is absolutely essential in
order to achieve the United Nations Sustainable Development Goals (SDGs) by
2030.2 Given that many developing countries still have not been able to deliver basic
services like education, health, water and sanitation, as well as infrastructures and
other utilities, to all their population regardless of who they are or where they live,
a much greater focus on ICT needs to be made given its extremely low cost, its
power of reach and its ability to be relatively rapidly rolled out (Ericsson and the
Earth Institute 2016). Thus, the aim in these countries is to ensure access for all
through the universal availability of such basic services. The more developed econ-
omies have, by and large, already realized universal access, so here the focus tends
to be on more advanced and personalized services enabled by ICT as the next step,
although there are also many examples of such services in developing countries as
this chapter demonstrates.
http://www.una.org.uk/content/global-development-goals-leaving-no-one-behind
1
http://www.un.org/sustainabledevelopment/sustainable-development-goals/
2
Technology Innovations in Public Service Delivery for Sustainable Development 243
service delivery. This chapter focuses largely on this issue, together with the use of
such mainstream ICT alongside and in support of the more traditional delivery
channels mentioned above.
Whereas some countries have exploited the full potential, there remain large dis-
parities in whether and how ICT is used for innovative public service delivery
between global regions, between and within countries, between different target
populations and for different types of services. In particular, many developing coun-
tries have hardly begun to exploit ICT in this way, despite the huge potential bene-
fits in doing so. In developing countries in general, there is a lack of reach and
quality of public services, a lack of efficiency and effectiveness in service provision,
and infrequent linking to issues of good governance. ICT can do much to directly
remedy these disparities which this chapter addresses.
The rest of this chapter examines, in turn, four main focus areas in which technol-
ogy innovations can significantly enhance public service delivery in the context of
sustainable development.
• Section 2: “Improving Access to Basic Public Services Using Innovative ICT”
• Section 3: Strengthening the Governance of Basic Services Through Innovative
Technology Solutions
• Section 4: Meeting the Social Needs of Target Populations
• Section 5: Enhancing the Policy and Strategic Framework for Basic Services
In each of these areas, the analysis and examples will focus mainly on the so-
called universal ‘basic’ public services of education, health and water, as well as on
infrastructure and utilities, as these are fundamental building blocks needed for sus-
tainable development to take place.
The analysis in each of the above sections consists of two main parts. First, an
examination of the challenges which need to be addressed. Despite the great poten-
tial of ICT for innovating public service delivery and the fact that many achieve-
ments have already been made, there remain many significant challenges to
overcome, especially inequity in digital infrastructures and services between and
within countries; insufficient use of technologies to improve governance; the highly
diverse social needs of target populations which need to be met; and inadequate
policy and strategic frameworks.
The second part in each chapter will examine how technology innovations can
significantly improve public service delivery for sustainable development.
Governments are by far the most important actor in public service delivery, although
this often takes place in collaboration with non-public actors, so their role is also
critical in meeting the SDGs by 2030. This means that, rather than relying only on
existing solutions which are clearly failing to make the progress needed, a signifi-
cant investment in ICT will be necessary. Research undertaken by Ericsson and the
Technology Innovations in Public Service Delivery for Sustainable Development 245
Earth Institute (2016) characterizes the SDGs as highly ambitious “stretch” goals
that “will require a transformation of societies that is far deeper and faster than in
the past. If they are to be achieved, these goals must leverage existing and widely
deployed technologies, such as broadband, but also require new innovative services
and improved reach of technological solutions.” “ICT will play a special role in
today’s low-income countries, a point strongly and cogently emphasized by the
UN’s Broadband Commission (Broadband Commission 2014). In essence, ICT are
“leapfrog” and transformational technologies, enabling all countries to close many
technology gaps at record speed.” According to the United Nations (2015), public
service delivery comprises a number of strategies. First, both enhancing both the
quality and reach of public services, as well as improving the efficiency of public
service delivery. Second, providing strong public sector leadership and government
structures, and collaborating with non-state actors, especially civil society, and shar-
ing information and resources. Third, improving the overall effectiveness of the
public sector and promoting good governance. All of these strategies can be signifi-
cantly enhanced by the use of ICT and other new technology, as will be demon-
strated for each of the four focus areas mentioned above.
The scope of this chapter is global given that the SDGs have been agreed by
virtually all member states of the United Nations in September 2015 as part of the
2030 Sustainable Development Agenda.3 Thus, the examples used to illustrate each
of the four areas are drawn from countries around the world. The final section
“Conclusions and Lessons Learnt” of the chapter provides overall conclusions and
lessons learnt.
Challenges to Be Addressed
There are significant disparities in digital infrastructure and services between coun-
tries, within countries and between various groups within a country. These consti-
tute serious barriers to the successful use of new technologies for the delivery of
public services especially in the developing countries which need it most. The ineq-
uity directly contributes to socio-economic inequalities generally, reflecting the
multivariate causes of poverty and deprivation in many countries and which seri-
ously challenges the potential for sustainable development (Millard 2015a).
For instance, new technologies including ICT, can assist in education and learn-
ing but its innovative use is limited in many countries. Even where Internet access
is available, many people in developing countries do not possess the necessary
basics of computer literacy. Figure 1 illustrates wide inequities among regions with
Africa as a region lagging far behind others.
http://www.un.org/sustainabledevelopment/development-agenda/
3
246 J. Millard
Fig. 2 Public Education Institutions with Computer-Assisted Instruction (CAI) in 2012 (UNESCO
Institute for Statistics (2014). ICT Infrastructure 1 – ISCED 1,2 and 3. Retrieved from: http://www.
uis.unesco.org/DataCentre/Pages/BrowseCommunication.aspx)
The challenge in many developing countries is both lack of opportunity for basic
education as well as limited programmes in schools for computer literacy. There are
also wide disparities in the access and use of ICT in education among developing
countries. For instance, as Fig. 2 shows, while 100% of the public institutions in
Malaysia provide computer-assisted instruction (CAI) at the secondary school level,
only 3% of public schools in Cambodia and Madagascar do so.
Technology Innovations in Public Service Delivery for Sustainable Development 247
4
Internet language data is from “The Usage of Content Language for Website Survey” from
W3Techs, http://w3techs.com/technologies/overview/content_language/all, retrieved February
2016; Internet user data is from “The Internet World User by Language Survey” from Internet
World Stats, http://www.internetworldstats.com/stats7.htm, retrieved February 2016.
248 J. Millard
functional literacy restricts access to these as well. The global e-learning market was
expected to reach $107 billion by 2015, with the global self-paced e-learning market
component of this reaching $49.9 billion in 2015.5 Although the largest global mar-
ket share remains concentrated in developed countries, some emerging economies
are now starting to grow much faster and to catch up (Docebo 2014). Figure 4 indi-
cates that the top ten growing countries consist largely of such emerging economies,
but again shows that Africa is largely missing from this development.
In terms of e-health, Fig. 5 shows the total global market size forecast for 2016
and, once again, underlines the dominance of developed countries and especially
the USA. However, the sheer size of China with the rapid growth in demand for
health and e-health services is also evident.
A large part of the lack of optimal use of new technologies for service delivery is
digital infrastructure disparities, including factors such as the cost and quality of
ICT connection and related services available to users. The comparison between
different ICT infrastructures across types of countries in Fig. 6, shows significant
gaps between developed countries on the one hand and developing and less devel-
oped countries on the other. However, this gap is least in relation to mobile cellular
access given that demand here is strongest because of the very large price, conve-
nience and ease-of-use advantages compared with the other infrastructures.
As noted above, the innovative use of ICT is needed to fully meet the SDGs by 2030
as this makes it possible for millions more people in developing countries who pre-
viously had no access to basic public services to be able to do so, for example in
Fig. 6 ICT access by development status, 2015 (International Telecommunications Union 2015)
250 J. Millard
under-served areas like shanty towns or rural and remote locations.6 “Leaving no
one behind”7 in this way has been a key feature of preparations for the SGDs and the
2030 Sustainable Development Agenda, and it is clear that the use of ICT is an
important component of this objective.8 For instance, in Tanzania, an innovative
poverty mapping tool using GIS data has been developed to improve geographical
identification of the poorest villages, which are beneficiaries of the country’s social
protection scheme.9 ICT can also improve the speed and convenience of public ser-
vices access through 24/7 availability, considerably reduce the cost of access, and in
some cases provide completely new types of public services. As mentioned in sec-
tion “Introduction”, the use of ICT can also enhance and complement the continued
use more traditional channels, such as over the counter, the postal service, telephone
call centers, or TV and radio, where these remain more appropriate forms of com-
munication and interaction with public officials or other service providers.
ICT can dramatically widen access to and the scope of education for anyone with
a computing device and access to the Internet or mobile network. It can provide
anywhere-anytime education, facilitate personalized education and new learning
environments, as well as provide data for learning analytics which can be used to
plan and monitor educational provision as well as precisely target services to spe-
cific needs and groups. ICT can also support crowd learning and citizen inquiry, and
make education more attractive and appealing using gaming approaches. The Quest
to Learn (USA)10 is a collaboration between the Institute of Play, New Visions for
Public Schools and the New York City Department of Education. It combines learn-
ing and gaming to meet the needs and interests of children who are anyway increas-
ingly engaging in digital media platforms. The Professor-Why project in Poland11
combines computer generated images with real images and introduces online stu-
dents to the world of science to be explored both at school and at home, as well
as supporting virtual experiments. The project has had significant impact due to
dissatisfaction with the current form of chemistry education, and the lack of
real experiments in schools which greatly impoverishes the study of chemistry.
6
Text of speech of Mahmoud Mohieldin. Corporate Secretary and President’s Special Envoy. The
Independent Commission on Multilateralism, New York, United States at the 2030 Agenda for
Sustainable Development and Addressing Climate Change. The World Bank. November 12, 2015.
http://www.worldbank.org/en/news/speech/2015/11/12/as-prepared-for-delivery-the-2030-agenda-
for-sustainable-development-and-addressing-climate-change-the-independent-commission-on-
multilateralism
7
http://www.una.org.uk/content/global-development-goals-leaving-no-one-behind
8
https://www.odi.org/sites/odi.org.uk/files/odi-assets/publications-opinion-files/9534.pdf
9
Text of speech of Mahmoud Mohieldin. Corporate Secretary and President’s Special Envoy. The
Independent Commission on Multilateralism, New York, United States at the 2030 Agenda for
Sustainable Development and Addressing Climate Change. The World Bank. November 12, 2015.
http://www.worldbank.org/en/news/speech/2015/11/12/as-prepared-for-delivery-the-2030-agenda-
for-sustainable-development-and-addressing-climate-change-the-independent-commission-on-
multilateralism
10
www.q2l.org
11
www.professor-why.pl
Technology Innovations in Public Service Delivery for Sustainable Development 251
12
http://izonenyc.org/ in New York.
13
Yuan, L (2013). MOOCs and Open Education: Implications for Higher Education. CETIS;
http://blog.coursera.org/post/29062736760/coursera-hits-1-million-students-across-196-coun-
tries; and http://www.nytimes.com/2013/11/01/education/us-plans-global-network-of-free-online-
courses.html?hpw&_r=3&
252 J. Millard
Leaving no one behind in health care is also an area being revolutionised by ICT
both through greater outreach, for example, by mobile phones, as well as in innovat-
ing public service delivery through service personalization and thereby more tar-
geted high quality and convenient health services for individual patients, alongside
traditional ‘warm hands’ interaction with public service staff. More sophisticated
ICT is also becoming more mainstream in many countries, like body sensors to
monitor a patient’s condition, or the analysis of a patient’s data over time in the
context of large numbers of similar patients which can speed up diagnosis and make
treatment more appropriate. The role of process, operational and organisational
innovations supported by ICT is also very important, as well as the benefits of emu-
lating innovations from commercial companies.
For example Ghana, with the support of international development organiza-
tions, is setting up an Integrated eHealth System to focus on developing founda-
tional systems and outreach to underserved communities in the country Ghana
Health Service 2016). The Patient Briefcase initiative in Denmark is a remote ser-
vice connecting the patient in her/his own home with professional medical and care
personnel through live video and audio channels over a broadband Internet link
(European Commission 2013). It is the result of collaboration between the public
and private sectors, originally supported by public innovation funds (both European
and Danish), as well as private investment by the company involved itself, and today
functions as a fully commercial operation. The service places strong focus on user-
friendliness and making it easy for patients to be ‘admitted to hospital’ in their own
homes. It is also an example of how the hospitals as service provider have been able
to significantly change they way they work to fully exploit the technology by caring
at a distance whilst routine aspects are automated. This frees up staff time to have
more personal contact with patients who need it, as well as dramatically decreases
transport costs and carbon emissions, whilst saving staff and patient waiting time.
The PatientsLikeMe service in the UK14is a free patient online network where
people can connect with each other to better understand their diseases, share condi-
tion and treatment information, and get support from peers to improve their health.
It is also a research platform for medical staff given that when patients report on
their disease experiences, they provide real-world insight into diseases and long-
term conditions based on their anonymized data. Those insights are shared with
companies, government organizations and others who use them to continuously
develop more effective products, pharmaceuticals, services and care.
More standard and inexpensive ICT can also have a significant impact on health
issues. mPedigree is an African-based for profit company spun out of a non-profit
organization founded by a Ghanaian social entrepreneur. Launched in 2007, it
works with mobile operators and pharmaceutical manufacturers to provide a mobile
phone-based drug verification system for addressing the issue of counterfeit drugs
in pharmacies at the point-of-sale in Ghana, Kenya, and Nigeria. The mPedigree
service is free to users and allows instant verification of whether a drug is real or
counterfeit by sending a unique code via simple SMS and getting an automated
response in appropriate language. The service relies on various partners across the
14
www.patientslikeme.com
Technology Innovations in Public Service Delivery for Sustainable Development 253
value chain, both private and public, while remaining simple to roll-out to new cus-
tomers and easy to access for the end-user.15
All these uses of ICT illustrate the many different ways ICT can be used to meet
the SDGs. There are also examples of how traditionally expensive and high cost
healthcare can deploy ICT to support advanced technology in meeting the SDG for
healthy lives and the promotion of well-being for all ages by providing cheap but very
effective services through innovations on the provider side and in the value chain.
Source: http://www.narayanahealth.org
One of Narayana Health’s main specialities is cardiac care which only
about 8% of the world’s population can normally afford. In the US, cardiac
surgeries can cost up to $50,000 compared to about $6000 in India, whilst
Narayana health has reduced this down to less than $3000, irrespective of the
complexity of the procedure. In serving the poor, Narayana health reaches
out to patients through a network of rural clinics and via telemedicine facili-
ties. Patients come to the Bangalore facility from more than 50 countries.
The philosophy is that no one will be denied treatment due to a lack of funds.
This dramatic reduction of costs has not reduced quality as the group has a
mortality rate of about 2% and hospital-acquired infection rate of 2.8 per
1000. This favorably compares to the best hospitals across the world.
Source: United Nations (2015)
http://www.eldis.org/go/topics/resource-guides/icts-for-development/icts-and-health&id=
15
61115&type=Document#.Vw7UqKT2aAw
254 J. Millard
16
ICT to improve water governance: World Water Week in Stockholm (2013): http://programme.
worldwaterweek.org/event/changing-relationships-ict-2882.
17
WASH is Water, Sanitation and Hygiene Services.
Technology Innovations in Public Service Delivery for Sustainable Development 255
Using a combination of smart phone mapping and data analysis using ICT, pro-
vides valuable means to pinpoint precise water source locations, as well as the
amounts and quality available. ICT also enables understanding of how these link
together and can best be exploited and monitored in a sustainable manner to provide
basic services. Both Water Point Mapping and monitoring of water delivery and
quality typically take place through public-private partnerships between govern-
ments and local authorities, on the one hand, and commercial and/or donor agencies
on the other providing both technology and expertise. Partnerships with civil society
organizations are often also important through local volunteers undertaking the
mapping and providing local knowledge and support on behalf of the community.
Infrastructures and utilities provide the basic physical and organizational facilities
needed for the operation of society and for the services necessary for an economy to
function, such as roads, water supply, sewers, power grids, and telecommunications.
The goal needs to be to provide a universal service of such basic infrastructure, as in
the Ethiopia example. ICT is itself an important basic infrastructure, but can also be
used both to provide access and to deliver better quality utilities in order to leave no
one behind, for example, smart power and water grids, road and congestion monitor-
ing and coordinating public transport. For example, a study on the use of mobile
devices in Kenya found that 25% of users could get more work and earn more money
because they were more ‘reachable’ (Crandall et al. 2012). According to the eTrans-
form Africa report (World Bank 2012), easier access via mobile and broadband “is
quickly changing lives, driving entrepreneurship fuelled in part by collaborative tech-
nology hubs, and delivering innovation and home-grown solutions for Africa.” The
report focuses on eight key areas: agriculture, climate change, education, financial
services, government, health, ICT competitiveness, and trade facilitation and regional
integration. It emphasizes the need to build a competitive ICT industry to promote
256 J. Millard
innovation, job creation, and boost the export potential of African companies. Part of
this is the flowering of technology hubs across Africa – such as iHub and NaiLab in
Kenya, Hive CoLab and AppLab in Uganda, Activspaces in Cameroon, BantaLabs in
Senegal, Kinu in Tanzania or infoDev’s mLabs in Kenya and South Africa. These
hubs are creating new spaces for collaboration, innovation, training, applications and
content development, and for pre-incubation of African firms (GSMA 2014).
Given the explosive growth of towns, and especially cities in developing countries,
there is also an increasing importance of so-called smart cities using ICT to provide,
interconnect and improve the efficiency and effectiveness of basic infrastructures.
in order to improve the management of urban flows and enable real time responses
to problems arising. Through the use of sensors integrated with real-time monitor-
ing systems, data is collected from both people and things, through the so-called
Internet-of-Things. The data is then processed and analyzed for the purpose of
enhancing the quality, performance and interactivity of urban services, and thereby
reduce both costs and resource consumption.
Challenges to Be Addressed
Fig. 8 The World Justice Project’s Open Government Index, 2015 (http://worldjusticeproject.org/
open-government-index)
Technology Innovations in Public Service Delivery for Sustainable Development 259
as the World Justice Project report shows that, while in general, high-income coun-
tries attain higher Open Government Index 2015 scores, in developing countries there
is no relationship between GDP per capita and open government. “This suggests that
the level of government openness is not necessarily driven by economic resources”,
and thus there is real opportunity for developing countries in particular to improve
their open government performance without linking this directly to economic growth.
And, as the WJP report points out, open government can itself be an important enabler
of economic growth as it engenders trust in the ability to invest and do business. Both
Figs. 7 and 8 show clear correspondence between measures of open government and
the control of corruption, on the one hand, with disparities in digital infrastructures
and services across countries and between global regions on the other.
The innovative use of technology can directly support good governance through
greater outreach, openness and effectiveness in the delivery of services. Such inno-
vations can arise from enhancements in the capacity of governments to enable new
technologies for a framework of good governance and public institutions that are
efficient, effective, transparent, accountable, inclusive and participatory. ICT also
provides a medium for building partnerships amongst all stakeholders for better
service delivery. In recent years, so-called ‘open government’ has become an impor-
tant feature of good governance and one of the main pillars of ICT-enabled public
sector innovation and public service deliver in many countries through the opening
up of government data, processes and services using ICT as an indispensable tool.
The US administration launched the Open Government Partnership in 2011 based
on the principles of transparency, participation and collaboration, and by early 2016
reached a total of 69 member countries from around the world.18
ICT can be used to increase social awareness, advocacy and feedback concerning
the lack or poor quality of basic utilities and services, for example using social
media and mobile devices. It can thereby help change the behaviour and attitudes of
both citizens and service providers alike. If citizens can provide feedback to govern-
ment about service delivery using the increasingly ubiquitous mobile channel, even
in places with little or poor infrastructure, and rate the quality of specific programs,
then government will have more information to prioritize services and should be
more accountable to citizens. A project in urban India uses mobile technology to
track how citizens experience water service delivery.19 Citizen feedback is collect-
ing and analyzed using innovative mobile applications, thereby providing a ‘reality
check‘ on service levels from the citizens’ standpoint. It gives city managers more
granular data at the sub-city level (ward/zone) which can facilitate improved moni-
toring and problem solving, and provides inputs into project planning processes for
service providers. Most importantly, the project provides a suitable platform to
http://www.opengovpartnership.org
18
http://www.wsp.org/FeaturesEvents/Features/using-technology-track-how-citizens-experience-
19
water-service-delivery-india
260 J. Millard
Nairobi water had poor response times when dealing with customer com-
plaints and it did not have strong, direct links with customers. Given there are
at least 30 million users of mobiles in Kenya, the company now enables cus-
tomers to report service exploitation and receive news updates on water sup-
ply using their mobile phone. Customers can be sent updates by text, including
photos from engineers when they repair a leak. As a result, the number of
reported leakages has doubled since the introduction of MajiVoice, so there
has been a much improved service performance through greater accountabil-
ity, which directly helps customers voice critical service issues more easily
without needing to visit an office. This also enables staff to process and
resolve complaints faster, and strengthens management and regulation through
better data based on the collection of customer service data.
Source: ICT to improve water governance: World Water Week in Stockholm
(2013): http://programme.worldwaterweek.org/event/changing-relationships-
ict-2882
ICT is an excellent tool for collecting, analyzing and updating so-called big data in
order to improve service efficiency and effectiveness, although it does rely on the data
being representative and of good quality. Open data implies that big data is open to oth-
ers to scrutinize and validate the data collected, for example, by governments or other
service providers, as well as themselves to contribute to the data and use it for their own
purposes. A health project in Cambodia combatting malaria is a good example of this.
Cambodia Malaria Information System (MIS)
Established in 2003, the malaria consortium is composed of national malaria
control programs, research institutions plus commercial and civil society
organizations, with the aim to share learning and discuss key issues. In
Cambodia prior to 2009, malaria case data came from a national system which
provided aggregate data at operational district level, but not down to village
level. In 2009 the malaria information system (MIS) was developed by the
malaria consortium, together with national program staff, to help process
malaria data from village malaria workers and village health facilities using
open source software for SMS reporting via mobile phones. This also provides
a tool for district staff to manage their activities, such as mosquito net distribu-
tion and stratification at village level. The MIS also draws on other data, such
as individual case data for all patients, mosquito net distribution data, demo-
graphic data on villages and data on the type and location of private sector
outlets such as clinics and pharmacies. The MIS incorporates a ‘drug stock
out’ system tracking drug stock levels in health centers and clinics around the
country reported every 2 weeks or when levels drop below a set threshold.
A feature of the MIS is comprehensive data reporting and graphing. It
allows exporting raw data and mapping of data using Google earth. The sys-
tem is now decentralized to all 44 targeted operational districts in Cambodia,
and also available for researchers and policy makers to improve malaria sur-
veillance and cooperation with other countries.
Stopthebribes, Nigeria
Stopthebribes is a crowdsourcing platform for accessing public feedbacks on
the conduct of Nigeria police force personnel. The platform receives com-
plaints through multiple channels, including mobile, SMS, twitter, Facebook,
email and direct reporting onto the website. Stopthebribes therefore promotes
inclusive policing by involve members of the public in oversight of the police,
thereby engendering public accountability and transparency.
Reports on the platform are closely monitored and acted upon by the
Nigeria police force responsible for ensuring internal accountability. The plat-
form is operated by the Nigeria police force and thus eliminates bureaucracies
and limitations that hitherto characterized making complaints or observations
on police conduct.
Source: http://www.stopthebribes.net/page/index/7
Technology Innovations in Public Service Delivery for Sustainable Development 263
The above examples demonstrate the clear benefits of close cooperation between
civil society and the public sector in improving public service delivery using
ICT. However, ICT can be a double-edged sword as it can also be used by corrupt
government officials who have access to databases and applications in government
back offices. Without adequate supervision and a robust code of conduct, they can
manipulate ICT systems for their own benefit. In order to increase awareness of this
challenge, a report and checklist has been prepared by the EU-supported Regional
School of Public Administration in the Western Balkans (ReSPA 2013).
ICT provides the communication tools for service users to directly participate in
the design and delivery of services, as many of the above examples have also demon-
strated. Another prominent example is participatory decision-making and budgeting,
264 J. Millard
an approach that allows citizens to discuss and vote on how some parts of a govern-
ment’s budget should be used. The archetypal example at Porto Alegre in Brazil is
recognized internationally as a ground-breaking initiative at the local level where the
state government has engaged over one million residents in its multi-channel (online
and offline) participatory decision-making in the provision of a whole range of public
services and utilities.20 There are also examples of participatory decision-making
using mobile technology for public services in Cameroon21 and in South Kivu in the
Democratic Republic of the Congo.22 In the latter example, communities were given
the chance to voice their basic service needs which the government responded to. As
a result, tax collection rates increased as people have come to believe that their gov-
ernment can actually deliver valuable services, and this may demonstrate one way to
increase tax collection in developing countries, where such rates are notably low.
Another case from the Basque Country in northern Spain shows how the government
is using ICT to take the initiative in involving citizens in decision making.
Both in the Porto Alegre and the Irekia examples, the government has itself taken
the initiative to use digital technology to make it possible for citizens and users of
public services to become involved in policy and decision making. Like the other
cases presented above, the use of the technology in this way, directly supports many
of the aspects of good governance as defined in chapter “European Strategies for
e-Governance to 2020 and Beyond” of this report, including accountability and con-
trol of corruption, openness and transparency, efficiency and effectiveness, as well
as participation and collaboration.
20
http://odta.net/post/technology-drives-citizen-participation-and-feedback-in-rio-grande-do-sul-brazil-0
21
http://odta.net/post/participatory-budgeting-cameroon
22
http://blogs.worldbank.org/ic4d/mobile-enhanced-participatory-budgeting-in-the-drc
Technology Innovations in Public Service Delivery for Sustainable Development 265
Challenges to Be Addressed
The social needs of target populations are directly reflected in the SDGs, ranging from
education, health, basic infrastructures, water and sanitation, as well as challenges
such as poverty, food, housing and employment. All of these need innovative public
service delivery if targets are to be achieved by 2030. However, the provision of public
service is increasingly challenged by the diversity of social needs across different loca-
tions and segments related to, such as ageing societies, digitally-savvy populations,
economic pressure, and unequal conditions for public service delivery existing within
and across countries. For example, the failure of public service delivery in many devel-
oping countries is not just due to the scarcity of resources but also to the problems of
incentives, accountability and governance that vary from one context to another.23
For example, there are important gender disparities between male and female
Internet usage which are much greater in developing and less developed countries
than in developed countries, as shown in Fig. 9.
Fig. 9 Gap in Internet usage between males and females and by development status, 2013 and
2015 (ITU 2015)
Fig. 10 Percent of individuals using the Internet by income group, 1997–2013 (World Economic
Forum 2015)
Figure 9 also shows that, although this gap generally fell between 2013 and
2015, it tended to reduce more slowly in the developing and less developed coun-
tries, indicating that progress is slower here. Such gender differences are important
for the use of ICT-enabled public services given that women, as prime users of basic
services in their role as mothers and caregivers, are often more severely affected
than men by poverty, lack of employment, lack of water, inadequate maternal child
health care and lack of education opportunities.
Household surveys by Research ICT Africa, conducted in 12 African countries in
2011, also reported a close relationship between Internet access differences by gen-
der, level of income, level of education, extent of disability and other variables
(Dean-Swarray et al. 2013). That income differences mark a sharp diversification in
usage of the Internet, is depicted in Fig. 10, which also shows that the lowest income
groups generally increase their take-up of the Internet more slowly than higher
income groups.
Furthermore, as depicted in Fig. 11, there are stark differences between urban and
rural dwellers in most countries as exemplified by mobile 3G coverage, generally
necessary for the operation of smart as opposed to dumb phones and thus access to
more sophisticated services and data. Even though global urban inhabitants surpassed
50% of the total population for the first time in 2009, the size of the rural population
will remain large for many years, and indeed there are important sustainable develop-
ment reasons for keeping as many people in these areas as possible. Appropriate
infrastructures and services outside of towns and cities are necessary to achieve this
goal, and indeed ICT provides relatively efficient and effective means of doing so.
Technology Innovations in Public Service Delivery for Sustainable Development 267
Fig. 11 Population coverage by 3G networks, urban and rural areas, 2015 (ITU 2015)
Overcoming the challenges discussed above requires in public service delivery that:
• creates and maintains an eco-system of government agencies, businesses, non-
profit organizations, universities, citizens and other actors that participate in the
provision, consumption and intermediation in public service delivery brings ser-
vices closer to the consumers through, e.g. the provision of multi-service centers
and the use of diverse delivery channels
• learns about public service provision locally and from around the world and
adapts the knowledge to the local contexts digitizes public services, tailors them
to individual needs, and delivers them through various digital channels using
new social and organizational innovation models.24
When meeting these challenges and to achieve the 2030 Agenda for Sustainable
Development, especially in relation to poor and marginalized individuals and com-
munities, new types of innovation are beginning to be deployed that are beyond but
build on conventional technology and top-down driven innovation. These prioritize
collaboration, diversity and a range of voices, skills, competencies and resources,
across all types of public sector activity, and especially for public service delivery.
OECD (2015b). The OECD Innovation Strategy 2015. Retrieved November 14, 2015, from
24
https://www.innovationpolicyplatform.org/sti/oecd-innovation-strategy-2015
268 J. Millard
Many of the new innovation forms are typified as ‘open innovation’ (Chesbrough
2003) in which all can be involved, where there are no supposed monopolies of
innovation talent and potential, and where the solutions become owned by as many
people as possible, which results in greater acceptance, trust and impact, such as
through co-creation. A specific variant of open innovation has being shown to be
highly relevant to public service design and delivery, i.e. ‘social innovation’, which
is becoming well embedded and recognized in many developed countries. This is
meeting a social need (for example for an education or health service) in a new way
that also involves collaboration with, and the empowerment of, the service user or
beneficiary. It works with them rather than just doing something to them as passive
recipients, thereby also developing their own capabilities around, and ownership of,
the service and thereby transforming their social relations and improving their
access to power and resources. In other words, social innovations are social both in
their means and in their ends (Millard et al. 2016). Many social innovations explic-
itly target the otherwise excluded, for example by adapting or developing a public
service which ensures they are not left out. It directly targets the needs of the low-
income or the base-of-the-pyramid (BoP) population (Prahalad 2004).
Social needs are highly diverse, so the public services designed to meet them
must respond accordingly. Different social groups require different types of public
services and these should be addressed in different ways depending on their unique
social needs. One size fits all public services, not tailored to specific needs, can miss
their mark and thus both waste public resources as well as prove ineffective. ICT is
a powerful tool enabling this to happen, both when used and initiated by the govern-
ment and other service providers, as well as when utilized directly by the users
themselves. For example, a highly successful initiative in Bangladesh supports peo-
ple with low incomes and low educational levels learn English.
BBC Janala: Free Interactive English Lessons for Low Income People in
Bangladesh
BBC Jamala provides interactive English language lessons to Bangladeshis in
accessible format over multiple platforms – Mobile, web, television and
newspapers – At affordable rates and completely free when necessary. The
service is delivered by the BBC world service trust to provide comprehensive
English language learning opportunities to Bangladeshis across multiple eco-
nomic and social strata. Starting in November 2009, the service has received
over 10 million calls from 3.8 million people with over 170,000 mobile inter-
net lesson downloads since launch, in addition to 20 million television and
newspaper users. The use of multiple ICT and other media platforms for
stand-alone lessons designed with hundreds of hours of testing (input and
feedback regarding Bengali language, accents, dialect, and currently specific
references), innovatively removes barriers to entry for low-income and low-
education users. Leveraging and coordinating the different platforms has led
to widespread use and high market saturation in a short period of time.
(continued)
Technology Innovations in Public Service Delivery for Sustainable Development 269
The BBC Janala case show shows a multi-channel approach, largely based on
relatively cheap ICT in the form of mobile phones as well as more traditional media
like TV, radio and newspapers. It can be hugely successful, also on a semi-commercial
basis if the business model is right, but in this case also relies on support from a
developed country aid budget. In some contrast, the Mondey project in Germany26
aims to improve early fostering and diagnosis of very young children with retarded
development by supporting parents, pedagogical staff and scientists with monitoring
and documentation of everyday situations. It provides advanced training for peda-
gogical personnel and parents in diagnostic skills using a blended educational
approach consisting of the Internet, tablets and face-to-face contact by professionals
and experts. It is an open and free educational service which uses standard tools, so
that parents, pedagogical personnel and experts can monitor and diagnose the devel-
opment of the children in their care. They can choose to document the development
of a child for themselves offline or use the interactive online database.
Developed by a social entrepreneur, the Buddy app case in the UK27 aims to
improve patients’ mental health so they feel less dependent on the therapist in the
clinic by using text messaging between therapy sessions. The dotHIV initiative in
Germany28 generates money to support HIV patients, raise awareness of the global
threat of HIV/AIDS and de-stigmatise HIV-positive people. It is an innovative
approach for raising awareness for a social problem, whilst also generating income
from sales of .hiv domain products and services that are forwarded to support projects
25
The Bottom-of-the Pyramid (BoP) population is the largest, but poorest socio-economic group
globally, at about 3 billion people who live on less than roughly $2.50 per day. (See 2004).
26
www.mondey.de
27
www.buddyapp.co.uk
28
https://click4life.hiv/de
270 J. Millard
and organizations addressing the condition. ICT is used both to raise awareness and
collect and allocate funds. A ground-breaking project in the USA has had a large
impact on expectant and young mothers from disadvantaged backgrounds.
A 2011 study showed “very high satisfaction with the service, increase in
users’ health knowledge, improved interaction with healthcare providers,
improved adherence to appointments and immunizations, and increased
access to health resources.” (National Latino Research Center 2011) partici-
pants rated text4baby as an 8.5 out of 10 overall, and indicated that 81% have
an annual household income under $40,000, 65% are either uninsured or
enrolled in California’s Medicaid program, and 75% said they learned and
followed up on a medical warning sign they didn’t know previously.
Source: https://www.text4baby.org
Technology Innovations in Public Service Delivery for Sustainable Development 271
29
Leveraging Information and Communication Technology for the Base Of the Pyramid ICT for
development in education, health, finance and agriculture: http://www.eldis.org/go/topics/
resource-guides/icts-for-development/icts-and-health&id=61115&type=Document#.
Vw7UqKT2aAw
30
E-Government as a driving force for institutional integration and effective service delivery in the
Bahamas. Presentation by Rowena Bethel, Director and CEO, The national Insurance Board/The
Bahamas Government; and D. Shane Gibson, Minister of the Public Service, Labour and National
Insurance, The Bahamas, at the Expert Group Meeting “Innovating Public Service Delivery for
Sustainable Development”, Medellin, Colombia, 23–26 June 2015.
272 J. Millard
Challenges to Be Addressed
Access to, and the quality of, public services can be vastly improved by appropriate
policies and strategic frameworks enabling the use of new technologies, both within
the public sector and between the government and citizens or businesses, and thereby
enable innovations in the delivery of public services. One of the biggest challenges
in this context is legacy policies, strategies and legal and regulatory systems, which
often reflect a pre-ICT period when all government business took place on paper or
in-person. Without changes making the use of new technologies possible, and which
can promote responsible innovation, the use of ICT will be severely curtailed.
In turn, the lack of formal legal and regulatory structures can retard necessary
changes in informal working and administrative cultures within the public sector,
and reinforce the often already embedded resistance to change. The entrenchment
of a ‘risk adverse culture’ and ‘business as usual procedures’ remains strong within
many government at all levels, creating an inherent obstacle to the introduction of
new processes, products, services and good governance that ICT enables.
Although issues like political will, leadership and resources are important in
effecting needed changes, the lack of relevant regulation frequently hampers the
adoption and use of new technologies in basic services like education health, water
and other infrastructure initiatives. Figure 12 shows that the global top 25 nations in
terms of laws related to the use of ICT, such as delivering online services, electronic
commerce, digital signatures and consumer protection, are almost only developed
countries. The only exceptions are two Gulf countries, the United Arab Emirates
and Qatar, plus Malaysia, each of which has invested heavily in ICT in government
Technology Innovations in Public Service Delivery for Sustainable Development 273
Fig. 12 Laws relating to ICT: top 25 countries (World Economic Forum 2015, p. 269)
and has enacted appropriate laws and regulations in support of both public and pri-
vate use of new technology. In contrast, the bottom 25 countries all comprise devel-
oping countries, with the exceptions of Argentina and Venezuela as emerging
economies (World Economic Forum 2015, p. 269).
there are also strong trends towards the decentralization of public service responsi-
bility and design to lower tier entities and especially cities. Trans-national jurisdic-
tions can also be important, as in the European Union where there is a long history
of cooperation and agreement on public services, especially related to e-government.
Many countries today have also entered into formal and informal learning and peer
exchange relationships with neighbouring, similar or lead nations, given that many
of the challenges are the same although contexts vary widely.
In order to meet the 2030 Sustainable Development Agenda and the SDGs,
Ericsson and the Earth Institute (2016) conclude that governments need to ensure
that the entire public sector, including service delivery in health, education, and
infrastructure, is fully supported by high-quality ICT infrastructure. This includes:
broadband connectivity of all public facilities; ICT training of all relevant public
officials and service providers; ICT-based delivery systems for healthcare, educa-
tion, and infrastructure; deployment of the Internet of Things (remote sensing and
control of connected devices) for the public infrastructure and environmental man-
agement; encouraging universities to scale up education and incubation of ICT solu-
tions, including through partnerships with the business sector; Public-Private
Partnerships (PPPs) for ICT-enabled systems; and deployment of an ICT-based
SDG information system that connects public services, public facilities, the business
sector, and the public. The accelerated uptake of ICT-based services is the key to
achieving the SDGs, not only because ICT empowers other technologies and ser-
vices, but because it itself is also one of the technologies that can accelerate uptake.
Public services delivery is one of the most expensive aspects of any govern-
ment’s budget, so it is extremely important to have the right policies and legal
frameworks for the specific context a jurisdiction finds itself in.
(continued)
Technology Innovations in Public Service Delivery for Sustainable Development 275
network utilizing ICT to set up citizen spaces for the provision of digitally
delivered services, with in-person assistance if required. This addresses the
fact that digital literacy is not at the same level everywhere in the country.
Portugal is also moving towards shared services as a means to improve
public service delivery through better use of resources. Pilots are in progress
in five action areas: Financial shared services integrating budget, property and
logistics; human resources management and the integrated system of manage-
ment and performance appraisal in the public administration; public procure-
ment; management of the state’s car fleet; and sharing the means and resources
for developing information systems and the rationalization of ICT infrastruc-
tures. Shared services provide a win-win: For citizens who can access public
entities more simply and solve minor problems faster and in a more personal-
ized way; and for the public administration that will be able to dedicate its
own resources to its core functions whilst benefitting from the common shared
resources which all entities need.
Source: United Nations (2015)
The Portuguese case demonstrates that fully embracing the digital revolution can
enable a new vision and provide better tools for service delivery, but should also go
hand-in-hand with understanding that the human element remains essential. In
order both to save money and resources as well as provide better quality services,
sharing across the public sector is needed, both of good practices and ideas but also
in terms of human, organizational and physical assets. Multiple service delivery
channels supported by the local authority as well as by local organizations and citi-
zens provide both better tailored and contextual services as well as improve the
inclusion of everybody. Blending physical, digital and voice channels addresses the
need for convenience and time saving, but also provides physical outreach to people
where they live. Human and organizational capacity building is essential for both
routine service delivery but also for promoting creativity, experimentation and inno-
vation in a continuous search for improvement.
(continued)
276 J. Millard
The Malaysian national policy and strategy for ICT and e-government is an
example showing the importance of the long-term development of the legal basis for
using ICT in e-government generally and public service delivery in particular
(Malaysia Government, undated). This is one of the reasons the country scores so
high in laws relating to ICT, as shown in Fig. 12. The case shows how policy, strat-
egy and a sound legal basis are a means for institutional capacity reinforcement for
transforming public services. It also illustrates how political will and resources,
sanctioned from the top are important, and how this also depends on long-term and
consistent commitment transcending changing political conditions.
Basic services like education, health, water and sanitation, as well as infrastructures
and other utilities, are essential for sustainable development strategies and for
improving people’s quality of life and prosperity. They need to be delivered univer-
sally in order not to leave anyone behind, as this is the only way that the SDGs can
be achieved by 2030. New technologies and ICT are essential to ensure this can
happen, both through enabling the significant widening of access as well as by pro-
viding large beneficial impacts for service users, at the same time as provider costs
can be reduced.
Technology Innovations in Public Service Delivery for Sustainable Development 277
New technology has the potential to assist in moving towards universality in the
access, reach, intensity and quality of basic public service delivery. The innovative
use of ICT can enable people to find and use basic public services in cases when
access was previously denied to them, for example in under-served areas like shanty
towns or rural and remote locations. It can also improve the speed and convenience
of public services access through 24/7 availability, considerably reduce the cost of
access, and in some cases provide completely new types of public services.
ICT can also enhance and complement the continued use of more traditional
channels, such as over the counter, the postal service, telephone call centers, or TV
and radio, where these remain appropriate forms of communication and interaction
with public officials or other service providers. The technology can lead to new
types of services based, for example, on the personalization of service offerings
through interaction with the service interface or direct with the service provider.
In many cases, close cooperation is beneficial with the private sector providing
investment and technical expertise as well as civil society organizations which are
close to service users both geographically in in terms of understanding their real
situation. Public, private and civil partnerships often provide good opportunities for
dramatically extending basic services to large numbers of people, as long as the
oversight and regulation are appropriate.’
The main lessons in summary are:
1. ICT enabled service delivery on a large scale can significantly reduces costs,
widen access and result in increased sustainable development impacts. ICT and
other technology innovations are necessary enablers and can be game changers,
but organizational, human resource and process innovations are also necessary,
the best and most suitable of which might be emulated from commercial compa-
nies or civil organizations.
278 J. Millard
2. Simple and relatively cheap technology such as mobile phones is a very flexible
tool that maximizes reach, is generally personalizable to the individual user and
enables two or multi-way interaction with the service provider as well as between
users themselves.
3. On a larger scale, more sophisticated and powerful ICT systems can knit together
other infrastructures and utilities, for example by deploying high capacity Internet,
sensors and databases to dramatically reduce costs and increase service integra-
tion and impact in real time, for example in smart city or smart neighborhood
initiatives. This can enhance the quality, performance and interactivity of services
as well as strengthen coordination through innovative technology solutions.
New technology has the clear potential to directly support good governance through
greater outreach, openness and innovations in the speed and delivery of services.
Such innovations can arise from enhancements in the capacity of governments to
enable new technologies for a framework of good governance and public institu-
tions that are efficient, effective, transparent, accountable, inclusive and participa-
tory. ICT also provides a medium for building partnerships amongst all stakeholders
for better service delivery.
ICT can increase accountability and strengthen the fight against corruption in
public service delivery. Citizens can report and compare their experiences on a web-
site, via mobile phones or social media, bypassing official channels when these are
not responsive. Bottom-up pressure can be applied and collated through responsible
intermediaries like civil society organizations or local government agencies if these
are open, cooperative and prepared to listen and be responsive.
One of the main ambitions of ICT-enabled good governance is to ensure that
public policy and public services focus on becoming more open and innovative as
well as efficient and effective, and indeed it is clear that these attributes are
complementary. It is becoming clear that the public sector cannot successfully
tackle service delivery challenges entirely on its own, but also needs to collaborate
with other actors, and especially the private and civil sectors, and a powerful tool in
this context is ICT.
The main lessons in summary are:
1. Data from service providers, users as well as other legitimate sources can be
used and made available as big and/or open data (as long as individual privacy is
protected) to increase the efficiency of public service delivery through the use of
ICT by highlighting where costs are incurred and whether processes can be ratio-
nalized or eliminated.
2. Data can also be used together with ICT tools to improve the effectiveness of
public service delivery by enabling the service provider to specifically target a
service to precise user requirements, and so that users themselves can personal-
ize a service to her/his own situation.
Technology Innovations in Public Service Delivery for Sustainable Development 279
3. New technologies like social media, mobile phones and other interactive ICT
can enable service provides to obtain feedback from users and the wider society
concerning corruption and mis-management, about specific services as well as
public policy issues more generally. These tools can also be used unilaterally by
users to address service providers and governments concerning a wide range of
legitimate public policy issues. It is important that the public sector listens,
learns and responds to these new forms of communication.
The social needs of people, communities and locations are highly diverse, so the
public services designed to meet them must be inclusive and respond according.
Different social groups require different types of public services and need to be
addressed in different ways depending on their unique social needs. One size fits all
public services, not tailored to specific needs, can miss their mark and thus both
waste public resources as well as prove ineffective. ICT is a powerful tool enabling
personalization to happen, both when used and initiated by the government and
other service providers, as well as when utilized directly by the users themselves.
Often different needs can be precisely tailored by using a multi-channel approach
consisting of different combinations of both ICT and traditional means. Especially
when targeting mainly poor people or those suffering from a range of disadvan-
tages, relatively cheap ICT in the form of mobile phones as well as more traditional
media like TV, radio and newspapers are highly effective. Such approaches can be
hugely successful, also on a commercial basis, if the business model is right.
The main lessons in summary are:
1. Collaboration both between the public and civil sectors and the public and pri-
vate sectors, or all three, is often highly productive as each can bring specific
competencies and assets to the table. However, the public interest, and especially
the specific needs of the users and their social needs must be constantly priori-
tized in an open and transparent manner.
2. Mobile technology is typically the most powerful tool to reach poor and disad-
vantaged people and provide them with high impact basic services. For example,
through the two-way collection of information and data from service users which
service providers can then analyze and actively use as a management tool to
organize and deploy their own resources, and to react rapidly to changing cir-
cumstances or emergencies.
3. ICT infrastructural policies should be directly embedded in broader national or
regional sustainable development policies and plans to meet the public service
needs of different societal groups. Often this will mean adopting a multi-channel
approach consisting of both ICT and more traditional channels to deliver basic
services and provide a safety net for the poorest and most vulnerable.
280 J. Millard
A strong focus is needed on a clear and long-term policy and strategic framework in
order to meet the challenge of an effective enabling environment for technology in
public service delivery. This will need to provide the overall setting, direction and
importance given to public services in support of sustainable development within a
specific legal and regulatory jurisdiction, whether this is local, municipal, regional
or national. Public service delivery is one of the most expensive aspects of any gov-
ernment’s budget, so it is extremely important to have the right policies for the
specific context a jurisdiction finds itself in. In particular, lack of relevant legal
provision and regulation hampers the adoption and use of new technologies in basic
public services. Developing a long-term strategy for ICT enhanced public service
design and delivery, underpinned by a sound legal basis, together with consistent
political will and resources, is critical.
The main lessons in summary are:
1. It is important to fully embrace the digital revolution and develop a new vision
around it which can provide better tools for service delivery, but also to recog-
nize the challenges and potential dangers this might reveal, for example in rela-
tion to security and privacy issues, as well as the uneven digital access and skills
people posses.
2. An ICT strategy should also go hand-in-hand with understanding that the human
element remains essential. Human and organizational capacity building is impor-
tant for both routine service delivery but also for promoting creativity, experi-
mentation and innovation in a continuous search for improvement.
3. In order both to save money and resources as well as provide better quality ser-
vices, ICT-enabled sharing across the public sector is needed, both of good prac-
tices and ideas but also in terms of human, organizational and physical assets.
Multiple service delivery channels supported by the local authority as well as by
local organizations and citizens provide both better tailored and contextual ser-
vices as well as improve the inclusion of everybody.
References
Broadband Commission (2014) Means of transformation: harnessing broadband for the post-2015
development agenda, a collaboration between UNESCO, ITU and Ericsson. http://www.broad-
bandcommission.org/documents/reports/tf-post2015-advocacy-2014.pdf
Chesbrough HW (2003) The era of open innovation. MIT Sloan Manag Rev 44(3):35–41
Crandall A, Otieno A, Mutuku L, Colaço J (2012) Mobile phone usage at the Kenyan base of the
pyramid. iHub Research, November 2012. https://blogs.worldbank.org/ic4d/files/ic4d/mobile_
phone_usage_kenyan_base_pyramid.pdf
Dean-Swarray M, Gillwald A, Morrell A (2013) Lifting the veil on ICT gender indicators in Africa.
Research ICT Africa. https://www.researchictafrica.net/publications/Evidence_for_ICT_
Policy_Action/Policy_Paper_13_-_Lifting_the_veil_on_gender_ICT_indicators_in_Africa.pdf
Technology Innovations in Public Service Delivery for Sustainable Development 281
Docebo (2014) E-Learning market trends & forecast 2014–2016 Report, March 2014. https://www.
docebo.com/landing/contactform/elearning-market-trends-and-forecast-2014-2016-docebo-
report.pdf. Retrieved April 2016
Ericsson and the Earth Institute (2016) How information and communications technology
can accelerate action on the sustainable development goals. http://www.ericsson.com/res/
docs/2016/ict-sdg.pdf
European Commission (2013) Study on business and financing models related to ICT for ageing
well. DG CONNECT, Danish Technological Institute, Ernst & Young. https://ec.europa.eu/
digital-single-market/en/news/study-business-and-financing-models-related-ict-ageing-well
Ghana Health Service (2016) Integrated national ICT for health and development forum,
Millennium Promise. http://1millionhealthworkers.org/files/2016/09/ICT_REPORT.pdf
Global Opportunity Network (2016) Global opportunity report 2016. Published by DNV GL AS,
the United Nations Global Compact, and Monday Morning Global Institute. http://www.globa-
lopportunitynetwork.org/report-2016/#.V0LvqTY-KqA
GSMA (Mobile Enabled Community Services) (2014) The synergies between mobile, energy
and water access: Africa. http://www.gsma.com/mobilefordevelopment/wp-content/uploads/
2014/04/MECS_Synergies-between-Mobile-Energy-and-Water-Access_Africa.pdf. Retrieved
April 2016
International Telecommunication Union (2013) Measuring the information society. Geneva.
http://www.itu.int/en/ITU-D/Statistics/Pages/publications/mis2013.aspx; and The ITU World
Telecommunication/ICT Indicators database. http://www.itu.int/en/ITU-D/Statistics/Pages/
stat/default.aspx. Retrieved April 2016
International Telecommunication Union (2015) Measuring the information society report 2015.
Geneva. http://www.itu.int/en/ITU-D/Statistics/Pages/publications/mis2015.aspx; and The
ITU World Telecommunication/ICT Indicators database. http://www.itu.int/en/ITU-D/
Statistics/Pages/stat/default.aspx. Retrieved April 2016
John JK, Nair MS, Selvanthan PJ, Kuppusamy M (2005) Using ICT as a catalyst for sustain-
able development: the role of national policy. In: Rahim RA, Waldburger D, Muinde GS
(ed) Access, empowerment and governance: creating a world of equal opportunities with
ICT. Global Knowledge Partnership, Kuala Lumpur
Malaysia Government (undated) The national telecommunication policy of Malaysia (1994–2020)
Millard J (2015a) The digital divide and the post-2015 development debate. In: Andreasson K
(ed) Digital divides: the new challenges and opportunities of e-inclusion. Taylor and Francis
Publishing Group, Abingdon
Millard J (2015b) Open governance systems: doing more with more. Gov Inf Q, 12 Sept 2015.
http://doi.org/10.1016/j.giq.2015.08.003
Millard J, Weerakkody V, Missi F, Kapoor K, Fernando G (2016) Social innovation for pov-
erty reduction and sustainable development: some governance and policy perspectives. In:
Proceedings of the 9th international conference on the theory and practice of electronic gover-
nance (ICEGOV2015–16), The ACM Press, Montevideo, 1–3 March 2016
Mohsin Bin Hj Ahmed (2007) Implementation of electronic government in Malaysia: the status
and potential for better service to the public. Public Sect ICT Manage Rev 1(1)
National Latino Research Center (2011) Text4baby preliminary study. California State University
and the University of California, San Diego
Prahalad CK (2004) The fortune at the bottom of the pyramid: eradicating poverty through profits.
Pearson Prentice Hall, New York
Regional School of Public Administration (2013) Abuse of Information Technology (IT) for cor-
ruption. ReSPA, Danilovgrad
UNDP Montenegro (2014) Increasing citizens’ engagement by using social innovation and social
media to enhance government’s transparency and accountability. Retrieved from https://info.
undp.org/docs/pdc/Documents/MNE/Pro Doc Citizen Engagement signed.pdf
United Nations (2013) Governance, public administration and information technology for post-
2015 development. United Nations Department of Social and Economic Affairs New York.
http://workspace.unpan.org/sites/Internet/Documents/Governance_PA_Report.pdf. Retrieved
April 2016
282 J. Millard
United Nations (2015) Innovative public service delivery: learning from best practice. United
Nations Department of Social and Economic Affairs New York. http://workspace.unpan.org/
sites/Internet/Documents/EGM%20Report%20on%20Innovative%20Public%20Service%20
Delivery%20Learning%20from%20Best%20Practices.docx.pdf. Retrieved April 2016
World Bank (2012) eTransform Africa – the transformational use of information and communica-
tions technologies in Africa. The World Bank and the African Development Bank with the sup-
port of the African Union, Washington, DC. www.eTransformAfrica.org. Retrieved April 2016
World Bank Group (2015) Unlocking the potential of information communications technology to
improve water and sanitation services. Water and Sanitation Program, Washington, DC
World Economic Forum (2015) The global information technology report 2015: ICTs for inclu-
sive growth. World Economic Forum and INSEAD, Geneva. http://www3.weforum.org/docs/
WEF_Global_IT_Report_2015.pdf. Retrieved April 2016
Jeremy Millard is Director of the consultancy Third Millennium Governance, as well as having
senior research positions at the Danish Technological Institute in Denmark and Bradford University
in the UK. He has over 40 years’ global experience on issues ranging from governance, ICT, open
and social innovation, participation, sustainable and socio-economic development, tackling pov-
erty and exclusion, the new economy, urbanization and nature-based solutions for growth, and has
published extensive in these and related fields. His many clients include governments, the European
Commission, United Nations, OECD and World Bank, as well as many non-profits and companies
around the world. Recent assignments in the area of e-government include on-going support to
both the United Nations regarding their biennial eGovernment Survey, and the European
Commission regarding e-government research and innovation, as well as a survey on back-office
developments in support of user-centred e-government strategies for the OECD. He has also
worked with the UN on ICT and governance issues for the 2030 sustainable development agenda,
and with ESCWA on integrated service delivery across the Arab Region.
Blockchain as a Next Generation Government
Information Infrastructure: A Review
of Initiatives in D5 Countries
Adegboyega Ojo and Samuel Adebayo
Introduction
A. Ojo (*)
Insight Centre for Data Analytics, National University of Ireland Galway,
Lower Dangan, Galway, Ireland
e-mail: adegboyega.ojo@insight-centre.org
S. Adebayo
Insight Centre for Data Analytics, National University of Ireland Galway (NUIG),
IDA Business Park, Lower Dangan, Galway, Ireland
e-mail: Samuel.Adebayo@insight-centre.org
the motives for adopting this technology is that it affords transparent real-time trans-
action settlement and auto-executing so-called smart contracts with business logic
encoded into the ledger (Wyman 2016). Another significant motive for the adoption
of this innovation is its extended capabilities to provide significant impacts to differ-
ent economics and social activities in the society (Taylor, 2016)
Blockchain according to findings can be used to address inefficiencies in current
systems and increase the effectiveness of public service activities (Drucker 2016). It
can also create a data network platform where citizens, private companies, and gov-
ernments can access for the verification of information (Oscar 2016). The adoption
of Blockchain in the public sector is expected to reduce the cost of operations par-
ticularly by eliminating fraud, error in payments, providing greater transparency of
transactions between government, other agencies and citizens. It strengthens citi-
zens data protection and encourages data sharing among entities (Taylor, 2016). In
general, government entities can perform the following activities on the Blockchain1:
(1) verification of documents such as licenses, proofs of records, transactions, pro-
cesses or events such as birth of a child, (2) movement of assets such as transferring
money from one entity to another after some work conditions are met, (3) asset
ownership registers such as land registries, property titles and other types of owner-
ship of physical assets and (4) management of identities like e-identities for citizens
and city residents.
Interestingly, while there is growing literature on Blockchain applications in the
private sector, the literature on possible applications of this new generation informa-
tion infrastructure in the government domain are few (Ølnes 2008). This chapter
addresses this knowledge gap by examining some of the flagship Blockchain initia-
tives in leading five digital champion (so-called D5) countries including United
Kingdom, United States, Estonia, New Zealand and Israel.
The rest of the chapter is organized as follows: section “Background” provides a
brief background on Blockchain and Distributed Ledger technology. The approach
for the study is presented in section “The Digital 5 Countries as Innovators” while
the case studies selected from the D5 countries are laid out in section “Cases”. We
provide some analysis of these initiatives in section “Discussion” and conclude in
section “Conclusion”.
Background
The blockchain is a digital ledger and a “database that can be shared across a net-
work of multiple sites, geographies or institutions” (Taylor 2016). It could also be
described as a database of secure transaction ledgers only accessible to all parties
involved in a distributed network. It has the capacity to record and save every trans-
action which occurs in the network and also create an irrevocable and auditable
transaction history (Finextra 2016). Other authors consider the Blockchain as
http://observer.com/2016/09/why-the-Blockchain-is-perfect-for-government-services/
1
Blockchain as a Next Generation Government Information Infrastructure… 285
2
https://www.ethereum.org/
3
http://www.gridcoin.us/
4
https://ripple.com/
286 A. Ojo and S. Adebayo
user-friendly and transactions made on the blockchain are not reversible, so genuine
errors cannot be corrected by any administrator.
The popularity of the blockchain technology is driven by a number of factors
(ODI 2016) including: (1) The capabilities of the platform to store data that is very
robust in nature and that cannot be tampered with; the highly distributed nature of
Blockchain platform comprising of nodes managed by different parties making col-
lusion to compromise the infrastructure difficult. Another driving factor of the
Blockchain technology is the optimization of cost and time efficiency in both public
and private sectors. For instance, it is now faster through this innovation to move
funds between two different institutions and geographical zones without any inter-
ference of intermediaries (Probst et al. 2016).
Finally, the Blockchain technology has potential to impact any industry or prod-
uct line that relies on the storage and verification of information or value. Blockchain
technology’s programmable aspects can also facilitate the development of indepen-
dent governance systems, contracts and legal constructs (e.g., “smart contract”) or
the ability of interrelated devices to interact with and even pay each other in the
“Internet of Things” (Kaye 2016).
We have chosen to review some past and ongoing Blockchain innovation in Digital
countries due to the strong commitment that these countries have for undertaking
digital transformations and serving as innovators and early adopters with respect to
emerging technologies. The Digital 5 or D5 is a networking group of leading digital
government countries with the objective of strengthening the digital economy.
There is a promise among the members to be open while they aim at how to trans-
form government’s relationship with technology through the espousal of open stan-
dards and open source software and also increasing the effectiveness of digital
government. Furthermore, these countries are also working towards encouraging
digital skills in-house and also short-term contracts with small and medium busi-
ness suppliers (Wikipedia 2016).
This network group was founded on the 9 December 2014. The founding mem-
bers of the group are Estonia, Israel, New Zealand, South Korea and the United
Kingdom. These countries possess mutual agreement to create this network group
and develop a platform where best practice will be shared and also collaborate on
common projects that will provide support in growing digital economies. Some of
the goals of the D5 countries according to (Palo et al. 2015) are:
• User needs – provide citizen-centric public services taking into consideration
specific needs of different segments of the citizenry.
• Open standards – employ technologies that are interoperability and show a clear
commitment to a credible royalty free open standards policy.
• Open source – ensure that future government systems, tradecraft, manuals, and
standards are created as open resources and shareable among members.
Blockchain as a Next Generation Government Information Infrastructure… 287
• Open markets – ensure true competition for companies regardless of size in gov-
ernment procurements, promote and support start-up culture as well as economic
growth through open markets.
• Open government (transparency) – be a member of the Open Government
Partnership and use open licences to produce and consume open data.
• Connectivity – develop an online population through comprehensive and high-
quality digital infrastructure.
• Teach children to code – show commitment to offer children the opportunity to
learn to code and acquire next generation skills.
• Assisted digital services– show commitment to supporting all its citizens to
access digital services.
• Commitment to share and learn – commit to work together to help solve each
other’s issues wherever they can.
Our study examined public sector innovation activities of these countries related
to the use of blockchains through exploratory desktop research. Information was
consolidated from scholarly and online articles and news on Blockchain technolo-
gies. We outline the identified initiatives in section “Discussion” and discuss them
in section “Conclusion”.
Cases
access.
Blockchain as a Next Generation Government Information Infrastructure… 291
invest in financial technology and an ICT-based start-up that can develop creative
ideas on innovation and change project based on the Blockchain technology. The
government and its partners have identified Blockchain innovation as a tool that
can be used for asset ownership and settlement management. The government
also believes that Blockchain will pave the way for new technologies and solu-
tions in the Fintech industry.
United Kingdom The country ranks in 1st position in the 2016 E-Government
Development Index and the 3rd place in the 2016 Global Innovation Index. UK
Government through its Office of Science published a report on Distributed Ledger
Technology: Beyond blockchain (Taylor 2016). The report expressed the transfor-
mational potential of distributed ledger and also advanced a number of technology,
governance, security and privacy, and trust and interoperability related recommen-
dations. Furthermore, the UK government believes that it stands in a good position
to leverage the benefits and address the challenges related to the use of distributed
ledgers in the public service and economy because of the digital capability, innova-
tive financial services, the effective research community and growing private ser-
vice. Some of the ongoing blockchain based initiatives in the UK include:
• Distributed ledger based Gross Settlement System (Peter 2016): the Bank of
England is currently working on replacing its current real-time gross settlement
(RTGS) system to be ready for future demands. Specifically, the future system
must address the following strategic RTGS requirements: (1) capability of
responding to the changing structure of the financial system; (2) recognising that
payment system users want simpler and more resilient pathways for their pay-
ments; (3) capability of interfacing with a range of new technologies being used
in the private sector, including distributed ledgers, if/when they achieve critical
mass; (4) to remain highly resilient to the increasingly diverse range of threats to
continuity of service, and (5) develop capacity to support the future evolution of
regulatory and monetary policy tools. From the bank point of view, the new sys-
tem will change a lot of features between the existing system which was built in
1996 and its successor. Some of these changes will include and enhanced secu-
rity, which could be provided through the use of distributed ledger/blockchain
solutions.
• Blockchain for benefit payment (Lynsey 2016): the government is currently
test-running a blockchain based social welfare payment mobile app. Claimants
in receipt of this payment are advised to download the app on their phones
which will enable them to receive and spend their benefit payments. With their
consent, their transactions are being recorded on a distributed ledger to sup-
port their financial management. This initiative focuses on adding an addi-
tional layer of richer data and identity onto payments so that a deeper and
more effective relationship can be established between the government and
claimants. The aim of this project is to identify the possibility for welfare pay-
ment to citizens to be sent through a secure app and also to see if people reli-
ant on welfare payments would benefit from this approach. This new system
consists of a mobile app and a Blockchain system that records payments sent
Blockchain as a Next Generation Government Information Infrastructure… 295
Discussion
have simply leveraged the infrastructure and services provided by local blockchain
start-ups to realise pilot initiatives. In other cases, the governments have sought to
focus on developing the blockchain ecosystems (e.g. Israel) by facilitating the inter-
action of local start-ups and investors.
In addition to the various type of goals that emerged from the cases and described
in Table 1, blockchain and distributed ledger technology could help in the specific
area of governance including (Hopping 2016): traceability of government spending,
protecting critical infrastructure, registering assets such as intellectual property,
wills, and health data as well as reducing waste resulting from benefit fraud.
To further develop and mature blockchain initiatives, the UK Government Chief
Scientific Office provided some recommendations in advancing blockchain innova-
tions in government and society, which include (Taylor 2016): (1) establishing a
ministerial level leadership to ensure that government provides the vision, leader-
ship and the platform for distributed ledger technology within government; (2) that
the research community invest in the research required to ensure that distributed
ledgers are scalable, secure and provide proof of correctness of their contents; (3)
that government supports the creation of distributed ledger demonstrators for local
government that consolidates all the elements necessary to test the technology and
its applications; (4) government should put in place the necessary regulatory frame-
work for distributed ledger; (5) that government works with academia and industry
to ensure that standards are set for the integrity, security and privacy of distributed
ledgers and their contents which should be reflected in both regulations and soft-
ware code; (6) that government works with academia and industry to ensure the
most effective and usable identification and authentication protocols are imple-
mented for organizations and individuals.
Similar recommendations have been advanced in other sources1. For instance, it
was recommended that Government leaders need to familiarise themselves with the
potentials and benefits of the blockchain as a digital transformation technology
before committing to exploring its potentials; and 3) commence experimentation
with blockchain technology via proofs of concepts and small projects.
As indicated in many of the recommendations above, government’s close col-
laboration with academia is critical to advancing research in blockchain and distrib-
uted ledger technology. From the different cases reviewed, we observe that a number
of interesting concepts are emerging from the interaction of blockchain technology
and governance. Some of the concepts that could redefine governance and definitely
worth examining further include (James et al. 2016): “Do-it-Yourself” Governance,
Decentralised Autonomous Organization, Decentralised Citizen Engagement,5
Provably Secure Governance, Provable Transparency, and Collaborative manage-
ment of jointly owned digital assets.
http://netfutures2016.eu/wp-content/uploads/2016/05/1-Project-presentation-net-futures-.pdf
5
Blockchain as a Next Generation Government Information Infrastructure… 297
Conclusion
This chapter has directly contributed to addressing the paucity of scholarly litera-
ture on the application of blockchain and distributed ledger technology in the gov-
ernment domain as highlighted in (Ølnes 2008). We have reviewed several initiatives
across the Digital 5 countries in which government has played various roles in
blockchain initiatives. While some of the reviewed initiatives show great promise,
most of these initiatives are far from operating at scale. At the same time, there are
a number of legal, regulatory, ethical as well as technical barriers that must be
addressed to fully harness the potentials of the blockchain and distributed ledger
technology in government.
References
Pilkington M (2016) Blockchain technology: principles and applications. In: Elgar F, Olleros
X, Zhegu M, Elgar E (eds) Research handbook on digital transformations. Edward Elgar,
Cheltenham, p 39
Probst L, Frideres L, Cambier B, Martinez-Diaz C (2016) Business innovation observatory: block-
chain applications & services. European Union, p 16
Shrier D, Iarossi J, Sharma D, Pentland A (2016) Markets and marketplaces, pp 1–19
Srisukvattananan Y (2016) Overview of blockchain and possible use cases in the Thai payment
system. Massachusetts Institute of Technology, pp 1–172
Taylor S (2016) Distributed ledger technology: beyond block chain
United Nations Department of Economic and Social Affairs (2016) UN E-government survey
2016. E-government in support of sustainable development. New York
Wikipedia (2016) Digital 5
Wyman O (2016) Blockchain in capital markets: the prize and the journey. Euro Clear
Dr. Adegboyega Ojo is Senior Research Fellow at the Insight Centre for Data Analytics, National
University of Ireland Galway (NUIG). He leads the E-Government Group at Insight Centre and
serves as Adjunct Lecturer at the College of Engineering and Informatics. His current research
interests include data driven innovations in government, Open data policies and Infrastructures,
data analytics and governance of smart cities. He is a member of the Editorial Boards of the
Government Information Quarterly and International Journal of Public Administration in the
Digital Age.
Samuel Adebayo a research assistant at Insight Centre for Data Analytics, NUI Galway. Samuel’s
research interests include Open Data, Social media and emerging technologies in government. He
joined the e-Government unit at Insight Centre in April 2015. Samuel holds a bachelor degree in
Business Management from University of Wales and a master’s degree in Information System
Management from National University Ireland Galway. His area of expertise includes project man-
agement, requirement life cycle management, analysis and design, solution evaluation, testing and
quality engineering.
Governance, Transparency
and the Collaborative Design of Open Data
Collaboration Platforms: Understanding
Barriers, Options, and Needs
for people to monitor and discuss policies and policy outcomes, participatory democ-
racy is a process emphasizing the broad participation of citizens and public adminis-
trators in the direction and operation of political systems and the co-creation of
public value. The era of participatory democracy is associated with widespread
availability of government documents and data on websites and open data portals
(Meijer 2015b), which opens the potential not only to monitor government activity,
but also deliberate and discuss policies in an informed manner, and participate and
collaborate in the formulation of policy and the co-creation of public value.
Research and innovation in the area of transparency enhancing technologies
emerges in parallel with different perspectives and conceptualisations of transpar-
ency. Different approaches to conceptualising transparency may influence open data
platform and software design. Consistent with the collective intelligence design
methods developed by Warfield (2006), which emphasise a stakeholder-driven
approach to design, and consistent with the principles of scenario-based design
(Caroll 2000), which emphasises the importance of understanding specific scenarios
of usage in the technology design process, our view is that understanding the context
of technology usage and the specific problems stakeholders are trying to resolve in
context is important for the design of transparency-enhancing technologies. As such,
we advocate contextualism as an orienting philosophy for conceptualising transpar-
ency and for understanding the technology-mediated activities that support transpar-
ency in context. In general, we believe that conceptualisations of transparency can be
understood by reference to different worldviews, or ways of understanding reality,
and different worldviews can influence the development of different frameworks
shaping research, design, and innovation (Hayes et al. 1988). Drawing upon Pepper’s
(1942) distinction between formism, mechanism, organicism and contextualism,
below we will briefly describe these worldviews in turn, and the rationale for adopt-
ing contextualism as an approach to technology design in the current project.
Formism, as defined by Pepper, involves the identification of forms, or aspects of
reality, that share common or similar characteristics. Heald (2006) highlights a vari-
ety of different forms of transparency. For example, Heald (2006) draws a distinc-
tion between nominal versus effective transparency. While a nominal form of
transparency might imply the availability of data on an open data platform, an effec-
tive form of transparency might involve data that is effectively used to shape valued
outcomes (Heald 2006). Similarly, Heald (2006) distinguishes between forms of
transparency that are based on an analysis of historical data (i.e., transparency in
retrospect) and forms of transparency that are based on an analysis of data that
reflects the current state of a system (i.e., transparency in real-time). As noted by
Pepper (1942) identifying different forms, or aspects of reality, can be an important
precursor to the development of more complex models, for example, mechanistic
models that describe causal relationships between different aspects of reality.
Similarly, formism may shape design thinking. For example, in the context of the
design and innovation of an open data platform, formist conceptualisations of trans-
parency may support design thinking in relation to specific aspects of technology
design, related to specific forms of transparency. For instance, drawing upon a dis-
tinction between transparency in retrospect and transparency in real-time, a technol-
ogy design team might include platform features that allow for a distinction to be
Governance, Transparency and the Collaborative Design of Open Data Collaboration… 303
made between current data and historical data, and possibly add prompts that help
users to make these distinctions. At the same time, formist conceptualisations of
transparency may limit design thinking in certain respects. While it might support
design thinking in relation to specific forms of transparency, a formist approach to
analysis does not generally emphasise a dedicated focus on activity in context. As
such, a formist approach to understanding transparency may neglect key aspects of
the context of transparency-related activities, or specific problem situations that
involve interactions between stakeholders who analyse, discuss, and make use of
open data in an effort to support transparency-related activities. In the absence of
this more contextual focus distinctions between different forms of transparency may
have limited value for the overall design of transparency-enhancing technologies.
In Pepper’s (1942) scheme, a mechanistic worldview may build upon formist
accounts by specifying how components parts of a system (or machine) work
together. From this view, different forms of transparency may be viewed as different
components of a system of interdependencies. For example, a mechanistic model
may be developed to explain how components of transparency work together to
produce trust in societies (Meijer 2009; Mei and Dewan 2014). Specific compo-
nents of transparency such as visibility (the degree to which information is complete
and easily located) and inferability (the degree to which information can be used to
draw verifiable inferences) may in turn be influenced by other components of a
system, and a mechanistic model of transparency may become increasingly com-
plex as more components of reality are identified and modelled. For example, stud-
ies report that increased demand drives up visibility; and demand is strongest for
issues that represent acute concerns of citizens, such as finance, health and security
(Piotrowski et al. 2011). Although complex mechanistic models of transparency
describing many component interdependencies can be developed to shed light on
specific issues relevant for transparency-enhancing technology design, by virtue of
their mechanistic structure, and the defined set of variables and components in the
model, mechanistic models may constrain the ability of a design team to consider
the varied actions and needs of users across different scenarios and contexts.
According to Pepper (1942), distinct from mechanism as a worldview is organi-
cism. From the perspective of organicism transparency would be viewed as part of a
living system that actively develops through various stages of maturity or functional
complexity. For instance, the Transparency Maturity Model (Cappelli et al. 2013)
characterises five levels of transparency – opaque, disclosed, comprehended, reli-
able, and participative. At the lowest level of maturity, the opaque level, the organiza-
tion provides information access to the external environment in a non-systematic
fashion. In the disclosed level, the organization provides information access to the
external environment, but not necessarily in a way that is easily comprehended or
responsive to feedback from external stakeholders. The comprehended level enables
access to understandable information and thus facilitates a higher level of transpar-
ency and engagement. At the reliable level, the organization allows for auditability of
the information provided. Finally, at the participative level the organisation allows
for ongoing dialogue with the external environment about the information provided.
As a worldview orientation, Pepper (1942) notes that organicism is linked to ideal-
ism, in the sense that there is an assumption that a system has the potential to develop
304 M. Hogan et al.
toward a more ideal state of functioning. However, these idealist assumptions may
not be aligned with the activity in context and thus by adopting organicism as a
worldview, designers of transparency-enhancing technologies may neglect the prob-
lems and activities of technology users in context and thus fail to develop technolo-
gies that are well suited to the problems users are working to resolve.
As an approach to analysis and design, Pepper (1942) notes that contextualism
emphasises a focus on activity in context. Contextualism allows for different strands
of enquiry in relation to different activities in context, each of which may be impor-
tant for successful workings, or the resolution of a specific problem in context. For
example, a contextualist might consider the activities of key stakeholders seeking to
access, understand, and use open data – the key barriers they face and the specific
information, social-collaborative, and decision-making needs they have across dif-
ferent problem solving scenarios. In a participatory democracy scenario, where
there is a focus on collaboration over open data in response to a specific political
and social problem, one strand of contextualist enquiry might focus on the qualities
of data, such as accessibility, usability, understandability, informativeness and
auditability of the data (Cappelli et al. 2013). A related strand of analysis might
focus on the social and organisational context within which data is sourced, includ-
ing who the information holders are, the relevance of different types of public sector
information, the availability of the information, and the distribution channels of
information (Deloitte 2013). An analysis of these and related issues may be essen-
tial to the success of the participatory democracy group working together in the
local problem situation. Notably, according to Pepper (1942), adopting the contex-
tualist approach to research and innovation implies a focus on the specific purpose
or goal(s) of actors in the problematic situation, and success is determined by the
extent to which their purpose or goal(s) are achieved.
Given our focus on the design of a new open data platform, and our focus on
developing system requirements that were matched to the context or scenario of
usage identified across our pilot sites, we adopted a contextual and collective intel-
ligence scenario-based approach to transparency research and innovation.
Specifically, in the current study, we draw upon the collective intelligence scenario-
based design thinking of stakeholders to define the scope of our analysis of open
data transparency and our approach to the design of a new open data platform that
may help to overcome barriers to accessing, understanding, and using open data and
fulfil the key needs of stakeholders working across a variety of scenarios.
The research findings reported in this paper emerge as part of an ongoing EU inno-
vation project, the “Route-to-PA” project (http://routetopa.eu/). Route-to-PA is
focused on the design and evaluation of an open data collaboration platform that can
be used by citizens and public administrators across a wide variety of usage sce-
narios. As the goal of the project is to design user-friendly transparency-enabling
technologies for public administrations across a range of EU countries categorised
Governance, Transparency and the Collaborative Design of Open Data Collaboration… 305
by the Open Data Barometer (2015)1 as high capacity (UK, France, and the
Netherlands) and emerging and advancing (Italy and Ireland), it was important to
understand the varied political and social contextswhere our design and innovation
is to be realised. This involved an analysis of the open data readiness of each coun-
try, and a mapping of the local open data context for specific usage scenarios that
reflect ongoing priorities of citizens and public administrations in each country (see
http://routetopa.eu/wp-content/uploads/2016/07/D7.1_Market_analysis.pdf/). To
maximize the socio-technical capabilities and vision of the design team, it was
essential to engage with key stakeholders and users in each pilot site to understand
the barrriers to accessing, understanding, andusing open data, options to overcome
these barriers, and the key needs and requirements of users across a range of moni-
torial, deliberative, and participatory democracy scenarios. Furthermore, as the goal
of the Route-to-PA project is the design of a flexible open data collaboration plat-
form that allows for a range of democratic activities, up to and including collabora-
tion and co-creation of public value, it was essential that the range of needs
stakeholders specified in response to scenarios include not only information needs,
but also social-collaborative and decision-making needs. In order words, the open
data platform needed to allow for collaboration, shared learning, and decision mak-
ing in the context of accessible, usable, understandable open data. As such, we
approached our contextual analysis and system design work using an integrative
collective intelligence scenario-based design approach. Below we describe our
approach to system design in more detail and present the results of our study, high-
lighting in particular the range of barriers, options, and needs our stakeholders iden-
tified and how we have grounded our open data platform design in this collective
intelligence work.
While it is widely recognised that open data platforms can foster democratic pro-
cesses by promoting transparency (Lourenço 2013; Dawes and Helbig 2010;
Janssen 2011), researchers have identified a range barriers that hamper effective
service design and the full potential of open data innovations. Barriers to effective
service design in the area of open data include limited organizational resources and
budget, legislative challenges, poor information quality, lack of usability and techni-
cal issues (Janssen et al. 2012; Attard et al. 2015; Meijer 2015a ). In working to
overcome these barriers researchers have proposed a range of generic user require-
ments (Lourenço 2013; Jaeger et al. 2012; Van Velzen et al. 2009) and assessment
frameworks for open data portals and policies (Sandoval-Almazan and Gil-Garcia
2012; Zuiderwijk and Janssen 2014; Lee and Kwak 2012). These approaches either
take users (both citizens and government) or open data portals as point of departure
for analysis. However, the unique context and scenarios of usage and the unique
perspectives of stakeholders in relation to information, social-collaborative, and
decision-making needs are less often considered in the literature and open data plat-
form design process (Dahlander et al. 2009). Focusing on specific scenarios of
usage and the specific needs of users may be important for adoption, uptake and use
of open data and open data platforms.
At a basic level, effective computer-mediated transparency implies that external
or receiving parties are capable of processing information that has been made avail-
able (Heald 2006). However, platforms for open-data enabled transparency are
often limited in this regard. Literature on open data portal software shows that social
media features are limited on existing or first generation open data portal software
or platforms (Alexopoulos et al. 2014). Specifically, these platforms do not provide
beyond features for sharing information about datasets on major social media plat-
forms, thus limiting the potential use of open data in participatory democracy sce-
narios. In addition, features for checking compliance with metadata standards and
good practices (Greiner et al. 2015) are very limited, thus limiting feedback from
users to data providers that may enhance the quality of data published online.
Understanding the unique perspectives of stakeholders and their unique scenarios of
usage is critical for the design of platforms and platform software features that are
responsive to user needs.
Central to our design work in the Route-to-PA project is the combination of col-
lective intelligence (Warfield 2006) with scenario-based design (Caroll 2000) and
agile user story (Cohn 2004) methods. Collective intelligence methods ensure input
from a diverse range of representative stakeholders in the design process and the use
of scenario-based design methods ensures that identified needs and requirements of
users are grounded in an understanding of specific political and social scenarios that
are relevant to stakeholders. Finally, the use of agile user stories allows for the
specification of user needs, and reasons for those needs, at a level of detail that
allows for agile software development of specific functionalities. Working across
four EU countries and five pilot sites, we used these methods in a series of carefully
designed workshops, one in each pilot site, for the purpose of developing a compre-
hensive set of user needs, as proposed by key stakeholders.
Each workshop brought together experts, academics, industry specialists, open
data practitioners, representatives of governments, open data researchers, and
potential users (including citizens, representatives of citizens and social service
institutes, various stakeholder groups, and journalists) to reflect on (a) barriers to
accessing, understanding and using open data, (b) options to overcome specific cat-
egories of barriers, and (c) specific user needs and requirements necessary for con-
sideration in the design of the Route-To-PA platform. More specifically, based on
John Warfield’s (1994) science of design, in the first phase of each workshop, we
used collective intelligence methodologies to understand barriers to accessing and
using open data, and options to overcome these barriers. Participants then worked to
develop scenario-based user needs (Rosson and Carroll 2002), which involved
profiling user needs in light of the barriers and options and high level scenarios of
Governance, Transparency and the Collaborative Design of Open Data Collaboration… 307
open data usage. This included a separate focus on (1) information needs – what
kinds of data do stakeholders want?; (2) social and collaborative interaction needs –
how do stakeholders want to use and interact with the data?; and (3) understand-
abilty, usability and decision-making needs – what kinds of decisions do stakeholders
want to make with the data and how would they like to use the data? High level
scenarios including multiple users were used to prompt idea writing and discussion
in relation to user needs. The scenarios addressed various contextual issues, relevant
to each workshop site, and aligned with the primary case focus and societal issue in
each pilot site. For example, the Dublin workshop focused on community network-
ing and opportunity creation; the Groningen workshop focused on the challenge of
population decline; the Den Haag workshop focused on employment and opportu-
nity creation; the Prato workshop focused on local policy and budget issues; and the
workshop in Issy-les-Moulineaux focused on the facilitation of start-up companies
and the digital economy. The research team conducted a meta-analysis of barriers,
options, and needs across all sites and used this analysis to inform the specific use-
case models and system requirements for the Route-to-PA platform. Below we
describe these methods and our results in more detail.
Method and Results
A total of 83 workshop participants across the five sites participated in the study.
Participants represented a broad variety of stakeholders with stakeholder represen-
tation distributed evenly across sites. Participants included representatives of stake-
holder groups, business representatives, NGO representatives, public administrators
and other government representatives, data experts, developers, and researchers.
See Fig. 1 for a breakdown of stakeholders across sites.
Workshops
Each pilot site ran a workshop following a common method. The workshop began
with a collective intelligence (CI) analysis of barriers to accessing, understanding
and using open data, followed by an analysis of options that may overcome these
barriers. Based on Warfield’s (1994) science of generic design, the CI process is a
facilitated problem solving methodology that helps groups to develop outcomes that
integrate contributions from individuals with diverse views, backgrounds, and per-
spectives. Established as a formal system of facilitation in 1980 after a developmen-
tal phase that started in 1974, CI was designed to assist groups in dealing with
complex issues. The CI approach carefully delineates content and process roles,
assigning to participants responsibility for contributing ideas and to the facilitator
308 M. Hogan et al.
50
40
Business
Stakeholder
30 groups/NGO/Journalists
Data Experts and
Technologists
20 Government
Researchers
10
0
Den Haag Dublin Groningen Issy Les Prato
(M=15; F=2) (M=11; F=7) (M=11; F=6) Molineaux (M=9; F=6)
(M=11; F=6)
with opportunity for individuals to add ideas as they read others’ papers; (e) discus-
sion and clarification of unique ideas; and (f) an oral report of the ideas generated by
each working group in a plenary session. In this plenary session, duplicate ideas across
the working groups are eliminated from the set and new ideas are added; the resulting
set of ideas is then ready for use in the next stage of the group’s work.
In the current application of CI, workshop participants first engaged in ideawrit-
ing in response to the question:
“What are barriers to accessing, understanding and using Open Data?”
Table 1 Scenarios
Pilot Context Actors involved Use of open data in scenarios
Dublin Deliberative • Public • Societal Issues
Democracy; Administrator • Improved Govemment
Participatory • Entrepreneur financial efficiency
Democracy • Citizen • Business development
• Local Activist • Community building
• Local Group • Citizen–Government
Coordinator communication
• Civic Hacker
Groningen Deliberative • Principal • overnment actions
G
Democracy; • Public monitoring and
Participatory Administrator collaboration
Democracy • Community • Business community
Activist collaboration
• Entrepreneur
• Local Business
Community
• Local Community
Members
Den Haag Deliberative • Public • ocial problem solving –
S
Democracy; Administrator unemployment of disabled
Participator • Business Owner
Democracy • Citizen
• Unemployed
• Entrepreneur
• Disabled Job
Seeker
Prato Monitorial • Public • itizen–Government
C
Democracy, Administrator communication
Deliberative • Student • Inclusive policy making
Democracy; • Citizen • Citizen collaboration and
• Community co-creation
Activist • Service improvement
• Journalist • Government actions
• Accountant monitoring
Issy-les- Deliberative • Entrepreneur • Social problem solving-
Moulineaux Democracy, • Local Community ecology, technology, and
Participatory Members mobility services
Democracy • Businesses
• Public
Administrator
• Domain Expert
Cost (N=9)
30
Data Applications (N=14)
and Usability; Technical, Infrastructure and Resources; and Cost. Finally, a set of
barrier categories related to training and engagement issues, including: Citizen
Engagement; Skills and Training; and Motivation. Table 2 presents a sample of bar-
riers from each category.
The frequency analysis – that is, an analysis of the number of barrier statements
generated by each site across the 12 categories, controlling for the total number of
ideas generated in each site – allows for comparison of the relative weight stake-
holders in each pilot site placed on the various barrier categories. Looking at Fig. 2,
it can be seen, for example, that 35% of all barriers generated in Prato related to
Data Quality, Accessibility, and Usability. As such, this category accounted for the
highest percentage of total barriers generated by stakeholders in Prato. Looking
across the pilot sites, it is also evident that the category Data Quality, Accessibility,
and Usability accounted for the highest or joint-highest percentage of total barriers
in Groningen, Issy-les-Moulineaux, and Den Haag.
Table 2 (continued)
Categories of barriers Sample barriers Sample options
Government and Failure to understand the Demonstrate the business
Organisational: Resistance to organisational benefits of case to local governments
open data initiatives releasing open data through case studies,
It will take a lot of effort to feedback and further
convince people to use open innovation outcomes
data Support and drive
organisational change
programs; Organisational
change management is
essential.
Motivation Failure to understand the Identify and publish data that
benefits that Open Data can is relevant and engaging
offer
Data publishing is not Promote the benefits of an
perceived as a “mission” in open data portal and give
administration’s point of good examples
view’
Privacy and Security Personal information Very clear data protocol and
accessed by public canlead guidance
to data protection
infringement
Some data is commercially Profiling of platform
sensitive members could support their
research without violating
personal information or
property rights
Skills and Training Inadequate technical Provide information, training
expertise to produce data in a and education, for all
usable format government agencies on the
benefits of an open data
portal
Users lack the skills to Provide open data FAQs for
process data and translate basic users
into information
Technical, Infrastructure, and Data is spread over different Pooling of public sector
Resources organizations and resources
departments Better curation and
Inadequate institutional maintenance of data quality
capacity to provide open data
services, to develop
standards and to provide
expertise
As was the case with regard to barriers, there were also differences in the relative
frequencies of options across sites. For example, while the Technical, Infrastructure
and Resources category accounted for a high percentage of total options generated
in Den Haag and Issy-les-Moulineaux, fewer options were generated in response to
this category in the other sites. Similarly, while Citizen Engagement received a high
314 M. Hogan et al.
User Needs
1. Information Needs
Stakeholders also highlighted specific needs of users in light of specific scenarios
of usage. Table 3 presents sample information needs for each category.
Given the range of scenarios, the user information needs generated across sites
were numerous and diverse, allowing for interesting comparisons (see Fig. 4). For
example, while the focus of the Den Haag workshop was on employment and
opportunity creation, resulting in a high proportion of information needs being
developed under the category Jobseekers Information, the Dublin workshop, which
focused on community engagement and planning generated information needs
across a much wider range, including: Community Information; Planning
Information; Services, Amenities and Event Information; Business and Financial
Information; and Child and Education-related. Also of note, for example, is the
high percentage of needs devoted to Business and Financial Data, in two pilot
Governance, Transparency and the Collaborative Design of Open Data Collaboration… 315
sites – Issy-les-Moulineaux, and Prato – two sites that have a focus on business and
local budgeting scenarios, respectively. It is likely that the information needs across
sites will develop further as each pilot site works to realise their scenarios by refer-
ence to the key open data that allows for effective collaboration between citizens
and public administrators.
316 M. Hogan et al.
2. Social-collaborative needs
Table 4 presents the category analysis for social-collaborative needs across sites.
Participants identified a range of social and collaborative needs, highlighting a num-
ber of forms of interaction for use over Open Data, as well as various considerations
and capabilities which would enhance the impact and appeal of the platform.
Participants highlighted the need for coaching and support, dialogue and discussion
spaces; feedback, moderation and maintenance of these spaces; platform tool capa-
bilities for interaction; varied forms of interaction over the data; and sharing and
requesting data.
Analysis of the relative frequencies of social and collaborative needs (see Fig. 5)
revealed that the Forms of interaction category accounted for a high percentage of
the total social and collaborative needs in three pilot sites: Den Haag, Dublin, and
Groningen. Coaching and support received the highest weighting in Issy-les-
Moulineaux, and it also received a high weighting in Den Haag (along with Forms
of interaction). Platform Tools and Capabilities for Interaction, which had the high-
est weighting in Prato, also received high relative weighting in Issy-les-Moulineaux
and Dublin. Examples from the three categories highlighted above emphasise, for
example, the need for flexibility of interaction: “there must be multiple modes”
(Forms of interaction); the need for support tools to be in place to “help users to
select the relevant data” (Coaching and support); and the ability to easily share data
analyses with others: “To be able to easily share graphs and reports obtained by TET
on social networks” (Platform Tool and Capabilities for Interaction).
3. Understandabilty, usability, and decision-making needs
Participants also used their scenarios to generate a set of understandabilty, usabil-
ity, and decision-making needs (see Table 5). Categories of needs here include:
Governance, Transparency and the Collaborative Design of Open Data Collaboration… 317
certification tools; decision-making support tools; guidance and support tools; abil-
ity to visualise and personalising data, and data analysis and reporting tools.
A relative frequency analysis of Understandability, Usability, and Decision-
making Needs (see Fig. 6) shows that, in four out of five pilot sites (Den Haag,
Dublin, Groningen, and Prato), the category The Ability to Visualise and Personalise
Data, generated the highest percentage of needs. This category included affordances
which would help users to understand and use open data, by allowing a degree of
flexibility and personal control over the way data is presented. Ideas in this category
referred to the need, for example, to “Filter data to my neighbourhood/interests”, to
“Return all data about my local area and visualize”, and “To be able to aggregate
geographic data belonging to different data sets on a new map”. Similarly, the cat-
egory Data Analysis and Reporting Tools included a high percentage of overall
needs across four pilot sites (Dublin, Groningen, Issy-les-Moulineaux, and Prato).
This category includes a number of needs which are important for deeper analysis
of open data, including: “Modelling tools that I can use with open data and citi-
zens”, “Data mining tools”, and the need “To build in real time graphics and visual
reports using Open Data”.
318 M. Hogan et al.
30
20
10
0
Den Haag Dublin Groningen Issy Les Prato
Molineaux
80
Data Analysis and Reporting Tools
(N=35)
Profiling (N=4)
20
The Ability to Visualise and Personalise
Data (N=49)
0
Den Haag Dublin Groningen Issy Les Prato
Molineaux
Discussion
Research and innovation focused on the design of open data platforms has the
potential to foster democratic processes by promoting transparency (Lourenço
2013; Dawes and Helbig 2010; Janssen 2011). A range of barriers have been identi-
fied that hamper effective service design and the full potential of open data plat-
forms, including poor information quality, lack of usability and technical issues,
limited organizational resources, and legislative challenges (Janssen et al. 2012;
Attard et al. 2015; Meijer 2015a). A range of generic user requirements have been
proposed to overcome barriers to effective open data platform design and service
delivery (Lourenço 2013; Jaeger et al. 2012; Van Velzen et al. 2009), but the unique
scenarios of usage and the unique needs of stakeholders are less often considered in
the open data platform design process (Dahlander et al. 2009).
The current study reflects a contextualist approach to conceptualising transpar-
ency and open data platform design, drawing in particular on the collective intelli-
gence scenario-based design ideas of stakeholders across five pilot sites in an effort
to analyse barriers to accessing, understanding, and using open data, options to
overcome these barriers, and the specific needs of open data platform users working
across a variety of scenarios. This research was conducted as part of an EU innova-
tion project, the Route-to-PA project. A primary goal of the project is the design of
an open data collaboration platform that can be flexibly used by citizens and public
administrators across a wide variety of usage scenarios that reflect a range of moni-
torial, deliberative, and participatory democracy activities. It was important for the
platform design team to understand the varied political and social contexts where
the open data platform is to be used, and the key needs of stakeholders. By using a
320 M. Hogan et al.
organizational resources and budget as major barriers to achieveing the full poten-
tial of open data platforms (Janssen et al. 2012; Attard et al. 2015; Meijer 2015a) .
Analysis of the relative frequencies of barriers across sites provided insight into
the relative weight stakeholders in each pilot site placed on the various barrier cat-
egories. For example, the high frequency of barriers in the Data Quality, Accessibility,
and Usability category overall reflected the fact that this category accounted for a
high percentage of total barriers generated by stakeholders in Prato, Groningen,
Issy-les-Moulineaux, Den Haag, and Dublin. Barrier statements in this category
were also phrased similarly across sites. For example, “Information is not presented
in a user friendly manner” (Den Haag); “Lack of user-friendly file-formats”
(Dublin); and “Data is published but cannot be found and does not have a user-
friendly format” (Groningen).
However, a number of differences were also observed across sites. For example,
the relatively stronger focus on data application barriers in Prato and Den Haag
could reflect the fact that both of these pilot sites and stakeholder groups are rela-
tively new to working with open data platforms. By contrast, Dublin, which has an
active open data platform, emphasised less data application barriers but highlighted
more barriers linked to skills and training. It may be that certain barriers and needs
(e.g., associated with the skilled used of platforms) will only arise after stakeholders
have had experience working with an evaluating existing platforms and services. A
key goal of the Route-to-PA project is to build upon existing platforms and provide
coaching and training in the use of key functionalities, working directly with stake-
holders in each pilot site. Work is ongoing to evaluate user experience of key func-
tionalities and the specific training needs that will be required as new platform users
are introduced to the platform. It is noteworthy that all pilot sites in the current study
emphasised technical, infrastructure, and resource barriers. Overcoming these bar-
riers may be essential to ensuring sustainable inputs in terms of quality data, itera-
tive design of platforms to enhance functionalities, and ongoing skills training to
increase the data competencies and collaboration skills of stakeholders and open
data platform users engaged in governance networks.
The different focus across sites is also evident in the absence of categories of
barriers in certain pilot sites. For example, Skills and Training is represented in all
pilot sites except Prato. Similarly, neither Resistance to Open Data Initiatives nor
Fear of Losing Control of Data are represented in the barrier categories in Den Haag
or Prato. The lower representation of barriers across categories in Prato is not sur-
prising, given that 79% of their total barriers fell into the three data-related catego-
ries: Data Quality, Accessibility, and Usability; Data Management/Policies; and
Data Applications. This suggests that stakeholders in Prato are primarily focused on
data-related barriers at this stage of their work together, and may not yet have
encountered organisational or training related barriers to the extent that other pilot
sites have.
There were also differences across pilot sites in the number and types of options
generated in response to barriers. For example, while Citizen Engagement received
a high percentage of generated options in Groningen (e.g. ask citizens which infor-
mation they find useful) and Issy-les-Moulineaux (e.g. allow citizens to make rec-
324 M. Hogan et al.
ommendations on the mode of data collection, the quantity of data and the
presentation format), it received relatively less attention in the other sites. Also,
whereas options in Dublin were spread across all categories, options were more
focused on a smaller set of specific categories in Den Haag and Prato. This suggests
that stakeholders in Den Haag and Prato, at least in their initial evaluation of the
problem situation, have identified a strong need for options in response to a select
number of barrier categories. Similarities and differences across sites provide useful
insights for the design team in terms of the possible focus of attention across pilot
sites when open data platform innovations are fully operational. They also highlight
key areas where flexible design of platform features needs to be combined with
broader strategies of political and social engagement with stakeholders and user
groups to ensure uptake and continued use of open data platform innovations.
Each pilot site in the current study focused on unique scenarios that reflect local
political and social priorities and thus stakeholders in each site had unique needs.
This was clearly reflected, in the first instance, in the range of open data information
needs across sites. For example, while the scenario in Den Haag focused on employ-
ment and opportunity creation, resulting in a high proportion of jobseekers informa-
tion needs, the Dublin scenario, which focused on community engagement and
planning, generated information needs across a much wider range, including com-
munity, planning, services, amenities, business, and education information. It is
likely that the information needs across sites will develop further as each pilot site
works to realise their scenarios and promote effective collaboration between citi-
zens and public administrators.
More generally, essential for the future success of open data portals is that more
varied high-quality open data is made available to stakeholders in an increasingly
accessible, understandable and usable manner. Societal challenges or problems,
including those that stakeholder focused on in the current study, are invariably com-
plex. A key goal of networked governance is to enhance our overall capacity to col-
laboratively resolve societal problems. However, as noted by Warfield (2006),
understanding societal problems always involves an effort to identify how problems in
the problem situation interact. Failure to recognise potential interactions between
problems in the problem situation can result in unexpected and often undesirable out-
comes. To the extent that networked governance arrangements involve collaboration
over open data in efforts to resolve societal problems, having access to sufficiently
varied, usable and understandable open data matched to the complexity of the prob-
lematic situation will be a core requirement of effective governance into the future.
Ongoing work by the Route-to-PA team has involved profiling the extent to which
open data is available, matched to, and useful for, the scenarios of interest to stake-
holders in each pilot site. This profiling of data is being used to feedback to public
administrators and key data providers to highlight some of the key gaps in the data.
Stakeholders across pilot sites in the current study also highlighted a range of
social and collaborative needs, in particular, the need for different forms of interac-
tion over open data, including dialogue and discussion spaces, moderation and main-
tenance of these spaces, feedback, sharing and requesting data, and also coaching
Governance, Transparency and the Collaborative Design of Open Data Collaboration… 325
There are a number of limitations to the current study. First, while reflecting the
different scenarios and contexts of usage identified as the starting point for the
Route-to-PA project, there was considerable variation in the stakeholders who par-
ticipated in the collective intelligence sessions across the different sites in the cur-
rent study. For example, Issy-les-Moulineaux was focused on a local enterprise
development scenario and thus the major citizen group in this context was stake-
holders in the business sector. The study results, and the range of information
needs identified across sites in particular, also vary as a function of the scenario
and the participants in the scenarios that stakeholders in each pilot site were using
to support idea generation at their respective workshops. At the same time, these
scenarios reflected the types of problems that stakeholders in each pilot site were
seeking to address, and thus the variation across sites is consistent with our contex-
tual approach to open data platform design. Future research should seek to exam-
ine the barriers, options and needs of different user groups across a range of
different scenarios, to further our understanding of the range of barriers, options,
and needs that will need to be considered in the future, in efforts to design
326 M. Hogan et al.
increasingly flexible and adaptable open data platforms that support the goals of
stakeholders across a range of different scenarios of usage. We speculate that per-
ceived barriers to accessing, understanding, and using open data, and options to
overcome these barriers, may show greater similarities across different contexts,
when compared with the range of information, social-collaborative, and decision-
making needs of users, as these barriers may reflect underlying political and social-
organisational challenges that are fundamental to the broader societal challenge of
supporting transparency and collaboration over open data. Furthermore, our con-
clusions regarding the specific needs of users are a function of the specific methods
we used, and future research should seek to combine our collective intelligence
scenario-based design methods with other user-centered methodologies to provide
more insight into the specific barriers, options, and needs of open data platform
users. For example, the use of remote user testing may be particularly useful in the
iterative design of open data platforms as they evolve and develop further.
Conclusions
In line with the approach adopted in the current study, Ojo and Mellouli (2016) note
that governments are increasingly engaging private sector organizations, civil soci-
ety and citizens to tackle complex policy challenges across a variety of networked
governance arrangements. Although evidence suggests that networks of non-state
actors are equally as important as networks of state actors in terms of their contribu-
tion to governance outcomes (Bodin and Crona 2009), networked governance
implies the need to develop a shared understanding of problems and solutions to
problems (Huppé et al. 2012). This implies the need for a collective intelligence
approach to the design of platforms that facilitate the deliberation of diverse gover-
nance networks over open data, and the co-creation of policies and projects that help
to resolve societal problems, increase trust in government, and empower increas-
ingly effective networked governance arrangements into the future.
As noted by Ojo and Mellouli (2016), the efficacy of governance networks is
contingent on the inclusion of citizen in the networks, and mobile social-media
platforms could constitute a key infrastructure for enabling citizen participation in
this regard. However, based on their case study analyses, they also note that these
networks are still largely steered by government and it remains important that gov-
ernments initiate and demonstrate deep commitments to partnerships with citizens
for collaborative governance networks to be effective. Ojo and Mellouli (2016)
note that government is ultimately responsible for building trust with partners and
are accountable for the overall outcome of the networked governance arrangement.
This implies ongoing investment and iterative design, innovation and experimenta-
tion with key infrastructures that may support networked governance. Considering
the specificity of the key understandability, usability, and decision-making needs
identified in the current study, it is clear that governments and citizens need to
Governance, Transparency and the Collaborative Design of Open Data Collaboration… 327
work with social scientists and technology experts to design open data platforms
that include a range of data analysis and decision-making affordances that support
collaborative societal problem solving and policy development. This needs to be
coupled with appropriate training in the use of these affordances. Based on their
case study analyses, Ojo and Mellouli (2016) highlight the need to effectively
motivate citizen participation in governance networks and align the divergent
views of the different actors collaborating in the network. From a contextualist
perspective, the collective intelligence scenario-based design thinking of stake-
holders in the current study highlights that motivating citizens may be contingent
on meeting their needs. This implies designing a socio-technical infrastructure that
supports their social-collaborative and decision-making needs, which will be criti-
cal to sustain motivation in the use of the platform.
Consequently, based on the outcomes from our study and related literature, we
conclude that: (1) the nature of barriers and needs of stakeholders can vary signifi-
cantly from one context to another and this needs to be considered in the develop-
ment of open data platforms that are designed explicitly for use across several local
authorities or contexts; (2) the iterative use of collective intelligence scenario-based
design methods employed in eliciting barriers, options and needs from different
stakeholders could be an effective approach for engaging stakeholders in the design
of open data platforms into the future, particularly if it can be effectively combined
with other user-centered methods; (3) continued engagement of stakeholders in the
design and development of open data platforms is contingent on the support pro-
vided by local authorities working with the stakeholders.
Appendix 1
Sample scenarios
Entrepreneur Annie is interested in starting a locally based café/food business and would like to
connect with public administrators and potential customers to find out if there is a demand for
this new business, what kind of premises or permissions she might need, what supports are
available and to connect with other people who might partner/work with her in starting this
business. She would like to use technology to build local social networks to connect with her
business peer network and build a local customer base.
Civic Joe is part of the civic hacker community and a member of an active citizen group. He is
a keen advocate for social equality and feels that citizens need a more participatory
democracy to create a better society for all. He is interested in open data as a means of opening
access to public information and promoting transparency. He wants to be able to interact with
public data to understand how public decisions are made, to give his views in an easy and
transparent way and receive feedback on them from public administrators who area leading
local projects, so that he feels he has been part of the decision and policy making process. Joe
also wants to be able to share ideas and data with other citizen groups, with a view to
collaborating on projects and common goals.
(continued)
328 M. Hogan et al.
Sample scenarios
Jane is a public administrator in a Dublin Local Authority. Jane is helping to prepare a new plan
to promote local community and economic development in Dublin and wants to explore how
technology might be used to engage a wider demographic and to facilitate bottom up
community building. Jane is particularly interested in consulting with young people and
people with a disability or other citizens who may not engage in more formal consultations.
Jane wants an easy to use platform to gather and give feedback to citizens on issues that matter
to them to inform policy and to build public trust. Jane also wants to be able to negotiate and
plan activities with other public administrators in her community development group in her
local authority public administration offices. She wants both citizens and her colleagues in the
local community development group to have some flexibility in the way they draw upon data
and information when working together to develop community projects. Jane is very passionate
about promoting local community and economic development in Dublin and she wants a
platform and set of services that will help her do good work.
Citizen Kay is interested in putting down more roots and getting involved in her local
community. She initially got involved in community issues when a group of her neighbours got
together to object to a big new development that would have caused a lot of disturbance in her
quiet street. As a concerned citizen she wants an easy way to put her issues on a public
platform, to share and find out about local news, to discuss with other local residents and have
an input into what is happening in her community. She would like a meaningful exchange with
public administrators and to build local social networks to highlight the good things that are
happening in her community and perhaps to start up a skillshare/ local volunteering exchange.
Kay wants to be able to access information on other similar local groups, so that she can get
advice on starting her own.
References
Cappelli C, Engiel P, De Araujo RM, Cesar J, Leite P (2013) Managing transparency guided by a
maturity model. In: 3rd global conference on transparency research, HEC, Paris, France, 24–26
October 2013, pp 1–17
Caroll J (2000) Five reasons for scenario-based design. Interacting with Computers 13:43–60
Christakis AN (1987) Correspondence: Systems profiles. Systems Research 4(1):53–58
Coke JG, Moore CM (1981) Coping with a budgetary crisis: Helping a city council decide where
expenditure cuts should be made. In: Burks SW, Wolf JF (eds) Building city council leader-
ship skills: A casebook of models and methods. National League of Cities, Washington, DC,
pp 72–85
Cohn M (2004) User stories applied for Agile software development. Addison-Wesley, Boston
Dahlander L, Fredriksen L, Rullani F (2009) Online Communities and Open Innovation. Ind Innov
15(2):115–123
Dawes S, Helbig N (2010) Information strategies for open government: challenges and prospects
for deriving public value from government transparency. Electron Gov 6228:50–60
Deloitte (2013) Market assessment for public sector information. Written for UK, Department for
Business, Innovation and Skills
Denis J, Goëta S (2014) Exploration, Extraction and “ Rawification ” The Shaping of Transparency
in the Back Rooms of Open Data. In: Neil Postman Graduate Conference. New York, pp 1–8
Dwyer C, Hogan M, Harney O, O'Reily J (2014) Using interactive management to facilitate a
student-centered conceptualisation of critical thinking: a case study. Educ Technol Res Dev
62(6):687–709
Fox J (2007) The uncertain relationship between transparency and accountability. Dev Pract 17(4–
5):663–671. doi:10.1080/09614520701469955
Feeg R (1988) Forum of the future of pediatric nursing: Looking toward the 21st century. Pediatric
Nursing 14:393–396
Ghaus-Pasha A (2007) Governance for the millenium development: core issues and good practices.
Building
Greiner A, Isaac A, Iglesias C, Laufer C, Guéret C, Stephan EG, Kauz E, Atemezing GA,
Bittencourt II, Almeida JP, Carrasco MT, Archer P, Albertoni R, Purohit S Córdova Y (2015)
Data on the Web best practices – W3C working draft 25 June 2015. Retrieved from http://www.
w3.org/TR/2015/WD-dwbp-20150625/
Grimmelikhuijsen SG, Welch EW (2012) Developing and testing a theoretical framework for
computer-mediated transparency of local governments. Public Adm Rev 72(4):562–571.
doi:10.1111/j.1540-6210.2011.02532.x
Groarke JM, Hogan MJ (2016) Enhancing wellbeing: An emerging model of the adaptive func-
tions of music listening. Psychology of Music 44(4):769–791
Hayes SC, Hayes LJ, Reese HW (1988) Finding the philosophical core: a review of Stephen
C. Pepper’s world hypotheses: a study in evidence. J Exp Anal Behav 1(1):97–111
Habermas J (1962) The structural transformation of the public sphere (1962, trans: Cambridge
Massachusetts, 1989, MIT Press)
Heald D (2006) Varieties of transparency. Proceedings-British Academy 25–43. doi:10.5871/
bacad/9780197263839.003.0002
Hilgers D, Ihl C (2010) Citizensourcing: applying the concept of open innovation to the public
sector. Int J Public Participation 4(1):67–88
Hogan MJ, Johnston H, Broome B, McMoreland C, Walsh J, Smale B et al (2015) Consulting
with citizens in the design of wellbeing measures and policies: lessons from a systems science
application. Social Indicators Research 123(3):857–877
Jaeger P, Bertot JC, Shilton K (2012) Information policy and social media: framing government-
citizen Web 2.0 interactions. In: Reddick CG, Aikins S (eds) WEb 2.0 Technologies and
Democratic Governance. Political, policy and management implications, Springer, New York,
pp 11–25
Janssen K (2011) The influence of the PSI directive on open government data: an overview of. Gov
Inf Q 28:446–456
330 M. Hogan et al.
Janssen M, Charalabidis Y, Zuiderwijk A (2012) Benefits, adoption barriers and myths of open
data and open government. Inf Syst Manag 29(4):258–268
Keever, D. B. (1989, April). Cultural complexities in the participative design of a computer-based
organization information system. Paper presented at the International Conference on Support,
Society and Culture: Mutual Uses of Cybernetics and Science, Amsterdam, The Netherlands
Lee G, Kwak Y (2012) An open government maturity model for social media-based public engage-
ment. Gov Inf Q 29(4):492–503
Lourenço RP (2013) Open government portals assessment: a transparency for accountability per-
spective. Lecture Notes in Computer Science (including Subseries Lecture Notes in Artificial
Intelligence and Lecture Notes in Bioinformatics), 8074 LNCS, 62–74. doi:10.1007/
978-3-642-40358-3-6
Mei CS, Dewan SM (2014) Towards conceptualizing information transparency and its role in inter-
net consumers’ concerns: a literature review
Meijer A (2009) Understanding modern transparency. Int Rev Adm Sci 75(2):255–269.
doi:10.1177/0020852309104175
Meijer A (2015a) E-governance innovation: barriers and strategies. Gov Inf Q 32:198/206
Meijer A (2015b) Government transparency in historical perspective: from the ancient regime to
open data in The Netherlands. Int J Public Adm 38(3):189–199. doi:10.1080/01900692.2014
.934837
Ojo A, Porwol L, Waqar M, Stasiewicz A, Osagie E, Hogan M, Harney O, Ahmadi-Zeleti F (2016)
Realizing the innovation potentials from open data: Stakeholders’ perspectives on the desired
affordances of open data environment, 17th IFIP working conference on virtual enterprises,
Porto, Portugal, 3–5 October 2016, Springer
Ojo A, Mellouli S (2016) Deploying governance networks for societal challenges. Gov Inf Q.
doi:10.1016/j.giq.2016.04.001
Peled A (2011) When transparency and collaboration collide: the USA open data program. J Am
Soc Inf Sci Technol 62(11):2085–2094
Pepper SC (1942) World hypotheses: a study in evidence. University of California Press
Piotrowski S, Conference G, Sasaki D (2011) Conceptualizing the Quality of Transparency Paper
prepared for the 1, 0–27
RezaeiZadeh M, Hogan M, O’Reilly J, Cunningham J, Murphy E (2017) Core entrepreneurial
competencies and their interdependencies: insights from a study of Irish and Iranian entrepre-
neurs, university students and academics. Int Entrep Manag J 13(1):35–73
Rosson MB, Carroll J (2002) Scenario-based design. In: Jacko J, Sears A (eds) The human-
computer interaction handbook: fundamentals, evolving technologies and emerging applica-
tions. Lawrence Erlbaum Associates, Mahwah, pp 1032–1050
Sandoval-Almazan R, Gil-Garcia JR (2012) Are government internet portals evolving towards
more interaction, participation, and collaboration? Revisiting the rhetoric of e-government
among municipalities. Government Information Quarterly 29:72–81
Sato T (1979) Determination of hierarchical networks of instructional units using the ISM method.
Educational Technology Research 3:67–75
Van Velzen L, Van der Geest T, Ter Hedde MD (2009) Requirements engineering for e-Government
services: a citizens-centric apporach and case study. Gov Inf Q 26:477–486
Warfield JN (1994) Science of generic design: managing complexity through systems design. Iowa
State Press, Ames
Warfield JN (2006) An introduction to systems science. World Scientific, Singapore
Zuiderwijk A, Janssen M (2014) Open data policies, their implementation and impact: a frame-
work for comparison. Gov Inf Q 31:17–29
Michael Hogan is a senior lecturer and researcher at NUI, Galway, Ireland. His research foci
include individual, social and technology factors contributing to adult learning, motivation, and
collaborative performance. He is a co-director of the Structured PhD in Perception, Cognition and
Action, co-director of the Structured PhD in Learning Sciences, and a co-leader of the Health and
Wellbeing theme at the Whitaker Institute for Innovation and Social Change, NUI, Galway.
Governance, Transparency and the Collaborative Design of Open Data Collaboration… 331
Dr. Adegboyega Ojo is Senior Research Fellow at the Insight Centre for Data Analytics, National
University of Ireland Galway (NUIG). He leads the E-Government Group at Insight Centre, serves
as Adjunct Lecturer at the College of Engineering and Informatics. His current research interests
include data driven innovations in government, Open data policies and Infrastructures, data analyt-
ics and governance of smart cities. He is member of the Editorial Boards of the Government
Information Quarterly and International Journal of Public Administration in the Digital Age.
Owen Harney is a PhD candidate (Learning Sciences) at the School of Psychology, NUI Galway.
His PhD research focuses on the integration of systems science methodologies into educational con-
texts. Outside of his PhD research, Owen has applied his knowledge of systems science and collec-
tive intelligence methodologies in a wide variety of research and organisational contexts, including:
collaborative learning, e-governance, chronic pain, personalised nutrition, and citizen engagement.
Erna Ruijer is a postdoctoral researcher at the Utrecht School of Governance in the Netherlands.
Her research focuses on open data, government transparency, open government and government
communication.
Albert Meijer is a Professor of Public Innovation at the Utrecht School of Governance in the
Netherlands. His research focusses on e-government, government transparency, co-production in
the public sector and public innovation.
Jerry Andriessen is a senior scientist and research director of Wise & Munro learning Research,
The Hague, Netherlands. Wise & Munro participates in many innovative European projects, on
technology, open data, collaboration, primary education, and cybersecurity. These projects add
societal relevance to their scientific implications. Jerry has more than 25 years of experience in
uses of technology for support of learning and collaboration in various settings. His current interest
is in the qualitative interpretation of discourse and action, including the roles of emotions.
Mirjam Pardijs is an independent educational scientist. She is co-director of Wise & Munro
Learning Research, the Hague, Netherlands. She specialises in the design, coaching, and interpre-
tation of collaborative learning in educational and professional contexts. Her main interests are
narratives as a tool for learning, appropriation of technology, and the role of pedagogical support
for innovation in learning settings.
Paolo Boscolo gained a Masters degree in Electronics Engineering in 1986. In 1992 he joined the
Comune di Prato as head of the Telecommunications group. Is new responsible for the general
coordination of the ICT service and for co-funded project in ICT field. During last 10 years has
been charged for the management of advanced eGovernment projects for the Prato City Council.
He is currently responsible for the definition of the Prato Smart City Plan. He’s the responsible for
Route-to-PA project operations at Comune di Prato.
Elena Palmisano received a degree in Civil Engineering in 1988 and a Ph. D. in Methods and
Technologies for Environmental Monitoring in 1994 at the University of Florence (Italy). She
works as a consultant with PIN Scrl (a public/private Consortium managing the University of
Florence site in Prato) and with the Municipality of Prato. She has taken part in many regional,
national and EU projects in the field of innovation technology, ICT applications in different con-
texts, such as infomobility, e-government, cultural heritage and education. She is currently col-
laborating with the Municipality of Prato in two H2020 projects: Route-to-PA (GA N. 645860),
where she acts as community manager in the pilot activity, and City.Risks (GA N. 653747) where
she contributes to the managemnt of dissemination activities. With PIN S.c.r.l. she is also involved
in the definition of the guideline for the development of the Smart City plan in Prato.
Matteo Satta is an international Political Sciences graduate in Turin (Italy), he is a senior project
manager that has always been involved, between France and Italy, in the field of R&D and
332 M. Hogan et al.
Innovation (EU and International level) with a particular focus on its valorisation and exploitation.
Since 2005, he has contributed to the management and the development of various European proj-
ects, such as e-Photon/ONe, the Researchers’ Night in Turin (Italy), RADICAL, OTN, ECIM,
Citadel… on the Move and ROUTE-TO-PA, and IPR Licensing programs, such as MPEG Audio
(MP3) and DVB-T. He is today EU Project Manager in Issy Média, a semi-public company of the
City of Issy-les-Moulineaux, specializing on Smart City subjects and projects with a particular
interest in citizens’ participation, Living Lab, Open Data and Smart Mobility.
Michael Baker is a Research Director (tenured Research Professor) in psychology and language
sciences, of the Centre National de la Recherche Scientifique (CNRS). He currently works in the
Social and Economics Science Departement of Telecom ParisTech, the French National
Telecommunications Engineering School. His research aims to analyse and model the processes of
knowledge elaboration in dialogues produced in learning and work situations, focussing on argu-
mentation and the role of technological mediation. Michael Baker recently published the following
book, with Baruch Schwarz: “Dialogue, Argumentation and Education: History, Theory and
Practice” (Cambridge University Press, New York, 2017).
Françoise Détienne is research director at CNRS, Paris. Her research in cognitive ergonomics
focusses on collaboration, design, creativity and online communities as well as on the role of
socio-technical systems. She is associate editor of International Journal of Human Computer
Studies, and member of the editorial board of Interacting With Computers and CoDesign.
Dr. Lukasz Porwol is a Postdoctoral Researcher and deputy unit leader at Insight Centre for Data
Analytics (formerly known as DERI), National University of Ireland, Galway. At Insight Centre
for Data Analytics, his research work focuses on leveraging social media, games and mobile tech-
nologies to support effective dialogue between citizens and decision-makers. His background
combines solid engineering and scientific knowledge (MSc in Computer Science and Engineering
in Information Technologies) with good management and PR skills.
Teresa Scassa and Amy Conroy
Abstract This paper explores strategies for balancing privacy with transparency in
the release of government data and information as part of the growing global open
government movement and within an evolving technological context. Government
data or information may contain many different types of personal information. In
some cases, transparency will require the release of this personal information; in
other cases, the release of personal information will not advance the goals of gov-
ernment transparency. The situation is complicated by the availability of technolo-
gies that facilitate widespread dissemination of information and that allow for the
mixing and mining of information in ways that may permit the reidentification of
individuals within anonymized data sets. This paper identifies a number of strate-
gies designed to assist in identifying whether data or information contains personal
information, whether it should be released notwithstanding the presence of personal
information, and what techniques might be used to minimize any possible adverse
privacy impacts.
Introduction
This paper explores strategies for balancing privacy with transparency in the release
of government information. It does so within the context of the global movement
towards more open and transparent government – a movement which encourages
the release of government data and information through open data and proactive
disclosure. It also does so within a rapidly evolving technological context and one
in which big data analytics plays an ever-increasing role. In this paper we identify
strategies for balancing privacy with transparency, although we do not set out to
establish what the outcome of that balance should be. The appropriate balance may
T. Scassa (*)
University of Ottawa, Faculty of Law, Ottawa, ON, Canada
e-mail: tscassa@uottawa.ca
A. Conroy
Carleton University, Department of Law and Legal Studies, Ottawa, ON, Canada
e-mail: Amy.Conroy@carleton.ca
Open data involves the release of government data in reusable digital formats and
under an open license (Kitchin 2014). While transparency is one of the goals of the
open data movement, other objectives include stimulating innovation and encourag-
ing citizen engagement (Janssen 2012). Vejkovic et al. (2014, 281–282) identify the
data sets most frequently released as falling within the categories of “Finance and
Economy, Environment, Health, Energy, Education, Transportation, Infrastructure,
Employment and Population.”
Proactive disclosure can be of data or other types of government information. The
goal of proactive disclosure is to anticipate and release the kinds of information most
frequently sought from governments and to ensure that this information is freely and
easily accessible. Rather than having to file individual freedom of information
requests in order to access government information, proactive disclosure can push
more information towards the public, minimizing cost and delays (Queensland 2013).
In many countries the principle that court proceedings should be open to the
public underlines the point that transparency is essential to a properly functioning
judicial system (McLachlin 2003). This principle of openness is generally extended
to the publication of court decisions by default, with exceptions made only in very
particular circumstances where the public interest requires it (Winn 2004). Many
courts now make their decisions freely available to the public online either through
their own websites or through ‘legal information institutes’ – online portals designed
to facilitate public access to court decisions (Greenleaf 2011). Some courts are also
extending public access to other documents (such as legal briefs, for example)
through online portals. Some administrative tribunals are following suit and making
their decisions publicly available online. While on the one hand, digital openness of
this kind can enhance transparency, it is not without its impacts on privacy. Online
and fully searchable dissemination of this information may cause harm to individu-
als that was not considered significant enough to warrant suppression of the infor-
mation when distribution of these materials was paper-based (Scassa 2014). The
disclosure of the personal information of litigants in digital format and on a global
scale may have certain positive impacts (for example increasing transparency in
cases of serial bankruptcy or fraud). The increased exposure of personal details
about litigants may also, however, have adverse impacts on the administration of
justice and on public confidence in judicial or administrative processes if details of
highly sensitive cases (for instance family law disputes) become searchable online.
Finally, some government information is made available to the public through
registries. Such registries frequently contain personal information. The public dis-
closure of this information may be mandated by law (as, for example, in the case of
public land titles registries, political campaign donor lists, or other information
deemed disclosable in the interests of transparency). As these registries move from
paper-based to online platforms, digital dissemination may change the nature of the
privacy impacts (for example making it easier and more convenient to access per-
sonal information out of curiosity, for malicious purposes, or for data mining or
profiling, rather than for the purpose for which the paper registry was created). As a
result, the change towards making this information more readily available online
may require some mitigation of potential privacy harms.
336 T. Scassa and A. Conroy
Technological Context
The technological context in which governments now release data (meaning raw
facts) and information (contextualized or interpreted knowledge) (Kitchin 2014)
has changed dramatically and these changes are ongoing. Indeed, the demand for
increased openness of governments is driven in part by this technological change.
The value of government data for research, analysis and innovation has greatly
increased as the technologies that enable reuse of this information have evolved and
shifted into the hands of individuals, civil society organizations and large and small
corporate entities. Not only are governments pressured to release more information,
they are asked to do so in formats that are machine-readable and easily reusable.
While the Internet has greatly facilitated the dissemination of information of all
kinds, including government data, available and evolving technologies also permit
rapid and low-cost storage, reuse, dissemination, copying, mining and analysis of
the data (Kitchin 2014). Big data analytics have become mainstream, with the anal-
ysis of vast stores of information being used across all sectors of decision-making
from weather prediction to consumer profiling, and from professional sports to
medical research. Governments are also taking advantage of big data analytics for
planning and resource management purposes, among many others (Mayer-
Schönberger and Cukier 2014).
It is into this evolving technological landscape that governments release infor-
mation and data under open government programs and under existing laws. While
much of this material may have no privacy implications and may contain no personal
information of any kind, other data sets or information do contain personally iden-
tifiable information. It is within this context that the balance between privacy and
The Privacy/Transparency Balance in Open Government 337
transparency must be assessed. While excessive concerns for privacy should not be
allowed to trump transparency (and while privacy should not be used as an excuse
to avoid transparency such as, for example, with respect to political campaign
financing or public contracting), it must also be taken into account that much of the
data may be sought not for its transparency value but for its potential for commer-
cial reuse (for instance to support consumer profiling or marketing). In this context
a lack of attention to privacy might undermine citizen confidence in government and
might lead to privacy harms, including for instance a chilling effect on people com-
municating with government for fear that their personal information will not be
properly protected (Borgesius et al. 2015).
Transparency and Privacy
According to Yu and Robinson (2012 at 186), “open government” has been used “pri-
marily as a synonym for public access to previously undisclosed government informa-
tion.” In this sense, open government is about transparency. The transparency
objectives of open government are often focused on making more information avail-
able at a lower cost. The reduced cost is not due only to the fact that information is
made available to the public without fees – lower costs can include reducing the time
or effort needed to access – or to reuse – the data or information (Candeub 2013). Both
open data and proactive disclosure reduce the time, cost and effort of access to data
and information, and in this sense, they increase government transparency. Whether
there are actual outcomes – such as increased accountability of government – will
depend on whether the information is used by anyone to serve these goals.
The concept of transparency in government is often linked to ‘openness’ (Schauer
2011). However, ‘transparency’ itself has little normative content. There is no single
standard for the degree or terms of openness required of governments. Further, as
Schauer notes, the concept of transparency does not dictate any particular outcomes;
a more transparent government does not necessarily mean one that is held more
strictly to account. Transparency merely requires that government information,
data, and processes should be publicly available or accessible – as appropriate in the
circumstances.
Used in this sense, the concept of transparency focuses on providing access, not
on specific results. By contrast, the open government movement is more results-
oriented. For example, the Open Government Declaration (Open Government
Partnership 2011) identifies a range of objectives for open government. Signatories
commit to “greater civic participation in public affairs, and seeking ways to make
their governments more transparent, responsive, accountable, and effective.” Open
data is also released in order to stimulate innovation by encouraging its reuse in the
private sector. Open government, therefore, is not only about transparency; it serves
a diversity of goals. Not only is it important to consider what may drive the demand
for certain types of data, it is also difficult to predict how data sets may be used in
order to achieve transparency.
338 T. Scassa and A. Conroy
These dynamics are important when balancing transparency with privacy. The
disclosure of information that is either personally identifying or capable of being
used to identify individuals may make governments more transparent (at the expense
of personal privacy) but it does not necessarily make them more accountable. The
disclosed information may be used for accountability purposes, but it may also be
used to serve purely private or commercial interests.
The disclosure of personally identifiable information in the hands of govern-
ment has an impact on individual privacy, although the nature or degree of impact
may not always be evident, and may depend upon the nature of the information
that is disclosed. Personally identifiable information may be used to attack or
stigmatize individuals (as, for example, when personal health information is
revealed in a tribunal decision posted online). If very specific identifiers are
revealed, they might be used in targeted attacks such as identity theft or imperson-
ation. In many cases, though, what will be disclosed is deidentified or anonymized
information; privacy impacts might therefore depend upon the existence of both a
desire to reidentify and the resources to do so. In some instances, various data sets
will be combined and used to contribute to profiles of individuals that may have
direct or indirect effects on them that are difficult to identify or quantify, as where
data is used in big data analytics to profile or target particular consumers.
How the balance is struck between transparency and privacy may depend upon
cultural norms within any given jurisdiction. These may include different views
on what constitutes sensitive personal information and, as a result, what type of
information should be released or withheld in an effort to ensure transparency in
government (Zuiderwijk and Janssen 2014; Article 29 Working Party 2013 at 18).
Social and political factors such as the level of secrecy in government or the level
of citizen engagement will also influence the understanding of transparency and
privacy as well as expectations about what information the government should be
permitted to release (Roy 2014).
Personal Information
The second category of information is that which does not on its own identify a
specific individual, but can be used to identify an individual when it is combined
with other available information. Increasingly advanced anonymization techniques
are needed to forestall reidentification as more and more information becomes
available for cross-referencing (El Emam and Fineberg 2009). In some cases, while
information may appear to have been anonymized, it may be relatively easy to iden-
tify the other pieces of information that could be used to identify an individual from
an anonymized dataset. However, in the big data environment it is becoming increas-
ingly difficult to know just what other data is available. This is not simply because
governments at all levels are releasing large volumes of data, but also because huge
stores of personally identifiable information are also in the hands of private sector
actors (Ohm 2010). Assessing reidentification risk can therefore be challenging. It
can also be a moving target as both the stores of available data (both personal infor-
mation and de-identified data) and analytic capacity increase (Ohm 2010; Schneier
2015). This means that the cost in terms of time and effort required to properly
anonymize information before it is released proactively or as open data can be
expected to rise. A relevant question to ask is how to fund these activities as part of
the overall effort to release more and more government data/information (particu-
larly given the potential for commercial gain as it is used in new and innovative
ways by private sector actors).
The privacy problems discussed here have been identified and addressed by data
commissioners in different contexts and at different times. Many responses have
evolved in the access to information/right to know context. There is therefore a
growing body of information on ways to protect privacy in the release of govern-
ment information. In this part of the paper, we distill some of these into a set of
strategies that offer ways in which government institutions can protect privacy in the
release of government information, whether it be through open data, open courts,
proactive disclosure or access to information requests. The strategies are adapted to
the Web 3.0 environment. They take into account the need to protect privacy while
meeting transparency goals. As a result, they require consideration to be given to the
balance to be struck between privacy and transparency. This balance may be differ-
ent in different contexts, and may depend upon factors such as: the degree of sensi-
tivity of the information, the circumstances under which the information was
provided to government, the risk of harm from reuse/misuse of the information, the
risk of reidentification (in the case of anonymized data) and the transparency value
of the information.
The first two strategies outlined below (data minimization and inter- and intra-
governmental consultation) address overall institutional practices. The third s trategy
is aimed at assessing the extent to which personally identifiable information is pres-
ent within any given dataset or document. The fourth, fifth and sixth strategies offer
340 T. Scassa and A. Conroy
Data Minimization
One way in which citizen privacy can be protected in the context of open govern-
ment is through the minimization of the amount of personal information that is col-
lected in the first place. Data minimization principles are already present in public
and private sector data protection laws. For example, the UK Data Protection Act
1998 provides in Schedule 1, Part I, art. 3 that: “Personal data shall be adequate,
relevant and not excessive in relation to the purpose or purposes for which they are
processed.” Canada’s Privacy Act (s. 4) restricts the government to collecting only
information that “relates directly to an operating program or activity of the institu-
tion.” Other countries have started implementing an “entry once” principle to
require reuse of data that has already been collected as opposed to collecting the
same information directly from the individual a second time (Meyerhoff-Nielsen
and Krimmer 2015 at 279). The idea behind data minimization is simple: only per-
sonal information genuinely required to meet the needs of a particular program
should be collected. In endorsing data minimization, Ontario’s former Information
and Privacy Commissioner (Cavoukian 2009 at 10) recommended that interactions
between government agencies and the public should “begin with non-identifiable
interactions and transactions as the default”. The minimization of personal informa-
tion collection will mean that there is less personal information to protect in con-
texts where disclosure is sought or where the decision is made to release information
through proactive disclosure or as open data. Efforts to ensure citizens are informed
of the information that the government holds about them and the purpose for which
it was collected and is used (see discussion of MyPage initiatives in Norway and
Denmark in Meyerhoff-Nielsen and Krimmer 2015) may potentially promote
increased attention to the need for data minimization in the government.
Although data minimization principles can shape what information is collected
by government agencies, they can also apply to the disclosure of information for
secondary purposes. Such purposes might include health or other types of research.
For example, El Emam and Fineberg (2009) argue that government agencies that
disclose health data to researchers should make use of anonymization techniques in
order to limit the amount of personally identifiable information that is released. In
the context of courts (and by extension administrative tribunals) data minimization
principles could apply at the release stage to ensure that only that personal informa-
tion necessary to serve the purposes of providing transparency in legal proceedings
should be disclosed in decisions in order to avoid privacy consequences for indi-
viduals when decisions are published online (Sherman 2013; Berzins 2008).
Data retention policies that require the purging of personal information that is no
longer required to meet the purposes for which it was collected are also a means by
which government agencies can limit the amount of personal information in their
The Privacy/Transparency Balance in Open Government 341
hands, and therefore the privacy risks to individuals (Office of the Privacy
Commissioner of Canada 2014; UK ICO 2016). Data minimization through limits
on retention presents somewhat of a challenge in the Web 3.0 environment as the
destruction or disposal of data goes against the ethos that more is better and that it
is impossible to know what data will be relevant or useful in analytics (Office of the
Privacy Commissioner of Canada 2014; Kitchin 2014). Nevertheless, as an element
of privacy protection and government accountability, limiting the retention of per-
sonal information and properly disposing of personal data when it is no longer
required protects privacy and limits the risk of improper disclosure of personal
information.
Data minimization is not, on its own, a solution to the privacy challenges identi-
fied in this paper. It is a strategy that can help reduce the amount of personal infor-
mation in the hands of government, thereby diminishing the possibility of
inappropriate or harmful release of personal information. The reality is, however,
that governments will always need to collect personal information in order to oper-
ate their many programs and services. Because of this, it should be used in combina-
tion with other strategies (including ensuring citizens are able to become informed
of the information that the government holds about them) as part of an overall infor-
mation management plan.
messaging to the public about open data may not match the realities experienced by
public servants who are charged with making data available as open data. Writing in
the Canadian context, Roy (2014) argues for a national strategy involving all levels
of government in order to co-ordinate the different open data initiatives. While nei-
ther Roy nor Zuiderwijk & Janssen specifically address balancing privacy and trans-
parency in open government, their arguments in favour of greater communication
and collaboration both within and between governments are relevant to the manage-
ment of privacy issues.
As with data minimization, communication and cooperation both within and
between governments is not a panacea for addressing the challenges of balancing
privacy with transparency. They are both broad strategies that can contribute to an
environment that improves the management of personal information and decision-
making around the coordinated release of government information and data.
However, within this environment, case-by-case decisions must still be made
regarding the public release of particular data sets and other types of government
information (including court decisions or information released as part of proactive
disclosure). The next four strategies are ones which are aimed at this case-by-case
decision-making process.
The PIA can be specifically adapted to the open government context. This context
is one in which the protection of privacy may have to be balanced against transparency
values. The Australian Information Commissioner’s Guide to Undertaking Privacy
Impact Assessments (2014) offers an example of an assessment process that is designed
to take into account competing and counterbalancing considerations. It provides that
a PIA is much more than a simple compliance check. It should ‘tell the full story’ of a
project from a privacy perspective, going beyond compliance to also consider the broader
privacy implications and risks, including whether the planned uses of personal information
in the project will be acceptable to the community. (Office of the Australian Information
Commissioner 2014 at 2)
This suggests that a PIA adapted to the open government context should take into
account not just the privacy impacts but also the transparency value of the data or
information that is to be released. If there is a high transparency value this may
favour release in spite of privacy implications; data or information with a low trans-
parency value might require stronger measures to protect personal information from
release or reidentification.
The PIA process for open data and proactive disclosure can incorporate a series
of specifically designed questions that explore both privacy risks and transparency
values. The list of questions below is adapted from Scassa and Conroy (2016).
1. What is the purpose of disclosure of the information or release of the dataset?
(Possible purposes might include furthering government transparency,
encouraging open engagement, or supporting innovation, or research)
2. Does the document or dataset contain any personal information?
(Personal information includes any specific identifiers such as the name of a
person, their address, or a unique identifying number)
3. If the answer to Question 2 is ‘yes’, is the personal information relevant to the
purpose(s) for disclosure?
4. Does the data set contain any personally identifiable information?
(This can be information which does not, on its own, identify specific indi-
viduals, but which, when linked to other information might lead to their identifi-
cation. Examples can include postal codes, gender, profession or other
demographic information. For example, if the information is in the form of court
decisions, the names of the parties may be relevant to the principle of open courts
that motivates the publication of the decisions.)
5. If the document or data set contains personally identifiable information, are any
individuals identifiable by reference to those variables?
(It is possible, for example, with small sample sizes that information such as
gender might lead to the identification of a specific individual)
6. If the answer to question #5 is ‘yes’, is the personally identifiable information
relevant to the purpose(s) for the disclosure?
(For example, a dataset disclosed that provides demographic information
about government employees by department might need to include information
about gender in order to be useful in assessing the extent of efforts to increase the
representation of women in these positions).
344 T. Scassa and A. Conroy
Anonymization of Data
Data sets that contain personal information can be anonymized by removing all
personal information (El Emam 2013; Fraser and Willison 2009). Further, data sets
that contain personally identifiable information can also be manipulated to elimi-
nate or reduce the risk of reidentification (UK ICO 2012). There are a number of
available techniques. Aggregation, a technique typically used with statistical data,
displays data as totals, averages or in ranges. The presentation of data in this way
may be suitable for some purposes, although it may not be sufficiently fine-grained
for other purposes. Randomization involves the scrambling of direct and indirect
identifiers in the database (UK ICO 2012). The coding or pseudonymization of data
involves the replacing of unique identifiers with codes or pseudonyms. This is a
technique commonly used in the context of research involving personal health
information (Emam and Fineberg 2009; UK ICO 2012). Where the information at
issue is qualitative, personal and personally identifying information can be removed
through redaction (UK ICO 2012).
In cases where the datasets contains personally identifiable information – in
other words information that is not itself identifying of individuals, but could be
used in combination with other information to identify specific individuals – tech-
niques such as heuristics or analytics may be used (Emam and Fineberg 2009).
Heuristics makes use of threshold rules to assess the risk that an individual might be
uniquely identifiable through any combination of quasi-identifiers in the data set; or
that outside sources of information might be combined with the data to identify
specific individuals within the dataset. Some data may be suppressed if it is deter-
mined that the reidentification risk will be too high if it remains.
Although anonymization and deidentification techniques can be useful to protect
privacy in the release of government information, the risk of reidentification is
increased by the vast amount of other data that is already available as open govern-
ment data or that is in the hands of the private sector. In the big data environment,
reidentification risk is real (Ohm 2010; Daries 2014). The EU Article 29 Data
Protection Working Party (2013) observes that there are a number of reasons why
an individual or an organization might attempt to reidentify individuals whose dei-
dentified information is released within the open government context. These might
include reidentification for commercial or law enforcement purposes, or to reveal
personal information that may be newsworthy or relevant in an adversarial political
setting, or simply to satisfy individual curiosity. The UK Information Commissioner
(2014) warns that organizations should periodically review their anonymization
practices in order to ensure that they are keeping up with reidentification risk within
the big data environment.
Techniques used to anonymize data or to protect against reidentification typically
have an impact on the quality of the data and its fitness for some purposes (Cavoukian,
et al. 2014; Daries 2014). The decision regarding what technique or techniques to
employ and in what circumstances may depend upon an assessment of the risk of
reidentification, the degree of sensitivity of the information, the purposes to which
346 T. Scassa and A. Conroy
the data may be put, and the transparency value of the data. Where the transparency
value is high, disclosure of anonymized information may be warranted even if there
is a risk of reidentification. The argument for disclosure may be less compelling if
the reidentification risk is high and the transparency value of the information is rela-
tively low. Of course, it is not always obvious what the transparency value of data
may be, as it might be used in unprecedented combination with other data to pro-
duce unanticipated results. Some data sets have a more obvious value for use in
holding the government to account. For other data sets, it may be that a case for their
value will need to be made.
License Restrictions
reidentify individuals; and (iii) require that licensees notify the licensor of any
reidentification that occurs (UK ICO 2014).
In cases where the release of government information that raises significant pri-
vacy concerns is being contemplated, customized licences that contain additional
terms and conditions could be used to provide an additional layer of privacy protec-
tion. For example, Australia has a Restrictive Licence Template (AusGoal 2011) for
circumstances where it is necessary to protect personal or confidential information
in a government data set. This is not really an open licence since the terms and con-
ditions that may be imposed go beyond what would be considered acceptable in an
open licence. Nevertheless, as part of a suite of possible licence templates it gives
some flexibility to government and allows for the release of data that might not
otherwise be released, albeit under more restrictive terms.
The EU Article 29 Data Protection Working Party (2013) notes that in order to
comply with data protection laws and the EU Directive on the re-use of public sec-
tor information, public sector bodies may attach specific conditions and safeguards
to personal information when making it available online for reuse. The opinion
notes, however, that such terms should not place unnecessary restrictions on reuse;
the key concern is viewed as the need to ensure that personal information is not used
for a purpose inconsistent with that for which the information was collected (Article
29 Working Party 2013 at 3 and 26).
One area where licence terms that restrict the use of personal information con-
tained in government data may be particularly useful is where the information
released by a government institution contains directly identifying personal informa-
tion. This is typically the case, for example, in court or administrative tribunal deci-
sions that are published online. Principles regarding the transparency of judicial
proceedings often require that the names of the parties to proceedings and of the
witnesses be made public (unless there is a compelling reason to provide anonym-
ity). Court and tribunal decisions may also contain a variety of other personal infor-
mation. Where these decisions are published online, any licence permitting
reproduction and reuse of these materials could place specific restrictions on uses of
the personal information contained in the decisions.
The use of licences to protect personal information is far from a perfect solution.
The investigation and enforcement of breaches of licence terms may simply not be
practicable (Daries 2014; Article 29 Working Party 2013; UK ICO 2012). Since the
licensor is the government, it would have to be sufficiently motivated to take legal
action against a licensee who uses personal information in a manner contrary to the
terms of the licence. In addition, in some cases, it may simply not be possible to
establish the source of certain personal information that is being misused. In other
words, it may be impossible to trace it back to the government data set, as opposed
to some other source. It may even be difficult to tell whether or how government
data sets were used in reidentification processes. Licence restrictions, particularly in
otherwise open licences may also make productive reuse of the licenced data more
complicated, since the use of this data in combination with other datasets made
available under different license terms can create headaches for downstream licens-
ing of end-products or services (Mewhort 2012).
348 T. Scassa and A. Conroy
Technological barriers to reuse are generally not consistent with open data, since one
of the goals of open data is to encourage reuse of the data provided and not to create
obstacles to reuse (Borgesius et al. 2015). However, technological barriers may be
useful in some circumstances where the goal is to provide access to government infor-
mation for transparency purposes but there is a need to limit reuse in the interests of
privacy. An example of such circumstances is in the online publication of court and
tribunal decisions. While transparency values may require publication of these materi-
als without redaction, and might be best served by broad and open dissemination, such
decisions may contain a considerable amount of often sensitive personal information
(Austin and Pelletier 2005). In addition, the availability of this sensitive personal
information online may increase reidentification risks elsewhere as it may be used to
identify individuals within other anonymized government data sets.
Technological barriers can be as simple as using restricted proprietary formats
such as PDF for the release of information. Such formats can make information
more difficult and time-intensive to reuse. Nevertheless, those determined to reuse
the data will find ways to circumvent technological barriers (Thompson 2014).
Online applications already exist that make it possible for the average user to defeat
many such technological barriers and to manipulate data into machine-readable for-
mats. Thus, where there is a high degree of interest in reuse of the data, technologi-
cal barriers will provide only a very limited protection for the privacy interests at
issue. At the same time, they may impose an undesirable transparency cost.
Another form of technological barrier is the use of software to prevent the index-
ing of web pages in search engines (Austin and Pelletier 2005). For example, the
Canadian Legal Information Institute (CanLII), a website which provides free
online access to Canadian court and tribunal decisions, states in its privacy policy:
CanLII adheres to the principle of openness and transparency of legislative and judicial
processes, and recognizes their fundamental importance in democratic societies. In order to
minimize the negative impact of such transparency on the privacy of those involved in cases
leading to judicial decisions, CanLII does not permit its case law collections to be indexed
by external search engines. (CanLII 2016)
The policy also states that external search engines are prohibited from indexing the
text and style of cause of court decisions published by CanLII. A similar approach
is adopted by the Australian Legal Information Institute (AUSTLII) and by the
British and Irish Legal Information Institute (BAILII), which provides in its privacy
policy:
BAILII does not consent to the contents of these databases being indexed by other web
sites. BAILII attempts to prevent such indexing occurring by placing these database outside
the permitted scope of web ‘crawlers’, ‘robots’ or ‘spiders’ that adhere to the voluntary
Robot Exclusion Standard (BAILII Privacy Policy 2016)
The usefulness of this type of technological barrier is limited by the fact that the
Robot Exclusion Standard is voluntary and depends upon the willingness of search
engines to comply with it. Major search engines such as Google, Bing and Yahoo
currently do so, but others may not.
The Privacy/Transparency Balance in Open Government 349
CanLII also places limits on bulk downloads of court decisions. This limitation
is both technological and contained in the site’s licence terms. Although there may
also be other reasons to prevent bulk downloads, doing so reinforces the other tech-
nological privacy protections since it prevents others from downloading court deci-
sions in bulk and publishing them elsewhere online in fully indexable and searchable
formats. The importance of this issue was highlighted when a complaint was made
to the Office of the Privacy Commissioner of Canada (OPCC) after an individual
found that detailed personal information from her past appeared that had been
reported in a Canadian court decision appeared in a Google search for her name.
The court decision had been one of a very large number of decisions that a Romanian-
based company named Globe24 had downloaded in bulk from different sources
including official court websites in Canada (Dobby 2015; PIPEDA 2015).
While the Globe24 case reveals the importance of the technological measures
adopted by CanLII and others to protect privacy of personal information originating
from government sources, it also reveals their shortcomings. These restrictions did
not prevent the information in question from becoming repurposed by another site
in a way that adversely impacted personal privacy. This strongly indicates that gov-
ernment institutions would be unwise to rely solely upon either license restrictions
or technological measures to protect the privacy of personal information.
In addition, the use of technological barriers to access government information
has been criticized. Some argue that anyone should be able to build a fully search-
able database of court decisions (Cameron-Huff 2014). Such a perspective denies
the obligation of governments to protect citizen privacy and favours free enterprise
over a balancing of interests.
Conclusion
In this paper, we have outlined strategies that can be used by those responsible for
the release of information through proactive disclosure or as open data. These strat-
egies are intended to assist in balancing transparency goals with personal privacy in
the release of government information. Achieving this balance is made more chal-
lenging by the fact that the presence of personally identifiable information may be
difficult to gauge, particularly where individuals may be reidentified in anonymized
data sets by combining that data with other available data from indeterminate
sources. The rapidly evolving big data environment is one in which massive quanti-
ties of data are already available and more is constantly being generated or released.
This makes it difficult to anticipate what data might be used in order to achieve
reidentification (Conroy and Scassa 2015).
We identify several strategies that can be adopted in order to balance privacy
with transparency in open government. The first two strategies are addressed to
general practices. Data minimization can reduce the amount of personal informa-
tion both by limiting collection only to that which is specifically necessary, and by
limiting retention only for as long as is necessary. Data minimization principles can
350 T. Scassa and A. Conroy
Acknowledgments The authors gratefully acknowledge the support of the Social Sciences and
Humanities Research Council of Canada for the Geothink project of which this research forms a part.
References
Article 29 Data Protection Working Party (2013) Opinion 06/2013 on Open Data and Public
Sector Information (‘PSI’) Reuse. Retrieved from: http://ec.europa.eu/justice/data-protection/
article-29/documentation/opinion-recommendation/files/2013/wp207_en.pdf
AusGoal (2011) Restrictive Licence Template. Retrieved from: http://www.ausgoal.gov.au/
restrictive-licence-template
The Privacy/Transparency Balance in Open Government 351
Austin LM, Pelletier F (2005) Synthesis of the comments on JTAC’s discussion paper on open
courts, electronic access to court records, and privacy (Prepared on Behalf of the Judges
Technology Advisory Committee for the Canadian Judicial Council, 2005). Retrieved from:
https://www.cjc-ccm.gc.ca/cmslib/general/news_pub_techissues_Synthesis_2005_en.pdf
BAILII Privacy Policy (2016) Retrieved from: http://www.bailii.org/bailii/privacy.html
Bennett CJ, Raab C (2006) The governance of privacy: policy instruments in global perspective.
MIT Press, Cambridge, MA
Berzins C (2008) Personal information in the adjudicative decisions of administrative tribunals: an
argument for limits. Advocates’ Q 34(3):261–284
Borgesius FZ, Gray J, van Eeechoud M (2015) Open data, privacy, and fair information principles:
towards a balancing framework. Berkeley Technol Law J 30(3):2073–2130
British Columbia (2016) Open information. Retrieved from: http://www2.gov.bc.ca/gov/content/
governments/about-the-bc-government/open-government/open-information
Cameron-Huff A (2014) Why Google can’t build a case law search engine in Ontario. Retrieved
from: http://www.cameronhuff.com/blog/ontario-case-law-private/index.html
Candeub A (2013) Transparency in the administrative state. Houston Law Rev 51(2):385–416
CanLII Privacy Policy (2016) Retrieved from: https://www.canlii.org/en/info/privacy.html
Cavoukian A (2009) Privacy and government 2.0: the implications of an open world. Ontario,
Information and Privacy Commissioner. Retrieved from: https://www.ipc.on.ca/images/
Resources/priv-gov-2.0.pdf
Cavoukian A, Dix A, El Emam K (2014) The unintended consequences of privacy paternalism.
Information and Privacy Commissioner, Toronto. Retrieved from: https://www.ipc.on.ca/
images/Resources/pbd-privacy_paternalism.pdf
Conroy A, Scassa T (2015) Promoting transparency while protecting privacy in open government
in Canada. Alberta Law Rev 53(1):175–206
Daries JP (2014) Privacy, anonymity, and big data in the social sciences. ACM Queue 12(7):1–12.
Retrieved from: http://queue.acm.org/detail.cfm?id=2661641
Davies TG (2014) Open data policies and practice: an international comparison (September 5,
2014). Available at SSRN: https://ssrn.com/abstract=2492520 or http://dx.doi.org/10.2139/
ssrn.2492520
Dobby C (2015) Canadians upset with Romanian Website that exposes court
case details. The Globe and Mail (4 January, 2015). Retrieved from: http://
w w w. t h eg l o b e a n d m a i l . c o m / r e p o r t - o n - bu s i n e s s / i n d u s t r y - n ew s / t h e - l aw - p a g e /
canadians-upset-over-romanian-website-that-exposes-court-case-details/article22284367/
El Emam K (2013) Guide to the de-identification of personal health information. CRC Press, Boca
Raton
El Emam K, Fineberg A (2009) An overview of techniques for de-identifying personal health
information. CHEO Research Institute, Ottawa
Fraser R, Willison D (2009) Tools for de-identification of personal health informa-
tion. Report Prepared for the Pan-Canadian Health Information Privacy (HIP)
Group. Retrieved from: https://www.infoway-inforoute.ca/en/component/edocman/
supporting-documents/500-tools-for-de-identification-of-personal-health-information
Government of Canada (2002) Privacy impact assessment policy. Canada, Treasury Board
Secretariat, Ottawa. Retrieved from: http://www.tbs-sct.gc.ca/pol/doc-eng.aspx?id=12450
Greenleaf G (2011) Free access to legal information, LIIs, and the free access to law movement. In:
Danner R, Winterton J (eds) IALL international handbook of legal information management.
Ashgate, Aldershot/Burlington VT, pp 201–228
Janssen K (2012) Open government data and the right to information: opportunities and obstacles.
J Community Informatics 8:2. online: http://ci-journal.net/index.php/ciej/article/view/952/954
Kitchin R (2014) The data revolution: big data, open data, data infrastructures & their conse-
quences. SAGE Publications Ltd., London
Mayer-Schönberger V, Cukier K (2014) Big data. Mariner Books, New York
McLachlin B (2003) Courts, transparency and public confidence – to the better administration of
justice. Deakin Law Rev 8:1–12
352 T. Scassa and A. Conroy
Mewhort K (2012) Creative commons licenses: options for Canadian open data providers. (2012)
Report for the Canadian Internet Policy and Public Interest Clinic. Retrieved from: https://
cippic.ca/sites/default/files/Creative%20Commons%20Licenses%20-%20Options%20for%20
Canadian%20Open%20Data%20Providers.pdf
Meyerhoff-Nielsen M, Krimmer R (2015) Reuse of data for personal and proactive service:
an opportunity not yet utilised. Retrieved from: https://www.researchgate.net/publica-
tion/278023437_Reuse_of_Data_for_Personal_and_Proactive_Service_An_Opportunity_
Not_Yet_Utilised
Office of the Australian Information Commissioner (2014) Guide to undertaking privacy impact
assessments. Retrieved from: https://www.oaic.gov.au/resources/agencies-and-organisations/
guides/guide-to-undertaking-privacy-impact-assessments.pdf
Office of the Privacy Commissioner of Canada (2014) Personal information retention and dis-
posal: principles and best practices. Retrieved from: https://www.priv.gc.ca/information/pub/
gd_rd_201406_e.asp
Ohm P (2010) Broken promises of privacy: responding to the surprising failure of anonymization.
UCLA Law Rev 57:1701–1777
Open Government Partnership (2011) Open Government Declaration. Retrieved from: https://
www.opengovpartnership.org/open-governmentdeclaration
PIPEDA Report of Findings No. 2015–002, [2015] C.P.C.S.F No. 2, [2015] S.C.C.P.V.P.C. No 2
Queensland, Office of the Information Commissioner (2013) Proactive disclosure and publi-
cation schemes. Retrieved from: https://www.oic.qld.gov.au/guidelines/for-government/
access-and-amendment/proactive-disclosure/proactive-disclosure-and-publication-schemes
Roy J (2014) Open data and open governance in Canada: a critical examination of new opportuni-
ties and old tensions. Future Internet 6(3):414–432. doi:10.3390/fi6030414
Scassa T (2014) Privacy and open government. Future Internet 6:397–413. doi:10.3390/fi6020397.
Available at: http://www.mdpi.com/1999-5903/6/2/397
Scassa T, Conroy A (2016) Strategies for protecting privacy in open data and proactive disclosure.
Canadian J Law Technol 14:215–262
Schauer F (2011) Transparency in three dimensions. Univ Ill Law Rev 4:1339–1358
Schneier B (2015) Data and goliath: the hidden battles to collect your data and control your world.
W.W. Norton & Co., New York
Sherman J (2013) Court information management: policy framework to accommodate the digital
environment. Canadian Judicial Council, Ottawa. Retrieved from: http://www.cjc-ccm.gc.ca/
cmslib/general/AJC/Policy%20Framework%20to%20Accommodate%20the%20Digital%20
Environment%202013-03.pdf
Solove DJ (2004) The digital person: technology and privacy in the information age. New York
University Press, New York
Thompson SA (2014) Unlocking Ontario public sector salary data. Media Canadian Assoc
Journalists 16(2):27
United Kingdom, Information Commissioner’s Office (UK ICO) (2012) Anonymisation: manag-
ing data risk code of practice. ICO, Cheshire. Retrieved from: https://ico.org.uk/media/for-
organisations/documents/1061/anonymisation-code.pdf
United Kingdom, Information Commissioner’s Office (UK ICO) (2014) Privacy Impact Assessment
Code of Practice. Retrieved from: https://ico.org.uk/media/for-organisations/documents/1595/
pia-code-of-practice.pdf
United Kingdom, Information Commissioner’s Office (UK ICO) (2016) Retaining Personal Data
(Principle 5). Retrieved from: https://ico.org.uk/for-organisations/guide-to-data-protection/
principle-5-retention/
Veljkovic N, Bogdanovic-Dinic S, Stoimenov L (2014) Benchmarking open government: an open
data perspective. Gov Inf Q 31(2):278–290
Winn PA (2004) Online court records: balancing judicial accountability and privacy in an age of
electronic information. Symposium – Technology, Values, and the Justice System. Washington
Law Rev 79:307–330
The Privacy/Transparency Balance in Open Government 353
Dr. Teresa Scassa is the Canada Research Chair in Information Law at the University of Ottawa,
where she is also a professor at the Faculty of Law. She is the author or co-author of several books
on intellectual property and technology law subjects. She is a past member of the External Advisory
Committee of the Office of the Privacy Commissioner of Canada, and of the Canadian Government
Advisory Committee on Open Government. She is a member of the GEOTHINK research partner-
ship, and has written widely in the areas of intellectual property law, law and technology, privacy,
and open government.
Dr. Amy Conroy recently completed her doctorate in the Faculty of Law at the University of
Ottawa. She is a part-time professor at Carleton University. She has co-authored several papers in
the area of privacy law with Teresa Scassa. Her doctoral dissertation, entitled “E-racing the Genetic
Family Tree: A Critical Race Analysis of the Impact of Familial DNA Searching on Canada’s
Aboriginal Peoples,” argues against using familial searching in Canada’s national DNA data bank
based on the discriminatory impact this would have on Aboriginal peoples in Canada.
Index
A B
Accenture five-stage model, 80 Bangladesh, 268
Administrative tribunals, 335 BBC Janala, 268, 269
Adoption of Blockchain, 284 BBC World Service Trust, 268
Advanced health care in India, 253 Benefit payment, blockchain, 294
Advanced process model repository Berrypicking, 70, 71
(APROMORE), 116, 117 Big data analytics, 336
Advantages of linked data, 176 Big data enabled policy cycle, 38, 51–53
African companies, ICT, 256 Big open linked data (BOLD), 19
Against method, 40 Bitcoin protocol, 285
Aggregation technique, 345 Blockchain applications, 284, 295
Agile Software Development process, 210 Blockchain-as-a-service, 295
Akomo Ntoso model, 30 Blockchain-based financial innovation, 293
Alhomod four-stage model, 88 Blockchain technology, 21, 22, 283,
Almazan & Gil-Garcia’s six-stage model, 85 285–287, 291
Alpha version of TET, 209 background, 284–286
Anonymization of data, 345 goal, 285
Anonymization techniques, 339 implementations, 285
Application programmable interfaces innovation, 285
(APIs), 179 network security, 285
Applications of waternomics project, 185–192 popularity, 286
goal-oriented accessing water, 191, 192 programmable aspects, 286
manager dashboard, 187, 188 cases, 287–295
observations control panel, 190 e-voting, e-resident shareholders, 291
public display, 186, 187 migration of government data, 287, 291
water retention time observer, 188–190 public notary to e-residents, 291
wearable info-centre, 190, 191 digital 5 countries as innovators, 286, 287
Aristotle University of Thessaloniki discussion, 295, 296
(AUTH), 181 government entities, 284
Artificial intelligence (AI), 19, 20, 56 in private sector, 284
Augmented reality (AR), 20 related initiatives across D5 countries,
Australian Legal Information Institute 288–290
(AUSTLII), 348 Bottom-up approach, 200
Australian National Audit Office’s (ANAO), 72 British Columbia Open Information
Awareness of water, 185 website, 334
British and Irish Legal Information Institute CKAN open data platform, 209
(BAILII), 348 CoinStack platform, 293
Building management system (BMS), 183 Coláiste na Coiribe (CnaC) School, 182
Business as usual procedures, 272 Collaboration, 197, 300, 304, 315, 323,
Business model, 220–223, 251 324, 326
components of, 222, 223 Collaborative economy, 18
existing, 223–225 Collaborative public management, 66
Hamel business model, 223 Collective intelligence (CI), 307, 309
Osterwalder and Pigneur, 221, 222 analysis, 201
Shafer, Smith and Linder, 222 scenario-based approach, 304
Business process modelling, 112, 113, scenario-based design approach, 305–307
120–129 workflow for, 202
configurable models, municipalities, Color-coded visualization, 193
125–129 Combinational theory, 48
compression rate and time evaluation, Commonwealth Bank of Australia (CBA), 292
127–129 Communication tools, ICT, 263
tool support, 126, 127 Community model, 213
reference, 119–125 Comprehensive Knowledge Archive Network
configurable models, 123–125 (CKAN) Platform, 208
hierarchical reference process models, Computer-assisted instruction (CAI), 246
121–123 Computer-mediated transparency, 306
placeholders refinement, 120, 121 Conceptualisations of transparency, 302
summary and discussion, 125 Conceptualising transparency, 319
repository Conceptualization, ODBMs, 224–228
APROMORE, 116–117 Consumption of water, 185
BP-suite, 115 Contextual water management, 175
process variant repository, 114 Contextualism, 301–304
semantic business process repositories, Contextualist approach, 319
115, 116 Controlled natural language (CNL)
summary and discussion, 117–119 approach, 32
reuse in public administration, 111 Coursera, MOOC platform, 251
methodology, 112, 113 Cross-governmental cooperation, 97
Crown Legislation Markup Language
(CLML), 30
C Cryptography, 292
Cambodia malaria information system Cyber laws, 276
(MIS), 261 Cybersecurity, 292
Canada’s Privacy Act (s. 4), 340
Canadian Legal Information Institute
(CanLII), 348 D
Capability maturity models (CMM), 67 Danish tax system, 5
Chandler & Emanuel four-stage model, 77 Data, 221
Cisco three-stage model, 84 integration, 178
Citizens, 199, 300, 305, 311, 317, 319, 320, minimization, 340, 341
322, 323, 340 retention policies, 340
engagement, 322, 323 sets, 345
feedback, 259 DatalEt-Ecosystem Provider (DEEP), 206, 208
improving engagement of, 199 Datalets, 207, 208
science, 42–45, 56 Dataspace support services, 178
Scientist, 53–56 Davison four-element model, 82
Citizen Weather Observer Program Decentralized transactions, 285
(CWOP), 55 Declarative Rule-based Agent Modelling
City budget management, 211 Software (DRAMS), 31
Civic crafting in urban planning, 53 Deidentification techniques, 342
Index 357
F I
Fast moving consumer goods (FMCGs), 271 Idea writing, 201
Fintech, 292 Ideawriting method, 308
Footprint, water, 176 India, advanced healthcare, 253
Forecasting systems, 185 Individual privacy, 338
Formism, 302 Inferability, 303
Formist approach, 303 Information and Communication Technology
Four waves of e-government evolution, 6 (ICT), 1–9, 12, 13, 15, 17, 18, 23,
Fourth industrial revolution, 18 63–69, 77, 82, 91, 93, 97, 101–103,
Free and open source software (FOSS), 207 241, 245–248, 257–276
Frequency analysis, 311 access by development status 2015, 249
for capacity building, 242
laws relating to, 273
G lessons learnt, 276–280
Gartner Group four state model, 72 meeting social needs of target populations
Genetically modified organisms challenges to be addressed, 265–267
(GMOs), 21 technology innovations in public
Global e-learning market, 248 service delivery, 267–272
Global functional digital literacy, 246 policy and strategic framework
Global innovation index, 293 challenges to be addressed, 272–273
Global open government partnership, 2 technology innovations in public
Global Opportunities Network (2016), 243 service delivery, 273–276
Goal-oriented accessing water, 191, 192 public services using innovative
Google Earth, 261 challenges to be addressed, 245–248
Governance, 300, 323, 324, 326 role in public service delivery, 242
Government information, 334, 336, 337 solutions, 242
Government-led WPM, Ethiopia, 255 strengthening governance
3G network, population coverage, 267 challenges to be addressed, 257–259
Graphical user interface (GUI), 206 challenges to be addressed, 257
Grey Economy initiative, 262 technology innovations in public
Groningen workshop, 201 service delivery, 259–264
Guide to undertaking privacy impact technologies, 243
assessments (2014), 343 technology innovations in public service
delivery, 248–257
Information holders, 304
H Information integration projects, 176, 177
Hamel business model, 223 Information needs, transparency, 314, 315
Hammurabi’s code, 27 Information systems (IS), 213
Heeks’ Manchester eGov-MM, 89 Information technologies, 203
Heuristics, 345 Innovations in public service delivery, 242, 244
Hiller & Bélanger’s five-stage maturity technology, 248–257, 259–264
models, 74 Innovators, digital 5 countries, 286, 287
Hodgkingson five-stage model, 77 Integrated eHealth System, 252
Holistic multidisciplinary approach, 215 Inter-governmental communication, 341, 342
Household surveys, ICT, 266 Internet, 198, 245, 250, 285
Howard three-stage maturity model, 74 usage between males and females, 265
Human computation service, 179 individuals using by income group, 266
Human-computer interaction Internet of Things (IoT), 19, 243
techniques, 192 Intra-governmental communication, 341, 342
Human Development Index (HDI), 243 Ipaidabribe.com, 263
Hydro-meteorological information, 192 Irekia open-government portal, 264
Hyper text transfer protocol IS management, 66, 67
(HTTP), 177 Israel, 286, 292
HyperLogLog-Algorithm, 50 Issue based information systems (IBIS), 31
Index 359
L O
Lambda architecture realization, 183 ODDC framework, 156, 157, 166, 167
Layne & Lee’s maturity model, 74 ODVC framework, 155
Lee & Kwak five-stage model, 88 Office of the Privacy Commissioner of Canada
Lee’s framework, 97 (OPCC), 349
LegalRuleML, 30 OGD maturity model, 100
License restrictions, 346, 347 O-Government, 10–12
Linate Airport, 180 framework and setting, 9–13
Lingua scientia, 38 open data, 10
Linked data, 174, 176, 177 open process, 12
Linked Real-time Dataspace for Waternomics open service, 11
project, 177–179 and global challenges, 3, 4
and market changes, 1–3
societal level perspective and new
M framework, 4–7
MajiVoice software, 260 Ontorion Controlled Natural Language
Malaria consortium, 261 (OCNL), 32
Malaria information system (MIS), 261 Open data (OD), 10, 198, 199, 203, 219–221,
Malaysian National Telecommunications, 275 335, 337
Massive open online courses (MOOCs), 251 Open data Barometer, 305
Maturity models, 67, 69, 71, 72, 74, 80, 82 Open data business models (ODBMs),
Modernization of public services, Portugal, 220–225, 233–235
274, 275 conceptualization, 224–228
Mondey project in Germany, 269 literature review
Moon’s five-stage model, 78 business model, 221–225
mPedigree service, 252 OD, 220, 221
Multi-disciplinary approach, 215 model elaboration, 228, 229
Municipality of Thermi, 181 partners and value disciplines, 229–236
Mwater, 254 categories, 234
patterns, 233, 234
value disciplines, 234, 235
N and value proposition categories, 236
Nairobi Water, 260 Open data-driven organizations (ODDOs),
Narayana Health hospital, 253, 254 135, 136
National Association of State Chief dynamic capabilities, 147–154
Information Officers’ (NASCIO), theoretical background, 136–154
84, 91, 92, 97, 99, 100, 104 value capabilities, 141–145, 147
National Telecommunications Policy Open data dynamic capabilities (ODDCs),
(NTP), 275 136, 161–163, 165
360 Index