A Guide To Data Governance
A Guide To Data Governance
Contents
What is Data Governance?................................................................................................ 3
Why Do We Need It?.............................................................................................................................................................................. 4
Metadata Lineage..................................................................................................................................................................................11
Conclusions.......................................................................................................................25
Data Governance is
In many companies today, data governance has become increasingly important but what
about ensuring that exactly is it? What does data governance mean?
the data being used
in your core business
operations, reports
There are several definitions of data governance available from various sources. These
and analyses is include the following quotations:
protected and can be
trusted
“Data Governance (DG) is defined as the exercise of authority and control (planning,
monitoring, and enforcement) over the management of data assets.”
Source: DAMA Data Management Body of Knowledge V2 (DMBOK2)
Looking at these, data governance is about ensuring that the data being used in your
core business operations, reports and analyses is discoverable, accurately defined, and is
totally trusted and can be protected. Additionally, data governance has become increasingly
important to a business because according to a prediction from IDC, digital data is expected
to grow to approximately 175 zettabytes by 2025.1
1 https://www.idc.com/getdoc.jsp?containerId=AP45214519
Data governance is
In many companies today, the expectation in the board room is that data and artificial
needed to improve intelligence (AI) will drive competitive advantage. Not surprisingly therefore, executives are
data quality so that
data is trusted. eager to sponsor AI initiatives in their determination to become data driven. However, for AI
Data quality if of
to become effective, the data it is using must be trusted. Otherwise decision accuracy may
utmost importance be compromised, decisions may be delayed, or actions missed which impacts on the bottom
because when
companies work with line. Companies do not want ‘garbage in, garbage out’. It might seem relatively straight
inferior data, this forward to fix data quality until you look at the impact that digital transformation has had on
negatively impacts
their downstream data in the last few years.
insights, analyses, and
recommendations. For most companies, the introduction of digital transformation has resulted in a more
Data quality must
entail the data is complex operating environment in comparison to just having a single data centre. Today,
complete, unique, most companies have created an operating environment that spans the edge, multiple
valid, timely, accurate,
and consistent. clouds and the data centre. Surveys over the last couple of years have shown this with one2
Data quality problems last year showing 81% of companies surveyed had systems running in multiple public clouds
can impact on and one or more private / dedicated clouds. That typically translates to meaning that both
business operations
causing process operational and analytical systems are running in the cloud and the data centre. Examples
errors, process of operational transaction processing systems running in the cloud include Microsoft
delays, unplanned
operational costs and Dynamics, Workday, Salesforce, ServiceNow and Marketo. Analytical systems running in
inaccurate decisions.
the cloud could include data warehouses, graph databases, data lakes being used by data
Data needs to be scientists and real-time IoT streaming analytic applications. The result is that companies are
governed across a
distributed computing now dealing with a hybrid environment with data in multiple different data stores that are
environment
scattered across all of this landscape similar to that shown in Figure 1.
1
https://blog.syncsort.com/2019/01/data-quality/data-integrity-vs-data-quality-different/
2
IDC’s Multi-cloud Management Survey 2019
Azure &
other cloud providers
Data center
Figure 1
This includes data stored in edge databases relational DBMSs, NoSQL DBMSs, Files, Cloud
storage, Hadoop systems and scalable messaging queuing systems (e.g. Kafka).
Digital transformation
The other major impact of digital transformation is that there are a lot of new data sources
has resulted in a lot of that business now wants to analyse beyond the traditional master data and transaction data
new data sources that
businesses want to found in data warehouses. This includes machine generated data such as clickstream data in
analyse web server log files, human generated data from social networks, inbound email, and open
government data. Also, unstructured content in various documents is in multiple locations.
Data is increasingly
With data increasingly spreading out across a hybrid multi-cloud, distributed data landscape,
spreading out across it is not surprising that people are struggling to know where it is in order to govern it. Yet,
data stores in the
data centre, multiple the business impact from ungoverned data can be considerable. Poor data quality impacts
clouds and at the business operations because data errors cause process errors and delays. Poor quality data
edge which makes
it harder to find and also impacts business decision making and the ability to remain compliant. Data governance
harder to govern needs to therefore include data discovery, data quality, policy creation, data sharing, and
metadata to help track and govern data activity.
Data needs to be
The other major driver for data governance is data protection. This is needed primarily to
protected to prevent remain compliant on data privacy with regulatory legislation such as the European Union
data breaches and
enables you to General Data Protection Regulation (GDPR) or the California Consumer Privacy Act (CCPA)
remain compliant and to prevent data breaches. Data privacy as well as the growing number of data breaches
on data privacy with
regulations and has made data protection a top priority in the boardroom. These breaches highlight the risk
legislation to sensitive data such as personally identifiable customer data. The consequences of data
privacy violation and / or a data security breach are numerous and include:
Also, in most cases, publicly quoted companies must declare these breaches.
If they happen, customers are more likely to blame the company first rather than the hacker
and may boycott the company for several months or may never return.
Failure to comply with regulatory legislation like GDPR and CCPA on data privacy may
also result in several very significant financial penalties. No one wants any of this and so
governing data to avoid such risks is well worth doing.
Given this backdrop, what then are the requirements to govern data in a modern enterprise?
• Data item and data entity definition to create a common business vocabulary in a
business glossary
• Data item and data entity identification / discovery
• Data governance classification to govern data access security, data privacy and data
retention
A common business
The first of these is critical. It is a common business vocabulary. The reason this is needed is
vocabulary is a to clear up ambiguity in the meaning of data caused by ambiguous data names in different
set of commonly
defined data names data stores, in reports and in data made available through APIs. For example, a business
and data definitions analyst may produce a report that shows “Total Sales”. Another report from a different
documented in a
business glossary analyst may show “Sales” and another uses “Gross Revenue”. Are they the same? Are they
within a data catalog different? Does Total Sales include sales tax or not? What about Sales? Does that include
It’s purpose is to sales tax or not? These are everyday ambiguities that need to be avoided. The answer lies
ensure that data is
consistently named in creating a common business vocabulary of common data names and data definitions for
and commonly data entities and their attributes clearly defined in a business glossary.
understood especially
when it is shared
Data needs to be
The other challenge is that data is being collected and stored in multiple places across the
governed across enterprise. This may include data collected and stored in different geographies and different
multiple geographies
that may be subjected legal jurisdictions. As a result, different legislation may apply to governing the same data
to different legislation in different jurisdictions. You therefore need to discover what data exists across the hybrid
multi-cloud distributed data landscape (including geographic location) to be able to:
1. Understand what data attributes, data entities and data relationships exist across the
distributed data landscape
2. Classify the data to know how to govern it
3. Define policies to specify how data should be governed for each type of governance
classification
4. Enforce data quality, data access security, data privacy and lifecycle management
policies across the distributed data landscape
Figure 2
Automating the data confidentiality and data retention classification process using the
classes defined in each scheme is needed to consistently label data across the distributed
data landscape to enable it to be consistently and correctly governed. Rules and policies
would then need to be defined for each class in the classification scheme to specify how to
govern data according to its classification.
• Who sets success metrics and monitors how well the data governance program is
working?
• Who are the data owners?
• Who defines and maintains a business glossary?
• Who creates and maintains policies on access security?
• Who is protecting PII data privacy for compliance with GDPR and CCPA?
• Who is looking after the quality of product data across all brochures and partner
websites?
• Who ensures customer data is consistent across all systems?
• Who is policing external subscription data usage Vs the license?
• Who is policing privileged users like DBAs and data scientists?
Is it a C-level executive? Is it a department head? Is it the head of governance, risk and
compliance? What about the legal department? Or is it IT’s responsibility? Roles and
responsibilities are needed to avoid confusion and to set the foundation upon which a data
culture can materialize.
Data governance
In addition to roles and responsibilities, processes are needed. For example, to:
processes are needed
to govern how data • Govern the definition and maintenance of a common business vocabulary
is defined, discovered
and classified • Discover and identify what data you have, what it means and where it is
• Classify data to know how to govern it
Data governance
• Govern the definition and maintenance of data access security policies
processes are also
needed to ensure
• Govern the definition and maintenance of data privacy policies
governance policies
and master data
• Detect data quality problems and remediate them
are created and
maintained correctly
• Apply policies to ensure action is taken for compliance
• Govern maintenance of master data
• Data integrity
• Data ingestion
Manufacturing systems
CRM = Customer Relationship Management
SCM = Supplier Chain Management
ERP = Enterprise Resource Planning
Figure 3
The same applies to reference data such as code sets and financial markets data. In this case
standardization and synchronization of code sets is known as reference data management
which is also a requirement.
Metadata Lineage
Finally, there is a requirement for metadata lineage (Figure 4). This is the need to provide
an audit trail to know where data originated and how it has been transformed on route to a
report or a data store. In addition, metadata is needed to trace who or what is maintaining
data (e.g. master data) including when and where this occurs.
Metadata lineage
provides an audit Tracks all activity
trail on where data performed on the
data by the user
originated, how it
was transformed on
its way to the point
of use and how it has
been maintained
Figure 4
Data
Data lifecycle
Create Protect Store Use Maintain Archive Destroy
Figure 5
Data
Data lifecycle
the program. These objectives should be aligned with strategic business objectives to show
contribution to common goals. The strategy also includes success metrics (KPIs) and targets
to be reached to monitor the progress. There are two types of metrics to be considered. The
A data governance first is risk management and compliance metrics designed to measure improvements in data
strategy should quality, security, privacy and retention. The second type is value creation metrics. These help
specify objectives,
success metrics and to monitor how data governance is contributing to improving business value through the
targets to be achieved creation and use of trusted data. Business value in this case could mean:
The data governance strategy should then include the projects / initiatives that are needed
to achieve the business objectives, meet the targets set and deliver the ROI identified in
business cases. In addition, it should include the budget allocated to these projects, who is
leading the data governance program and who is accountable for achieving them. It may
also include some data principles. Two examples here are that data should be treated as an
asset and that data is the property of the company and should be shared.
Data
Data lifecycle
Data
Data lifecycle
Role Responsibility
Executive sponsor Senior business stakeholder with authority and budget who is accountable for
(e.g. CFO / CIO) ensuring data governance is established
Data Governance The person with overall accountability and responsibility for implementing the
program leader (e.g. CDO data governance program.
or appointed lead)
Data Governance Control Includes data governance lead and data owners. Sets success metrics, owns
Board the data governance roadmap, selects working groups, holds the budget for
the data governance program, arbitrates when conflicts occur on priorities and
definitions of cross functional data
Data Governance Plan and progress data definition and improvement of a specific data domain
Working Group (E.g. Customer or Supplier), update Data Governance Control Board on progress,
manage stewardship across the enterprise for a specific domain
Data owner Senior business stakeholder with authority and budget who is accountable for
overseeing the quality and protection of a specific data subject area or data
entity across the enterprise and make decisions on who has the right to access
and maintain that data and on how it is used
Business data steward Business professional responsible for overseeing the quality and protection of
a data subject area or data entity. They are typically experts in the data domain
and work in a team with other data stewards across the enterprise to monitor
and make decisions to ensure data quality is maintained
Data Protection Officer Senior business stakeholder with authority and budget who is accountable
(DPO) for the protection of personal data specific to compliance legislation in all
jurisdictions that the company operates
Data security team Responsible and accountable for data access security and data privacy policy
enforcement
Data Publishing Manager Responsible and accountable for quality assurance checking and publishing of
newly created trusted data assets in a data marketplace for consumers to find
and use
Roles, responsibilities The objective is to organise in a way that allows you to take a ‘divide and conquer’ approach
and organisational to governing data throughout its lifecycle across a hybrid computing environment. One way
structure set out how
to organise people to of doing this is to have multiple working groups reporting into a Data Governance Control
successfully govern Board (Figure 6) with each working group responsible for a particular data domain / data
data
entity (e.g. Customer) or a data subject area that consists of multiple data entities.
Executive
Figure 6
People
Data governance vision and strategy
Data discovery processes • A data and data entity discovery, mapping and cataloguing process
(to understand the data • A data profiling discovery process to determine the quality of data
landscape) • A sensitive data discovery and governance classification process
• A data maintenance discovery process for CRUD3 analysis (e.g. from log files)
to understand usage and maintenance of data (e.g. master data) across the
enterprise
Data governance definition • Create and maintain a common business vocabulary in a business glossary.
processes This involves defining data entities (including master data), data attributes
names, data integrity rules and valid formats
• Define reference data to standardise code sets across the enterprise
• Define data governance classifications schemes to label data to determine
how to govern it
• Define data governance policies and rules to govern data entity and
document lifecycles
• Define success metrics and threshold
Data governance policy • A process to automate application / enforcement of data governance policies
and rule enforcement and rules
processes • A process to manually apply and enforce policies and rules
• Event-driven, on-demand and timer-driven (batch) data governance
processes published as services that can be invoked to govern:
• Data ingestion - cataloguing, classification, owner assignment, and storing
• Data quality
• Data access security
• Data privacy
• Data usage e.g. including sharing and to ensure licensed data is only used
for approved purposes
• Data maintenance e.g. of master data
• Data retention
• Master data and reference data synchronisation
Monitoring processes • Monitor and audit data usage activity, data quality, data access security, data
privacy, data maintenance and data retention
• Monitor policy rule violation detection and resolution
3
CRUD = Create, Read, Update, Delete
Governance processes
The common business vocabulary should be defined in a business glossary within a data
set out how to catalog. Following on from the discussion on data governance working groups, each
discover, define,
enforce and monitor working group should take responsibility for defining a specific data entity or data subject
area (multiple related entities). Therefore, multiple data entities in the vocabulary, along with
the policies and rules, can be worked on in parallel (see Figure 7).
Business Glossary
Integration of the catalog business glossary with other technologies such as ETL tools,
data modelling tools, BI tools, DBMSs, master data management, data virtualisation tools,
software development tools etc.., is then needed to get consistent common data names into
all technologies.
A data concept A good practice to quick start the creation of a common business vocabulary is to create
model is a good top a data concept model. This top down approach gets you started because it identifies data
down approach to
identify data entities concepts that can be used as data entities in a common business vocabulary. It is then
to get your common possible to assign a different data governance working group to each data concept (entity)
business vocabulary
started or group of related data concepts (subject area). In this way different working groups
Different data are assigned to govern different data entities across the landscape. During the build of
governance working a common business vocabulary, it should be possible to use data catalog software to
groups can work on
different data entities automatically discover what data is out there across multiple data stores to help identify all
using automatic the attributes associated with specific data entities. This is a bottom-up approach. By using a
data discovery in the
catalog to identify top down approach of a data concept model to get you started and a bottom up automated
attributes to add
to the common
data discovery approach to identify the attributes of a data entity, it should be possible
business vocabulary for multiple working groups to incrementally build up a common business vocabulary
to describe those data
entities reasonably quickly.
Using a data catalog for automated data discovery enables the mapping of disparate data
to a common vocabulary to understand where the data for each particular data entity in the
business glossary is actually located across the enterprise.
Data
Data lifecycle
Looking back to the data governance classification schemes in Figure 2, the data
confidentiality classification scheme has five classification levels. These are Public, Internal
Use Only, Confidential, Sensitive PII and Restricted.
Once policies and rules are defined in a data catalog for each class in a data governance
classification scheme, they can be passed to other technologies from a data catalog (via
APIs) for them to enforce. Alternatively, a common data management platform (data fabric)
that can connect to multiple data stores could potentially enforce them.
It should then be possible to monitor data quality, privacy, access security, usage,
maintenance and retention of specific data entities through their lifecycle.
People
Data governance vision and strategy
• A business glossary
• Automated data discovery, profiling, tagging, cataloguing and mapping to a glossary
• Automated sensitive data detection and governance classification
• Interoperability with other catalogs, tools and applications to share metadata via
APIs and open standards
Enterprise data
In the context of technology needed for end-to-end data governance, Microsoft provides its
catalog, and Azure own technologies and also partner technologies on Azure.
Data Factory are key
technologies to help
you govern data
Microsoft provides the following technology components to assist you in governing data:
The Microsoft Common Data Model (CDM) is an open, pre-built set of common business
entities and activities used across a business that can be used to shortcut the creation of
your common business vocabulary.
Figure 8
Visualize & report Ingest, prepare, Train & predict Model & serve
transform & enrich
Ingest
Figure 10
Microsoft’s strategic
Microsoft Azure Data Factory is a fully managed, pay-as-you-use, hybrid data integration
pay-as-you-use service for highly scalable ETL and ELT processing. It uses Spark to process and analyse data
data management
platform (data fabric) in parallel and in memory to maximise throughput.
for cleaning and
integrating data is It supports over 80 connectors to external data sources and databases and has templates
Azure Data Factory
(ADF) for common data integration tasks. A visual front-end browser-based GUI enables non-
ADF allows you to
programmers to create and run process pipelines to ingest, transform and load data, while
build scalable data more experienced programmers have the option to incorporate custom code if required
integration pipelines
code free (e.g. Python programs).
ADF enables
collaborative
development between
business and IT
professionals in the
creation of reusable
trusted data assets
Figure 11
ADF can be used to implement and manage a hybrid environment, which includes
connectivity to on-premise, cloud, edge streaming and SaaS data (e.g. from applications
such as Salesforce), in a secure and consistent way.
ADF wrangling data flows enables business users to make use of the platform to visually
discover, explore and prepare data at scale without writing code. This easy to use ADF
capability is similar to Microsoft Excel Power Query or Microsoft Power BI Dataflows where
business users use a spreadsheet style user interface with drop-down transforms to prepare
and integrate data.
Data in disparate
Data catalog
registered data
Common vocabulary, data quality, data privacy, data access security, data retention
sources across the
Glossary
data landscape can be
Azure Data Factory (Enterprise data fabric)
ingested into Azure
Data Lake Storage CDM
and integrated
using Azure Data
Factory to create
trusted, commonly
understood, reusable
Edge devices Data center Azure AWS Google Cloud
CDM data assets
that can be persisted
back in the data lake Figure 12
published in Azure
Data Catalog
Everything that is
underpinned by Enterprise data catalog Data marketplace on enterprise data catalog
ADLS in Figure 10 (register, profile & tag sources) (register, profile & tag trusted data assets)
can then make use of
trusted, commonly
understood CDM
described data assets
Ingest Prepare, transform, Publish
The objective is build & analyze
once, publish in a data
marketplace (Azure
Data Catalog) and
reuse everywhere
Azure Data
Data sources Azure Data Lake Factory Azure Data Lake Data consumption
ingestion zone trusted zone
Azure
Databricks
Figure 13
Microsoft partners
Azure Marketplace Partners for data governance include:
also offer technology
on Azure to help with • Tamr (for ETL processing)
data governance
• Talend (for ETL processing and Data Cataloguing)
• Informatica (for Enterprise Data Cataloguing)
• Qlik Data Integration
• Semarchy (master data management)
• Profisee (master data management)
A master data
Central to any data governance program is master data management. Creating trusted
management (MDM) master data is therefore critical. This can be done by defining master data entities in the
system is a core
component needed in business glossary within Azure Data Catalog and then using the data catalog to register data
data governance sources and discover where disparate master data is located across multiple data stores in
Master data entities the distributed data landscape.
can be defined in the
Azure Data Catalog
business glossary as
By mapping the physical data names of discovered disparate master data to the common
part of a common business vocabulary in Azure Data Catalog, it then becomes possible to know how to clean,
business vocabulary
match and integrate the data discovered to create golden master data records stored in
Disparate master
data can then be
a central MDM system. This can be done using Azure Data Lake Storage and Azure Data
ingested into ADLS Factory as shown in Figure 13. Once created and stored centrally, master data can then be
from where it can be
cleaned, matched and synchronised with all other systems that need it to make sure they are consistent.
integrated using ADF
to populate an MDM In addition, master data maintenance needs to be governed. The challenge is therefore to
system
identify in which tasks of which business processes that maintenance occurs. This can be
Master data
maintenance also
done using business process identification and CRUD analysis. However, it is often a manual
needs to be governed task to work this out but is now helped by the emergence of process mining and analysing
database log files. Once the tasks within a process that maintain master data have been
identified, it can be governed.
No roles and responsibilities Roles and responsibilities defined Roles and responsibilities defined Roles and responsibilities defined
defined
No DG control board DG control board in place but no DG control board in place with data DG control board in place with data
ability
No DG working groups No DG working groups Some DG working groups in place All DG working groups in place
No data owners accountable for No data owners accountable for Some data owners in place All data owners in place
data data
No data stewards appointed with Some data stewards in place for Data stewards in place and Data stewards in place assigned
People
responsibility for data quality DQ but scope too broad e.g. whole assigned to DG working groups for to DG working groups for specific
dept specific data data
No one accountable for data No one accountable for data CPO accountable for privacy (no CPO accountable for privacy with
privacy privacy tools) tools
No one accountable for access IT accountable for access security IT Sec accountable for access IT Sec accountable for access
security security security & responsible for enforcing
privacy
No one to produce trusted data Data publisher identified and Data publisher identified and Data publisher identified and
assets accountable for producing trusted accountable for producing trusted accountable for producing trusted
data data data
No SMEs identified for data entities Some SMEs identified but not SMEs identified & in DG working SMEs identified & in DG working
Benchmark your engaged groups groups
company on this data No common business vocabulary Common biz vocabulary started in Common business vocabulary Common business vocabulary
governance maturity a glossary established complete
model to gauge your No way to know where data is Data catalog auto data discovery, Data catalog auto data discovery, Data catalog auto data discovery,
progress located, its data quality or if it is profiling & sensitive data detection profiling & sensitive data detection profiling & sensitive data detection
sensitive data on some systems on all structured data on structured & unstructured in all
systems w/ full auto tagging
No process to govern authoring or Governance of data access security Governance of data access Governance of data access
maintenance of policies and rules policy authoring & maintenance on security, privacy & retention policy security, privacy & retention policy
some systems authoring & maintenance authoring & maintenance
No way to enforce policies & rules Piecemeal enforcement of data Enforcement of data access Enforcement of data access
access security policies & rules security and privacy policies and security, privacy & retention
across systems with no catalog rules across systems with catalog policies and rules across all systems
integration integration
Process
No processes to monitor data Some ability to monitor data Monitoring and stewardship of DQ Monitoring and stewardship of DQ
quality, data privacy or data access quality & data privacy on core systems & data privacy on all systems with
security with DBMS masking dynamic masking
Some ability to monitor privacy
(e.g. queries)
No availability of fully trusted data Dev started on a small set of Several core trusted data assets Continuous delivery of trusted
assets trusted data assets using data created using data fabric data assets with enterprise data
fabric software marketplace
No way to know if a policy violation Data access security violation Data access security violation Data access security violation
occurred or process to act if it did detection in some systems detection in all systems detection in all systems
No vulnerability testing process Limited vulnerability testing Vulnerability testing process on all Vulnerability testing process on all
process systems systems
No common process for master MDM with common master data MDM with common master data MDM with common master data
data creation, maintenance & sync CRUD & sync processes for single CRUD & sync processes for some CRUD & sync processes for all
entity data entities master data entities complete
No data governance classification Data governance classification Data governance classification Data governance classification
schemes on confidentiality & scheme for confidentiality scheme for both confidentiality and scheme for both confidentiality and
retention retention retention
No policies & rules to govern data Policies & rules to govern data Policies & rules to govern data Policies & rules to govern data
quality quality started in common quality defined in common quality defined in common
vocabulary in business glossary vocabulary in catalog biz glossary vocabulary in catalog biz glossary
No policies & rules to govern data Some policies & rules to govern Policies & rules to govern data Policies & rules to govern data
access security data access security created in access security & data privacy access security, data privacy and
different technologies consolidated in the data catalog retention consolidated in the
using classification scheme data catalog using classification
schemes and enforced everywhere
Policies No policies & rules to govern data Some policies & rules to govern Policies & rules to govern data Policies & rules to govern data
privacy data privacy access security & data privacy access security, data privacy and
consolidated in the data catalog retention consolidated in the
using classification scheme data catalog using classification
schemes and enforced everywhere
No policies & rules to govern data No policies & rules to govern data Some policies & rules to govern Policies & rules to govern data
retention retention data retention access security, data privacy and
retention consolidated in the
data catalog using classification
schemes and enforced everywhere
No policies & rules to govern Policies & rules to govern master Policies & rules to govern master Policies & rules to govern master
master data maintenance data maintenance for a single data maintenance for some master data maintenance for all master
master data entity data entities data entities
No data catalog with auto data Data catalog with auto data Data catalog with auto data Data catalog with auto data
discovery, profiling & sensitive data discovery, profiling & sensitive data discovery, profiling & sensitive data discovery, profiling & sensitive data
detection detection purchased detection purchased detection purchased
No data fabric software with Data fabric software with multi- Data fabric software with multi- Data fabric software with multi-
multi-cloud edge and data centre cloud edge and data centre cloud edge and data centre cloud edge and data centre
connectivity connectivity & catalog integration connectivity & catalog integration connectivity & catalog integration
purchased purchased purchased
Technology
No metadata lineage Metadata lineage available in data Metadata lineage available in data Metadata lineage available in data
catalog on trusted assets being catalog on trusted assets being catalog on trusted assets being
developed using fabric developed using fabric developed using fabric
No data stewardship tools Data stewardship tools available as Data stewardship tools available as Data stewardship tools available as
part of the data fabric software part of the data fabric software part of the data fabric software
No data access security tool Data access security in multiple Data access security in multiple Data access security enforced in
technologies technologies all systems
No data privacy enforcement No data privacy enforcement Data privacy enforcement in some Data privacy enforcement in all
software software DBMSs data stores
No master data management Single entity master data Multi-entity master data Multi-entity master data
system management system management system management system
Conclusions
Benchmark your
The key to successful data governance is to break structured data down into data entities
company on this data and data subject areas and then make use of a data governance solution to surround
governance maturity
model to gauge your specific data entities and data subject areas with people, processes, policies and technology
progress to govern the lifecycle of each of those data entities. This can be done by establishing a
common business vocabulary in a business glossary within a data catalog.
Creating data governance classification schemes such as the examples shown in Figure
Confidentiality
and retention 2 provide different levels of governance classification. These need to be defined in the
data governance
classification schemes
data catalog. At this point, policies and rules can then be created in the data catalog and
guide the creation of associated with different levels of governance classification.
policies and rules to
govern data
It should then be possible to label (or tag) data attributes in the business glossary
with confidentiality and retention classes to specify how to govern it. And because the
data catalog already knows the mappings of physical data attributes in different data
stores to attributes in business glossary, then labelling an attribute in the glossary
automatically determines how to govern data mapped to it in underlying data stores.
Multiple technologies that integrate with the data catalog can then access this metadata
to consistently enforce these policies and rules across all data stores in a distributed
data landscape. The exact same governance classification labels can also be applied to
unstructured data.
MDM is also
important because Master data entities are critical because this data is so widely shared. It is also frequently
master data is so
widely shared across associated with documents. For example, a customer and an invoice, a supplier and a
both operational contract, an asset and an operating manual. Therefore, master data values (e.g. supplier
transaction processing
systems and analytical name) can be used to tag related documents to ensure that relationships between
systems structured and unstructured data are preserved.
Using the common vocabulary data entities defined in the data catalog, and the mappings
Creating trusted,
reusable data assets discovered, it should then be possible to create pipelines using data fabric to create trusted
descibed using a
common business
data assets that can be published in a data marketplace for all to share. The key point about
vocabulary and data governance is that there are methods here to get your data under control and once
published in a data
marketplace enables trusted, to then use it to drive value. Success will be determined by how well you organise
trusted data to be and collaborate to do it. This Microsoft Data Governance Guide is provided to assist with
widely shared
that so that you can systematically make use of people, processes, policies and technology
Data governance
helps to systematically to get your data into a trusted well governed state to eradicate data quality problems and
create trusted, and the impact they have, uphold privacy, secure access and drive business value.
protected data