[go: up one dir, main page]

0% found this document useful (0 votes)
174 views330 pages

Integrating Data and Managing Spaces in SAP

The document provides a comprehensive guide on integrating data and managing spaces within SAP Datasphere, detailing various connection types and methods for data integration. It includes sections on creating, editing, validating, and managing connections to multiple data sources, as well as managing data exchange with SAP SQL Data Warehousing and SAP HANA Cloud Data Lake. Additionally, it covers monitoring and managing data integration processes effectively.

Uploaded by

fjaimesilva
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
174 views330 pages

Integrating Data and Managing Spaces in SAP

The document provides a comprehensive guide on integrating data and managing spaces within SAP Datasphere, detailing various connection types and methods for data integration. It includes sections on creating, editing, validating, and managing connections to multiple data sources, as well as managing data exchange with SAP SQL Data Warehousing and SAP HANA Cloud Data Lake. Additionally, it covers monitoring and managing data integration processes effectively.

Uploaded by

fjaimesilva
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 330

PUBLIC

Document Version: 2025.6 – 2025-03-18

Integrating Data and Managing Spaces in SAP


Datasphere
© 2025 SAP SE or an SAP affiliate company. All rights reserved.

THE BEST RUN


Content

1 Integrating Data and Managing Spaces in SAP Datasphere. . . . . . . . . . . . . . . . . . . . . . . . . . . . 6

2 Integrating Data via Connections. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8


2.1 Create a Connection. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16
Adverity Connections. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .20
Amazon Athena Connections. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21
Amazon Redshift Connections. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22
Amazon Simple Storage Service Connections. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 23
Apache Kafka Connections. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 26
Business Data Product Connections. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 30
Confluent Connections. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 30
Cloud Data Integration Connections. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 37
Generic JDBC Connections. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40
Generic OData Connections. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 42
Generic SFTP Connections. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 45
Google BigQuery Connections. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .47
Google Cloud Storage Connections. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 48
Hadoop Distributed File System Connections. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .50
Microsoft Azure Blob Storage Connections. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 52
Microsoft Azure Data Lake Store Gen1 Connections (Deprecated). . . . . . . . . . . . . . . . . . . . . . . 53
Microsoft Azure Data Lake Store Gen2 Connections. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 54
Microsoft Azure SQL Database Connections. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 56
Microsoft SQL Server Connections. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 58
Open Connectors Connections. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 61
Oracle Connections . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .63
Precog Connections. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .66
SAP ABAP Connections. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 67
SAP BW Bridge Connection. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .77
SAP BW Connections. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 77
SAP BW∕4HANA Model Transfer Connections. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 83
SAP ECC Connections. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 85
SAP Fieldglass Connections. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 90
SAP HANA Connections. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 92
SAP HANA Cloud, Data Lake Relational Engine Connections. . . . . . . . . . . . . . . . . . . . . . . . . . 100
SAP HANA Cloud, Data Lake Files Connections. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 101
SAP Marketing Cloud Connections. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .103
SAP SuccessFactors Connections. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 104

Integrating Data and Managing Spaces in SAP Datasphere


2 PUBLIC Content
SAP S/4HANA Cloud Connections. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 107
SAP S/4HANA On-Premise Connections. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 112
Connections to Partner Tools. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .122
2.2 Edit a Connection. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 123
2.3 Validate a Connection. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 125
2.4 Delete a Connection. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 126
2.5 Pause Real-Time Replication for a Connection . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 126
2.6 Filter Connection List. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 127
2.7 Importing SAP BW Queries – A Comparison of the Different Options. . . . . . . . . . . . . . . . . . . . . . . 128
2.8 Manage Connectivity via REST APIs. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 129

3 Integrating Data From SAP BW Bridge. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 136

4 Integrating Data from SAP Business Data Cloud. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 137


4.1 Reviewing Installed Insight Apps. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 140
4.2 Applying Row-Level Security to Data Delivered through Insight Apps. . . . . . . . . . . . . . . . . . . . . . . .141
4.3 Extending Insight Apps. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 142

5 Integrating Data via Database Users/Open SQL Schemas. . . . . . . . . . . . . . . . . . . . . . . . . . . .145


5.1 Create a Database User. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 147
5.2 Connect to Your Open SQL Schema. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 148
5.3 Allow the Space to Access the Open SQL Schema. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 149

6 Exchanging Data with SAP SQL Data Warehousing HDI Containers. . . . . . . . . . . . . . . . . . . . . 152
6.1 Prepare Your HDI Project for Exchanging Data with Your Space. . . . . . . . . . . . . . . . . . . . . . . . . . . 153
6.2 Add an HDI Container and Access its Objects in Your Space. . . . . . . . . . . . . . . . . . . . . . . . . . . . . .155
6.3 Consume Space Objects in Your HDI Container. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 156
6.4 Allow Your Space to Write to Your HDI Container. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 157

7 Integrating Data to and From SAP HANA Cloud Data Lake. . . . . . . . . . . . . . . . . . . . . . . . . . . 158
7.1 Working with Data Lake. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 159
7.2 Data Lake API. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 161

8 Managing and Monitoring Data Integration. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .163


8.1 Navigating in the Monitor Tabs. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 164
8.2 Understanding Statuses and Substatuses. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 165
8.3 Remote Tables in Data Access Remote Only Vs Data Access Remote and Replication. . . . . . . . . . . . 167
8.4 Replicating Data and Monitoring Remote Tables. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 168
Replicate Full Set of Data. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .179
Replicate Data Changes in Real-Time. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 181
Pausing and Restarting Real-Time Replication. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 183
Resuming Real-Time Replication After a Fail. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 183
Disabling Real-Time Data Replication. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 184
Partitioning Remote Table Data Loads. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 184

Integrating Data and Managing Spaces in SAP Datasphere


Content PUBLIC 3
8.5 Monitoring Local Tables. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 189
Performing Data Deletion for Local Tables . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .191
8.6 Monitoring Local Tables (File). . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 195
8.7 Persisting and Monitoring Views. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 196
Data Persistence and Run Mode . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 204
Persisted Views and Data Access Control. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 205
Creating Partitions for Your Persisted Views. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 210
Persisted Views and Memory Consumption. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 214
Data Persistence, Detailed Logs and Remote Statements. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 216
Understanding the Data Persistence Metrics. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .218
Exploring Views with View Analyzer. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 221
8.8 Monitoring Flows. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 239
Working With Existing Replication Flow Runs. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 242
Cancel a Transformation Flow Run. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 245
Watermarks. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 245
Change the Run Mode for a Transformation Flow. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .247
Metrics for Transformation Flows. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .247
Explore Transformation Flows. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 248
8.9 Monitoring Remote Queries. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 249
Creating Statistics for Your Remote Tables. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 251
8.10 Monitoring Task Chains. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 252
8.11 Scheduling Data Integration Tasks. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 256
Schedule a Data Integration Task (Simple Schedule). . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 256
Schedule a Data Integration Task (with Cron Expression). . . . . . . . . . . . . . . . . . . . . . . . . . . . . 261
Take Over the Ownership of a Schedule. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 265
Pause or Resume a Scheduled Task. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 266
8.12 Deleting Task Logs to Reduce Storage Consumption. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 266

9 Securing Data with Data Access Controls. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .268


9.1 Create a "Single Values" Data Access Control. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 271
9.2 Create an "Operator and Values" Data Access Control. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .274
9.3 Create a "Hierarchy" Data Access Control. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 277
9.4 Import SAP BW and SAP BW∕4HANA Analysis Authorizations. . . . . . . . . . . . . . . . . . . . . . . . . . . . 281
9.5 Process Source Changes in the Data Access Control Editor. . . . . . . . . . . . . . . . . . . . . . . . . . . . . .284

10 Importing SAP and Partner Business Content from the Content Network. . . . . . . . . . . . . . . 286

11 Transporting Content Between Tenants. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 290


11.1 Creating Packages to Export. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 292
11.2 Exporting Content for Sharing with Other Tenants. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 296
Adding Sharing Destinations. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 300
Managing Exported Content. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 301

Integrating Data and Managing Spaces in SAP Datasphere


4 PUBLIC Content
11.3 Importing Content from Another Tenant. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 301
11.4 Transporting Your Content through SAP Cloud Transport Management. . . . . . . . . . . . . . . . . . . . . 305

12 Managing Your Space. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 308


12.1 Control User Access to Your Space. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 308
12.2 Create Time Data and Dimensions. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .311
12.3 Logging Read and Change Actions for Audit. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 313
12.4 Translating Metadata for SAP Analytics Cloud. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 315
12.5 Save and Deploy Your Space. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .319
12.6 Review Your Space Status. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 320
12.7 Unlock a Locked Space. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 321
12.8 Monitor Your Space Storage Consumption. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 323
12.9 Manage Your Space via the Command Line. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 324
12.10 Export Your Space Data. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 325
12.11 Delete Your Space. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 325

Integrating Data and Managing Spaces in SAP Datasphere


Content PUBLIC 5
1 Integrating Data and Managing Spaces in
SAP Datasphere

Users with a space administrator or integrator role can create connections to source systems and databases
and can schedule and monitor data replication and other data integration tasks. Space administrators are,
additionally, responsible for controlling user access to their space, creating data access controls to secure
data, enabling other forms of data integration, transporting content between tenants, and monitoring and
otherwise managing the space.

This topic contains the following sections:

• Prepare and Manage Your Space [page 6]


• Create Connections to Source Systems [page 7]
• Integrate Other Data Sources [page 7]
• Import Business and Sample Content [page 7]
• Prepare Row-Level Security for Data [page 7]
• Manage and Monitor Data Integration Tasks [page 7]

 Tip

The English version of this guide is open for contributions and feedback using GitHub. This allows you
to get in contact with responsible authors of SAP Help Portal pages and the development team to
discuss documentation-related issues. To contribute to this guide, or to provide feedback, choose the
corresponding option on SAP Help Portal:

•  Feedback Edit page : Contribute to a documentation page. This option opens a pull request on
GitHub.
•  Feedback Create issue : Provide feedback about a documentation page. This option opens an
issue on GitHub.

You need a GitHub account to use these options.

More information:

• Contribution Guidelines
• Introduction Video: Open Documentation Initiative
• Blog Post: Introducing the Open Documentation Initiative

Prepare and Manage Your Space

An administrator will assign you the DW Space Administrator role, create your space, and assign you to it. Once
this is done, you can prepare your space as follows:

• Assign SAP Datasphere users to your space (see Control User Access to Your Space [page 308]).
• Transport objects securely to and from your space (see Transporting Content Between Tenants [page
290]).

Integrating Data and Managing Spaces in SAP Datasphere


6 PUBLIC Integrating Data and Managing Spaces in SAP Datasphere
• Use various monitoring and logging tools to manage your space (see Managing Your Space [page 308]).

Create Connections to Source Systems

Space administrators and integrators can create connections to source systems to allow space users to
acquire data from those systems (see Integrating Data via Connections [page 8]).

Integrate Other Data Sources

Space administrators can integrate data from other sources:

• Create database users to allow external tools to connect to the space and write data to Open SQL schemas
associated with the space (see Integrating Data via Database Users/Open SQL Schemas [page 145]).
• If your space has access to the SAP HANA Cloud, data lake, you can access it via an Open SQL schema
(see Integrating Data to and From SAP HANA Cloud Data Lake [page 158]).
• Add SAP HDI containers to your space (see Exchanging Data with SAP SQL Data Warehousing HDI
Containers [page 152])
• Generate a time table and associated time dimension views for use in the space (see Create Time Data and
Dimensions [page 311]).

Import Business and Sample Content

SAP and our partners provide business content to support end-to-end business scenarios for various
industries and lines of business via the Content Network and our Community Content site (see Importing
SAP and Partner Business Content from the Content Network [page 286]).

Prepare Row-Level Security for Data

We recommend that you create data access controls, which can be applied to views to provide row-level
filtering of your space data (see Securing Data with Data Access Controls [page 268]).

Manage and Monitor Data Integration Tasks

You can enable, run, schedule, and monitor data replication tasks in the  (Data Integration Monitor) (see
Managing and Monitoring Data Integration [page 163]).

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data and Managing Spaces in SAP Datasphere PUBLIC 7
2 Integrating Data via Connections

Users with a space administrator or integrator role can create connections to SAP and non-SAP source
systems, including cloud and on-premise systems and partner tools, and to target systems for outbound
replication flows. Users with modeler roles can import data via connections for preparation and modeling in
SAP Datasphere.

This topic contains the following sections:

• Supported Connection Type Features [page 9]


• Connection Types Overview [page 10]
• Managing Connections [page 14]

Integrating Data and Managing Spaces in SAP Datasphere


8 PUBLIC Integrating Data via Connections
Supported Connection Type Features

Each connection type supports a defined set of features. Depending on the connection type and the
connection configuration, a connection can be used for one or more of the following features:

Feature Description

Remote Tables Allows modelers to import tables from a source when creating graphical and SQL
views or entity-relationship models. During the import, the tables are deployed as
remote tables. Depending on the connection type, you can use remote tables to:

• directly access data in the source (remote access)


• copy the full set of data (snapshot or scheduled replication)
• copy data changes in real-time (real-time replication)

SAP Datasphere is using two types of adaptors to connect to remote tables:

• SAP HANA smart data integration (used in connections with Data


Provisioning option = Data Provisioning Agent).
• SAP HANA smart data access adaptors (used in connections with no Data
Provisioning option or Data Provisioning option = Cloud Connector or Direct).

 Note
If your source data comes from an SAP HANA On-Premise system, select the
adaptor following your use case:

• You want to access the data remotely: SAP HANA smart data access
(Data Provisioning Option: Direct) would be the recommended adaptor
to read the data. It allows higher degree of query pushdown to the
remote database, leading to better response times and less resource
consumption.
• You want to replicate the data into SAP Datasphere: The preferred option
for this is to use Replication Flows, see Creating a Replication Flow. In
case you require replication for remote tables, SAP HANA smart data
integration (Data Provisioning Option: Data Provisioning Agent) is the
recommended adaptor to push the data. It offers more options when
loading the data, such as applying filter conditions or data partitioning.

For more information on these adaptors, see Connecting SAP HANA Cloud,
SAP HANA Database to Remote Data Sources.

For more information on remote tables, see:

• Importing Tables and Views from Sources


• Replicating Data and Monitoring Remote Tables [page 168]

Data Flows Allows modelers to use objects from a connected source when creating data
flows to extract, transform and load data.

For more information, see Creating a Data Flow.

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 9
Feature Description

Model Import Allows modelers to import objects from any supporting SAP system with rich
metadata and, if supported, with their dependencies. This way, you can leverage
your existing investment in your customer system landscape without having to
rebuild the objects manually.

For more information, see Importing SAP BW∕4HANA Models.

Replication Flows Allows modelers to replicate data from multiple source objects from a connected
source to a target including simple projection and mapping capabilities.

You can enable a single initial load or request initial and delta loads to load
changes at regular intervals. It depends on the connection type if you can use the
connection as source or target connection.

For more information, see Creating a Replication Flow.

 Note

Some connection types, such as SAP S/4HANA Cloud and on-premise or SAP ABAP, support replication
with both remote tables via Data Provisioning Agent (SAP HANA Smart Data Integration) and replication
flows (for on-premise systems via Cloud Connector).

Regardless of whether you use the same connection or two separate connections, for replicating data from
a dedicated source object we strongly recommend to only use remote tables or replication flows and not
both.

To allow space users using a feature, the feature must be enabled in the connection configuration. You can find
more information about which features are supported by which connection type and what prerequisites and
connection properties are required to enable (or disable) a feature from the links in the overview below.

Connection Types Overview

To connect to remote systems, applications, databases, or storages, SAP Datasphere provides different
connection types. The following connection types are available with SAP Datasphere:

Remote Tables

(Remote Ac-
cess and Snap-
shot or Sched-

Connection uled Replica- Replication


Type tion) Flows Data Flows Model Import Category Sources

Adverity Con- * no no no Cloud Partner Tools


nections [page
20]

Integrating Data and Managing Spaces in SAP Datasphere


10 PUBLIC Integrating Data via Connections
Remote Tables

(Remote Ac-
cess and Snap-
shot or Sched-

Connection uled Replica- Replication


Type tion) Flows Data Flows Model Import Category Sources

Amazon Athena yes no no no Cloud Non SAP


Connections
[page 21]

Amazon Red- yes no yes no Cloud Non SAP


shift Connec-
tions [page 22]

Amazon Simple no yes (source and yes no Cloud Non SAP


Storage Service target)
Connections
[page 23]

Apache Kafka no yes (target) no no Cloud Non SAP


Connections
[page 26]

Business Data not applicable not applicable not applicable not applicable not applicable not applicable
Product Con-
(replication
nections [page
30] flows are cre-
ated and de-
[created when
ployed via the
installing an
data product in-
SAP Business
stallation)
Data Cloud data
product from
the connected
system for the
first time and
only available
within an SAP-
managed inges-
tion space]

Cloud Data In- yes (including no yes no Cloud SAP


tegration Con- real-time repli-
nections [page cation**)
37]

Confluent Con- no yes (source and no no On-Premise, Non SAP


nections [page target) Cloud
30]

Generic JDBC yes no no no On-Premise, Non SAP


Connections Cloud
[page 40]

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 11
Remote Tables

(Remote Ac-
cess and Snap-
shot or Sched-

Connection uled Replica- Replication


Type tion) Flows Data Flows Model Import Category Sources

Generic OData yes no yes no On-Premise, Non SAP


Connections Cloud
[page 42]

Generic SFTP no no yes no On-Premise, Non SAP


Connections Cloud
[page 45]

Google Big- yes yes (target) yes no Cloud Non SAP


Query Connec-
tions [page 47]

Google Cloud no yes (source and yes no Cloud Non SAP


Storage Con- target)
nections [page
48]

Hadoop Distrib- no no yes no Cloud Non SAP


uted File Sys-
tem Connec-
tions [page 50]

Microsoft Azure no no yes no Cloud Non SAP


Blob Storage
Connections
[page 52]

Microsoft Azure no no yes no Cloud Non SAP


Data Lake Store
Gen1 Connec-
tions (Depre-
cated) [page
53]

Microsoft Azure no yes (source and yes no Cloud Non SAP


Data Lake Store target)
Gen2 Connec-
tions [page 54]

Microsoft Azure yes (including yes (source) yes no Cloud Non SAP
SQL Database real-time repli-
Connections cation**)
[page 56]

Microsoft SQL yes (including no yes no On-Premise Non SAP


Server Connec- real-time repli-
tions [page 58] cation**)

Open Connec- no no yes no Cloud SAP


tors Connec-
tions [page 61]

Integrating Data and Managing Spaces in SAP Datasphere


12 PUBLIC Integrating Data via Connections
Remote Tables

(Remote Ac-
cess and Snap-
shot or Sched-

Connection uled Replica- Replication


Type tion) Flows Data Flows Model Import Category Sources

Oracle Connec- yes (including no yes no On-Premise Non SAP


tions [page 63] real-time repli-
cation**)

Precog Connec- * no no no Cloud Partner Tools


tions [page 66]

SAP ABAP Con- yes (including yes (source) yes no On-Premise, SAP
nections [page real-time repli- Cloud
67] cation**)

SAP BW Con- yes (including via connection yes no On-Premise SAP


nections [page real-time repli- type SAP ABAP
77] cation**) Connections
[page 67]
(source)

SAP Data- yes no no yes Cloud SAP


sphere, SAP
BW bridge Con-
nection

[generated and
only available
within an SAP
BW bridge
space if SAP
BW bridge has
been provi-
sioned]

SAP no no no yes On-Premise SAP


BW∕4HANA
Model Transfer
Connections
[page 83]

SAP ECC Con- yes (including via connection yes no On-Premise SAP
nections [page real-time repli- type SAP ABAP
85] cation**) Connections
[page 67]
(source)

SAP Fieldglass yes (including no yes no Cloud SAP


Connections real-time repli-
[page 90] cation**)

SAP HANA yes (including yes (source and yes no On-Premise, SAP
Connections real-time repli- target) Cloud
[page 92] cation**)

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 13
Remote Tables

(Remote Ac-
cess and Snap-
shot or Sched-

Connection uled Replica- Replication


Type tion) Flows Data Flows Model Import Category Sources

SAP HANA no yes (source and yes no Cloud SAP


Cloud, Data target)
Lake Files Con-
nections [page
101]

SAP HANA yes no yes no Cloud SAP


Cloud, Data
Lake Relational
Engine Connec-
tions [page 100]

SAP Marketing yes (including no yes no Cloud SAP


Cloud Connec- real-time repli-
tions [page 103] cation**)

SAP Success- yes no yes no Cloud SAP


Factors Con-
nections [page
104]

SAP S/4HANA yes (including yes (source) yes yes Cloud SAP
Cloud Connec- real-time repli-
tions [page 107] cation**)

SAP S/4HANA yes (including yes (source) yes yes On-Premise SAP
On-Premise real-time repli-
Connections cation**)
[page 112]

*Data is pushed into the SAP Datasphere space by using the database user (Open SQL schema) SQL Interface.
Typically, the push action is triggered from the partner tool. When data is pushed by the source, you can import
a table from the Sources tab of the Source Browser panel in the Data Builder which is then deployed as local
table (see Connections to Partner Tools [page 122]).

** For more information about any constraints for using real-time replication, see Replicate Data Changes in
Real-Time [page 181].

***Connections marked as "On-Premise" can be used to connect to sources located in the cloud, too. Note
that in order to connect to a cloud-based SAP ABAP or Microsoft SQL Server, for example, you have to connect
through the on-premise installed SAP HANA smart data integration Data Provisioning Agent.

Managing Connections

In the  (Connections) app, you get an overview of all connections created in your space.

Integrating Data and Managing Spaces in SAP Datasphere


14 PUBLIC Integrating Data via Connections
Use the following tools to create and manage connections:

Tool Description

 (Add Connection) Create Create a connection to allow users assigned to the space to use the connected
remote system for data modeling and data access in SAP Datasphere.
Connection
For more information, see Create a Connection [page 16].

Edit Select and edit a connection to change its properties. Warnings might indicate
that you need to edit a connection.

For more information, see Edit a Connection [page 123].

Delete Select and delete one or more connections if they are not used anymore.

For more information, see Delete a Connection [page 126].

Validate Select and validate a connection to get detailed status information and make
sure that it can be used for data modeling and data access. Always validate a
connection after you have created or edited it.

For more information, see Validate a Connection [page 125].

Pause/Restart For connections that connect to a remote system through SAP HANA Smart Data
Integration and its Data Provisioning Agent, you can pause and restart real-time
replication for selected connections, if required.

For more information, see Pause Real-Time Replication for a Connection [page
126].

 (Reload Connection List) Reload the connections list to include the latest updates into the list.

 (Sort Connections) Open the Sort dialog to control the ordering of the connections list.

By default, the list is sorted by Business Name. To sort on a specific column,


select a Sort Order and a Sort By column, and then click OK to apply them.

 (Filter Connections) Select one or more filter values to restrict the connection list according to your
needs.

For more information, see Filter Connection List [page 127].

 (Select Columns) Open the Columns dialog to control the display of columns in the results table.

Modify the column list in any of the following ways, and then click OK to apply
your changes:

• To select a column for display, select its checkbox. To hide a column deselect
its checkbox.
• Click on a column token to highlight it and use the arrow buttons to move it in
the list.
• Click Reset to go back to the default column display.

Search Enter one or more characters in the Search field to restrict the list to connections
containing the string.

Search considers the Technical Name, Business Name, Real-Time Replication


Status, and Owner column.

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 15
Tool Description

When there is one or more messages for your connections, a button is displayed
on the top right corner of the Connections app specifying the number of warning
messages for all connections in the list.
(Warning Messages)
Click the button to open the list of messages. Clicking a message title selects the
corresponding connection in the list. Clicking  (Navigation) for a message opens
a more detailed message containing guidance on how to solve the issue.

In the connections list, connections with warning messages are highlighted in


yellow.

2.1 Create a Connection

Create a connection to allow users assigned to a space to use the connected source or target for data modeling
and data access in SAP Datasphere.

Prerequisites

See Preparing Connectivity for Connections.

Procedure

1. In the side navigation area, click  (Connections) and select a space if necessary.

2. On the Connections tab, click  (Add Connection) Create Connection to open the connection
creation wizard.
3. Click an appropriate connection type tile for your source.

One or more of the following options help you to quickly find the connection type you need:

Option Description

Search In the search field, enter a search string to only show


connection types that contain the string.

Sort Use  (Sort) to sort alphabetically.

Integrating Data and Managing Spaces in SAP Datasphere


16 PUBLIC Integrating Data via Connections
Option Description

Filter In the Filters section, use one of the following filters:

• Features that a connection type supports (Remote


Tables, Data Flows, Replication Flows, Model Import)
For more information about the available features
with SAP Datasphere, see Supported Connection
Type Features [page 9].
• Categories that the corresponding source belongs to
(On-Premise, Cloud)
• Sources that you would like to connect (SAP, Non-
SAP, Partner Tools)

 Note

When hovering over a connection type, an Information button appears that you can click to get
information about the features supported by the connection type. This includes information about
whether the connected system can be used as source and/or target.

In the information, a warning icon indicates if a feature is not available, and clicking the icon tells you
what to do in that case. Also, if a connection type is inactive, the information tells you what to do to be
able to use it.

4. Complete the properties in the following sections and click Next Step.

Property Section Description

Connection Details Enter or select properties that are needed to locate the
source system or database.

Various Enter or select properties in additional sections depending


on the connection type.

Credentials Enter the credentials that are required to access the


source.

Features Select additional connection properties to enabe a fea-


ture, if required.

You can click the info button next to the Enabled or


Disabled information to get more information about what
is required to enable the feature.

 Note

If you connect to a partner tool, you will be redirected to an embedded partner UI to there enter
credentials and required connection properties, if applicable.

For an overview of all connection types including links to detailed property documentation, see Connection
Types Overview [page 10].

5. [optional] For some connection types that use a Data Provisioning Agent to connect to the source, for
example SAP ABAP, SAP HANA, or Microsoft SQL Server, you can define advanced properties which

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 17
give you more control over your connectivity setup. If required, override default values according to your
customer scenario’s needs, and then click Next Step.
6. Complete the following properties:

Property Description

Business Name Enter a descriptive name to help users identify the object.
This name can be changed at any time.

Technical Name Displays the name used in scripts and code, synchronized
by default with the Business Name.

The technical name can only contain alphanumeric char-


acters and underscores (_). Underscore (_) must not be
used at the start or end of the name. The maximum length
is 40 characters. The name must be unique within the
space.

 Note
Once the object is saved, the technical name can no
longer be modified.

Package Select the package to which the connection belongs.

Users with the DW Space Administrator role (or equivalent


privileges) can create packages in the Packages editor.
Once a package is created in your space, you can select it
here.

Packages are used to group related objects in order to


facilitate their transport between tenants.

 Note
Once a package is selected, it cannot be changed
here. Only a user with the DW Space Administrator
role (or equivalent privileges) can modify a package
assignment in the Packages editor.

For more information, see Creating Packages to Export


[page 292].

Description Provide more information to help users understand the


object.

7. If you connect to a partner tool such as Precog, in the Additional Configuration section of this wizard step,
you'll find information about IP addresses that you can copy in order to add them to the IP allowlist in SAP
Datasphere.
For more information, see Manage IP Allowlist.
8. Click Create Connection to create the connection and add it to the overview of available connections.

Integrating Data and Managing Spaces in SAP Datasphere


18 PUBLIC Integrating Data via Connections
 Note

When creating a connection to a partner tool such as Precog, the system checks whether the required
IP address or addresses are already available in the IP allowlist. Note that connection creation takes
longer because checking the allowlist in the database requires some time.

9. Select the connection and click  (Validate Connection) to check if your connection is valid and can be
used for the enabled features.

 Note

When creating a connection to SAP HANA on-premise using SAP HANA smart data access via Cloud
Connector, the system checks for a required internal service. In the connection overview, a warning
icon indicates that the service might not be ready yet. This happens for the first connection you create
or when you create a connection after the service has been disabled (after an automated weekly check
returning that there is no such connection anymore). Getting the service ready might take up to 45
minutes. Validate the connection and check the message details for more information.

Results

Depending on the connection type and the features that you have enabled for your connection:

• You can use the connection either to import tables when creating views or ER models, to create data flows,
or to create replication flows, or all three.
For more information, see:
• Importing Tables and Views from Sources
• Creating a Data Flow
• Creating a Replication Flow

 Note

Generally, when importing a table, it is created as remote table. However, tables imported from partner
tool connections are created as local tables (see Connections to Partner Tools [page 122]).

• You can use remote tables imported from the connection either to access data directly live in the
connected source (federation) or to copy the data into SAP Datasphere (replication).
For more information, see Replicating Data and Monitoring Remote Tables [page 168].
• You can use the connection to import entities from your source.
For more information, see:
• SAP BW∕4HANA Model Transfer Connections [page 83]
• Importing SAP BW∕4HANA Models

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 19
2.1.1 Adverity Connections

Extend connectivity beyond SAP Datasphere standard remote connectivity and cover additional data sources
that are available with Adverity.

Adverity provides connectivity to a variety of data sources, including marketing data sources, databases and
others.

When creating a connection to a partner tool, an Open SQL schema is generated in SAP Datasphere. To be
able to send data to the Open SQL schema, the IP addresses of the partner tool need to be added to the
SAP Datasphere IP allowlist, and during connection creation the partner tool on their side needs to establish a
connection to the schema.

For more information, see Connections to Partner Tools [page 122].

Prerequisites

See: Prepare Connectivity to Adverity

Using the Connection

The connection type supports creating views and entity-relationship models.

Configuring Partner Connection

Wizard step Configure Partner Connection shows an embedded Adverity UI. Here, you log on to the regional
Adverity server and perform the following steps:

1. Enter the URL for your regional Adverity server.


2. Choose Next.

 Note

As a prerequisite to proceed to the next step in the wizard, your browser needs to be enabled to store
cookies. This is not the case if you run the browser in incognito mode.

3. Enter your credentials.


4. Select the Adverity workspace that you want to use to create the destination to the SAP Datasphere Open
SQL schema in.

 Note

If there is no appropriate workspace available yet, you can create a new one.

5. Click Save to move to the next connection creation wizard step in SAP Datasphere.

Integrating Data and Managing Spaces in SAP Datasphere


20 PUBLIC Integrating Data via Connections
For more information, see Create a Connection [page 16].

What happens in Adverity after you have created the connection

You can now see the destination in the Adverity workspace. There, you can define a prefix for the tables created
for this destination to more easily find them later on in the SAP Datasphere Data Builder.

2.1.2 Amazon Athena Connections

Use an Amazon Athena connection to access data from Amazon Athena, an interactive query service which can
be used to analyze data in Amazon S3 using standard SQL.

Prerequisites

See: Prepare Connectivity to Amazon Athena

Using the Connection

The connection type supports the remote table feature.

Supported data access methods for remote tables:

• Remote
• Replication (snapshot)

For more information, see Replicating Data and Monitoring Remote Tables [page 168].

Configuring Connection Properties

Connection Details

Property Description

Region Enter the AWS region in your Amazon Athena regional end-
point that you use to make your requests. For example, us-
west-2.

Workgroup Enter the name of the workgroup. Amazon Athena uses


workgroups to control query access and costs. The default
workgroup is primary.

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 21
Credentials

Property Description

Access Key Enter the access key ID of the user that the application must
use to authenticate.

Secret Key Enter the secret access key of the user that the application
must use to authenticate.

Features

Remote Tables are enabled without the need to set any additional connection properties.

2.1.3 Amazon Redshift Connections

Use an Amazon Redshift connection to access data from Amazon Redshift 8.x databases.

Prerequisites

See: Prepare Connectivity to Amazon Redshift

Using the Connection

The connection type supports the remote table as well as the data flow feature.

Supported data access methods for remote tables:

• Remote
• Replication (snapshot)

For more information, see Replicating Data and Monitoring Remote Tables [page 168].

 Note

• In the view editor of the Data Builder, you can see technical artifacts from the source that you cannot
use for modeling. Affected artifacts are tables containing "pkey" or tables from technical schemas
pg_catalog, pg_internal, and information_schema, for example. Please ignore these artifacts.

Integrating Data and Managing Spaces in SAP Datasphere


22 PUBLIC Integrating Data via Connections
Configuring Connection Properties

Connection Details

Property Description

Host Enter the host name of the Redshift server.

Port Displays the default port 5439. You can overwrite the port, if required.

Database Name Enter the name of the database to which you want to connect.

Security

Property Description

Use SSL Select the SSL certificate verification mode to use when connecting. The default is false.

[if Use SSL = true] Validate Select if the server certficate should be validated (true) or not (false). The default is false.
Server Certificate
If you select true, then the host name in the certificate must match the host name specified
in the Host field.

Credentials

Property Description

User Name Enter the name of the Redshift user that has privileges to connect to the database.

Password Enter the password of the Redshift user.

Features

To enable Remote Tables, select a Data Provisioning Agent.

Data Flows are enabled without the need to set any additional connection properties.

2.1.4 Amazon Simple Storage Service Connections

Use an Amazon Simple Storage Service connection to connect to and access data from objects in Amazon S3
buckets.

This topic contains the following sections:

• Supported Features [page 24]


• Configuring Connection Properties [page 24]

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 23
Supported Features

Feature Additional Information

Data Flows You can use the connection to add source objects to a data flow.

Replication Flows You can use the connection to add source and target objects to a replication flow.

For more information, see:

• Cloud Storage Provider Sources


• Cloud Storage Provider Targets

 Note
You can only use a non-SAP target for a replication flow if your admin has assigned
capacity units to Premium Outbound Integration. For more information, see Premium
Outbound Integration and Configure the Size of Your SAP Datasphere Tenant.

Configuring Connection Properties

Connection Details

Property Description

Endpoint Enter the endpoint URL of the Amazon S3 server, for example s3.amazonaws.com. The
protocol prefix is not required.

 Note
When using Assume Role, you must enter the regional endpoint, for example s3.us-
west-2.amazonaws.com.

Protocol Select the protocol. The default value is HTTPS. The value that you provide overwrites the
value from the endpoint, if already set.

Root Path [optional] Enter the root path name for browsing objects. The value starts with the charac-
ter slash. For example, /My Folder/MySubfolder.

If you have specified the root path, then any path used with this connection is prefixed with
the root path.

Server Side Encryption

Property Description

Use Server Side Encryption [optional] Select true (default) if you want to use S3 objects encrypted through server side
encryption.

Integrating Data and Managing Spaces in SAP Datasphere


24 PUBLIC Integrating Data via Connections
Property Description

Encryption Option [optional] You can select:

• Encryption with Amazon S3 Managed Keys (SSE-S3) (default) if your objects are en-
crypted using the default encryption configuration for objects in your Amazon S3
buckets.
• Encryption with AWS Key Management Service Keys (SSE-KMS) if your Amazon S3
buckets are configured to use AWS KMS keys.

[if Encryption Option = Enter the KMS key Amazon Resource Name (ARN) for the customer managed key which
Encryption with AWS Key has been created to encrypt the objects in your Amazon S3 buckets.
Management Service Keys
(SSE-KMS)] KMS Key ARN

Assume Role

Property Description

Use Assume Role [optional] Select true if you want to use temporary security credentials and restrict access
to your Amazon S3 buckets based on an Identity and Access Management role (IAM role).
The default value is false.

[if Use Assume Role = true] Enter the role Amazon Resource Name (ARN) of the assumed IAM role.
Role ARN

[if Use Assume Role = true] [optional] Enter the name to uniquely identify the assumed role session.
Role Session Name

[if Use Assume Role = true] [optional] Enter the duration. The default duration is 3600 seconds.
Duration of Role Session (in
Seconds)

[if Use Assume Role = true] [optional] If an external ID is used to make sure that SAP Datasphere as a specified third
External ID party can assume the role, enter the ID.

[if Use Assume Role = true] [optional] If you want to use a session policy, enter the respective IAM policy in JSON
Role Policy format.

Credentials

Property Description

Access Key Enter the access key ID of the user that is used to authenticate to Amazon S3.

Secret Key Enter the secret access key of the user that is used to authenticate to Amazon S3.

Features

Feature Description

Data Flows Data Flows are enabled without the need to set any additional connection properties.

Replication Flows Replication Flows are enabled without the need to set any additional connection properties.

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 25
2.1.5 Apache Kafka Connections

Use the connection to connect to an Apache Kafka cluster.

This topic contains the following sections:

• Supported Features [page 26]


• Configuring Connection Properties [page 26]

For information about the required prerequisites, see Prepare Connectivity to Apache Kafka.

Supported Features

Feature Additional Information

Replication Flows You can use the connection to add target objects to a replication flow.

For more information, see Apache Kafka Targets.

 Note
You can only use a non-SAP target for a replication flow if your admin has assigned
capacity units to Premium Outbound Integration. For more information, see Premium
Outbound Integration and Configure the Size of Your SAP Datasphere Tenant.

Configuring Connection Properties

Connection Details

Property Description

Kafka Brokers Enter a comma-separated list of brokers in the format <host>:<port>.

Cloud Connector

Property Description

Use Cloud Connector [optional] Set to true if your source is an on-premise source and you want to use the
connection for replication flows. The default is false.

Integrating Data and Managing Spaces in SAP Datasphere


26 PUBLIC Integrating Data via Connections
Property Description

[if Use Cloud Connector = [optional] Select a location ID.


true] Location
 Note
To select another location ID than the default location, Connection.Read privilege is
required. The privilege is neither included in the DW Integrator nor in the DW Space
Administrator role. If you need to select a location ID, ask your tenant administrator to
either assign your user to a global role that is based on the DW Administrator role or
to assign your user to a custom global role (with license type SAP Datasphere) that
includes the required Connection.Read privilege.

[if Use Cloud Connector = [optional] Select Derive Virtual Host and Port from Connection Details or Enter Virtual Host
true] Virtual Destination and Port in Separate Fields.

If host and port entered in the connection details match the virtual host and port from the
Cloud Connector configuration, you can select Derive Virtual Host and Port from Connection
Details and don't need to enter the values manually.

[if Virtual Destination = Enter Enter the virtual host that you defined during Cloud Connector configuration.
Virtual Host and Port in
Separate Fields] Virtual Host

[if Virtual Destination = Enter Enter the virtual port that you defined during Cloud Connector configuration.
Virtual Host and Port in
Separate Fields] Virtual Port

Authentication

Property Description

Authentication Type Select the authentication type to use to connect to the Kafka brokers.

You can select:

• No Authentication
• User Name And Password

 Note
For basic authentication with user name and password, we recommend to use TLS
encryption to ensure secure communication.

• Salted Challenge Response Authentication Mechanism (256)


• Salted Challenge Response Authentication Mechanism (512) (default)
• Kerberos with Username and Password
• Kerberos with Keytab File

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 27
Property Description

SASL Authentication Type [read-only] Displays the Simple Authentication and Security Layer (SASL) authentication
mechanism to use depending on your selection for Authentication Type.

The mechanisms are:

• PLAIN: plaintext password defined in RFC 4616.


• SCRAM-256: Salted Challenge Response Authentication Mechanism (SCRAM) pass-
word-based challenge-response authentication from a user to a server -based on SHA
256.
• SCRAM-512: SCRAM password-based challenge-response authentication from a user
to a server - based on SHA 512.
• GSSAPI: Generic Security Service Application Program Interface authentication appli-
cable for Kerberos V5.

Kerberos

If Authentication Type = Kerberos with Username and Password or Kerberos with Keytab File:

Property Description

Kafka Kerberos Service Name Enter the name of the Kerberos service used by the Kafka broker.

Kafka Kerberos Realm Enter the realm defined for the Kafka Kerberos broker.

Kafka Kerberos Config Upload the content of the krb5.conf configuration file.

Choose  (Browse) and select the file from your download location.

Once uploaded, you can check the configuration file by clicking the  (inspection) button.

Credentials (SCRAM)

If Authentication Type = Salted Challenge Response Authentication Mechanism (512) or Salted Challenge
Response Authentication Mechanism (256):

Property Description

Kafka SASL User Name Enter the user name that is used for Kafka SASL SCRAM authentication.

Kafka SASL Password Enter the Kafka SASL SCRAM connection password.

Credentials (User Name and Password)

If Authentication Type = User Name And Password:

Property Description

User Name Enter the user name that is used for Kafka SASL PLAIN authentication.

Password Enter the Kafka SASL PLAIN connection password.

Integrating Data and Managing Spaces in SAP Datasphere


28 PUBLIC Integrating Data via Connections
Credentials (Kerberos with Keytab File)
If Authentication Type = Kerberos with Keytab File:

Property Description

User Name Enter the name of the user that is used to connect to the Kerberos service.

Keytab File Upload the content of the keytab file.

Choose  (Browse) and select the file from your download location.

Once uploaded, you can check the keytab file by clicking the  (inspection) button.

Credentials (Kerberos with User Name And Password)


If Authentication Type = Kerberos with Username and Password:

Property Description

User Name Enter the name of the user that is used to connect to the Kerberos service.

Password Enter the Kerberos password.

Security

Property Description

Use TLS Select true (default) to use TLS encryption.

Validate Server Certificate Select true (default) to validate the TLS server certificate.

[if Use TLS = true and Select true to use mutual authentication with validating both client and server certificates.
Validate Server Certificate = The default is false.
true:]Use mTLS

[if Use mTLS = true:] Client Upload the client key.


Key
Choose  (Browse) and select the file from your download location.

 Note
The supported filename extensions for the key are .pem (privacy-enhanced mail)
and .key.

[if Use mTLS = true:] Client Upload the client certificate.


Certificate
Choose  (Browse) and select the file from your download location.

 Note
The supported filename extensions for the certificate are .pem (privacy-enhanced mail)
and .crt.

Features

Feature Description

Replication Flows Replication Flows are enabled without the need to set any additional connection properties.

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 29
2.1.6 Business Data Product Connections

If your SAP Datasphere tenant is included in an SAP Business Data Cloud formation, Business Data Product
connections are used to connect to other SAP systems included in the same formation.

Business Data Product connections are managed differently compared to other connections:

When your SAP Datasphere tenant is added to an SAP Business Data Cloud formation, the connections to the
source systems of the formation become available in SAP Datasphere with predefined technichal and business
names.

An SAP Datasphere administrator can find the systems and corresponding connections on the  (System)
 (Configuration) Business Data Products page and can change the business name of a connection
if required. For a system and its corresponding connection, the administrator can authorize the spaces to
which data products provided by the connected system can be installed (see Authorize Spaces to Install SAP
Business Data Cloud Data Products).

When a data product is installed to one or more authorized target spaces, the Business Data Product
connection is created in an SAP-managed ingestion space (if it does not already exist). In the ingestion
space, limited functionality is available for the connection, such as validating the connection or changing its
business name (on the Edit dialog). The connection does not support the standard connection features but
only supports the replication flows that have been created and deployed during data product installation.

 Note

Business Data Product connections do not appear in the target spaces, neither in the Connections app nor
in the Source Browser panel of the Data Builder.

Related Information

Integrating Data from SAP Business Data Cloud [page 137]


Evaluating and Installing SAP Business Data Cloud Data Products

2.1.7 Confluent Connections

Use the connection to connect to Apache Kafka hosted on either the Confluent Platform or Confluent Cloud.
The connection type has two endpoints: the Kafka brokers and the Schema Registry.

This topic contains the following sections:

• Supported Features [page 31]


• Configuring Connection Properties For Confluent Platform [page 31]
• Configuring Connection Properties For Confluent Cloud [page 36]

For information about the required prerequisites, see Prepare Connectivity to Confluent.

Integrating Data and Managing Spaces in SAP Datasphere


30 PUBLIC Integrating Data via Connections
Supported Features

Feature Additional Information

Replication Flows You can use the connection to add source and target objects to a replication flow.

 Note
You can only use a non-SAP target for a replication flow if your admin has assigned
capacity units to Premium Outbound Integration. For more information, see Premium
Outbound Integration and Configure the Size of Your SAP Datasphere Tenant.

With Schema Registry, replication flows use a schema reference when serializing a Kafka
topic message instead of providing the schema definition in every message. In the target
settings of the replication flow configuration, you can decide to use Schema Registry.

For more information, see Confluent Kafka Targets.

2.1.7.1 Configuring Connection Properties For Confluent


Platform

Connection Details

Property Description

System Type Select Confluent Platform (default).

Kafka Brokers Enter a comma-separated list of brokers in the format <host>:<port>.

Cloud Connector

Property Description

Use Cloud Connector Set the property to true if your platform is on-premise. The default is false.

[if Use Cloud Connector = Select the location ID for the Cloud Connector instance that is set up for connecting to the
true] Location Kafka brokers.

 Note
To select another location ID than the default location, Connection.Read privilege is
required. The privilege is neither included in the DW Integrator nor in the DW Space
Administrator role. If you need to select a location ID, ask your tenant administrator to
either assign your user to a global role that is based on the DW Administrator role or
to assign your user to a custom global role (with license type SAP Datasphere) that
includes the required Connection.Read privilege.

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 31
Property Description

[if Use Cloud Connector = [optional] Select Derive Virtual Host and Port from Connection Details or Enter Virtual Host
true] Virtual Destination and Port in Separate Fields.

If host and port entered in the connection details match the virtual host and port from the
Cloud Connector configuration, you can select Derive Virtual Host and Port from Connection
Details and don't need to enter the values manually.

[if Virtual Destination = Enter Enter the virtual host that you defined during Cloud Connector configuration.
Virtual Host and Port in
Separate Fields] Virtual Host

[if Virtual Destination = Enter Enter the virtual port that you defined during Cloud Connector configuration.
Virtual Host and Port in
Separate Fields] Virtual Port

Authentication

Property Description

Authentication Type Select the authentication type to use to connect to the Kafka brokers.

You can select:

• No Authentication
• User Name And Password

 Note
For basic authentication with user name and password, we recommend to use TLS
encryption to ensure secure communication.

• Salted Challenge Response Authentication Mechanism (256)


• Salted Challenge Response Authentication Mechanism (512) (default)
• Kerberos with Username and Password
• Kerberos with Keytab File

SASL Authentication Type [read-only] Displays the Simple Authentication and Security Layer (SASL) authentication
mechanism to use depending on your selection for Authentication Type.

The mechanisms are:

• PLAIN: plaintext password defined in RFC 4616.


• SCRAM-256: Salted Challenge Response Authentication Mechanism (SCRAM) pass-
word-based challenge-response authentication from a user to a server -based on SHA
256.
• SCRAM-512: SCRAM password-based challenge-response authentication from a user
to a server - based on SHA 512.
• GSSAPI: Generic Security Service Application Program Interface authentication appli-
cable for Kerberos V5.

Integrating Data and Managing Spaces in SAP Datasphere


32 PUBLIC Integrating Data via Connections
Kerberos
If Authentication Type = Kerberos with Username and Password or Kerberos with Keytab File:

Property Description

Kafka Kerberos Service Name Enter the name of the Kerberos service used by the Kafka broker.

Kafka Kerberos Realm Enter the realm defined for the Kafka Kerberos broker.

Kafka Kerberos Config Upload the content of the krb5.conf configuration file.

Choose  (Browse) and select the file from your download location.

Once uploaded, you can check the configuration file by clicking the  (inspection) button.

Credentials (SCRAM)
If Authentication Type = Salted Challenge Response Authentication Mechanism (512) or Salted Challenge
Response Authentication Mechanism (256):

Property Description

Kafka SASL User Name Enter the user name that is used for Kafka SASL SCRAM authentication.

Kafka SASL Password Enter the Kafka SASL SCRAM connection password.

Credentials (User Name and Password)


If Authentication Type = User Name And Password:

Property Description

User Name Enter the user name that is used for Kafka SASL PLAIN authentication.

Password Enter the Kafka SASL PLAIN connection password.

Credentials (Kerberos with Keytab File)


If Authentication Type = Kerberos with Keytab File:

Property Description

User Name Enter the name of the user that is used to connect to the Kerberos service.

Keytab File Upload the content of the keytab file.

Choose  (Browse) and select the file from your download location.

Once uploaded, you can check the keytab file by clicking the  (inspection) button.

Credentials (Kerberos with User Name And Password)


If Authentication Type = Kerberos with Username and Password:

Property Description

User Name Enter the name of the user that is used to connect to the Kerberos service.

Password Enter the Kerberos password.

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 33
Security

Property Description

Use TLS Select true (default) to use TLS encryption.

Validate Server Certificate Select true (default) to validate the TLS server certificate.

[if Use TLS = true and Select true to use mutual authentication with validating both client and server certificates.
Validate Server Certificate = The default is false.
true:]Use mTLS

[if Use mTLS = true:] Client Upload the client key.


Key
Choose  (Browse) and select the file from your download location.

 Note
The supported filename extensions for the key are .pem (privacy-enhanced mail)
and .key.

[if Use mTLS = true:] Client Upload the client certificate.


Certificate
Choose  (Browse) and select the file from your download location.

 Note
The supported filename extensions for the certificate are .pem (privacy-enhanced mail)
and .crt.

Schema Registry

Property Description

URL Enter the URL of the Schema Registry service. The required format is <protocol>://
<host>:<port>.

Authentication Type Select the authentication type to be used to connect to the Schema Registry.

You can select:

• User Name And Password (default)

 Note
We recommended that you configure Schema Registry to use HTTPS for secure
communication, because the basic protocol passes user name and password in
plain text.

• No Authentication

User Name [if Authentication Type = User Name And Password] Enter the name of the user used to
connect to the Schema Registry.

Password [if Authentication Type = User Name And Password] Enter the password.

Integrating Data and Managing Spaces in SAP Datasphere


34 PUBLIC Integrating Data via Connections
Cloud Connector For Schema Registry

Property Description

Use Cloud Connector Select the location ID for the Cloud Connector instance that is set up for connecting to the
Schema Registry.

 Note
Since Schema Registry might be used in another location than the Kafka brokers, you
have to enter the Cloud Connector properties for Schema Registry separately from the
properties for the Kafka brokers.

[if Use Cloud Connector = [optional] Select a location ID.


true] Location
 Note
To select another location ID than the default location, Connection.Read privilege is
required. The privilege is neither included in the DW Integrator nor in the DW Space
Administrator role. If you need to select a location ID, ask your tenant administrator to
either assign your user to a global role that is based on the DW Administrator role or
to assign your user to a custom global role (with license type SAP Datasphere) that
includes the required Connection.Read privilege.

[if Use Cloud Connector = [optional] Select Derive Virtual Host and Port from Connection Details or Enter Virtual Host
true] Virtual Destination and Port in Separate Fields.

If host and port entered in the connection details match the virtual host and port from the
Cloud Connector configuration, you can select Derive Virtual Host and Port from Connection
Details and don't need to enter the values manually.

[if Virtual Destination = Enter Enter the virtual host that you defined during Cloud Connector configuration.
Virtual Host and Port in
Separate Fields] Virtual Host

[if Virtual Destination = Enter Enter the virtual port that you defined during Cloud Connector configuration.
Virtual Host and Port in
Separate Fields] Virtual Port

Security For Schema Registry

Property Description

Use TLS Select true (default) to use TLS encryption.

Validate Server Certificate Select true (default) to validate the TLS server certificate.

Features

Feature Description

Replication Flows Replication Flows are enabled without the need to set any additional connection properties.

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 35
2.1.7.2 Configuring Connection Properties For Confluent
Cloud

Connection Details

Property Description

System Type Select Confluent Cloud.

Kafka Brokers Enter a comma-separated list of brokers in the format <host>:<port>.

Authentication

Property Description

Authentication Type [read-only] Displays API Key and Secrect for authentication based on API keys.

Credentials (API Key and Secret)

Property Description

API Key Enter the Confluent Cloud API key that is used to control access to Confluent Cloud.

Secret Enter the Confluent Cloud API secrect.

Security

Property Description

Use TLS Select true (default) to use TLS encryption.

Validate Server Certificate Select true (default) to validate the TLS server certificate.

Schema Registry

Property Description

URL Enter the URL of the Schema Registry service. The required format is <protocol>://
<host>:<port>.

Authentication Type Select the authentication type to be used to connect to the Schema Registry.

You must select:

• User Name And Password (default)

 Note
We recommended that you configure Schema Registry to use HTTPS for secure
communication, because the basic protocol passes user name and password in
plain text.

User Name Enter the Confluent Cloud API key that is used to control access to the Schema Registry.

Password Enter the Confluent Cloud API secret.

Integrating Data and Managing Spaces in SAP Datasphere


36 PUBLIC Integrating Data via Connections
Security For Schema Registry

Property Description

Use TLS Select true (default) to use TLS encryption.

Validate Server Certificate Select true (default) to validate the TLS server certificate.

Features

Feature Description

Replication Flows Replication Flows are enabled without the need to set any additional connection properties.

2.1.8 Cloud Data Integration Connections

Use a Cloud Data Integration connection to access data from SAP cloud applications which provide OData-
based APIs for data integration and have a Cloud Data Integration (CDI) provider service implemented.

Prerequisites

See: Prepare Connectivity for Cloud Data Integration

Using the Connection

The connection type supports the remote table as well as the data flow feature.

Supported data access methods for remote tables:

• Remote
• Replication (snapshot)
• Replication (real-time)

For more information, see Replicating Data and Monitoring Remote Tables [page 168].

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 37
Configuring Connection Properties

Connection Details

Property Description

URL Enter the URL with the following syntax: <protocol>://


<host><service path>

Protocol: HTTP or HTTPS. The default value is HTTPS.

Host: Host for accessing the cloud OData service

Service path: Relative path (without host and port) to the


Cloud Data Integration service endpoint (where the CDI pro-
vider service is running at the cloud provider side). The value
must start with a forward slash ( / ).

Root Path [optional] Enter the root path name to restrict browsing to a
certain CDI namespace or provider.

Cloud Connector

Property Description

Use Cloud Connector [optional] Set to true if your source is an on-premise source and you want to use the
connection for data flows. The default is false.

[if Use Cloud Connector = [optional] Select a location ID.


true] Location
 Note
To select another location ID than the default location, Connection.Read privilege is
required. The privilege is neither included in the DW Integrator nor in the DW Space
Administrator role. If you need to select a location ID, ask your tenant administrator to
either assign your user to a global role that is based on the DW Administrator role or
to assign your user to a custom global role (with license type SAP Datasphere) that
includes the required Connection.Read privilege.

[if Use Cloud Connector = [optional] Select Derive Virtual Host and Port from Connection Details or Enter Virtual Host
true] Virtual Destination and Port in Separate Fields.

If host and port entered in the connection details match the virtual host and port from the
Cloud Connector configuration, you can select Derive Virtual Host and Port from Connection
Details and don't need to enter the values manually.

[if Virtual Destination = Enter Enter the virtual host that you defined during Cloud Connector configuration.
Virtual Host and Port in
Separate Fields] Virtual Host

[if Virtual Destination = Enter Enter the virtual port that you defined during Cloud Connector configuration.
Virtual Host and Port in
Separate Fields] Virtual Port

Integrating Data and Managing Spaces in SAP Datasphere


38 PUBLIC Integrating Data via Connections
Authentication

Property Description

Authentication Type Select the authentication type to use to connect to the Cloud
Data Integration service endpoint.

You can select:

• X.509 Client Certificate


• OAuth 2.0
• User Name And Password for basic authentication
• No Authentication for no authentication

The default is X.509 Client Certificate.

Credentials (X.509 Client Certificate)


If Authentication Type = X.509 Client Certificate:

Property Description

X.509 Client Certificate To upload the certificate or certificate chain that is used to authenticate to the remote
system, click  (Browse) and select the file from your download location.

X.509 Client Private Key To upload the private key, click  (Browse) and select the file from your download location.

X.509 Client Private Key [optional] If the private key is encrypted, enter the password required for decryption.
Password

OAuth 2.0
If Authentication Type = OAuth 2.0:

Property Description

OAuth Grant Type Select the grant type.

You can select:

• Client Credentials
• User Name and Password

OAuth Token Endpoint Enter the token endpoint that the application must use.

OAuth Scope Enter the OAuth scope value.

OAuth Resource Enter the OAuth resource.

OAuth Response Type Select the OAuth response type. The values are token or
none.

OAuth Token Request Content Type Select the value for the content-type HTTP header that the
application must use when requesting a token. The values
are URLEncoded or JSON.

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 39
Credentials (OAuth 2.0)
If OAuth Grant Type = Client Credentials:

Property Description

Client ID Enter the client ID.

Client Secret Enter the client secret.

If OAuth Grant Type = User Name And Password:

Property Description

User Name Enter the name of the OAuth user.

Password (OAuth 2.0) Enter the OAuth password.

Credentials (User Name and Password)


If Authentication Type = User Name And Password:

Property Description

User Name Enter the user name that the application must use for authentication.

Password Enter the password for authentication.

Features

To enable Remote Tables, select a Data Provisioning Agent.

Data Flows are enabled without the need to set any additional connection properties.

2.1.9 Generic JDBC Connections

Use a Generic JDBC connection to access data from tables and views in any supported data source for which a
JDBC driver is available.

You can use this connection type to connect to most databases for which SAP Datasphere does not already
provide a connection type. In general, the connection type supports most databases that have SQL-based data
types and functions, and a JDBC driver. For latest information about supported data sources and versions, see
the SAP HANA smart data integration Product Availability Matrix (PAM) .

 Note

For information about unsupported data sources, see SAP Note 3130999 .

Prerequisites

See: Prepare Connectivity for Generic JDBC

Integrating Data and Managing Spaces in SAP Datasphere


40 PUBLIC Integrating Data via Connections
Using the Connection

The connection type supports the remote table feature.

Supported data access methods for remote tables:

• Remote
• Replication (snapshot)

For more information, see Replicating Data and Monitoring Remote Tables [page 168].

Configuring Connection Properties

Connection Details

Property Description

JDBC Driver Class Enter the JDBC driver class for the database you are using.

JDBC URL Enter the URL for the JDBC driver.

 Note
The JDBC URL is stored as plain text and is not stored in
encrypted form. It is visible when editing the connection
or when accessing the remote source with a database
analysis user. Therefore, do not provide any sensitive
information like passwords in the URL.

Credentials

Property Description

User Name Enter the database user name.

Password Enter the password for the database user.

Features

To enable Remote Tables, select a Data Provisioning Agent.

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 41
2.1.10 Generic OData Connections

Use a Generic OData connection to access data from an OData service.

Prerequisites

See: Prepare Connectivity for Generic OData

Using the Connection

The connection type supports the remote table as well as the data flow feature.

Supported data access methods for remote tables:

• Remote
• Replication (snapshot)

For more information, see Replicating Data and Monitoring Remote Tables [page 168].

Configuring Connection Properties

Connection Details

Property Description

URL Enter the OData service provider URL.

 Note
The OData service URL needs to be publicly available.

Version Select the OData version used to implement the OData service (V2 or V4).

Cloud Connector

Property Description

Use Cloud Connector [optional] Set to true if your source is an on-premise source and you want to use the
connection for data flows. The default is false.

Integrating Data and Managing Spaces in SAP Datasphere


42 PUBLIC Integrating Data via Connections
Property Description

[if Use Cloud Connector = [optional] Select a location ID.


true] Location
 Note
To select another location ID than the default location, Connection.Read privilege is
required. The privilege is neither included in the DW Integrator nor in the DW Space
Administrator role. If you need to select a location ID, ask your tenant administrator to
either assign your user to a global role that is based on the DW Administrator role or
to assign your user to a custom global role (with license type SAP Datasphere) that
includes the required Connection.Read privilege.

[if Use Cloud Connector = [optional] Select Derive Virtual Host and Port from Connection Details or Enter Virtual Host
true] Virtual Destination and Port in Separate Fields.

If host and port entered in the connection details match the virtual host and port from the
Cloud Connector configuration, you can select Derive Virtual Host and Port from Connection
Details and don't need to enter the values manually.

[if Virtual Destination = Enter Enter the virtual host that you defined during Cloud Connector configuration.
Virtual Host and Port in
Separate Fields] Virtual Host

[if Virtual Destination = Enter Enter the virtual port that you defined during Cloud Connector configuration.
Virtual Host and Port in
Separate Fields] Virtual Port

Authentication

Property Description

Authentication Type Select the authentication type to use to connect to the


OData endpoint.

You can select:

• No Authentication
• User Name And Password for basic authentication
• OAuth 2.0 (default value)

OAuth 2.0
If Authentication Type = OAuth 2.0:

Property Description

OAuth Grant Type Displays Client Credentials as grant type used to retrieve an
access token.

OAuth Token Endpoint Enter the API endpoint to use to request an access token.

OAuth Scope [optional] Enter the OAuth scope, if applicable.

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 43
Property Description

OAuth Token Request Content Type [optional] Enter the content type of the OAuth2 token re-
quest.

You can select:

• URL Encoded (default value)


OAuth2 token request parameters will be url-encoded
and included in the HTTP request body.
• JSON
OAuth2 token request parameters will be in JSON for-
mat and included in the HTTP request body.

Credentials (OAuth 2.0)


If Authentication Type = OAuth 2.0:

Property Description

Client ID Enter the client ID to authenticate SAP Datasphere to the


authorization server.

Client Secret Enter the client secret to authenticate SAP Datasphere to


the authorziation server.

Credentials (User Name And Password)


If Authentication Type = User Name And Password:

Property Description

User Name Enter the user name.

Password Enter the password.

Features

Remote Tables are enabled without the need to set any additional connection properties.

Data Flows are enabled without the need to set any additional connection properties. If your source is an
on-premise source, make sure you have maintained the properties in the Cloud Connector section.

Enter Custom HTTP Headers

Different OData services support different sets of custom headers to pass parameters to the service when
using the connection. You can specify one or more custom HTTP headers to add them to the OData request.

Click Add to add name and value for a custom HTTP header field.

 Note

• Security-relevant headers such as "authorization" are not allowed.

Integrating Data and Managing Spaces in SAP Datasphere


44 PUBLIC Integrating Data via Connections
2.1.11 Generic SFTP Connections

Use a Generic SFTP connection to access files on a Secure File Transfer Protocol (SFTP) server.

Using the Connection

The connection type supports the data flow feature.

Supported file types:

• JSON, JSONL
• Parquet
• ORC
• CSV
• XLS, XLSX

For information about the required prerequisites, see Prepare Connectivity for Generic SFTP.

Configuring Connection Properties

Connection Details

Property Description

Category Select the category. Select Cloud if you want to connect to an SFTP server in the public
cloud, or select On-Premise if you want to connect to an SFTP server in your local network.

Host Enter the host name of the SFTP server.

Port Enter the port number of the SFTP server.

Host Key Enter the public SSH host key (public key of the SFTP server, not the key fingerprint).

Choose  (Browse) and select the file from your download location.

If you don’t know your host key, ask your administrator or use command line tools like ssh-
keyscan to obtain the host key (only through a trustable channel). For more information, see
Prepare Connectivity for Generic SFTP.

Root Path [optional] Enter the root path name for browsing objects. The value starts with the charac-
ter slash. For example, /My Folder/MySubfolder.

If you have entered root path, then any path used with this connection is prefixed with the
root path.

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 45
Cloud Connector

Property Description

Use Cloud Connector [optional] Set to true if your source is an on-premise source and you want to use the
connection for data flows. The default is false.

[if Use Cloud Connector = [optional] Select a location ID.


true] Location
 Note
To select another location ID than the default location, Connection.Read privilege is
required. The privilege is neither included in the DW Integrator nor in the DW Space
Administrator role. If you need to select a location ID, ask your tenant administrator to
either assign your user to a global role that is based on the DW Administrator role or
to assign your user to a custom global role (with license type SAP Datasphere) that
includes the required Connection.Read privilege.

[if Use Cloud Connector = [optional] Select Derive Virtual Host and Port from Connection Details or Enter Virtual Host
true] Virtual Destination and Port in Separate Fields.

If host and port entered in the connection details match the virtual host and port from the
Cloud Connector configuration, you can select Derive Virtual Host and Port from Connection
Details and don't need to enter the values manually.

[if Virtual Destination = Enter Enter the virtual host that you defined during Cloud Connector configuration.
Virtual Host and Port in
Separate Fields] Virtual Host

[if Virtual Destination = Enter Enter the virtual port that you defined during Cloud Connector configuration.
Virtual Host and Port in
Separate Fields] Virtual Port

Authentication

Property Description

Authentication Type Select the authentication type to be used.

Credentials (SSH)
If Authentication Type = SSH:

Property Description

User Name Enter the user who is accessing the SFTP server.

Private Key Enter the user SSH private key used for SSH key authentication. The server must know the
user SSH public key.

Choose  (Browse) and select the file from your download location.

Passphrase Enter the passphrase needed to decrypt the private key in SSH key authentication.

Integrating Data and Managing Spaces in SAP Datasphere


46 PUBLIC Integrating Data via Connections
Credentials (User Name and Password)

If Authentication Type = User Name and Password:

Property Description

User Name Enter the user accessing the SFTP server.

Password Enter the user's password used for authentication.

Features

Data Flows are enabled without the need to set any additional connection properties. If your source is an
on-premise source, make sure you have maintained the properties in the Cloud Connector section.

2.1.12 Google BigQuery Connections

Use the connection to connect to and access data from Google BigQuery.

This topic contains the following sections:

• Supported Features [page 47]


• Configuring Connection Properties [page 48]

For information about the required prerequisites, see Prepare Connectivity to Google BigQuery.

Supported Features

Feature Additional Information

Remote Tables You can use remote tables imported from the connection either to access data directly live
in the connected source (federation) or to copy the data into SAP Datasphere (snapshot
replication).

For more information, see Replicating Data and Monitoring Remote Tables [page 168].

Data Flows You can use the connection to add source objects to a data flow.

Replication Flows You can use the connection to add target objects to a replication flow.

For more information, see Google BigQuery Targets.

 Note
You can only use a non-SAP target for a replication flow if your admin has assigned
capacity units to Premium Outbound Integration. For more information, see Premium
Outbound Integration and Configure the Size of Your SAP Datasphere Tenant.

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 47
Configuring Connection Properties

Connection Details

Property Description

Project Enter the ID of the Google Cloud project to which you want
to connect. You can find the project ID in the json key file that
you need for the credentials.

Location [optional] Enter an additional location. By default, SAP


Datasphere connects to BigQuery datasets only from Goo-
gle's default location - US. Datasets from any other location
will only be available in the Data Builder if you enter the
additional location here.

A location can be a region, for example us-west1or


asia-east1, or a multi-region, for example EU.

 Note
If you provide an invalid location, the connection valida-
tion still passes, but failures will happen when you're
using the connection in the Data Builder.

Credentials

Property Description

Key Upload the json key file that is used for authentication.

Choose  (Browse) and select the file from your download


location.

Features

Feature Description

Remote Tables Remote Tables are enabled without the need to set any additional connection properties.

Data Flows Data Flows are enabled without the need to set any additional connection properties.

Replication Flows Replication Flows are enabled without the need to set any additional connection properties.

2.1.13 Google Cloud Storage Connections

Use the connection to connect to and access objects from Google Cloud Storage.

This topic contains the following sections:

• Supported Features [page 49]


• Configuring Connection Properties [page 49]

Integrating Data and Managing Spaces in SAP Datasphere


48 PUBLIC Integrating Data via Connections
Supported Features

Feature Additional Information

Data Flows You can use the connection to add source objects to a data flow.

Replication Flows You can use the connection to add source and target objects to a replication flow.

For more information, see:

• Cloud Storage Provider Sources


• Cloud Storage Provider Targets

 Note
You can only use a non-SAP target for a replication flow if your admin has assigned
capacity units to Premium Outbound Integration. For more information, see Premium
Outbound Integration and Configure the Size of Your SAP Datasphere Tenant.

Configuring Connection Properties

Connection Details

Property Description

Project Enter the ID of the Google Cloud Storage project to which


you want to connect.

Root Path [optional] Enter the root path name for browsing objects.
The value starts with the character slash. For example, /My
Folder/MySubfolder.

If you have specified the root path, then any path used with
this connection is prefixed with the root path.

Credentials

Property Description

Key Enter the content of the json key file that is used for authen-
tication.

Choose  (Browse) and select the file from your download


location.

Features

Feature Description

Data Flows Data Flows are enabled without the need to set any additional connection properties.

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 49
Feature Description

Replication Flows Replication Flows are enabled without the need to set any additional connection properties.

2.1.14 Hadoop Distributed File System Connections

Use the connection to access objects from an Hadoop Distributed File System (HDFS) server.

Using the Connection

The connection type supports the data flow feature.

Configuring Connection Properties

Connection Details

Property Description

Endpoint Select RPC or WEBHDFS

 Note
Along with Remote Procedure Call (RPC), HDFS can also extend connections with
WebHDFS.

If you select RPC, make sure that both the NameNode and the DataNodes are exposed
and can be connected from SAP Datasphere. For more information, see Finding SAP
Datasphere IP addresses.

Host Enter the hostname or the IP address of the HDFS namenode.

Port Enter the port of the HDFS namenode. If you do not provide any value, the default value for
the selected protocol is used.

Root Path [optional] Enter the root path name for browsing objects. The value starts with the charac-
ter slash. For example, /My Folder/MySubfolder.

If you have specified the root path, then any path used with this connection is prefixed with
the root path.

Custom Parameters

Property Description

Use Custom Parameters Select true to use an HDFS custom parameter.

Integrating Data and Managing Spaces in SAP Datasphere


50 PUBLIC Integrating Data via Connections
Property Description

[if Use Custom Parameters = Enter the HDFS custom parameter.


true] HDFS Configuration

[if Use Custom Parameters = Enter a value.


true] Value

Authentication

Property Description

Authentication Type Select the authentication type that the application must use.

To connect to HDFS the application supports User Name And Password, Simple, or Kerberos
authentication mechanisms.

Credentials (User Name and Password)


If Authentication Type = User Name And Password:

Property Description

User Name Enter the user name of the HDFS user.

Password Enter the password of the HDFS user.

Credentials (Simple)
If Authentication Type = Simple:

Property Description

User Name Enter the user name of the HDFS user.

Credentials (Kerberos)
If Authentication Type = Kerberos:

Property Description

User Name Enter the user principal.

keytab Enter the content of the keytab file of the user from the local system.

Choose  (Browse) and select the file from your download location.

Configuration
If Authentication Type = Kerberos:

Property Description

krb5.config Enter the content of the krb5.conf configuration file from your local system.

Choose  (Browse) and select the file from your download location.

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 51
Features

Data Flows are enabled without the need to set any additional connection properties.

2.1.15 Microsoft Azure Blob Storage Connections

Use the connection to access objects in Microsoft Azure Blob Storage.

Using the Connection

The connection type supports the data flow feature.

Configuring Connection Properties

Connection Details

Property Description

Protocol Select the protocol (wasb for unencrypted access or wasbs for TLS-encrypted access). The
default is wasbs.

Account Name Enter the name of the Microsoft Azure storage account through which Microsoft Azure Blob
Storage is provisioned.

Root Path [optional] Enter the root path name for browsing objects. The value starts with the charac-
ter slash. For example, /My Folder/MySubfolder.

If you have entered root path, then any path used with this connection is prefixed with the
root path.

Endpoint Suffix Displays the default value for the endpoint suffix which is "core.windows.net". You can
override the default value according to the domain assigned to the endpoint of your blob
service.

Credentials (Shared Key)

Property Description

Account Key Enter the storage account access key (shared key authorization).

Features

Data Flows are enabled without the need to set any additional connection properties.

Integrating Data and Managing Spaces in SAP Datasphere


52 PUBLIC Integrating Data via Connections
2.1.16 Microsoft Azure Data Lake Store Gen1 Connections
(Deprecated)

Use the connection to access objects in Microsoft Azure Data Lake Gen1 (ADL Gen1).

 Caution

Microsoft Azure Data Lake Storage Gen1 will be retired on February


29, 2024 (see https://azure.microsoft.com/en-us/updates/action-required-switch-to-azure-data-lake-
storage-gen2-by-29-february-2024/ ). After this date, it might not be possible to access data with Gen1
connections. Therefore, the connection type Microsoft Azure Data Lake Storage Gen1 is deprecated and
may be removed in future releases.

After your Microsoft Azure Data Lake Storage Gen1 has been migrated to Microsoft Azure Data Lake
Storage Gen2, recreate existing Gen1 connections with the Gen2 connection type and adapt your models
and processes to the new connections.

Using the Connection

The connection type supports the data flow feature.

Configuring Connection Properties

Connection Details

Property Description

Storage Account Name Enter the name of the Azure Data Lake Storage Gen1 account.

Tenant ID Enter the ID of the Azure Data Lake Storage Gen1 tenant.

Root Path [optional] Enter the root path name for browsing. It starts with a slash and the file system
name. For example /MyFileSystem/MyFolder. The file system must be provided.

Credentials

Property Description

Client ID Enter the client ID.

Client Key Enter the client key (also referred to as client secret or authentication key).

Features

Data Flows are enabled without the need to set any additional connection properties.

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 53
2.1.17 Microsoft Azure Data Lake Store Gen2 Connections

Use the connection to connect to and access objects in Microsoft Azure Data Lake Gen2 (ADL Gen2).

This topic contains the following sections:

• Supported Features [page 54]


• Configuring Connection Properties [page 54]

For information about the required prerequisites, see Prepare Connectivity to Microsoft Azure Data Lake Store
Gen2.

Supported Features

Feature Additional Information

Data Flows You can use the connection to add source objects to a data flow.

Replication Flows You can use the connection to add source and target objects to a replication flow.

For more information, see:

• Cloud Storage Provider Sources


• Cloud Storage Provider Targets

 Note
You can only use a non-SAP target for a replication flow if your admin has assigned
capacity units to Premium Outbound Integration. For more information, see Premium
Outbound Integration and Configure the Size of Your SAP Datasphere Tenant.

Configuring Connection Properties

Connection Details

Property Description

Storage Account Name Enter the name of the Azure Data Lake Storage Gen2 used for authentication.

Root Path [optional] Enter the root path name for browsing. It starts with a slash and the file system
name. For example /MyFileSystem/MyFolder. The file system must be provided. Any
path used with this connection will be prefixed with this root path.

Integrating Data and Managing Spaces in SAP Datasphere


54 PUBLIC Integrating Data via Connections
Authentication

Property Description

Authentication Type Select the authentication type to be used:

You can select:

• Shared Key (default)


• Shared Access Signature
• OAuth 2.0

Shared key provides full access to your storage account while with shared access signature
you can provide secure delegate access to the storage account resources.

Credentials (Shared Key)

If Authentication Type = Shared Key:

Property Description

Account Key Enter the account key used in the shared key authorization.

Credentials (Shared Access Signature)

If Authentication Type = Shared Access Signature:

Property Description

SAS Token Enter the shared access signature token (SAS token) used in shared access signature
authentication.

OAuth 2.0

If Authentication Type = OAuth 2.0:

Property Description

OAuth Grant Type Select the grant type.

You can select:

• Client Credentials with X.509 Client Certificate


• Client Credentials
• User Name and Password

[if OAuth Grant Type = Enter the token endpoint that the application must use to get the access token.
Client Credentials with X.509
Client Certificate or Client
Credentials] OAuth Token
Endpoint

[if OAuth Grant Type = User Enter the client endpoint to get the access token for authorization method User Name and
Name and Password] OAuth Password.
Client Endpoint

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 55
Credentials (OAuth 2.0 with X.509 Client Certificate)

Property Description

Client ID Enter the client ID.

X.509 Client Certificate To upload the certificate or certificate chain that is used to
authenticate to the remote system, click  (Browse) and
select the file from your download location.

X.509 Client Private Key To upload the private key, click  (Browse) and select the
file from your download location.

Credentials (OAuth 2.0)

If OAuth Grant Type = Client Credentials:

Property Description

Client ID Enter the client ID.

Client Secret Enter the client secret.

If OAuth Grant Type = User Name And Password:

Property Description

User Name Enter the name of the OAuth user.

Password (OAuth 2.0) Enter the OAuth password.

Features

Feature Description

Data Flows Data Flows are enabled without the need to set any additional connection properties.

Replication Flows Replication Flows are enabled without the need to set any additional connection properties.

2.1.18 Microsoft Azure SQL Database Connections

Use a Microsoft Azure SQL Database connection to access table data from a Microsoft Azure SQL database.

This topic contains the following sections:

• Supported Features [page 57]


• Configuring Connection Properties [page 57]

For information about the required prerequisites, see Prepare Connectivity to Microsoft Azure SQL Database.

Integrating Data and Managing Spaces in SAP Datasphere


56 PUBLIC Integrating Data via Connections
Supported Features

Feature Additional Information

Remote Tables You can use remote tables imported from the connection either to access data directly live
in the connected source (federation) or to copy the data into SAP Datasphere (replication).

For remote tables, real-time replication is supported. For information about any constraints,
see Replicate Data Changes in Real-Time [page 181].

Data Flows You can use the connection to add source objects to a data flow.

Replication Flows You can use the connection to add source objects to a replication flow.

 Note

The connection type supports replication with both remote tables via Data Provisioning Agent (SAP HANA
Smart Data Integration) and replication flows.

Whether you use the same connection or two separate connections, we strongly recommend using only
remote tables or only replication flows for replicating data from a dedicated source object, but not both.

Configuring Connection Properties

Connection Details

Property Description

Server Name Enter the host name of the Azure server.

Port Enter the port number of the Azure server. The default is 1433.

Database Name Enter the name of the database to which you want to connect.

Version Select the version.

Security

Property Description

Host Name in Server Enter *.database.windows.net .


Certificate

Credentials

Property Description

User Name Enter the name of the Azure user that has privileges to connect to the database.

Password Enter the password of the Azure user.

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 57
Features

Feature Description

Remote Tables To enable Remote Tables, select a Data Provisioning Agent.

Data Flows Data Flows are enabled without the need to set any additional connection properties.

Replication Flows Replication Flows are enabled without the need to set any additional connection properties.

2.1.19 Microsoft SQL Server Connections

Use a Microsoft SQL Server connection to access data from a Microsoft SQL Server database (on-premise).

Prerequisites

See: Prepare Connectivity to Microsoft SQL Server

Using the Connection

The connection type supports the remote table as well as the data flow feature.

Supported data access methods for remote tables:

• Remote
• Replication (snapshot)
• Replication (real-time)

 Note

SAP Datasphere uses trigger-based replication. For more information, see Microsoft SQL Server
Trigger-Based Replication in the SAP HANA Smart Data Integration and SAP HANA Smart Data Quality
Installation and Configuration Guide.

For more information, see Replicating Data and Monitoring Remote Tables [page 168].

Integrating Data and Managing Spaces in SAP Datasphere


58 PUBLIC Integrating Data via Connections
Configuring Connection Properties

Connection Details

Property Description

Server Name Enter the Microsoft SQL Server name.

 Note
Connecting via instance name is not supported.

Port Enter the Microsoft SQL Server port number. The value range is 1–65535.

 Note
Dynamic ports are not supported.

Database Name Enter the Microsoft SQL Server database name.

Version Select the Microsoft SQL Server version. Supported versions are Microsoft SQL Server
2012, 2014, 2016, 2017, 2019, and 2022 (default is 2022).

 Note
Microsoft SQL Server 2012 is not supported for remote tables.

Cloud Connector

 Note

Cloud Connector is not required if your Microsoft SQL Server database is available on the public internet.

Property Description

Use Cloud Connector [optional] Set to true if your source is an on-premise source and you want to use the
connection for data flows. The default is false.

[if Use Cloud Connector = [optional] Select a location ID.


true] Location
 Note
To select another location ID than the default location, Connection.Read privilege is
required. The privilege is neither included in the DW Integrator nor in the DW Space
Administrator role. If you need to select a location ID, ask your tenant administrator to
either assign your user to a global role that is based on the DW Administrator role or
to assign your user to a custom global role (with license type SAP Datasphere) that
includes the required Connection.Read privilege.

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 59
Property Description

[if Use Cloud Connector = [optional] Select Derive Virtual Host and Port from Connection Details or Enter Virtual Host
true] Virtual Destination and Port in Separate Fields.

If host and port entered in the connection details match the virtual host and port from the
Cloud Connector configuration, you can select Derive Virtual Host and Port from Connection
Details and don't need to enter the values manually.

[if Virtual Destination = Enter Enter the virtual host that you defined during Cloud Connector configuration.
Virtual Host and Port in
Separate Fields] Virtual Host

[if Virtual Destination = Enter Enter the virtual port that you defined during Cloud Connector configuration.
Virtual Host and Port in
Separate Fields] Virtual Port

Security

Property Description

Use SSL Select whether you’re using SSL. The default value is true.

[if Use SSL = true] Host Enter the host name that is in the SSL certificate.
Name in Certificate

Credentials

Property Description

User Name Enter the Microsoft SQL Server user name.

Password Enter the Microsoft SQL Server user password.

Features

To enable Remote Tables, select a Data Provisioning Agent.

Data Flows are enabled without the need to set any additional connection properties. Make sure you have
maintained the properties in the Cloud Connector section.

Advanced Connection Properties

When you've selected a Data Provisioning Agent, you can configure advanced connection properties to enable
specific customer scenarios. In the connection creation wizard, you configure advanced properties in an
additional wizard step. When editing the connection to configure the properties, scroll down and click the Show
Advanced Properties button. Advanced properties have default values that you can override according to your
customer scenario’s needs.

Available properties:

• Additional JDBC Connection Properties


• Include Table and Columns Remarks
• Allowlist Table in Remote Database

Integrating Data and Managing Spaces in SAP Datasphere


60 PUBLIC Integrating Data via Connections
• Schema Alias
• Schema Alias Replacement
• Use Windows Authentication
• Schema to Create System Objects
• Enable ABAP Manageable Trigger Namespace
• [if Enable ABAP Manageable Trigger Namespace = true] ABAP Manageable Trigger Namespace
• Connection Pool Size
• Minimum Scan Interval in Seconds
• Maximum Scan Interval in Seconds
• Maximum Batch Size
• Batch Queue Size
• Maximum Transaction Count in Scan
• Maximum Scan Size
• Enable Compound Triggers
• Triggers Record Primary Keys Only
• [if Triggers Record Primary Keys Only = true] Triggers Capture Before and After Images
• [if Enable Compound Triggers = false and Triggers Record Primary Keys Only = false] Transmit Data in
Compact Mode
• Enable Transaction Merge

For more information, see Microsoft SQL Server Log Reader Remote Source Configuration in the SAP HANA
Smart Data Integration and SAP HANA Smart Data Quality Installation and Configuration Guide.

2.1.20 Open Connectors Connections

Use an Open Connectors connection to access data from sources that are connected to the SAP Open
Connectors account that is integrated with your space.

Leveraging SAP Open Connectors for SAP Datasphere Connectivity

You can use the broad SAP Open Connectors connectivity available from SAP Business Technology Platform
directly within your SAP Datasphere space.

 Note

Using SAP Open Connectors may incur additional charges. You can review your usage in your SAP BTP
cockpit.

With SAP Open Connectors, (API) hubs provide uniform APIs to access a collection of resources enabling
one-to-many API access to multiple API providers. With integrating SAP Open Connectors SAP Datasphere
gets access to third-party cloud applications that are available with SAP Open Connectors hub categories.
Each hub category contains connectors each representing a pre-built API integration that enables a connection
into a specific API Provider endpoint, for example Salesforce, Snowflake, or Microsoft SharePoint. For any

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 61
connector, connector instances represent a single authenticated connection between SAP and an account at
the API provider.

SAP Datasphere supports the following SAP Open Connectors hub categories:

• CRM
• DB
• General
• Marketing
• Social
• Documents

 Note

• For information about SAP Open Connectors availability in data centers, see SAP Note 2903776 .
• Some SAP Open Connectors types might not work in case the API hasn't been fully implemented.
We recommend to check the connector on SAP Open Connectors side first before using it in
SAP Datasphere. For more information, see Working with Connectors in the SAP Open Connectors
documentation.
• For sources which SAP Datasphere natively supports with dedicated connection types, such as
Amazon S3 or Microsoft SQL Server, use the corresponding SAP Datasphere connection type. We
do not recommend using SAP Open Connectors connectivity in this case.
• Input parameters are not supported for data flows. For resources with mandatory parameters the
parameters have to be replaced in SAP Open Connectors, or the respective connector has to be
cloned and adjusted to work without any mandatory parameter. For more information, see Setup
Configuration and Parameters (for custom connectors) in the SAP Open Connectors documentation.

Prerequisites

See: Prepare Connectivity to SAP Open Connectors

Using the Connection

The connection type supports the data flow feature.

Configuring Connection Properties

Connection Details

Property Description

Open Connectors Instance Select the instance to which you want to connect.

Integrating Data and Managing Spaces in SAP Datasphere


62 PUBLIC Integrating Data via Connections
Features

Data Flows are enabled without the need to set any additional connection properties.

2.1.21 Oracle Connections

Use the connection to connect to and access data from an Oracle database (on-premise).

This topic contains the following sections:

• Supported Features [page 63]


• Configuring Connection Properties [page 63]

For information about the required prerequisites, see Prepare Connectivity to Oracle.

Supported Features

Feature Additional Information

Remote Tables You can use remote tables imported from the connection either to access data directly live
in the connected source (federation) or to copy the data into SAP Datasphere (replication).

For remote tables, real-time replication is supported. For information about any constraints,
see Replicate Data Changes in Real-Time [page 181].

 Note
SAP Datasphere uses trigger-based replication. For more information, see Oracle Trig-
ger-Based Replication in the SAP HANA Smart Data Integration and SAP HANA Smart
Data Quality Installation and Configuration Guide.

Data Flows You can use the connection to add source objects to a data flow.

Configuring Connection Properties

Connection Details

Property Description

Host Enter the host name or IP address on which the remote Oracle database is running.

Port Enter the Oracle database server port number.

 Note
The port number can vary depending on whether SSL is used or not.

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 63
Property Description

Database Name/SID Enter the Oracle database name

Service Name Enter the service name of Oracle database. When creating a remote source, you must set
only one of the following parameters: Database Name and Service Name.

If you set both, the Data Provisioning Agent connects to Oracle by the service name as the
first choice.

Version Select a version. Supported versions are Oracle 12c, Oracle 18c, and Oracle 19c. Default
version is Oracle 19c.

Cloud Connector

 Note

Cloud Connector is not required if your Oracle database is available on the public internet.

Property Description

Use Cloud Connector [optional] Set to true if your source is an on-premise source and you want to use the
connection for data flows. The default is false.

[if Use Cloud Connector = [optional] Select a location ID.


true] Location
 Note
To select another location ID than the default location, Connection.Read privilege is
required. The privilege is neither included in the DW Integrator nor in the DW Space
Administrator role. If you need to select a location ID, ask your tenant administrator to
either assign your user to a global role that is based on the DW Administrator role or
to assign your user to a custom global role (with license type SAP Datasphere) that
includes the required Connection.Read privilege.

[if Use Cloud Connector = [optional] Select Derive Virtual Host and Port from Connection Details or Enter Virtual Host
true] Virtual Destination and Port in Separate Fields.

If host and port entered in the connection details match the virtual host and port from the
Cloud Connector configuration, you can select Derive Virtual Host and Port from Connection
Details and don't need to enter the values manually.

[if Virtual Destination = Enter Enter the virtual host that you defined during Cloud Connector configuration.
Virtual Host and Port in
Separate Fields] Virtual Host

[if Virtual Destination = Enter Enter the virtual port that you defined during Cloud Connector configuration.
Virtual Host and Port in
Separate Fields] Virtual Port

Integrating Data and Managing Spaces in SAP Datasphere


64 PUBLIC Integrating Data via Connections
Security

Property Description

Use SSL Select whether you’re using SSL. The default value is true.

 Note
If you select SSL and you want to use the connection for data flows, in addition to
the ODBC driver file, additional files must have been uploaded to SAP Datasphere (see
Upload Third-Party ODBC Drivers (Required for Data Flows)).

[if Use SSL = true] Enter the distinguished name (DN) of the primary data server certificate.
Distinguished Name (DN) in
Certificate  Note
If this parameter is set, the DN field in the server certificate is verified to match this
parameter. If it doesn’t match, the connection to the primary data server fails.

Credentials

Property Description

User Name Enter the Oracle user name (case-sensitive)

Password Enter the Oracle user password

Features

To enable Remote Tables, select a Data Provisioning Agent.

Data Flows are enabled without the need to set any additional connection properties.

Advanced Connection Properties

When you've selected a Data Provisioning Agent, you can configure advanced connection properties to enable
specific customer scenarios. In the connection creation wizard, you configure advanced properties in an
additional wizard step. When editing the connection to configure the properties, scroll down and click the Show
Advanced Properties button. Advanced properties have default values that you can override according to your
customer scenario’s needs.

Available properties:

• Map INTEGER/NUMBER to DECIMAL(38,0)


• Time Zone Format
• Use Oracle TNSNAMES File
• Allowlist Table in Remote Database
• Use LDAP Authentication
• Schema Alias
• Schema Alias Replacement
• Include Table and Columns Remarks

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 65
• Enable ABAP Manageable Trigger Namespace
• Sequence Cache Size
• Connection Pool Size
• Minimum Scan Interval in Seconds
• Maximum Scan Interval in Seconds
• Maximum Batch Size
• Batch Queue Size
• Maximum Transaction Count in Scan
• Maximum Scan Size
• Enable Compound Triggers
• [if Enable Compound Triggers = true] Flush Threshold of Compound Triggers
• [if Enable Compound Triggers = false] Triggers Record Primary Keys Only
• [if Enable Compound Triggers = false and Triggers Record Primary Keys Only = false] Triggers Record ROWID
Pseudo Column Only
• [if Enable Compound Triggers = false and Triggers Record Primary Keys Only = true] Triggers Capture Before
and After Images
• [if Enable Compound Triggers = false and Triggers Record Primary Keys Only = false] Merge Before and After
Images of Updates into One Row
• [if Triggers Record Primary Keys Only = false and Triggers Record ROWID Pseudo Column Only = false]
Triggers Record LOB Values
• Enable Shadow Table Partition
• Enable Trigger Queue Table Partition
• Transmit Data in Compact Mode
• Enable Transaction Merge

For more information, see Oracle Log Reader Remote Source Configuration in the SAP HANA Smart Data
Integration and SAP HANA Smart Data Quality Installation and Configuration Guide.

2.1.22 Precog Connections

Extend connectivity beyond SAP Datasphere standard remote connectivity and cover additional data sources
that are available with Precog.

Precog provides connectivity to a variety of data sources, including marketing data sources, databases and
others.

When creating a connection to a partner tool, an Open SQL schema is generated in SAP Datasphere. To be
able to send data to the Open SQL schema, the IP addresses of the partner tool need to be added to the
SAP Datasphere IP allowlist, and during connection creation the partner tool on their side needs to establish a
connection to the schema.

For more information, see Connections to Partner Tools [page 122].

Integrating Data and Managing Spaces in SAP Datasphere


66 PUBLIC Integrating Data via Connections
Prerequisites

See: Prepare Connectivity to Precog

Using the Connection

The connection type supports creating views and entity-relationship models.

Configuring Partner Connection

Wizard step Configure partner settings shows an embedded Precog UI. Here, you log on to the regional Precog
server and perform the following steps:

1. Enter the URL for your regional Precog server.


2. Choose Next.

 Note

As a prerequisite to proceed to the next step in the wizard, your browser needs to be enabled to store
cookies. This is not the case if you run the browser in incognito mode.

3. Enter your credentials.


4. Click Send to move to the next connection creation wizard step in SAP Datasphere.

For more information, see Create a Connection [page 16].

What happens in Precog after you have created the connection

You can now see the destination with the name of the generated Open SQL schema in Precog and start loading
data for the tables available with Precog.

2.1.23 SAP ABAP Connections

Use an SAP ABAP connection to access data from SAP ABAP on-premise systems through RFC or to access
data from cloud source systems such as SAP S/4HANA Cloud through Web Socket RFC.

This topic contains the following sections:

• Supported Features [page 68]


• Configuring Connection Properties [page 72]

For information about the required prerequisites, see Prepare Connectivity to SAP ABAP Systems.

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 67
Supported Features

Feature Additional Information

Remote Tables You can use remote tables imported from the connection either to access data directly live
in the connected source (federation) or to copy the data into SAP Datasphere (replication).

You can access the following data:

• Data from virtual tables through RFC for ODP sources (extractors):
• Extraction context ABAP_CDS provides access to ABAP Core
Data Services (CDS) Views that include the annotation
@Analytics.dataextraction.enabled: true (in SAP ABAP on-prem-
ise systems) and do not contain the annotation @Analytics.Query: true.
For information about which ABAP CDS Views are available for extraction, see CDS
Views Enabled for Data Extraction in the SAP S/4HANA documentation.

 Note
For SAP S/4HANA on-premise, ABAP CDS Views are the preferred method
over SAPI extractors when integrating data with SAP Datasphere.

• Extraction context BW provides access to InfoProviders in SAP BW or SAP


BW∕4HANA systems:
• DataStore objects (advanced or classic)
• CompositeProviders
• Queries as InfoProviders (with a subset of OLAP capabilities) - can be enabled
in the runtime properties of queries in the BW Modeling tools
• InfoObjects (characteristics) - can be enabled in the general settings of the
InfoObject editor in the BW Modeling tools

 Note
For accessing CompositeProviders and Queries in SAP BW∕4HANA we
strongly recommended to use the model import with the specific connec-
tion type SAP BW/4HANA Model Transfer. For more information, see SAP
BW∕4HANA Model Transfer Connections [page 83].

 Note
To import BW Hierarchies and deploy them as remote tables in SAP
Datasphere, use the /BI*/H<InfoObject name> tables with data access via
SAP ABAP Dictionary tables instead of using ODP extraction context BW. In
the Data Builder, you can find the tables in the ABAPTABLES folder of the
connection.

• Extraction context SAPI provides access to Service API (SAPI) DataSources

Integrating Data and Managing Spaces in SAP Datasphere


68 PUBLIC Integrating Data via Connections
Feature Additional Information

 Note
Importing SAPI DataSources into SAP Datasphere is not supported if the
DataSource contains mandatory selection fields. You will receive an error mes-
sage when trying to import tables for such DataSources.

• Data from SAP ABAP Dictionary tables in SAP ABAP on-premise systems

Real-time replication is supported for ODP sources. For information about any constraints,
see Replicate Data Changes in Real-Time [page 181].

 Note
Certain SAPI DataSources (from FI-AA Business Content, for example) may send
duplicate records during snapshot replication or real-time initialization. Remote table
replication runs with optimized INSERT processing on DataSource key level which leads
to unique constraint violation errors. In such cases, we recommend to use a data flow
which uses the remote table as source and for which the target table has selected the
Append mode with the Update Records By Primary Key (UPSERT) option.

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 69
Feature Additional Information

Data Flows You can use the connection to add source objects to a data flow.

You can replicate the following data:

• extraction-enabled ABAP CDS views that are C1-released, that is views with annotation
@Analytics.dataextraction.enabled: true and that are available in the
connected system (access via ABAP Pipeline Engine)
For more information, see:
• in the SAP S/4HANA Cloud documentation: Data Integration and CDS Views Ena-
bled for Data Extraction
• in the SAP S/4HANA documentation: CDS Views Enabled for Data Extraction
• Tables from SAP Landscape Transformation Replication Server (SAP LT Replication
Server) sources, for which an appropriate configuration has been created in SAP LT
Replication Server. In the Sources of the data flow editor, you then can find the source
objects in an SLT - SAP LT Replication Server folder of the connection.

 Note
• The availability of the data flow feature depends on the used version and Support
Package level of SAP S/4HANA or the DMIS addon in the source. Make sure your
source systems meet the required minimum versions. We recommend to use the
latest available version of SAP S/4HANA and the DMIS add-on where possible and
have the latest SAP notes and TCI notes implemented in your systems.
For more information about required versions, recommended system landscape,
considerations for the supported source objects, and more, see SAP Note
2890171 .
• Data flows currently support one-time loading of data (also known as initial load)
only.
• The data preview in the data flow editor of the Data Builder is not available, except
for:
• CDS views if the version of the connected system is SAP S/4HANA on-prem-
ise 1909 or SAP S/4HANA 1909 or a higher version respectively.
• SAP Landscape Transformation Replication Server objects if ABAP Add-on
DMIS 2018 SP08 or DMIS 2020 SP04 is installed or a higher DMIS version.

Integrating Data and Managing Spaces in SAP Datasphere


70 PUBLIC Integrating Data via Connections
Feature Additional Information

Replication Flows You can use the connection to add source objects to a replication flow.

You can access the following data:

• extraction-enabled ABAP CDS views that are C1-released, that is views with annotation
@Analytics.dataextraction.enabled: true and that are available in the
connected system (access via ABAP Pipeline Engine)
For more information, see:
• in the SAP S/4HANA Cloud documentation: Data Integration and CDS Views Ena-
bled for Data Extraction
• in the SAP S/4HANA documentation: CDS Views Enabled for Data Extraction

 Note
When adding source objects in replication flows, you can find the CDS views in the
CDS_EXTRACTION container.

• standard and custom CDS view entities that are exposed using the SQL service from
SAP BTP, ABAP environment, orSAP S/4HANA Cloud, respectively
For more information, see:
• in the SAP S/4HANA Cloud documentation: Data Consumption Using SAP
Datasphere
• in the SAP BTP documentation: Data Consumption Using SAP Datasphere

 Note
When adding source objects in replication flows, you can find the CDS view entities
in the SQL_SERVICE container.

• Tables from SAP Landscape Transformation Replication Server (SAP LT Replication


Server) sources, for which an appropriate configuration has been created in SAP LT
Replication Server.
• Data from ODP sources (extraction contexts SAPI and BW)

 Note
• The availability of the replication flow feature depends on the used version and
Support Package level of SAP S/4HANA or the DMIS addon in the source. Make
sure your source systems meet the required minimum versions. We recommend to
use the latest available version of SAP S/4HANA and the DMIS add-on where pos-
sible and have the latest SAP notes and TCI notes implemented in your systems.
For more information about required versions, recommended system landscape,
considerations for the supported source objects, and more, see SAP Note
2890171 .
• The SAP BW and SAP ECC connection types don't support replication flows. In-
stead, you can use the SAP ABAP connection type to use replication flows for
replication from SAP BW, SAP BW/4HANA or SAP ECC systems.

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 71
 Note

The connection type supports replication with both remote tables via Data Provisioning Agent (SAP HANA
Smart Data Integration) and replication flows (for on-premise systems via Cloud Connector).

Regardless of whether you use the same connection or two separate connections, for replicating data from
a dedicated source object in the source we strongly recommend to only use remote tables or replication
flows and not both.

Best Practices for Integrating Data

See SAP Data Warehouse Cloud - First Guidance: Data Integration for ABAP Source Systems (published
February 2022) for general best practices for integrating data from ABAP source systems with SAP
Datasphere.

Configuring Connection Properties

Connection Details

Property Description

Protocol Select RFC (Use for on Premise ABAP Systems) to connect to an ABAP on-premise system.
If you want to use the on-premise system for building data flows, you additionally need to
set the Cloud Connector properties.

Select Web Socket RFC (Use for S/4HANA Cloud Systems) if you want to use the connection
for data flows only and your source is a cloud source such as SAP S/4HANA Cloud.

The default is RFC (Use for on Premise ABAP Systems).

[if Protocol = RFC (Use for Select Application Server to connect to a single application server, or select Message Server
on Premise ABAP Systems)] to use load balancing and connect to a message server that distributes the load to different
SAP Logon Connection Type application servers.

 Note
The SAP Logon connection type you select here must match the connection type in
the system mapping in the Cloud Connector configuration (load balancing logon or
connecting to a specific application server).

[if SAP Logon Connection Enter the name of the application server to which you want to connect to.
Type = Application Server]
Application Server

[if SAP Logon Connection Enter the SAP ABAP system instance number.
Type = Application Server]
System Number

[if SAP Logon Connection Enter the name of the message server to which you want to connect to.
Type = Message Server]
Message Server

Integrating Data and Managing Spaces in SAP Datasphere


72 PUBLIC Integrating Data via Connections
Property Description

[if SAP Logon Connection Enter the message server port (numerical).
Type = Message Server]
Message Server Port

[if SAP Logon Connection Enter the message server group.


Type = Message Server]
Message Server Group

[if Protocol = Web Socket Enter the name of the application server to which you want to connect to.
RFC (Use for S/4HANA Cloud
Systems)] Application Server

[if Protocol = Web Socket Enter the port number of the WebSocket RFC connection endpoint.
RFC (Use for S/4HANA Cloud
Systems)] Port

Client Enter the system client number.

System ID Enter the system ID.

Language Enter the two-digit ISO language code, EN for English or DE for German, for example. In the
Data Builder, object and field descriptions are then shown in the specified language. If not
set, the application uses the default logon language of the ABAP system for data flows, and
EN is used for remote tables and views.

Cloud Connector

Property Description

Use Cloud Connector Set to true if your source is an on-premise source and you want to use the connection for
data flows or replication flows. The default is false.

[if Use Cloud Connector = Select a location ID.


true] Location
 Note
To select another location ID than the default location, Connection.Read privilege is
required. The privilege is neither included in the DW Integrator nor in the DW Space
Administrator role. If you need to select a location ID, ask your tenant administrator to
either assign your user to a global role that is based on the DW Administrator role or
to assign your user to a custom global role (with license type SAP Datasphere) that
includes the required Connection.Read privilege.

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 73
Property Description

[if Use Cloud Connector = Select Derive Virtual Host and Port from Connection Details or Enter Virtual Host and Port in
true] Virtual Destination Separate Fields.

If host and port entered in the connection details match the virtual host and port from the
Cloud Connector configuration, you can select Derive Virtual Host and Port from Connection
Details and don't need to enter the values manually.

 Note
If the SAP Logon Connection Type for your connection is Application Server: When
you select Derive Virtual Host and Port from Connection Details for ABAP-based connec-
tions, virtual port is set to sapgw<system number> to make sure that it matches
the virtual port defined in the Cloud Connector configuration.

[if Virtual Destination = Enter Enter the virtual host that you defined during Cloud Connector configuration.
Virtual Host and Port in
Separate Fields] Virtual Host

[if Virtual Destination = Enter Enter the virtual port that you defined during Cloud Connector configuration.
Virtual Host and Port in
Separate Fields] Virtual Port  Note
• If the SAP Logon Connection Type for your connection is Application Server: Enter
sapgw<system number> to make sure that it matches the virtual port defined
in the Cloud Connector configuration.
• If the SAP Logon Connection Type for your connection is Message Server: Enter
a numerical port and make sure it is the port defined in the Cloud Connector
configuration.

Authentication

Property Description

Authentication Type Select the authentication type to use to connect to the SAP ABAP system.

You can select:

• User Name And Password for basic authentication (default value) - This option is read-
only if you set Cloud Connector to false.
• OAuth 2.0 - You can select this option only if you have selected the RFC protocol and
if you have set Cloud Connector to true to enable replication flows and data flows.
Remote tables currently are not supported with OAuth authentication.

OAuth 2.0

Property Description

OAuth Grant Type Displays Client Credentials as grant type used to retrieve an
access token.

OAuth Token Endpoint Enter the API endpoint to use to request an access token.

OAuth Scope [optional] Enter the OAuth scope, if applicable.

Integrating Data and Managing Spaces in SAP Datasphere


74 PUBLIC Integrating Data via Connections
Property Description

OAuth Token Request Content Type [optional] Enter the content type of the OAuth2 token re-
quest.

You can select:

• URL Encoded (default value)


OAuth2 token request parameters will be url-encoded
and included in the HTTP request body.
• JSON
OAuth2 token request parameters will be in JSON for-
mat and included in the HTTP request body.

Credentials (OAuth 2.0)

Property Description

Client ID Enter the client ID to authenticate SAP Datasphere to the


SAP ABAP system.

Client Secret Enter the client secret to authenticate SAP Datasphere to


the SAP ABAP system.

Credentials (User Name and Password)

Property Description

User Name Enter the user name that is used to connect to the SAP ABAP system.

Password Enter the user password.

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 75
Features

Feature Description

Remote Tables To enable Remote Tables, complete the following properties:

Property Description

Data Provisioning Select an agent if you want to use the connection to access data via
Agent imported remote tables and to build views.

[if you selected an Set to On if you want to use ABAP RFC streaming for loading tables
agent] Streaming based on small batches. Set to Off to use non-RFC streaming. The
Read default is Off.

[if Streaming Read = Enter the gateway host where the ABAP adapter would register
On] Gateway Host an RFC server instance to receive callbacks with the table data
batches. Usually, the gateway host is the same as the target ABAP
system host.

[if Streaming Read = Enter the port number of the gateway server where the ABAP
On] Gateway Port adapter would register an RFC server instance to receive call-
backs with the table data batches. You can find the port
number for sapgw<ABAP_instance_number> in file /etc/
services on the ABAP host. The default port number is
33<ABAP_instance_number>. If you have a file /etc/
services with this mapping on the Data Provisioning Agent’s
host, you can also enter sapgw<ABAP_instance_number>
instead of the port number.

[if Streaming = On] Enter the name of the RFC destination that you have created in
RFC Destination the source. For more information, see Prerequisites for ABAP RFC
Streaming.

Data Flows Data Flows are enabled without the need to set any additional connection properties. If your
source is an on-premise source, make sure you have maintained the properties in the Cloud
Connector section.

Replication Flows Replication Flows are enabled without the need to set any additional connection properties.
If your source is an on-premise source, make sure you have maintained the properties in the
Cloud Connector section.

Advanced Connection Properties

When you've selected a Data Provisioning Agent, you can configure advanced connection properties to enable
specific customer scenarios. In the connection creation wizard, you configure advanced properties in an
additional wizard step. When editing the connection to configure the properties, scroll down and click the Show
Advanced Properties button. Advanced properties have default values that you can override according to your
customer scenario’s needs.

Available properties:

• Connection Pool Size

Integrating Data and Managing Spaces in SAP Datasphere


76 PUBLIC Integrating Data via Connections
• Connection Limit
• Extractor’s ODP Version
• CDC Batch Size, MB
• RFC Unicode
• [if RFC Unicode = No] RFC Character Encoding
• SNC Mode
• [if SNC Mode = On] SNC Library
• [if SNC Mode = On] SNC Name of Client
• [if SNC Mode = On] SNC Name of SAP Server
• [if SNC Mode = On] SNC SSO
• [if SNC Mode = On] SNC Quality of Protection
• RFC Serialization (requires Data Provisioning Agent version 2.3.7.2 or higher)
• [if Streaming Read = On] Batch Size, MB
• [if Streaming Read = On] Batch Receive Timeout
• [if Streaming Read = On] RFC Trace

For more information, see SAP ABAP Adapter Remote Source Configuration in the SAP HANA Smart Data
Integration and SAP HANA Smart Data Quality Installation and Configuration Guide.

2.1.24 SAP BW Bridge Connection


In the SAP BW bridge space, a SAP BW bridge connection is being generated.

Technically, it comprises two connections:

• An HTTP ABAP connection to retrieve metadata from the SAP BW bridge


• A SAP HANA smart data access connection to retrieve the data from the SAP BW bridge

You can only display the connection. Here, you can copy the SAP BW service key, which you will need to enter
when you create an SAP BW bridge project in the SAP BW Bridge Modeling Tools. For more information, see
Copy the SAP BW Service Key.

Restrictions

The connection type SAP BW bridge connection doesn’t show up in any other space, it cannot be modified, and
you cannot create such a connection yourself.

SAP BW bridge connections only support remote tables and model import. Data flows and replication flows are
not supported.

2.1.25 SAP BW Connections


Use an SAP BW connection to access data from virtual tables through RFC for ODP sources (extractors) and
ABAP Dictionary tables from SAP Business Warehouse (SAP BW) or SAP BW∕4HANA systems. For SAP BW

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 77
systems that don't have the ABAP Pipeline Engine extension installed, ODP extractors can be used as sources
in data flows.

Prerequisites

See: Prepare Connectivity to SAP BW

Using the Connection

The connection type supports the remote table as well as the data flow feature.

 Note

If you want to provide SAP BW∕4HANA business semantics to SAP Datasphere, you can use connection
type SAP BW/4HANA Model Import and import SAP BW∕4HANA models based on analytic queries. During
the import, the semantics of the underlying SAP BW∕4HANA models are translated into native SAP
Datasphere entities that you can use to access the data stored in SAP BW∕4HANA, to enhance the data in
SAP Datasphere, or to consume them by analytic clients. For more information, see SAP BW∕4HANA Model
Transfer Connections [page 83].

 Note

This connection type doesn't support replication flows. Instead, we recommend using the SAP ABAP
connection type for replication flows.

Remote Tables

You can use the connection type to access:

• data from virtual tables through RFC for ODP sources (extractors):
• Extraction context ABAP_CDS provides access to ABAP Core Data Services (CDS) Views that include
the annotation @Analytics.dataextraction.enabled: true
For information about which ABAP CDS Views are available for extraction, see CDS Views Enabled for
Data Extraction in the SAP S/4HANA documentation.
• Extraction context BW provides access to InfoProviders in SAP BW or SAP BW∕4HANA systems:
• DataStore objects (advanced or classic)
• CompositeProviders
• Queries as InfoProviders (with a subset of OLAP capabilities) - can be enabled in the runtime
properties of queries in the BW Modeling tools
• InfoObjects (characteristics) - can be enabled in the general settings of the InfoObject editor in the
BW Modeling tools
• Extraction context SAPI provides access to Service API (SAPI) DataSources

Integrating Data and Managing Spaces in SAP Datasphere


78 PUBLIC Integrating Data via Connections
 Note

Importing SAPI DataSources into SAP Datasphere is not supported if the DataSource contains
mandatory selection fields. You will receive an error message when trying to import tables for such
DataSources.

• data from SAP ABAP Dictionary tables in SAP S/4HANA systems

Supported data access methods for remote tables:

• Remote
• Replication (snapshot)
• For ODP sources: Replication (real-time)
For more information about the prerequisites, see Replicate Data Changes in Real-Time [page 181].

For more information, see Replicating Data and Monitoring Remote Tables [page 168].

 Note

Certain SAPI DataSources (from FI-AA Business Content, for example) may send duplicate records during
snapshot replication or real-time initialization. Remote table replication runs with optimized INSERT
processing on DataSource key level which leads to unique constraint violation errors. In such cases we
recommend to use a data flow which uses the remote table as source and for which the target table has
selected the Append mode with the Update Records By Primary Key (UPSERT) option.

Data Flows
For legacy SAP BW systems that do not have the ABAP Pipeline Engine extension or DMIS Addon installed,
you can leverage Operational Data Provisioning (ODP) connectivity und use ODP extractors as sources in data
flows. Note that ABAP Dictionary tables are not supported as sources in data flows.

 Note

The data preview in the data flow editor of the Data Builder is not available for SAP BW sources.

Best Practices for Integrating Data


See SAP Data Warehouse Cloud - First Guidance: Data Integration for ABAP Source Systems (published
February 2022) for general best practices for integrating data from ABAP source systems with SAP
Datasphere.

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 79
Configuring Connection Properties

Connection Details

Property Description

SAP Logon Connection Type Select Application Server to connect to a single application server, or select Message Server
to use load balancing and connect to a message server that distributes the load to different
application servers.

 Note
The SAP Logon connection type you select here must match the connection type in
the system mapping in the Cloud Connector configuration (load balancing logon or
connecting to a specific application server).

[if SAP Logon Connection Enter the name of the application server to which you want to connect to.
Type = Application Server]
Application Server

[if SAP Logon Connection Enter the SAP ABAP system instance number.
Type = Application Server]
System Number

[if SAP Logon Connection Enter the name of the message server to which you want to connect to.
Type = Message Server]
Message Server

[if SAP Logon Connection Enter the message server port (numerical).
Type = Message Server]
Message Server Port

[if SAP Logon Connection Enter the message server group.


Type = Message Server]
Message Server Group

Client Enter the system client number.

System ID Enter the system ID.

Language Enter the two-digit ISO language code, EN for English or DE for German, for example. In the
Data Builder, object and field descriptions are then shown in the specified language. If not
set, the application uses the default logon language of the ABAP system for data flows, and
EN is used for remote tables and views.

Cloud Connector

Property Description

Use Cloud Connector Set to true if your source is an on-premise source and you want to use the connection for
data flows. The default is false.

Integrating Data and Managing Spaces in SAP Datasphere


80 PUBLIC Integrating Data via Connections
Property Description

[if Use Cloud Connector = Select a location.


true] Location
 Note
To select another location ID than the default location, Connection.Read privilege is
required. The privilege is neither included in the DW Integrator nor in the DW Space
Administrator role. If you need to select a location ID, ask your tenant administrator to
either assign your user to a global role that is based on the DW Administrator role or
to assign your user to a custom global role (with license type SAP Datasphere) that
includes the required Connection.Read privilege.

[if Use Cloud Connector = Select Derive Virtual Host and Port from Connection Details or Enter Virtual Host and Port in
true] Virtual Destination Separate Fields.

If host and port entered in the connection details match the virtual host and port from the
Cloud Connector configuration, you can select Derive Virtual Host and Port from Connection
Details and don't need to enter the values manually.

 Note
When you select Derive Virtual Host and Port from Connection Details for ABAP-based
connections, virtual port is set to sapgw<system number> to make sure that it
matches the virtual port defined in the Cloud Connector configuration.

[if Virtual Destination = Enter Enter the virtual host that you defined during Cloud Connector configuration.
Virtual Host and Port in
Separate Fields] Virtual Host

[if Virtual Destination = Enter Enter the virtual port that you defined during Cloud Connector configuration.
Virtual Host and Port in
Separate Fields] Virtual Port  Note
• If the SAP Logon Connection Type for your connection is Application Server: Enter
sapgw<system number> to make sure that it matches the virtual port defined
in the Cloud Connector configuration.
• If the SAP Logon Connection Type for your connection is Message Server: Enter
a numerical port and make sure it is the port defined in the Cloud Connector
configuration.

Credentials

Property Description

User Name Enter the user name that is used to connect to the SAP ABAP system.

Password Enter the user password.

Features

To enable Remote Tables:

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 81
Property Description

Data Provisioning Agent Select an agent if you want to use the connection to access data via imported remote tables
and to build views.

[if you selected an agent] Set to On if you want to use ABAP RFC streaming for loading tables based on small batches.
Streaming Read Set to Off to use non-RFC streaming. The default is Off.

[if Streaming Read = On] Enter the gateway host where the ABAP adapter would register an RFC server instance to
Gateway Host receive callbacks with the table data batches. Usually, the gateway host is the same as the
target ABAP system host.

[if Streaming Read = On] Enter the port number of the gateway server where the ABAP adapter would register an
Gateway Port RFC server instance to receive callbacks with the table data batches. You can find the
port number for sapgw<ABAP_instance_number> in file /etc/services on the
ABAP host. The default port number is 33<ABAP_instance_number>. If you have a
file /etc/services with this mapping on the Data Provisioning Agent’s host, you can
also enter sapgw<ABAP_instance_number> instead of the port number.

[if Streaming = On] RFC Enter the name of the RFC destination that you have created in the source. For more
Destination information, see Prerequisites for ABAP RFC Streaming.

Data Flows are enabled without the need to set any additional connection properties. Make sure you have
maintained the properties in the Cloud Connector section.

Advanced Connection Properties

When you've selected a Data Provisioning Agent, you can configure advanced connection properties to enable
specific customer scenarios. In the connection creation wizard, you configure advanced properties in an
additional wizard step. When editing the connection to configure the properties, scroll down and click the Show
Advanced Properties button. Advanced properties have default values that you can override according to your
customer scenario’s needs.

Available properties:

• Connection Pool Size


• Connection Limit
• Extractor’s ODP Version
• CDC Batch Size, MB
• RFC Unicode
• [if RFC Unicode = No] RFC Character Encoding
• SNC Mode
• [if SNC Mode = On] SNC Library
• [if SNC Mode = On] SNC Name of Client
• [if SNC Mode = On] SNC Name of SAP Server
• [if SNC Mode = On] SNC SSO
• [if SNC Mode = On] SNC Quality of Protection
• RFC Serialization (requires Data Provisioning Agent version 2.3.7.2 or higher)
• [if Streaming Read = On] Batch Size, MB
• [if Streaming Read = On] Batch Receive Timeout

Integrating Data and Managing Spaces in SAP Datasphere


82 PUBLIC Integrating Data via Connections
• [if Streaming Read = On] RFC Trace

For more information, see SAP ABAP Adapter Remote Source Configuration in the SAP HANA Smart Data
Integration and SAP HANA Smart Data Quality Installation and Configuration Guide.

2.1.26 SAP BW∕4HANA Model Transfer Connections

Use an SAP BW/4HANA Model Transfer connection to import analytic queries from SAP BW∕4HANA with their
Composite Providers and InfoObjects.

Prerequisites

See: Preparing SAP BW/4HANA Model Transfer Connectivity

Using the Connection

The connection type supports the model import feature.

For more information, see Importing SAP BW∕4HANA Models.

Configuring Connection Properties

Live Data Connection (Tunnel)

Property Description

Name Select the relevant live data connection of type tunnel.

A live data connection of type tunnel is required for a secure


connection to SAP BW/4HANA with Cloud Connector and
for accessing SAP BW/4HANA metadata with http requests.
With Cloud Connector, the SAP BW/4HANA system doesn't
need to be exposed to the internet in order to make the
system accessible.

If the required connection is not available for selection, you


or an administrator need to create the tunnel connection
before you can create the SAP BW/4HANA Model Transfer
connection.

For more information, see Create Live Data Connection of


Type Tunnel.

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 83
SAP HANA SQL Access

Property Description

Host Displays the SAP HANA host retrieved from SAP BW∕4HANA
via the selected tunnel connection.

Port Displays the SAP HANA port retrieved from SAP BW∕4HANA
via the selected tunnel connection.

Schema Displays the default schema in which DDIC tables are


located in the underlying SAP HANA database of SAP
BW∕4HANA. The information is retrieved from SAP
BW∕4HANA via the selected tunnel connection.

 Note

If required in exceptional cases, you can change the values in the Host, Port, and Schema fields.

In some use cases, for example when SAP BW∕4HANA is hosted by a third-party cloud provider, host, port,
and schema can't be retrieved and you must enter them manually.

In case of an error when retrieving host, port, and schema for example because of missing authorizations,
because of any of the services necessary for the tunnel connection not being active, or because of any
other issue with the tunnel connection, you can enter host, port, and schema manually to be able to
proceed with connection creation before validating and fixing the tunnel connection. We recommend,
however, to validate and fix the tunnel connection first.

You can reset the values for the Host, Port, and Schema fields to the values retrieved from SAP BW∕4HANA
by clicking Reset.

SAP HANA Security

Property Description

Enable SSL Encryption Specify whether to enable SSL encryption on connections


to the remote SAP HANA database of the SAP BW/4HANA
system. The default value is true.

To use SSL encryption with the remote SAP HANA database,


the Data Provisioning Agent must already be correctly con-
figured for SSL support.

Validate Server Certificate Specify whether to validate the certificate of the remote SAP
HANA server. The default value is true.

 Note
A certificate signed by a certificate authority (CA) is
needed here. If the certificate is self-signed and not CA
signed, select false.

Integrating Data and Managing Spaces in SAP Datasphere


84 PUBLIC Integrating Data via Connections
SAP HANA Credentials

Property Description

User Name Enter the SAP HANA database user name (case sensitive).

 Note
The SAP HANA user needs read privileges for the ABAP
schema in the SAP HANA database.

Password Enter the appropriate corresponding password.

Features

To enable Model Import, select a Data Provisioning Agent.

What happens when you've created the connection

On saving the connection, SAP Datasphere retrieves SAP HANA host and port via the SAP Analytics Cloud
tunnel connection.

You can now find your new connection in the list of the available connections for the space and can use it
in the Business Builder to import analytic queries from your connected SAP BW∕4HANA system into SAP
Datasphere. During the model import alongside other objects we will create and deploy remote tables in the
repository. The remote tables can be used in the following places:

• In the remote table monitor to replicate data.


• In the view editor of the data builder to build views on top (by dragging the object in from the repository
browser)
• In the data flow editor of the data builder to use them as sources (by dragging the object in from the
repository browser).

 Note

In the view or data flow editor in the data builder, SAP BW/4HANA Model Transfer connections with their
remote tables are not available as sources in the Sources tab of the source tree. Instead, as described
above, the remote tables are available from the Repository tab.

2.1.27 SAP ECC Connections

Use an SAP ECC connection to access data from virtual tables through RFC for ODP sources (extractors)
and ABAP Dictionary tables from SAP ERP Central Component (SAP ECC) systems (on-premise). For source

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 85
systems that don't have the ABAP Pipeline Engine extension installed, ODP extractors can be used as sources
in data flows.

Prerequisites

See: Prepare Connectivity to SAP ECC

Using the Connection

The connection type supports the remote table as well as the data flow feature.For more information, see
Supported Connection Type Features [page 9].

 Note

This connection type doesn't support replication flows. Instead, we recommend using the SAP ABAP
connection type for replication flows.

Remote Tables
You can use the connection type to access:

• data from virtual tables through RFC for ODP sources (extractors):
• Extraction context ABAP_CDS provides access to ABAP Core Data Services (CDS) Views that include
the annotation @Analytics.dataextraction.enabled: true
For information about which ABAP CDS Views are available for extraction, see CDS Views Enabled for
Data Extraction in the SAP S/4HANA documentation.
• Extraction context BW provides access to InfoProviders in SAP BW or SAP BW∕4HANA systems:
• DataStore objects (advanced or classic)
• CompositeProviders
• Queries as InfoProviders (with a subset of OLAP capabilities) - can be enabled in the runtime
properties of queries in the BW Modeling tools
• InfoObjects (characteristics) - can be enabled in the general settings of the InfoObject editor in the
BW Modeling tools
• Extraction context SAPI provides access to Service API (SAPI) DataSources

 Note

Importing SAPI DataSources into SAP Datasphere is not supported if the DataSource contains
mandatory selection fields. You will receive an error message when trying to import tables for such
DataSources.

• data from SAP ABAP Dictionary tables in SAP S/4HANA systems

Supported data access methods for remote tables:

• Remote
• Replication (snapshot)

Integrating Data and Managing Spaces in SAP Datasphere


86 PUBLIC Integrating Data via Connections
• For ODP sources: Replication (real-time)
For more information about the prerequisites, see Replicate Data Changes in Real-Time [page 181].

For more information, see Replicating Data and Monitoring Remote Tables [page 168].

 Note

Certain SAPI DataSources (from FI-AA Business Content, for example) may send duplicate records during
snapshot replication or real-time initialization. Remote table replication runs with optimized INSERT
processing on DataSource key level which leads to unique constraint violation errors. In such cases we
recommend to use a data flow which uses the remote table as source and for which the target table has
selected the Append mode with the Update Records By Primary Key (UPSERT) option.

Data Flows

For legacy SAP ECC systems that do not have the ABAP Pipeline Engine extension or DMIS Addon installed,
you can leverage Operational Data Provisioning (ODP) connectivity und use ODP extractors as sources in data
flows. Also ODP extractors from SAP S/4HANA on-premise systems that have a version lower than 1909 are
supported. Note that ABAP Dictionary tables are not supported as sources in data flows.

 Note

The data preview in the data flow editor of the Data Builder is not available for SAP ECC sources.

Best Practices for Integrating Data

See SAP Data Warehouse Cloud - First Guidance: Data Integration for ABAP Source Systems (published
February 2022) for general best practices for integrating data from ABAP source systems with SAP
Datasphere.

Configuring Connection Properties

Connection Details

Property Description

SAP Logon Connection Type Select Application Server to connect to a single application server, or select Message Server
to use load balancing and connect to a message server that distributes the load to different
application servers.

 Note
The SAP Logon connection type you select here must match the connection type in
the system mapping in the Cloud Connector configuration (load balancing logon or
connecting to a specific application server).

[if SAP Logon Connection Enter the name of the application server to which you want to connect to.
Type = Application Server]
Application Server

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 87
Property Description

[if SAP Logon Connection Enter the SAP ABAP system instance number.
Type = Application Server]
System Number

[if SAP Logon Connection Enter the name of the message server to which you want to connect to.
Type = Message Server]
Message Server

[if SAP Logon Connection Enter the message server port (numerical).
Type = Message Server]
Message Server Port

[if SAP Logon Connection Enter the message server group.


Type = Message Server]
Message Server Group

Client Enter the system client number.

System ID Enter the system ID.

Language Enter the two-digit ISO language code, EN for English or DE for German, for example. In the
Data Builder, object and field descriptions are then shown in the specified language. If not
set, the application uses the default logon language of the ABAP system for data flows, and
EN is used for remote tables and views.

Cloud Connector

Property Description

Use Cloud Connector Set to true if your source is an on-premise source and you want to use the connection for
data flows. The default is false.

[if Use Cloud Connector = Select a location.


true] Location
 Note
To select another location ID than the default location, Connection.Read privilege is
required. The privilege is neither included in the DW Integrator nor in the DW Space
Administrator role. If you need to select a location ID, ask your tenant administrator to
either assign your user to a global role that is based on the DW Administrator role or
to assign your user to a custom global role (with license type SAP Datasphere) that
includes the required Connection.Read privilege.

[if Use Cloud Connector = Select Derive Virtual Host and Port from Connection Details or Enter Virtual Host and Port in
true] Virtual Destination Separate Fields.

If host and port entered in the connection details match the virtual host and port from the
Cloud Connector configuration, you can select Derive Virtual Host and Port from Connection
Details and don't need to enter the values manually.

 Note
When you select Derive Virtual Host and Port from Connection Details for ABAP-based
connections, virtual port is set to sapgw<system number> to make sure that it
matches the virtual port defined in the Cloud Connector configuration.

Integrating Data and Managing Spaces in SAP Datasphere


88 PUBLIC Integrating Data via Connections
Property Description

[if Virtual Destination = Enter Enter the virtual host that you defined during Cloud Connector configuration.
Virtual Host and Port in
Separate Fields] Virtual Host

[if Virtual Destination = Enter Enter the virtual port that you defined during Cloud Connector configuration.
Virtual Host and Port in
Separate Fields] Virtual Port  Note
• If the SAP Logon Connection Type for your connection is Application Server: Enter
sapgw<system number> to make sure that it matches the virtual port defined
in the Cloud Connector configuration.
• If the SAP Logon Connection Type for your connection is Message Server: Enter
a numerical port and make sure it is the port defined in the Cloud Connector
configuration.

Credentials

Property Description

User Name Enter the user name that is used to connect to the SAP ABAP system.

Password Enter the user password.

Features

To enable Remote Tables:

Property Description

Data Provisioning Agent Select an agent if you want to use the connection to access data via imported remote tables
and to build views.

[if you selected an agent] Set to On if you want to use ABAP RFC streaming for loading tables based on small batches.
Streaming Read Set to Off to use non-RFC streaming. The default is Off.

[if Streaming Read = On] Enter the gateway host where the ABAP adapter would register an RFC server instance to
Gateway Host receive callbacks with the table data batches. Usually, the gateway host is the same as the
target ABAP system host.

[if Streaming Read = On] Enter the port number of the gateway server where the ABAP adapter would register an
Gateway Port RFC server instance to receive callbacks with the table data batches. You can find the
port number for sapgw<ABAP_instance_number> in file /etc/services on the
ABAP host. The default port number is 33<ABAP_instance_number>. If you have a
file /etc/services with this mapping on the Data Provisioning Agent’s host, you can
also enter sapgw<ABAP_instance_number> instead of the port number.

[if Streaming = On] RFC Enter the name of the RFC destination that you have created in the source. For more
Destination information, see Prerequisites for ABAP RFC Streaming.

Data Flows are enabled without the need to set any additional connection properties. Make sure you have
maintained the properties in the Cloud Connector section.

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 89
Advanced Connection Properties

When you've selected a Data Provisioning Agent, you can configure advanced connection properties to enable
specific customer scenarios. In the connection creation wizard, you configure advanced properties in an
additional wizard step. When editing the connection to configure the properties, scroll down and click the Show
Advanced Properties button. Advanced properties have default values that you can override according to your
customer scenario’s needs.

Available properties:

• Connection Pool Size


• Connection Limit
• Extractor’s ODP Version
• CDC Batch Size, MB
• RFC Unicode
• [if RFC Unicode = No] RFC Character Encoding
• SNC Mode
• [if SNC Mode = On] SNC Library
• [if SNC Mode = On] SNC Name of Client
• [if SNC Mode = On] SNC Name of SAP Server
• [if SNC Mode = On] SNC SSO
• [if SNC Mode = On] SNC Quality of Protection
• RFC Serialization (requires Data Provisioning Agent version 2.3.7.2 or higher)
• [if Streaming Read = On] Batch Size, MB
• [if Streaming Read = On] Batch Receive Timeout
• [if Streaming Read = On] RFC Trace

For more information, see SAP ABAP Adapter Remote Source Configuration in the SAP HANA Smart Data
Integration and SAP HANA Smart Data Quality Installation and Configuration Guide.

2.1.28 SAP Fieldglass Connections

Use an SAP Fieldglass connection to retrieve data from SAP Fieldglass.

Prerequisites

See: Prepare Connectivity to SAP Fieldglass

Using the Connection

The connection type supports the remote table as well as the data flow feature.

Integrating Data and Managing Spaces in SAP Datasphere


90 PUBLIC Integrating Data via Connections
Supported data access methods for remote tables:

• Remote
• Replication (snapshot)
• Replication (real-time)

For more information, see Replicating Data and Monitoring Remote Tables [page 168].

Configuring Connection Properties

Connection Details

Property Description

URL Enter the URL with the following syntax: <protocol>://


<host><service path>

Protocol: HTTP or HTTPS. The default value is HTTPS.

Host: Host for accessing the cloud OData service

Service path: Relative path (without host and port) to the


Cloud Data Integration service endpoint (where the CDI pro-
vider service is running at the cloud provider side). The value
must start with a forward slash ( / ).

Root Path [optional] Enter the root path name to restrict browsing to a
certain CDI namespace or provider.

OAuth 2.0

Property Description

OAuth Grant Type Select the grant type.

You can select:

• Client Credentials
• User Name and Password

OAuth Token Endpoint Enter the token endpoint that the application must use.

OAuth Scope Enter the OAuth scope value.

OAuth Resource Enter the OAuth resource.

OAuth Response Type Select the OAuth response type. The values are token or
none.

OAuth Token Request Content Type Select the value for the content-type HTTP header that the
application must use when requesting a token. The values
are URLEncoded or JSON.

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 91
Credentials (OAuth 2.0)

If OAuth Grant Type = Client Credentials:

Property Description

Client ID Enter the client ID.

Client Secret Enter the client secret.

If OAuth Grant Type = User Name And Password:

Property Description

User Name Enter the name of the OAuth user.

Password (OAuth 2.0) Enter the OAuth password.

Features

To enable Remote Tables, select a Data Provisioning Agent.

Data Flows are enabled without the need to set any additional connection properties.

2.1.29 SAP HANA Connections

Use an SAP HANA connection to access data from a remote SAP HANA database (on-premise or cloud).

This topic contains the following sections:

• Supported Features [page 92]


• Configuring Connection Properties (SAP HANA Cloud) [page 93]
• Configuring Connection Properties (SAP HANA on-premise) [page 94]

For information about the required prerequisites, see Prepare Connectivity to SAP HANA.

Supported Features

Feature Additional Information

Remote Tables You can use remote tables imported from the connection either to access data directly live
in the connected source (federation) or to copy the data into SAP Datasphere (replication).

For remote tables, real-time replication is supported. For information about any constraints,
see Replicate Data Changes in Real-Time [page 181].

Data Flows You can use the connection to add source objects to a data flow.

Replication Flows You can use the connection to add source and target objects to a replication flow.

Integrating Data and Managing Spaces in SAP Datasphere


92 PUBLIC Integrating Data via Connections
 Note

The connection type supports replication with both remote tables via Data Provisioning Agent (SAP HANA
Smart Data Integration) and replication flows (for on-premise systems via Cloud Connector).

Regardless of whether you use the same connection or two separate connections, for replicating data from
a dedicated source object in the source we strongly recommend to only use remote tables or replication
flows and not both.

2.1.29.1 Configuring Connection Properties (SAP HANA


Cloud)

Connection Details

Property Description

Category Select Cloud to connect to an SAP HANA Cloud instance.

Host Enter the fully qualified host name or IP address on which the remote SAP HANA server is
running.

Port Enter the SQL port number of the SAP HANA Cloud instance: 443.

Authentication

Property Description

Authentication Type Select the authentication type to use to connect to SAP HANA.

You can select:

• X.509 Client Certificate (default)


• User Name And Password for basic authentication

Credentials (X.509 Client Certificate)

If Authentication Type = X.509 Client Certificate:

Property Description

X.509 Client Certificate To upload the certificate or certificate chain that is used to authenticate to the remote
system, click  (Browse) and select the file from your download location.

X.509 Client Private Key To upload the private key, click  (Browse) and select the file from your download location.

X.509 Client Private Key [optional] If the private key is encrypted, enter the password required for decryption.
Password

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 93
Credentials (User Name And Password)

If Authentication Type = User Name And Password:

Property Description

User Name Enter the database user name (case sensitive).

Password Enter the password of the SAP HANA database user.

Features

Feature Description

Remote Tables Remote Tables are enabled without the need to set any additional connection properties.

Complete the following properties:

Property Description

Data Provisioning [read-only] Displays the midleware used to connect to the on-
Option premise system.

The option for SAP HANA Cloud is always set to Direct because no
middleware is required.

Data Access Select how you want to access data from source objects:

• Select Remote Only if you want to federate data from the


source objects of the connection and access them remotely
in the source system.
• Select Remote and Replication (default) if you want to allow
copying the data from the source object into SAP Datasphere.

For more information about the use cases for the options, see Re-
mote Tables in Data Access Remote Only Vs Data Access Remote
and Replication [page 167].

Data Flows Data Flows are enabled without the need to set any additional connection properties.

Replication Flows Replication Flows are enabled without the need to set any additional connection properties.

2.1.29.2 Configuring Connection Properties (SAP HANA on-


premise)

Connection Details

Property Description

Category Select On-Premise to connect to SAP HANA (on-premise).

Integrating Data and Managing Spaces in SAP Datasphere


94 PUBLIC Integrating Data via Connections
Property Description

Host Enter the fully qualified host name or IP address on which the remote SAP HANA server is
running.

Port Enter the TCP SQL port number of the remote SAP HANA server.

For more information about port assignment for SAP HANA, see Port Assignment in Tenant
Databases in the SAP HANA Platform documentation.

Cloud Connector

Property Description

Use Cloud Connector Set to true if your source is an on-premise source and you want to use the connection for
one of the following features:

• remote tables with SAP HANA smart data access


• data flows
• replication flows

The default is false.

[if Use Cloud Connector = [optional] Select a location ID.


true] Location
 Note
To select another location ID than the default location, Connection.Read privilege is
required. The privilege is neither included in the DW Integrator nor in the DW Space
Administrator role. If you need to select a location ID, ask your tenant administrator to
either assign your user to a global role that is based on the DW Administrator role or
to assign your user to a custom global role (with license type SAP Datasphere) that
includes the required Connection.Read privilege.

[if Use Cloud Connector = [optional] Select Derive Virtual Host and Port from Connection Details or Enter Virtual Host
true] Virtual Destination and Port in Separate Fields.

If host and port entered in the connection details match the virtual host and port from the
Cloud Connector configuration, you can select Derive Virtual Host and Port from Connection
Details and don't need to enter the values manually.

[if Virtual Destination = Enter Enter the virtual host that you defined during Cloud Connector configuration.
Virtual Host and Port in
Separate Fields] Virtual Host

[if Virtual Destination = Enter Enter the virtual port that you defined during Cloud Connector configuration.
Virtual Host and Port in
Separate Fields] Virtual Port

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 95
Security

Property Description

Enable SSL encryption Select whether to enable SSL encryption on the connection to the remote SAP HANA
database. The default value is true.

 Note
To use SSL encryption with a remote SAP HANA database connected via SAP HANA
smart data integration, the Data Provisioning Agent must already be correctly config-
ured for SSL support.

For more information, see Configure SSL for SAP HANA On-Premise [Manual Steps] in
the SAP HANA Smart Data Integration and SAP HANA Smart Data Quality documenta-
tion.

[if Enable SSL encryption Select whether to validate the certificate of the remote SAP HANA server. The default value
= true] Validate Server is true.
Certificate
If set to false, the host name used for the connection is used for verification.

 Note
• SSL is name-based; connecting to an IP address, or to “localhost” is unlikely to
work.
• When using SAP HANA smart data access via Cloud Connector for remote tables:
To validate the server certificate, the certificate must have been uploaded to SAP
Datasphere.
For more information, see Manage Certificates for Connections.

[if Validate Server Certificate Verify the host name field of the server certificate:
= true] Host Name in Server
Certificate • If the string is “*”, any name matches.
• If the string starts with “CN=”, it is treated as a common name, and the textual repre-
sentation of the common name entry in the certificate must be exactly the same.

Authentication

Property Description

Authentication Type Select the authentication type to use to connect to SAP HANA.

You can select:

• X.509 Client Certificate (default)


• User Name And Password for basic authentication

Integrating Data and Managing Spaces in SAP Datasphere


96 PUBLIC Integrating Data via Connections
Credentials (X.509 Client Certificate)
If Authentication Type = X.509 Client Certificate:

Property Description

X.509 Client Certificate To upload the certificate or certificate chain that is used to authenticate to the remote
system, click  (Browse) and select the file from your download location.

X.509 Client Private Key To upload the private key, click  (Browse) and select the file from your download location.

X.509 Client Private Key [optional] If the private key is encrypted, enter the password required for decryption.
Password

Credentials (User Name And Password)


If Authentication Type = User Name And Password:

Property Description

User Name Enter the database user name (case sensitive).

Password Enter the password of the SAP HANA database user.

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 97
Features

Feature Description

Remote Tables To enable Remote Tables, complete the following properties:

Property Description

Data Provisioning Select the midleware to use when connecting to and accessing the
Option on-premise system:

• None (default) if you don't want to use remote tables.


• Cloud Connector if you want to access the data via SAP HANA
smart data access.
• Data Provisioning Agent if you want to access the data via SAP
HANA smart data integration.

 Note
When creating a connection to SAP HANA on-premise using
SAP HANA smart data access via Cloud Connector, the system
checks for a required internal service. In the connection over-
view, a warning icon indicates that the service might not be
ready yet. This happens for the first connection you create or
when you create a connection after the service has been disa-
bled (after an automated weekly check returning that there is
no such connection anymore). Getting the service ready might
take up to 45 minutes. Validate the connection and check the
message details for more information.

For troubleshooting the Cloud Connector, see Troubleshooting


SAP HANA Smart Data Access via Cloud Connector.

[if Data Provisioning Select an agent from the list of agents configured for SAP
Option = Data
Datasphere.
Provisioning Agent]
Data Provisioning
Agent

[if Data Provisioning Select how you want to access data from source objects:
Option = Cloud
Connector or Data • Select Remote Only if you want to federate data from the
Provisioning Agent] source objects of the connection and access them remotely
Data Access in the source system.
• Select Remote and Replication (default) if you want to allow
copying the data from the source object into SAP Datasphere.

For more information about the use cases for the options, see Re-
mote Tables in Data Access Remote Only Vs Data Access Remote
and Replication [page 167].

Data Flows Data Flows are enabled without the need to set any additional connection properties. Make
sure you have maintained the properties in the Cloud Connector section.

Integrating Data and Managing Spaces in SAP Datasphere


98 PUBLIC Integrating Data via Connections
Feature Description

Replication Flows Replication Flows are enabled without the need to set any additional connection properties.
Make sure you have maintained the properties in the Cloud Connector section.

Advanced Connection Properties (SAP HANA on-premise)

When you've selected a Data Provisioning Agent, you can configure advanced connection properties to enable
specific customer scenarios. In the connection creation wizard, you configure advanced properties in an
additional wizard step. When editing the connection to configure the properties, scroll down and click the Show
Advanced Properties button. Advanced properties have default values that you can override according to your
customer scenario’s needs.

Available properties:

• Enable ABAP Manageable Trigger Namespace


• [if Enable ABAP Manageable Trigger Namespace = true] ABAP Manageable Trigger Namespace
• Connection Pool Size
• Minimum Scan Interval in Seconds
• Maximum Scan Interval in Seconds
• Maximum Batch Size
• DDL Scan Interval in Minutes
• Batch Queue Size
• Maximum Transaction Count in Scan
• Maximum Scan Size
• Enable Statement-level Triggers
• [if Enable Statement-level Triggers = true] Create Statement-level Insert Triggers
• [if Enable Statement-level Triggers = true] Create Statement-level Delete Triggers
• [if Enable Statement-level Triggers = true] Create Statement-level Update Triggers
• Triggers Record Primary Keys Only
• [if Enable Statement-level Triggers = false and Triggers Record Primary Keys Only = true] Triggers Upsert
Shadow Tables
• [if Triggers Record Primary Keys Only = true] Triggers Capture Before and After Images
• [if Triggers Record Primary Keys Only = true] Triggers Capture Full Before Images
• Shadow Table Type
• Trigger Queue Table Type
• Source Data Pattern Analysis
• Transmit Data in Compact Mode
• Enable Transaction Merge
• Allowlist Table in Remote Database
• Schema
• JDBC Connection Properties
• Retrieve Last Modified Dates for Objects in Dictionary

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 99
• Schema Alias
• Schema Alias Replacement

For more information, see SAP HANA Remote Source Configuration in the SAP HANA Smart Data Integration
and SAP HANA Smart Data Quality Installation and Configuration Guide.

2.1.30 SAP HANA Cloud, Data Lake Relational Engine


Connections

Use an SAP HANA Cloud, Data Lake Relational Engine connection to access table data from the Relational
Engine component of a standalone SAP HANA Cloud, data lake.

Standalone SAP HANA Cloud, data lake is a standalone component in the SAP Business Technology Platform
Cockpit. It is composed of the default data lake Files component and the data lake Relational Engine
component, and it is not integrated with SAP HANA database. For more information, see What is SAP HANA
Cloud, Data Lake in the SAP HANA Cloud, Data Lake documentation.

Using the Connection

The connection type supports the remote table as well as the data flow feature.

Supported data access methods for remote tables:

• Remote
• Replication (snapshot)

 Note

• With connection type SAP HANA Cloud, Data Lake Relational Engine you cannot connect to the SAP
HANA Cloud, data lake instance that is available with SAP Datasphere. To connect to this instance, use
your Open SQL schema in the space which you have selected to access the data lake.
For more information, see Integrating Data to and From SAP HANA Cloud Data Lake [page 158].
• SAP HANA Cloud, data lake Relational Engine connections are secure by default. They are encrypted
without the need of configuring any TLS settings. Also, uploading server certificates isn't required.
• To allow SAP Datasphere to access SAP HANA Cloud, data lake Relational Engine, it might be required
to add the SAP Datasphere outbound IP address and HANA IP addresses to the allowed IP addresses
for the data lake instance in SAP HANA Cloud Central.
For more information, see:
• Data Lake Connections in the SAP HANA Cloud documentation
• Finding SAP Datasphere IP addresses

Integrating Data and Managing Spaces in SAP Datasphere


100 PUBLIC Integrating Data via Connections
Configuring Connection Properties

Connection Details

Property Description

Host Enter the host of the SAP HANA Cloud, data lake Relational Engine server.

Port Enter the port number. The default port is 443.

Credentials

Property Description

User Name Enter the name of the SAP HANA Cloud, data lake Relational
Engine user.

 Note
This user needs to have corresponding privileges to
access the tables or views in SAP HANA Cloud, data
lake Relational Engine that you want to use in SAP
Datasphere. For more information, see SAP HANA
Cloud, Data Lake User Management for Data Lake IQ
in the SAP HANA Cloud, Data Lake documentation.

Password Enter the password.

Features

Remote Tables are enabled without the need to set any additional connection properties.

Data Flows are enabled without the need to set any additional connection properties.

2.1.31 SAP HANA Cloud, Data Lake Files Connections

Use an SAP HANA Cloud, Data Lake Files connection to access data from the Files component of a standalone
SAP HANA Cloud, data lake.

This topic contains the following sections:

• Context [page 101]


• Supported Features [page 102]
• Configuring Connection Properties [page 102]

Context

Standalone SAP HANA Cloud, data lake is a standalone component in the SAP Business Technology Platform
Cockpit. It is composed of the default data lake Files component and the data lake Relational Engine

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 101
component, and it is not integrated with SAP HANA database. For more information, see What is SAP HANA
Cloud, Data Lake in the SAP HANA Cloud, Data Lake documentation.

 Note

• With connection type SAP HANA Cloud, Data Lake Files you cannot connect to the SAP HANA Cloud,
data lake instance that is available with SAP Datasphere. To connect to this instance, use your Open
SQL schema in the space which you have selected to access the data lake.
For more information, see Integrating Data to and From SAP HANA Cloud Data Lake [page 158].
• SAP HANA Cloud, data lake Files connections are secure by default. They are encrypted without the
need of configuring any TLS settings. Also, uploading server certificates isn't required.
• To allow SAP Datasphere to access SAP HANA Cloud, data lake Files storage, it might be required to
add the SAP Datasphere outbound IP address and HANA IP addresses to the allowed IP addresses for
the data lake instance in SAP HANA Cloud Central.
For more information, see:
• Data Lake Connections in the SAP HANA Cloud documentation
• Finding SAP Datasphere IP addresses

Supported Features

Feature Additional Information

Data Flows You can use the connection to add source objects to a data flow.

Replication Flows You can use the connection to add source and target objects to a replication flow.

For more information, see:

• Cloud Storage Provider Sources


• Cloud Storage Provider Targets

Configuring Connection Properties

Connection Details

Property Description

Host Enter the SAP HANA Cloud, data lake Files REST API endpoint, for example <instance
id>files.hdl.prod-eu10.hanacloud.ondemand.com.

Root Path [optional] Enter a root path for browsing files. The value starts with the character slash. For
example, /My Folder/MySubfolder.

If you have specified the root path, then any path used with this connection is prefixed with
the root path.

Integrating Data and Managing Spaces in SAP Datasphere


102 PUBLIC Integrating Data via Connections
Credentials

Property Description

Keystore File Enter the Base64-encoded value of the client keystore file. The supported format is
PCKS#12 (P12 and PFX files).

You can also choose  Browse and select the file from your download location.

Keystore Password Enter the client keystore password.

Features

Feature Description

Data Flows Data Flows are enabled without the need to set any additional connection properties.

Replication Flows Replication Flows are enabled without the need to set any additional connection properties.

2.1.32 SAP Marketing Cloud Connections

Use an SAP Marketing Cloud connection to access data from SAP Marketing Cloud via its OData-based APIs for
data integration and SAP HANA Smart Data Integration.

Prerequisites

See: Prepare Connectivity to SAP Marketing Cloud

Using the Connection

The connection type supports the remote table as well as the data flow feature.

Supported data access methods for remote tables:

• Remote
• Replication (snapshot)
• Replication (real-time)

For more information, see Replicating Data and Monitoring Remote Tables [page 168].

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 103
Configuring Connection Properties

Connection Details

Property Description

URL Enter the URL with the following syntax: <protocol>://


<host><service path>

Protocol: HTTP or HTTPS. The default value is HTTPS.

Host: Host for accessing the cloud OData service

Service path: Relative path (without host and port) to the


Cloud Data Integration service endpoint (where the CDI pro-
vider service is running at the cloud provider side). The value
must start with a forward slash ( / ).

Root Path [optional] Enter the root path name to restrict browsing to a
certain CDI namespace or provider.

Credentials (User Name and Password)

Property Description

User Name Enter the user name that the application must use for authentication.

Password Enter the password for authentication.

Features

To enable Remote Tables, select a Data Provisioning Agent.

Data Flows are enabled without the need to set any additional connection properties.

2.1.33 SAP SuccessFactors Connections

Use an SAP SuccessFactors connection to access employee-related data in SAP SuccessFactors.

This topic contains the following sections:

• Supported Features [page 105]


• Configuring Connection Properties [page 105]

For information about the required prerequisites, see Prepare Connectivity to SAP SuccessFactors.

Integrating Data and Managing Spaces in SAP Datasphere


104 PUBLIC Integrating Data via Connections
Supported Features

Feature Additional Information

Remote Tables You can use remote tables imported from the connection either to access data directly live
in the connected source (federation) or to copy the data into SAP Datasphere (replication).

 Note
• We strongly recommend to use remote tables for replication only and not to use it
for federated access because of the following reason:
When using remote tables to replicate data from SAP SuccessFactors, the SAP
SuccessFactors connection type uses snapshot-based pagination for SAP Success-
Factors entities that support this method and this way ensures data consistency
(see Snapshot-Based Pagination in the SAP SuccessFactors Platform documenta-
tion). For federated access, snapshot-based pagination is not supported because
of its service limits (see Service Limits for Snapshot-Based Pagination in the SAP
SuccessFactors Platform documentation). Instead, federated access uses client-
side pagination, which potentially can lead to data loss or data duplication (see
Client-Side Pagination in the SAP SuccessFactors Platform documentation).
Please note that since the Generic OData connection type doesn’t use snapshot-
based pagination at all, we do not recommend to use it to federate or replicate SAP
SuccessFactors data with remote tables.
• When replicating employee-related data, independently of their validity not only
current data is replicated but all data including historical data.
• Picklists used in SAP SuccessFactors are not available in remote tables because
they are exposed as navigation properties and not as exposed properties of the
entity. For more information on picklists, see SAP Note 2088679 .

Data Flows You can use the connection to add source objects to a data flow.

Configuring Connection Properties

Connection Details

Property Description

URL Enter the OData service provider URL of the SAP SuccessFactors service that you want to
access.

The syntax for the URL is:

• For V2: <SAP SuccessFactors API Server>/odata/v2 (providing a sup-


ported SAP SuccessFactors service group /<service group> is optional)
• For V4: <SAP SuccessFactors API Server>/odatav4/<supported
SAP SuccessFactors service group>

Version [read-only] Displays the OData version used to implement the SAP SuccessFactors OData
service.

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 105
Authentication

Property Description

Authentication Type Select the authentication type to use to connect to the


OData endpoint.

You can select:

• User Name And Password for basic authentication


• OAuth 2.0

The default is OAuth 2.0.

 Note
HTTP basic authentication in SAP SuccessFactors will
soon be retired. For more information, see Deprecation
of HTTP Basic Authentication in SAP SuccessFactors
What's New Viewer.

OAuth 2.0
If Authentication Type = OAuth 2.0:

Property Description

OAuth Grant Type [read-only] Displays SAML Bearer as the grant type used to
retrieve an access token.

OAuth Token Endpoint Enter the API endpoint to use to request an access token:
<SAP SuccessFactors API Server>/oauth/
token.

OAuth Scope [optional] Enter the OAuth scope, if applicable.

OAuth Company ID Enter the SAP SuccessFactors company ID (identifying the


SAP SuccessFactors system on the SAP SuccessFactors API
server) to use to request an access token.

 Note

When editing a connection that has been created before entering the SAML assertion in the credentials was
required, you can set the Provide SAML Assertion property in the OAuth 2.0 section to true to switch to the
new method and then enter a valid SAML assertion in the Credentials section.

Note that once you change the Provide SAML Assertion property to true, enter the SAML assertion and save
the connection, you cannot turn back to use the /oauth/idp API to generate the SAML assertion.

Integrating Data and Managing Spaces in SAP Datasphere


106 PUBLIC Integrating Data via Connections
Credentials (OAuth 2.0)

If Authentication Type = OAuth 2.0:

Property Description

Client ID Enter the API key received when registering SAP Datasphere
as OAuth2 client application in SAP SuccessFactors.

SAML Assertion Enter a valid SAML assertion that has been generated for
authentication.

 Note
If the SAML assertion expires, the connection gets in-
valid until you update the connection with a new valid
SAML assertion.

Credentials (User Name And Password)

If Authentication Type = User Name And Password:

Property Description

User Name Enter the user name in <username@companyID> format.

Password Enter the password.

Features

Feature Description

Remote Tables Remote Tables are enabled without the need to set any additional connection properties.

Data Flows Data Flows are enabled without the need to set any additional connection properties.

2.1.34 SAP S/4HANA Cloud Connections

Use an SAP S/4HANA Cloud connection to access or import extraction-enabled ABAP Core Data Services
views (ABAP CDS views) from SAP S/4HANA Cloud.

This topic contains the following sections:

• Supported Features [page 108]


• Configuring Connection Properties [page 109]

For information about the required prerequisites in the connected systems and SAP Datasphere, see Prepare
Connectivity to SAP S/4HANA Cloud.

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 107
Supported Features

Feature Additional Information

Remote Tables You can use remote tables imported from the connection either to access data directly live
in the connected source (federation) or to copy the data into SAP Datasphere (replication).

You can access the following data:

• for federation: standard and custom CDS view entities that are exposed using the
ABAP SQL service from SAP S/4HANA Cloud (recommended for federation scenar-
ios)
For more information, see Using ABAP SQL Services for Accessing Data from SAP
S/4HANA Cloud.

 Note
On the Sources tab of the remote-table-related Data Builder editors in SAP
Datasphere, the service binding name from the SQL_SCHEMA authorization field is
visible as (virtual) schema.

• for federation and replication: ABAP CDS Views exposed as OData services for data
extraction. Data is accessed via Cloud Data Integration (legacy)

 Note
This legacy option is still supported, however we recommend using the ABAP SQL
service for federation (if available) and replication flows for replication.

For remote tables, real-time replication is supported. For information about any con-
straints, see Replicate Data Changes in Real-Time [page 181].

Model Import You can use the connection to import semantically-rich entities from SAP S/4HANA Cloud
and all their sources and dependencies. For more information, see Importing Entities with
Semantics from SAP S/4HANA.

Data Flows You can use the connection to add source objects to a data flow.

You can access the following data:

• extraction-enabled ABAP CDS views that are C1-released, that is views with annotation
@Analytics.dataextraction.enabled: true and that are available in the
connected system (access via ABAP Pipeline Engine)
For more information, see:
• Data Integration in the SAP S/4HANA Cloud documentation
• CDS Views Enabled for Data Extraction in the SAP S/4HANA Cloud documentation

 Note
• Data flows currently support one-time loading of data (also known as initial load)
only.
• The data preview in the data flow editor of the Data Builder is only available if the
version of the connected system is SAP S/4HANA Cloud 2302 or higher.

Integrating Data and Managing Spaces in SAP Datasphere


108 PUBLIC Integrating Data via Connections
Feature Additional Information

Replication Flows You can use the connection to add source objects to a replication flow. (recommended for
replication scenarios)

You can replicate the following data:

• standard and custom CDS view entities that are exposed using the ABAP SQL service
from SAP S/4HANA Cloud
For more information, see Data Consumption Using SAP Datasphere.

 Note
When adding source objects in replication flows, you can find the CDS view entities
in the SQL_SERVICE container.

• extraction-enabled ABAP CDS views that are C1-released, that is views with annotation
@Analytics.dataextraction.enabled: true and that are available in the
connected system (access via ABAP Pipeline Engine)
For more information, see:
• Data Integration in the SAP S/4HANA Cloud documentation
• CDS Views Enabled for Data Extraction in the SAP S/4HANA Cloud documentation

 Note
When adding source objects in replication flows, you can find the CDS views in the
CDS_EXTRACTION container.

 Note

The connection type supports replication with both remote tables via Data Provisioning Agent (SAP HANA
Smart Data Integration) and replication flows.

Regardless of whether you use the same connection or two separate connections, for replicating data from
a dedicated source object in the source we strongly recommend to only use remote tables or replication
flows and not both. Generally, for replication scenarios, we recommend to use replication flows.

Configuring Connection Properties

Connection Details

Property Description

Application Server Enter the name of the API-URL from the communication arrangements that
you set up in SAP S/4HANA Cloud (without https://), for example myXXXXX-
api.s4hana.ondemand.com or myXXXXX-api.s4hana.cloud.sap.

For more information about the required communication arrangements, see Prepare Con-
nectivity to SAP S/4HANA Cloud.

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 109
Property Description

Client Enter the system client number.

System ID Enter the system ID of the system to which you want to connect.

Language Enter a two-digit ISO language code, EN for English or DE for German, for example. Object
and field descriptions in the data flow editor of the Data Builder are then shown in the
specified language. If you don't enter any language code, the application uses the default
logon language of the SAP S/4HANA system.

Port Enter the port number of the WebSocket RFC connection endpoint. The default port num-
ber is 443.

Authentication

Property Description

Authentication Type Select the authentication type to use to connect to SAP S/4HANA Cloud.

You can select:

• X.509 Client Certificate


• User Name And Password for basic authentication

 Note
For federated access with remote tables using the ABAP SQL service, authentica-
tion with User Name and Password is required.

The default is X.509 Client Certificate.

Credentials (X.509 Client Certificate)

If Authentication Type = X.509 Client Certificate:

Property Description

X.509 Client Certificate To upload the certificate or certificate chain that is used to authenticate to the remote
system, click  (Browse) and select the file from your download location.

X.509 Client Private Key To upload the private key, click  (Browse) and select the file from your download location.

X.509 Client Private Key [optional] If the private key is encrypted, enter the password required for decryption.
Password

Credentials (User Name and Password)

If Authentication Type = User Name And Password:

Property Description

User Name Enter the user name that is used to connect to the SAP S/4HANA system.

Password Enter the user password.

Integrating Data and Managing Spaces in SAP Datasphere


110 PUBLIC Integrating Data via Connections
Remote Tables

Property Description

Data Provisioning Option Select the middleware to use when connecting to and accessing SAP S/4HANA Cloud:

• None (default): if you don't want to use remote tables.


• Data Provisioning Agent: if you want to use federation and replication via SAP HANA
smart data integration CloudDataIntegrationAdapter (legacy).

 Note
This legacy option is still supported, however we recommend using the ABAP SQL
service for federation (if available) and replication flows for replication.

• Direct: if you want to federate data from the source objects of the connection and ac-
cess them remotely in the source system using the ABAP SQL service (recommended
for federation scenarios).

 Note
This option is supported for authentication type User Name And Password.

[if Data Provisioning Option [read-only] Displays how data from source objects can be accessed:
= Data Provisioning Agent or
Direct] Data Access • Remote and Replication: if you selected data provisioning option Data Provisioning
Agent (using the SAP HANA smart data integration CloudDataIntegrationAdapter).
• Remote Only: if you selected data provisioning option Direct (using the ABAP SQL
service).

[if Data Provisioning Option Select an agent from the list of agents configured for SAP Datasphere.
= Data Provisioning Agent]
Data Provisioning Agent

[if Data Provisioning Option = Specify how ABAP data types are mapped to SAP HANA data types:
Direct] Type Map
• native: no conversion of ABAP data types
• semantic: conversion of the ABAP data types to an ABAP-independent data format
• semanticDatsTimsAsWchar: like semantic, except that the ABAP types DATS and TIMS
are mapped to the ODBC type SQL_WCHAR to allow for lossless conversion of date
literals and time literals

Features

Feature Description

Remote Tables To enable Remote Tables, complete the connection properties in the Remote Tables section.

 Note
For federation scenarios, we recommend using the ABAP SQL service (data provision-
ing option Direct).

Model Import To enable Model Import, select a Data Provisioning Agent.

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 111
Feature Description

Data Flows Data Flows are enabled without the need to set any additional connection properties.

Replication Flows Replication Flows are enabled without the need to set any additional connection properties.

 Note
For replication scenarios, we recommend using replication flows.

2.1.35 SAP S/4HANA On-Premise Connections

Use an SAP S/4HANA On-Premise connection to access data from SAP S/4HANA on-premise systems.

This topic contains the following sections:

• Supported Features [page 113]


• Configuring Connection Properties [page 117]

For information about the required prerequisites in the connected systems and SAP Datasphere, and about
supported source system versions, see Prepare Connectivity to SAP S/4HANA On-Premise.

Integrating Data and Managing Spaces in SAP Datasphere


112 PUBLIC Integrating Data via Connections
Supported Features

Feature Additional Information

Remote Tables You can use remote tables imported from the connection either to access data directly live
in the connected source (federation) or to copy the data into SAP Datasphere (replication).

You can access the following data:

• for federation: standard and custom CDS view entities that are exposed using the
ABAP SQL service from SAP S/4HANA (recommended for federation scenarios)
For more information, see Using ABAP SQL Services for Accessing Data from SAP
S/4HANA.

 Note
On the Sources tab of the remote-table-related Data Builder editors in SAP
Datasphere, the service binding name from the SQL_SCHEMA authorization field is
visible as (virtual) schema.

• for federation and replication: (legacy)


• Data from virtual tables through RFC for ODP sources (extractors):
• Extraction context ABAP_CDS provides access to ABAP Core
Data Services (CDS) Views that include the annotation
@Analytics.dataextraction.enabled: true
For information about which ABAP CDS Views are available for extraction, see
CDS Views Enabled for Data Extraction in the SAP S/4HANA documentation.

 Note
ABAP CDS Views are the preferred method over SAPI extractors when
integrating SAP S/4HANA data with SAP Datasphere.

• Extraction context BW provides access to InfoProviders in SAP BW or SAP


BW∕4HANA systems:
• DataStore objects (advanced or classic)
• CompositeProviders
• Queries as InfoProviders (with a subset of OLAP capabilities) - can be
enabled in the runtime properties of queries in the BW Modeling tools
• InfoObjects (characteristics) - can be enabled in the general settings of
the InfoObject editor in the BW Modeling tools
• Extraction context SAPI provides access to Service API (SAPI) DataSources

 Note
Importing SAPI DataSources into SAP Datasphere is not supported if the
DataSource contains mandatory selection fields. You will receive an error
message when trying to import tables for such DataSources.

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 113
Feature Additional Information

 Note
Certain SAPI DataSources (from FI-AA Business Content, for example)
may send duplicate records during snapshot replication or real-time initi-
alization. Remote table replication runs with optimized INSERT process-
ing on DataSource key level which leads to unique constraint violation
errors. In such cases, we recommend to use a data flow which uses the
remote table as source and for which the target table has selected the
Append mode with the Update Records By Primary Key (UPSERT) option.

• Data from SAP ABAP Dictionary tables in SAP S/4HANA systems


Real-time replication is supported for ODP sources. For information about any con-
straints, see Replicate Data Changes in Real-Time [page 181].

 Note
This legacy option is still supported, however we recommend using the ABAP SQL
service for federation (if available) and replication flows for replication.

Model Import You can use the connection to import semantically-rich entities from an SAP S/4HANA
system which you have selected in the Import Entities wizard. For more information, see
Importing Entities with Semantics from SAP S/4HANA.

You can import extraction-enabled ABAP CDS views that are C1-released, that is views with
annotation @Analytics.dataextraction.enabled: true and that are available
in the connected system. For more information, see CDS Views Enabled for Data Extraction
in the SAP S/4HANA documentation.

The options to access the data depend on the version of the SAP S/4HANA system. You
can see the data access option when you have selected a connection in the Import Entities
wizard:

• Replication Flow to Local Tables - available if the connected system is SAP S/4HANA
2021 or higher (SAP_BASIS 756 and higher). It's the default and recommended option
for these system versions.
• Remote Tables - available if the connected system is SAP S/4HANA 1809 or higher
(SAP_BASIS 753 and higher)
In the Import Entities wizard, you can change from the default Replication Flow to Local
Tables to Remote Tables if your system is SAP S/4HANA 2021 or higher (SAP_BASIS
756 and higher).

 Note
Note that associated entities can only be included in the import when data access
is Replication Flow to Local Tables. For more information, see Importing Entities with
Semantics from SAP S/4HANA.

Integrating Data and Managing Spaces in SAP Datasphere


114 PUBLIC Integrating Data via Connections
Feature Additional Information

Data Flows You can use the connection to add source objects to a data flow.

You can replicate the following data:

• extraction-enabled ABAP CDS views that are C1-released, that is views with annotation
@Analytics.dataextraction.enabled: true and that are available in the
connected system (access via ABAP Pipeline Engine)
For more information, see CDS Views Enabled for Data Extraction in the SAP S/4HANA
documentation.
• Tables from SAP Landscape Transformation Replication Server (SAP LT Replication
Server) sources, for which an appropriate configuration has been created in SAP LT
Replication Server.

 Note
• The availability of the data flow feature depends on the used version and Support
Package level of SAP S/4HANA or the DMIS addon in the source. Make sure your
source systems meet the required minimum versions. We recommend to use the
latest available version of SAP S/4HANA and the DMIS add-on where possible and
have the latest SAP notes and TCI notes implemented in your systems.
For more information about required versions, recommended system landscape,
considerations for the supported source objects, and more, see SAP Note
2890171 .
• Data flows currently support one-time loading of data (also known as initial load)
only.
• The data preview in the data flow editor of the Data Builder is not available, except
for:
• CDS views if the version of the connected system is SAP S/4HANA 1909 or
higher.
• SAP Landscape Transformation Replication Server objects if ABAP Add-on
DMIS 2018 SP08 or DMIS 2020 SP04 is installed or a higher DMIS version.

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 115
Feature Additional Information

Replication Flows You can use the connection to add source objects to a replication flow.

You can replicate the following data:

• extraction-enabled ABAP CDS views that are C1-released, that is views with annotation
@Analytics.dataextraction.enabled: true and that are available in the
connected system (access via ABAP Pipeline Engine)
For more information, see CDS Views Enabled for Data Extraction in the SAP S/4HANA
documentation.
• Data from ODP sources (extraction contexts SAPI and BW)

 Note
The availability of the replication flow feature depends on the used version and Support
Package level of SAP S/4HANA or the DMIS addon in the source. Make sure your
source systems meet the required minimum versions. We recommend to use the latest
available version of SAP S/4HANA and the DMIS add-on where possible and have the
latest SAP notes and TCI notes implemented in your systems.

For more information about required versions, recommended system landscape, con-
siderations for the supported source objects, and more, see SAP Note 2890171 .

 Note

The connection type supports replication with both remote tables via Data Provisioning Agent (SAP HANA
Smart Data Integration) and replication flows (for on-premise systems via Cloud Connector).

Regardless of whether you use the same connection or two separate connections, for replicating data from
a dedicated source object in the source we strongly recommend to only use remote tables or replication
flows and not both.

Best Practices for Integrating Data


See SAP Data Warehouse Cloud - First Guidance: Data Integration for ABAP Source Systems (published
February 2022) for general best practices for integrating data from ABAP source systems with SAP
Datasphere.

Integrating Data and Managing Spaces in SAP Datasphere


116 PUBLIC Integrating Data via Connections
Configuring Connection Properties

Connection Details

Property Description

SAP Logon Connection Type Select Application Server to connect to a single application server, or select Message Server
to use load balancing and connect to a message server that distributes the load to different
application servers.

 Note
The SAP Logon connection type you select here must match the connection type in
the system mapping in the Cloud Connector configuration (load balancing logon or
connecting to a specific application server).

[if SAP Logon Connection Enter the name of the application server to which you want to connect to.
Type = Application Server]
Application Server

[if SAP Logon Connection Enter the SAP ABAP system instance number.
Type = Application Server]
System Number

[if SAP Logon Connection Enter the name of the message server to which you want to connect to.
Type = Message Server]
Message Server

[if SAP Logon Connection Enter the message server port (numerical).
Type = Message Server]
Message Server Port

[if SAP Logon Connection Enter the message server group.


Type = Message Server]
Message Server Group

Client Enter the system client number.

System ID Enter the system ID.

Language Enter the two-digit ISO language code, EN for English or DE for German, for example. In the
Data Builder, object and field descriptions are then shown in the specified language. If not
set, the application uses the default logon language of the ABAP system for data flows, and
EN is used for remote tables and views.

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 117
Cloud Connector

Property Description

Use Cloud Connector Set to true if you want to use the connection for one of the following features:

• remote tables for data federation via the ABAP SQL service
• data flows
• replication flows
• model import

The default is false.

 Note
When the connection is configured for using the ABAP SQL service for data federation
with remote tables, you can't use the same connection for model import.

[if Use Cloud Connector = [optional] Select a location ID.


true] Location
 Note
To select another location ID than the default location, Connection.Read privilege is
required. The privilege is neither included in the DW Integrator nor in the DW Space
Administrator role. If you need to select a location ID, ask your tenant administrator to
either assign your user to a global role that is based on the DW Administrator role or
to assign your user to a custom global role (with license type SAP Datasphere) that
includes the required Connection.Read privilege.

[if Use Cloud Connector = [optional] Select Derive Virtual Host and Port from Connection Details or Enter Virtual Host
true] Virtual Destination and Port in Separate Fields.

If host and port entered in the connection details match the virtual host and port from the
Cloud Connector configuration, you can select Derive Virtual Host and Port from Connection
Details and don't need to enter the values manually.

 Note
When you select Derive Virtual Host and Port from Connection Details for ABAP-based
connections, virtual port is set to sapgw<system number> to make sure that it
matches the virtual port defined in the Cloud Connector configuration.

 Note
When the connection is configured for using the ABAP SQL service for data federation
with remote tables, you must enter virtual host and port manually. Deriving virtual host
and port from the connection details is not supported because the Cloud Connector
configuration requires two system mappings having different virtual ports (one for RFC
protocol for data flows and replication flows, and one for HTTP/HTTPS protocoll for
remote tables using the ABAP SQL service).

[if Virtual Destination = Enter Enter the virtual host that you defined during Cloud Connector configuration.
Virtual Host and Port in
Separate Fields] Virtual Host

Integrating Data and Managing Spaces in SAP Datasphere


118 PUBLIC Integrating Data via Connections
Property Description

[if Virtual Destination = Enter Enter the virtual port that you defined during Cloud Connector configuration.
Virtual Host and Port in
Separate Fields] Virtual Port  Note
• If the SAP Logon Connection Type for your connection is Application Server: Enter
sapgw<system number> to make sure that it matches the virtual port defined
in the Cloud Connector configuration.
• If the SAP Logon Connection Type for your connection is Message Server: Enter
a numerical port and make sure it is the port defined in the Cloud Connector
configuration.

Authentication

Property Description

Authentication Type Select the authentication type to use to connect to SAP S/4HANA.

You can select:

• User Name And Password for basic authentication (default value) - This option is read-
only if you set Cloud Connector to false.
• OAuth 2.0 - You can select this option only if you have set Cloud Connector to true to
enable replication flows and data flows. Remote tables and model import currently are
not supported with OAuth authentication.

OAuth 2.0

Property Description

OAuth Grant Type Displays Client Credentials as grant type used to retrieve an
access token.

OAuth Token Endpoint Enter the API endpoint to use to request an access token.

OAuth Scope [optional] Enter the OAuth scope, if applicable.

OAuth Token Request Content Type [optional] Enter the content type of the OAuth2 token re-
quest.

You can select:

• URL Encoded (default value)


OAuth2 token request parameters will be url-encoded
and included in the HTTP request body.
• JSON
OAuth2 token request parameters will be in JSON for-
mat and included in the HTTP request body.

Credentials (User Name And Password)

Property Description

User Name Enter the user name that is used to connect to the SAP ABAP system.

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 119
Property Description

Password Enter the user password.

Credentials (OAuth 2.0)

Property Description

Client ID Enter the client ID to authenticate SAP Datasphere to the


SAP S/4HANA system.

Client Secret Enter the client secret to authenticate SAP Datasphere to


the SAP S/4HANA system.

Remote Tables

Property Description

Data Provisioning Option Select the middleware to use when connecting to and accessing SAP S/4HANA:

• None (default): if you don't want to use remote tables.


• Data Provisioning Agent: if you want to use federation and replication via SAP HANA
smart data integration ABAPAdapter (legacy).

 Note
This legacy option is still supported, however we recommend using the ABAP SQL
service for federation (if available) and replication flows for replication.

• Cloud Connector: if you want to federate data from the source objects of the connec-
tion and access them remotely in the source system using the ABAP SQL service
(recommended for federation scenarios).

[if Data Provisioning Option [read-only] Displays how data from source objects can be accessed:
= Data Provisioning Agent
or Cloud Connector] Data • Remote and Replication: if you selected data provisioning option Data Provisioning
Access Agent (using the SAP HANA smart data integration ABAPAdapter).
• Remote Only: if you selected data provisioning option Cloud Connector (using the ABAP
SQL service).

[if Data Provisioning Option Select an agent from the list of agents configured for SAP Datasphere.
= Data Provisioning Agent]
Data Provisioning Agent

[if you selected an agent] Set to On if you want to use ABAP RFC streaming for loading tables based on small batches.
Streaming Read Set to Off to use non-RFC streaming. The default is Off.

[if Streaming Read = On] Enter the gateway host where the ABAP adapter would register an RFC server instance to
Gateway Host receive callbacks with the table data batches. Usually, the gateway host is the same as the
target ABAP system host.

[if Streaming Read = On] Enter the port number of the gateway server where the ABAP adapter would register an
Gateway Port RFC server instance to receive callbacks with the table data batches. You can find the
port number for sapgw<ABAP_instance_number> in file /etc/services on the
ABAP host. The default port number is 33<ABAP_instance_number>. If you have a
file /etc/services with this mapping on the Data Provisioning Agent’s host, you can
also enter sapgw<ABAP_instance_number> instead of the port number.

Integrating Data and Managing Spaces in SAP Datasphere


120 PUBLIC Integrating Data via Connections
Property Description

[if Streaming = On] RFC Enter the name of the RFC destination that you have created in the source. For more
Destination information, see Prerequisites for ABAP RFC Streaming.

[if Data Provisioning Option = Specify how ABAP data types are mapped to SAP HANA data types:
Cloud Connector] Type Map
• native: no conversion of ABAP data types
• semantic: conversion of the ABAP data types to an ABAP-independent data format
• semanticDatsTimsAsWchar: like semantic, except that the ABAP types DATS and TIMS
are mapped to the ODBC type SQL_WCHAR to allow for lossless conversion of date
literals and time literals

Replication Flows

Property Description

Fast Serialization Turn on RFC fast serialization to help improve replication flow performance by avoiding
unnecessary, costly conversions in ABAP. This can be useful particularly in cases of large
data volumes. The default is On.

 Note
• Fast serialization has no impacts on data flows.
• Fast Serialization for replication flows is not related to RFC Serialization in the
advanced properties of connections using a Data Provisioning Agent for remote
tables.
• For information about the prerequisites for using fast serialization in SAP
Datasphere, see Prepare Connectivity to SAP S/4HANA On-Premise.
Use the connection validation to check if the prerequisites in the source system for
using fast serialization are met.

Features

Feature Description

Remote Tables To enable Remote Tables, complete the connection properties in the Remote Tables section.

Data Flows Data Flows are enabled without the need to set any additional connection properties. Make
sure you have maintained the properties in the Cloud Connector section.

Replication Flows Replication Flows are enabled without the need to set any additional connection properties.
Make sure you have maintained the properties in the Cloud Connector section.

Model Import To enable Model Import, select the live data connection that connects to your SAP S/
4HANA system and select a Data Provisioning Agent.

 Note
When the connection is configured for using the ABAP SQL service for data federation
with remote tables, you can't use the same connection for model import.

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 121
Advanced Connection Properties

When you've selected a Data Provisioning Agent, you can configure advanced connection properties to enable
specific customer scenarios. In the connection creation wizard, you configure advanced properties in an
additional wizard step. When editing the connection to configure the properties, scroll down and click the Show
Advanced Properties button. Advanced properties have default values that you can override according to your
customer scenario’s needs.

Available properties:

• Connection Pool Size


• Connection Limit
• Extractor’s ODP Version
• CDC Batch Size, MB
• RFC Unicode
• [if RFC Unicode = No] RFC Character Encoding
• SNC Mode
• [if SNC Mode = On] SNC Library
• [if SNC Mode = On] SNC Name of Client
• [if SNC Mode = On] SNC Name of SAP Server
• [if SNC Mode = On] SNC SSO
• [if SNC Mode = On] SNC Quality of Protection
• RFC Serialization (requires Data Provisioning Agent version 2.3.7.2 or higher)
• [if Streaming Read = On] Batch Size, MB
• [if Streaming Read = On] Batch Receive Timeout
• [if Streaming Read = On] RFC Trace

For more information, see SAP ABAP Adapter Remote Source Configuration in the SAP HANA Smart Data
Integration and SAP HANA Smart Data Quality Installation and Configuration Guide.

2.1.36 Connections to Partner Tools

Extend connectivity beyond SAP Datasphere standard remote connectivity and cover additional data sources
that are available with partner tools.

Connections to partner tools are created differently compared to other connections in SAP Datasphere.

When creating a connection to a partner tool, an Open SQL schema is generated in SAP Datasphere and
information about host, port and credentials of the Open SQL schema is transferred to the partner tool. To be
able to send data to the Open SQL schema, during the connection creation process the partner tool on their
side establishes a connection to the Open SQL schema.

To successfully validate a new connection and before data can be written to the Open SQL schema and you can
use it in the Data Builder, the IP address or addresses of the partner tool need to be added to the IP allowlist in
SAP Datasphere. To get the relevant IP addresses, depending on the partner tool you might need to contact the
Account Manager or the support team (for Adverity, for example), or you can find and copy the IP addresses in
the last step of the connection cration wizard (for Precog, for example).

Integrating Data and Managing Spaces in SAP Datasphere


122 PUBLIC Integrating Data via Connections
After the connection has been created in the SAP Datasphere space, you can use it in the Data Builder to create
a view.

 Note

• The Open SQL schema (respectively the database user for which the schema has been created)
doesn't appear in the Connections app but only in the Database Users section in Space Management.
• In the Data Builder, unlike other Open SQL schemas, the Open SQL schema generated with the partner
connection doesn't appear in Source Browser Sources . Instead, you can find the corresponding
partner connection in Source Browser Sources Connections .
Once the partner tool fetches data from a source, in the Data Builder you can find the corresponding
table in Source Browser Sources Connections <partner connection name> . When importing
the table into the Data Builder, a local table is created and deployed. You can use it the same way as any
other table originating from an Open SQL schema.

You can access the Open SQL Schema that has been generated for the partner connection from a third-party
BI client via ODBC. For more information, see Consume Data in Power BI and Other Clients, Tools, and Apps via
ODBC/JDBC.

Related Information

Adverity Connections [page 20]


Precog Connections [page 66]

2.2 Edit a Connection

Edit a connection, for example if the setup of the connected remote system changed, if the connection has
been imported from another tenant, or if a new feature is available for the connection type after you have
created your connection. A warning in the connections list might indicate the need to edit a connection.

This topic contains the following sections:

• Use Cases for Editing a Connection [page 123]


• Restrictions When Editing Connections [page 124]
• Edit a Connection [page 125]
• Add a Connection to a Package [page 125]

Use Cases for Editing a Connection

You need to edit connections in various cases, for example:

• In case of password or credential rotation.

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 123
• When there's changes in the setup of the connected remote system
Changes in the remote system setup must also be made in the connection by updating the corresponding
properties, for example changes in the authentication method, host changes, changes in security settings,
or changes in the Data Provisioning Agent.
• When the connection has been imported
When a connection has been imported into the tenant, it cannot directly be used. You must first complete
the connection configuration and provide the credentials to create a runtime version for the connection
before you can use it.
• When you want to add the connection to a package to facilitate transport between tenants
• When a new feature is supported by the connection type
For an existing connection that has been created before the connection type supported a new feature, you
must re-enter the credentials before you can use the new feature.
• When you want to change the business name or description

Restrictions When Editing Connections

Connections to Partner Tools


You cannot edit connections to partner tools.

Connections Using a Data Provisioning Agent


Consider the following when editing a connection which uses a Data Provisioning Agent:

• When the connection isn't used and its real-time replication status is Inactive, you can change any property
except for the technical name.
• When you use the connection and its real-time replication status is Active, you can only change the
business name and description. If no package has been assigned yet, you can select a package. Note that
credential changes are not considered when saving the connection.
If you want to change other connection properties, pause real-time replication in the Connections app, or if
required disable real-time replication for all affected remote tables in the Data Integration Monitor.
• When the connection has real-time replication status Active but the Data Provisioning Agent is
disconnected, you can only change the business name and description. If no package has been assigned
yet, you can select a package. Note that credential changes are not considered when saving the
connection.
If you want to change other connection properties, you must first reconnect the agent and pause real-time
replication in the Connections app, or if required disable real-time replication for all affected remote tables
in the Data Integration Monitor.
• When the connection has real-time replication status Paused, you can only change a subset of the
properties. It depends on the connection type if any properties can be changed at all and if yes which
properties can be changed. Note that the user name cannot be changed in this case.
When the Data Provisioning Agent is disconnected, you need to reconnect the agent or use a different
agent to enable editing specific connection properties.

Integrating Data and Managing Spaces in SAP Datasphere


124 PUBLIC Integrating Data via Connections
Edit a Connection

1. In the side navigation area, click  (Connections) and select a space if necessary.
2. Select the relevant connection and click Edit.

 Tip

For imported connections, for example, and for other connections with warning messages, you can
click the warning button on the top right corner of the app to select the connection by then clicking the
corresponding message title.

3. Make the required changes.


4. For empty and editable credential fields: Re-enter your credentials.
5. Click Save.
6. Select the connection and click Validate to receive detailed status information.

Add a Connection to a Package

Users with the DW Space Administrator role (or equivalent privileges) can create packages in the Packages
editor. Once a package is created in your space, you can select it when editing a connection.

Packages are used to group related objects in order to facilitate their transport between tenants.

 Note

Once a package is selected, it cannot be changed here. Only a user with the DW Space Administrator role
(or equivalent privileges) can modify a package assignment in the Packages editor.

For more information, see Creating Packages to Export [page 292].

2.3 Validate a Connection

Select a connection and open the validation message to get detailed status information.

1. In the side navigation area, click  (Connections) and select a space if necessary.
2. Select the relevant connection and click Validate.
The message informs you about the availability of supported features and provides details in case of errors
in the connection.

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 125
2.4 Delete a Connection

Select one or more connections and delete them if they are not used.

1. In the side navigation area, click  (Connections) and select a space if necessary.
2. Select one or more connections and click Delete.

If the connection is still used for accessing data via remote tables, for importing models, or for data flows or
replication flows, you cannot delete it. A message informs you of any dependent objects.

2.5 Pause Real-Time Replication for a Connection

Pause real-time replication for a connection while updating or maintaining the source. After you have finished
with your maintenance activities you can restart real-time replication.

Context

You can pause real-time replication for connections that connect to a source through SAP HANA Smart Data
Integration and its Data Provisioning Agent.

If you need to upgrade or patch your Data Provisioning Agent, or you need to apply any changes to its
configuration, you can pause all connections for the Data Provisioning Agent. For more information, see Pause
Real-Time Replication for an Agent.

Procedure

1. In the side navigation area, click  (Connections) and select a space if necessary.
2. In the list of connections, you can see the Real-Time Replication Status for each connection. Depending on
the status you can perform different actions.

Real-Time Replication status Possible actions

Not Applicable none

Shown when the connection type doesn't support real-time replication.

Shown when the Remote Tables feature for an SAP HANA connection is set to Data
Access = Remote Only.

Integrating Data and Managing Spaces in SAP Datasphere


126 PUBLIC Integrating Data via Connections
Real-Time Replication status Possible actions

Active Pause

Shown when the connection type supports real-time replication and for the connection
at least one table is replicated via real-time replication (even if the status in the Remote
Tables monitor is Error).

Paused Restart

Shown when the connection type supports real-time replication and for the connection
at least for one table real-time replication is paused.

Inactive Pause

Shown when the connection type supports real-time replication and for the connection
currently there is no table replicating via real-time replication.

Disconnected none

Shown when the connection uses a Data Provisioning Agent that currently has status
Disconnected.

3. From the list of connections, select the connection in question and click Pause.

The status in the list of the connections changes from Active or Inactive to Paused.

In the Remote Tables monitor, the status for affected tables changes to Paused and actions related to
real-time replication are not available for these tables. Also, you cannot start real-time replication for any
table of a paused connection.
4. You can now perform maintenance activities in the source.
5. Once your maintenance activities finished, restart real-time replication. Select the connection in question
and click Restart.

The status in the list of the connections as well as in the Remote Tables monitor changes accordingly and
you can again perform real-time related actions for the tables or start real-time replication.

2.6 Filter Connection List

Select one or more filter values to restrict the connection list according to your needs.

1. In the side navigation area, click  (Connections) and select a space if necessary.
2. Click  (Filter Connections) and select one or more filter values to get a better overview of the
connections in the list.

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 127
The following filter categories and values are available:

Filter Category Values

Features that a connection type supports Data Flows

Model Import

Remote Tables

Replication Flows

Categories that the corresponding source belongs to Cloud

On-Premise

Sources that you would like to connect Non-SAP

Partner Tools

SAP

2.7 Importing SAP BW Queries – A Comparison of the


Different Options
In SAP Datasphere, there are different ways to import SAP BW queries.

You may ask yourself, when to use which option. Here is a summary of the differences.

These are the different options to import a SAP BW query:

• queries as InfoProviders using a SAP BW connection


• queries as InfoProviders using a SAP ABAP connection
• queries using a SAP BW/4HANA Model Transfer connection

In all three cases, remote tables are created.

These are the differences:

• For both the SAP BW connection and the SAP ABAP connection, only queries as InfoProviders are
supported. For this type of queries, there are some restrictions. For more information, see Query as
InfoProvider.
• The difference between the SAP BW connection and the SAP ABAP connection is, that the SAP BW
connection is a subset of the SAP ABAP connection, being restricted to SAP BW as a source whereas the
SAP ABAP connection is more comprehensive to other on-premise source systems.
• When you use the SAP BW/4HANA Model Transfer connection, you can import standard analytic queries,
and you don’t have to deal with the restrictions of queries as InfoProviders. But it is only available for SAP
BW/4HANA, not for SAP BW.
• For SAP BW, the SAP BW connection and the SAP ABAP connection can be used, but not the SAP BW/
4HANA Model Transfer connection. This means for SAP BW, you can only import queries as InfoProviders.
• For SAP BW/4HANA, you can use all three options: the SAP BW connection, the SAP ABAP connection,
and the SAP BW/4HANA Model Transfer connection.

Integrating Data and Managing Spaces in SAP Datasphere


128 PUBLIC Integrating Data via Connections
• If you want to provide SAP BW/4HANA business semantics to SAP Datasphere, you can use connection
type SAP BW/4HANA Model Import and import SAP BW/4HANA models based on analytic queries.
During the import, the semantics of the underlying SAP BW/4HANA models are translated into native SAP
Datasphere entities.

2.8 Manage Connectivity via REST APIs

You can manage TLS server certificates and connections via the Certificates and Connections REST APIs.
Creating and editing connections via the API is supported for SAP SuccessFactors connections only.

This topic contains the following sections:

• Introduction [page 129]


• Log in with an OAuth Client [page 130]
• Obtain a CSRF Token [page 130]
• List TLS Server Certificates [page 131]
• Upload TLS Server Certificates [page 131]
• Delete TLS Server Certificates [page 131]
• List Connections in a Space [page 132]
• Read Connection Details [page 132]
• Create Connection to SAP SuccessFactors [page 132]
• Validate Connections [page 134]
• Edit Connection to SAP SuccessFactors [page 135]
• Delete Connections [page 135]

Introduction

SAP Datasphere exposes REST APIs that allow you to programmatically manage TLS server certificates and
connections using dedicated endpoints.

The API specifications are available at the SAP Business Accelerator Hub .

Using the Certificates API, you can perform the following actions:

• List TLS server certificates


• Upload and delete TLS server certificates

To manage certificates via the REST API, you must have an administrator role or equivalent privileges:

• System Information (-RU-----)

Using the Connections API, you can perform the following actions:

• List connections in a space


• Read connection details
• Validate and delete connections

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 129
• Create and edit connections to SAP SuccessFactors

To manage connections via the REST API, you must have a scoped role based on the DW Integrator role or
equivalent privileges:

• Data Warehouse Connection (CRUD----)

You must, in addition, obtain the following parameters for an OAuth client created in your SAP Datasphere
tenant (with Purpose set to Interactive Usage and the Redirect URI set to the URI provided by the client, tool, or
app that you want to connect):

• Client ID
• Secret
• Authorization URL
• Token URL.

For more information about obtaining the OAuth client parameters, see Create OAuth2.0 Clients to
Authenticate Against SAP Datasphere.

Log in with an OAuth Client

With the OAuth Client created in your SAP Datasphere tenant, your client, tool, or application will be able to
authenticate against the system’s OAuth service and obtain authorization to authenticate against the API.

To obtain an OAuth access token, send a GET request to one of the API endpoints (see below) including
Authorization: Auth Type <OAuth 2.0> and passing the <client ID>, <secret>, <authorization
URL>, and <token URL> values.

You must pass the access token in the Authorization header of all requests that you make to the API.

Example syntax of the GET request:

 Sample Code

https://<tenant_url>/api/v1/datasphere/configuration/security/certificates

Obtain a CSRF Token

You must have a valid x-csrf-token before creating a POST, PUT, or DELETE request to the API endpoints.

You can get a token by sending a GET request to one of the API endpoints (see below) including the x-csrf-
token: fetch header.

The CSRF token is returned in the x-csrf-token response header. You must pass the token in the x-csrf-
token:<token> header of all POST, PUT, or DELETE requests that you make to the API.

Example syntax of the GET request:

Integrating Data and Managing Spaces in SAP Datasphere


130 PUBLIC Integrating Data via Connections
 Sample Code

https://<tenant_url>/api/v1/datasphere/configuration/security/certificates

List TLS Server Certificates

To retrieve a list of TLS server certificates available in the tenant, use the GET request with the/api/v1/
datasphere/configuration/security/certificates endpoint and enter:

https://<tenant_url>/api/v1/datasphere/configuration/security/certificates

Upload TLS Server Certificates

To upload a TLS server certificate, use the POST request with the/api/v1/datasphere/configuration/
security/certificates endpoint and enter:

https://<tenant_url>/api/v1/datasphere/configuration/security/certificates

 Note

The file that you upload must have a supported file extension: .pem (privacy-enhanced mail), .crt, or .cer.

Delete TLS Server Certificates

To delete a TLS server certificate from the tenant, use the DELETE request with the/api/v1/datasphere/
configuration/security/certificates/<fingerprint> endpoint and enter:

https://<tenant_url>/api/v1/datasphere/configuration/security/certificates/
<fingerprint>

 Note

You can retrieve the fingerprint from the list of certificates.

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 131
List Connections in a Space

To list all connections of a space, use the GET request with the/api/v1/datasphere/spaces/<space_id>/
connections endpoint and enter:

https://<tenant_url>/api/v1/datasphere/spaces/<space_id>/connections

Read Connection Details

To read the JSON definition of a connection (without its credentials) in a space , use the GET request
with the/api/v1/datasphere/spaces/<space_id>/connections/<connection_technical_name>
endpoint and enter:

https://<tenant_url>/api/v1/datasphere/spaces/<space_id>/connections/
<connection_technical_name>

Create Connection to SAP SuccessFactors

To create an SAP SuccessFactors connection in a space by providing the connection definition in stringified
json format, use the POST request with the/api/v1/datasphere/spaces/<space_id>/connections
endpoint and enter:

https://<tenant_url>/api/v1/datasphere/spaces/<space_id>/connections

The following example shows how to create a connection to SAP SuccessFactors for OData V4 and basic
authentication:

 Sample Code

{
"name": "<technical name>",
"businessName" : "<business name>",
"description":"<description>",
"authType": "Basic",
"url": "https://<SAP SuccessFactors API Server>/odatav4/<supported SAP
SuccessFactors service group>",
"version" :"V4",
"username": "<username>",
"password":"<password>"
}

The following example shows how to create a connection to SAP SuccessFactors for OData V2 and OAuth2
authentication:

 Sample Code

{
"name":"<technical name>",

Integrating Data and Managing Spaces in SAP Datasphere


132 PUBLIC Integrating Data via Connections
"businessName":"<business name>",
"description":"<description>",
"authType": "<OAuth2>",
"url": "https://<SAP SuccessFactors API Server>/odata/v2/",
"version": "V2",
"oauth2GrantType":"saml_bearer",
"oauth2TokenEndpoint": "https://<oauth2TokenEndpoint.com>/oauth/token",
"oauth2CompanyId": "<SAP SuccessFactors company ID>",
"clientId": "<client id>",
"clientSecret":"<SAML assertion>"
}

Connection parameters are set as follows:

Parameter Connection Property Description

name Technical Name [required] Enter the technical name of the connection. The technical
name can only contain alphanumeric characters and underscores (_).
Underscore (_) must not be used at the start or end of the name. The
maximum length is 40 characters. The name must be unique within the
space.

 Note
Once the object is saved, the technical name can no longer be modi-
fied.

businessName Business Name [optional] Enter a descriptive name to help users identify the object. This
name can be changed at any time.

description Description [optional] Provide more information to help users understand the object.

package Package [optional] Default value: none

authType Authentication Type [required] Enter the authentication type to use to connect to the OData
endpoint.

You can enter:

• Basic for basic authentication with user name and password


• OAuth2

 Note
HTTP basic authentication in SAP SuccessFactors will soon be re-
tired. For more information, see Deprecation of HTTP Basic Authen-
tication in SAP SuccessFactors What's New Viewer.

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 133
Parameter Connection Property Description

url URL [required] Enter the OData service provider URL of the SAP SuccessFac-
tors service that you want to access.

The syntax for the URL is:

• For V2: <SAP SuccessFactors API Server>/odata/v2


(providing a supported SAP SuccessFactors service group /
<service group> is optional)
• For V4: <SAP SuccessFactors API Server>/odatav4/
<supported SAP SuccessFactors service group>

version Version [required] Enter the OData version used to implement the SAP Success-
Factors OData service (V2 or V4).

oauth2GrantType OAuth Grant Type [required] Enter SAML Bearer as the grant type used to retrieve an
access token.

oauth2TokenEndp OAuth Token Endpoint [required] Enter the API endpoint to use to request an access token:
oint <SAP SuccessFactors API Server>/oauth/token.

OAuth Scope [optional] Enter the OAuth scope, if applicable.

oauth2CompanyId OAuth Company ID [required] Enter the SAP SuccessFactors company ID (identifying the
SAP SuccessFactors system on the SAP SuccessFactors API server) to
use to request an access token.

clientId Client ID [required] Enter the API key received when registering SAP Datasphere
as OAuth2 client application in SAP SuccessFactors.

clientSecret SAML Assertion [required] Enter a valid SAML assertion that has been generated for
authentication.

 Note
If the SAML assertion expires, the connection gets invalid until you
update the connection with a new valid SAML assertion.

Validate Connections

To validate a connection in a space, use the GET request with the/api/v1/datasphere/spaces/


<space_id>/connections/<connection_technical_name>/validation endpoint and enter:

https://<tenant_url>/api/v1/datasphere/spaces/<spaceId>/connections/
<connection_technical_name>/validation

Integrating Data and Managing Spaces in SAP Datasphere


134 PUBLIC Integrating Data via Connections
Edit Connection to SAP SuccessFactors

To edit an SAP SuccessFactors connection in a space by providing a new definition, use the PUT request
with the/api/v1/datasphere/spaces/<space_id>/connections/<connection_technical_name>
endpoint and enter:

https://<tenant_url>/api/v1/datasphere/spaces/<space-id>/connections/
<connection_technical_name>

Delete Connections

To delete a connection from a space, use the DELETE request with the/api/v1/datasphere/spaces/
<spaceId>/connections/<connection_technical_name> endpoint and enter:

https://<tenant_url>/api/v1/datasphere/spaces/<spaceId>/connections/
<connection_technical_name>

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Connections PUBLIC 135
3 Integrating Data From SAP BW Bridge

Users with a modeler role and who have access to SAP BW bridge, can import SAP BW bridge objects into SAP
Datasphere as remote tables.

SAP BW bridge enables you to use SAP BW functionality in the public cloud and to import SAP BW bridge data
models into SAP Datasphere.

The SAP BW bridge connection differs from other connection types, as it can't be created or modified. It is
being generated when SAP BW bridge is provisioned.

To learn more about SAP BW bridge, see Acquiring and Combining Data.

Integrating Data and Managing Spaces in SAP Datasphere


136 PUBLIC Integrating Data From SAP BW Bridge
4 Integrating Data from SAP Business Data
Cloud

SAP Business Data Cloud is a fully managed SaaS solution that unifies and governs all SAP data and
seamlessly connects with third-party data—giving line-of-business leaders context to make even more
impactful decisions.

SAP Datasphere in SAP Business Data Cloud

If your SAP Datasphere tenant is part of an SAP Business Data Cloud formation, then the SAP Business Data
Cloud administrator can:

• Install insight apps to the SAP Datasphere and SAP Analytics Cloud tenants in the formation.
• Activate data packages to allow the contained data products to be installed in SAP Datasphere.

For detailed information, see the SAP Business Data Cloud documentation.

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data from SAP Business Data Cloud PUBLIC 137
SAP Business Data Cloud Insight Apps

An SAP Business Data Cloud can install insight apps to the SAP Datasphere and SAP Analytics Cloud tenants
in the formation (see Installing Insight Apps in the SAP Business Data Cloud documentation).

When an insight app is installed:

• SAP-managed spaces are created in SAP Datasphere to contain the insight app content.
• Replication flows, tables, views, and analytic models are created in these spaces to ingest, prepare and
expose the required data to SAP Analytics Cloud.

SAP Datasphere users can work with insight app content in the following ways:

• Review the installed content (see Reviewing Installed Insight Apps [page 140]).
• Upload permissions records to control access to the data (see Applying Row-Level Security to Data
Delivered through Insight Apps [page 141]).
• Build on top of the delivered data products and content to extend the app (see Extending Insight Apps
[page 142])

SAP Business Data Cloud Data Products

An SAP Business Data Cloud can activate data packages to allow the contained data products to be installed in
SAP Datasphere (see Activating Data Packages in the SAP Business Data Cloud documentation).

Integrating Data and Managing Spaces in SAP Datasphere


138 PUBLIC Integrating Data from SAP Business Data Cloud
When a data package is activated, SAP Datasphere users can work with data products in the following ways:

• An SAP Datasphere administrator must choose the spaces to which data products from the originating
SAP system can be installed (see Authorize Spaces to Install SAP Business Data Cloud Data Products).
• SAP Datasphere modelers can install data products to their space for use in their modeling projects (see
Evaluating and Installing SAP Business Data Cloud Data Products.

SAP-Managed Spaces

SAP-managed spaces are automatically created to contain the data products and content installed from SAP
Business Data Cloud:

• By default, no users are granted access to these spaces, but a SAP Datasphere administrator can add
users to the spaces to monitor data integration and review the content.
• All the delivered content is SAP-managed, read-only, and cannot be modified as it is protected by a
namespace (see Namespaces).
• No user can create any objects in SAP-managed spaces.

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data from SAP Business Data Cloud PUBLIC 139
4.1 Reviewing Installed Insight Apps

When an insight app is installed, data is loaded from the source system into SAP Datasphere, combined and
prepared for analytics, and then exposed for consumption in SAP Analytics Cloud, where business users can
consume it as stories.

Installation and Data Ingestion

Data is loaded from the source system into SAP Datasphere, where the following spaces are created:

• Ingestion space - contains the data products as local tables, and the replication flows that load data to
them.
• Preparation space - contains views built on top of the data products to prepare them for consumption.
• App space - contain analytic models built on top of the views to expose the data for consumption in SAP
Analytics Cloud. This space has the name of the insight app.

 Note

These spaces are SAP-managed. You cannot create objects in them, share objects to or from them or
otherwise import or export content to or from them.

Integrating Data and Managing Spaces in SAP Datasphere


140 PUBLIC Integrating Data from SAP Business Data Cloud
Viewing Insight App Objects

The spaces created for the insight app and the objects they contain are SAP-managed, and cannot be edited.

However, users who are members of the relevant spaces can view these objects in the standard editors.

Running and Scheduling Insight App Task Chains

If the insight app contains one or more task chains, then a user with access to the space and a DW Integrator
role (or equivalent privileges) must run each task chain at least once, and should then create a schedule to run
the task chain regularly in the future. For more information, see:

• Monitoring Task Chains [page 252]


• Scheduling Data Integration Tasks [page 256]

4.2 Applying Row-Level Security to Data Delivered through


Insight Apps

All insight apps delivered through SAP Business Data Cloud are secured through data access controls. By
default, no user can view any data presented in the provide SAP Analytics Cloud stories. To make data available
to the users authorized to see it, you must upload your authorizations to the provided permissions table.

 Note

When data is delivered to SAP Datasphere through SAP Business Data Cloud, any authorizations that
are applied to the data in the source system are not imported. For general information about providing
row-level security, see Securing Data with Data Access Controls.

Procedure

1. In your preparation space, identify the fact views and the permissions table.
2. Prepare your permissions records in the operator and values format to provide row-level access to the
data in your facts (see Create an "Operator and Values" Data Access Control).

 Note

• By default, users cannot see any data in the protected views. To enable them to see any data, you
must have at least one permissions record in the permissions table.
• To show all records to a selected user, use the ALL or * operator.
• For performance reasons, avoid having more than 5,000 permissions records for a single user.

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data from SAP Business Data Cloud PUBLIC 141
3. Upload the appropriately formatted operator and values permissions data as a CSV file (see Load or
Delete Local Table Data).
You can also maintain permissions data manually using the data editor (see Maintain Local Table Data) .
4. [optional] Review the data that are accessible to different users, using the View as User dialog (see
Viewing Object Data).
5. Maintain the permissions table as necessary.

4.3 Extending Insight Apps

The data products installed via SAP Business Data Cloud as part of an insight app do not include any
extensions defined in your source system. However, you can modify the data products to add any required
custom columns, and adjust the delivered views and analytic models to consume them.

Context

If your organization has extended the SAP Business Data Cloud tables exposed via data products as part
of your insight app, you will need to copy the relevant data products and associated content from an SAP-
managed space to a new space to add these custom columns.

 Note

If SAP updates the data products and content, these updates are not copied over automatically to your
copied space and therefore, you would need to repeat this process for any updated data products.

Procedure

1. Identify all the relevant spaces which contain your data product or depend on it.
In this example, two data products are consumed by views and eventually exposed via analytic models:

Integrating Data and Managing Spaces in SAP Datasphere


142 PUBLIC Integrating Data from SAP Business Data Cloud
2. Request a user with the DW Administrator role (or equivalent privileges) to copy each of these spaces.
This will create editable versions of all objects by removing them from the protective namespace,
transforming sap.s4.entity technical names to sap_s4_entity.
For more information, see Copy a Space.
In our example, the spaces are copied but the preparation and app spaces are still consuming data from
the original ingestion space:

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data from SAP Business Data Cloud PUBLIC 143
3. A user with the DW Modeler role (or equivalent privileges) adjusts the sources of all the relevant objects, so
that they now consume the data products or views in the copied spaces.
For more information, see Replace a Source.
In our example, the copied spaces consume the data correctly from the copied ingestion space, but data
replication is still active in the original ingestion space:

4. Update the data products in the new ingestion space to include the extension columns by re-importing
them using the Import Entities wizard. This should update the local entities and replication flow with the
new columns, and optionally start the replication flow.
For more information, see Importing Entities with Semantics from SAP S/4HANA.
5. Stop replication into the original ingestion space and start replication into the copied space, if not yet done.
For more information, see Running a Flow.
In our example, replication is now active only in the copied ingestion space:

6. Ensure that the data access controls applied to the fact views are still protecting data appropriately.
For more information, see Applying Row-Level Security to Data Delivered through Insight Apps [page 141].
7. Modify the objects in the preparation and app spaces, to take into account the new extension columns
added in the ingestion space.
For more information, see Process Source Changes in the Graphical View Editor or Process Source
Changes in the SQL View Editor.

Integrating Data and Managing Spaces in SAP Datasphere


144 PUBLIC Integrating Data from SAP Business Data Cloud
5 Integrating Data via Database Users/
Open SQL Schemas

Users with a space administrator role can create database users to read data exposed by the space and to
write data to Open SQL schemas attached to space, providing a secure method for exchanging data with the
space via ODBC access to the run-time SAP HANA Cloud database.

Depending on the privileges they are granted, a database user can use their Open SQL schema to:

• Create tables and views and write data to them (see Connect to Your Open SQL Schema [page 148]).
Users assigned to a space can use these objects as sources when creating views and data flows (see Using
the Source Browser).

 Note

Views with parameters cannot be imported to SAP Datasphere through an Open SQL schema.

• Access the space schema and read data from the space.

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Database Users/Open SQL Schemas PUBLIC 145
 Note

Database users can only read data from views that are created in the space with which they are
associated if the view has the Expose for Consumption property enabled (see Exposing Data For
Consumption).

• Create a table to act as a target to receive data written from a data flow (see Allow the Space to Access the
Open SQL Schema [page 149]).

A database user is required in order to:

• Connect an analytics client or other tool to SAP Datasphere (see Consume Data in Power BI and Other
Clients, Tools, and Apps via an OData Service).
• Consume space data in an SAP SQL Data Warehousing HDI container (see Consume Space Objects in Your
HDI Container [page 156]).
• Work with SAP HANA Cloud, data lake (see Integrating Data to and From SAP HANA Cloud Data Lake [page
158]).

 Note

Users with the DW Administrator role can create database user groups, which are similar to database
users, but which are not assigned to a single space and can allow multiple spaces to read from and write to
their schemas (see Creating a Database User Group).

Integrating Data and Managing Spaces in SAP Datasphere


146 PUBLIC Integrating Data via Database Users/Open SQL Schemas
5.1 Create a Database User

Users with the DW Space Administrator role can create database users, granting them privileges to read from
and/or write to an Open SQL schema with restricted access to the space schema.

Procedure

1. In the side navigation area, click (Space Management), locate your space tile, and click Edit to open it.
2. In the Database Users section, click Create to open the Create Database User dialog.
3. Complete the properties as appropriate and click Create to create the user:

Property Description

Database User Name Suffix Enter the name of your database user, which will be appended to the space name to give
the name of the Open SQL schema.

Can contain a maximum of (40 minus the space name) uppercase letters or numbers
and must not contain spaces or special characters other than _ (underscore).

Enable Password Policy Require the database user to change their password with the frequency defined in the
password policy (see Set a Password Policy for Database Users).

Enable Automated Predic- Allow the user to access the SAP HANA Cloud machine learning libraries.
tive Library and Predictive
For information about enabling and using these libraries, see Enable the SAP HANA
Analysis Library
Cloud Script Server on Your SAP Datasphere Tenant

Enable Read Access (SQL) Allow the user to read all views that have Expose for Consumption enabled via their Open
SQL schema.

You can also enable:


• With Grant Option - Allow the user to grant read access to other database users.
• Enable HDI Consumption - Allow the creation of a user-defined service to give read
access to an HDI container added to the space (see Consume Space Objects in Your
HDI Container [page 156]).

Enable Write Access (SQL, Allow the user to create objects and write data to their Open SQL schema. These objects
DDL, DML) are available to any modeler working in the space for use as sources for their views and
data flows (see Using the Source Browser).

Enable Audit Logs for Read Enable logging of all read and change operations performed by the user.
Operations / Enable Audit
For more information, see Logging Read and Change Actions for Audit [page 313].
Logs for Change Operations

4. In your space page, click Deploy to deploy your space and create the database user in the run-time
database.

You will receive a notification when the deployment of the space is complete.
5. Click the  button for your user in the Database Users list to open the Database User Details dialog.
6. Request a password for your database user:

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Database Users/Open SQL Schemas PUBLIC 147
a. Click Request New Password.
b. Click Show to display the new password and enable the Copy Password button.
c. Click Copy Password.

If you want to work with the SAP HANA database explorer, you will need to enter your password to
grant the explorer access to your Open SQL schema. When connecting to your Open SQL schema with
other tools, you should additionally note the following properties:
• Database User Name
• Host Name
• Port
7. If you are using your database user to consume space data in an SAP SQL Data Warehousing HDI
container, click the Copy Full Credentials button to copy the json code that you can use to initialize your
user-defined service (see Consume Space Objects in Your HDI Container [page 156]).
8. Click Close to return to your space page.

You can now use your database user (see Connect to Your Open SQL Schema [page 148]).

5.2 Connect to Your Open SQL Schema

When you have created a database user, you can connect to your Open SQL schema with the SAP HANA
database explorer or with other tools.

Obtaining Your Open SQL Schema Connection Information

To obtain the host name, port, and user name required to connect to your Open SQL schema, click the
 button for your user in the Database Users list.

For security reasons, the password is not displayed in this dialog. If you have forgotten your password, you can
generate a new one by clicking Request New Password.

Connecting with SAP HANA Database Explorer

To open your Open SQL schema directly in the SAP HANA database explorer from your space page, select your
user in the Database Users list, and then click Open Database Explorer. You will need to enter your password in
order to add your schema to the list of databases.

To browse the objects in your schema, expand its node, expand Catalog, and click on an object category. To
browse the space schema, click Choose Schema and select the space schema in the list.

To import or export data, open the SQL console, or perform other actions on your schema, right-click it in the
list of databases and select the appropriate command.

For detailed information about working with the SAP HANA database explorer, see the SAP HANA Database
Explorer guide in the SAP HANA Cockpit documentation.

Integrating Data and Managing Spaces in SAP Datasphere


148 PUBLIC Integrating Data via Database Users/Open SQL Schemas
Connecting with Third-Party Tools

You can connect to your Open SQL schema with third-party ETL or analytics tools, via ODBC or JDBC. You will
need to download and install the SAP HANA client or the JDBC driver and provide your connection information.

For more information, see Connecting to SAP HANA Cloud in the SAP HANA Cloud, SAP HANA Database
documentation.

Working with Your Open SQL Schema

For information about importing data from cloud storage services, see Importing and Exporting Data in the SAP
HANA Cloud, SAP HANA Database documentation.

If the SAP HANA Cloud script server is enabled (see Enable the SAP HANA Cloud Script Server on Your SAP
Datasphere Tenant and your database user has the Enable Automated Predictive Library and Predictive Analysis
Library option enabled, you can work with the SAP HANA Automated Predictive Library (APL) and SAP HANA
Predictive Analysis Library (PAL) machine learning libraries.

For detailed information about using the machine learning libraries, see:

• SAP HANA Automated Predictive Library Developer Guide


• SAP HANA Cloud Predictive Analysis Library (PAL)

To list the privileges granted to your user, run the following statement in the SQL console:

select * from effective_privileges where user_name = current_user;

5.3 Allow the Space to Access the Open SQL Schema

To grant the space write privileges in the Open SQL schema and the ability to write data to target tables in the
schema, use the GRANT_PRIVILEGE_TO_SPACE stored procedure. Once this is done, data flows running in the
space can select tables in the Open SQL schema as targets and write data to them, and task chains can run
procedures in the schema.

Prerequisites

Only the database user for the open SQL schema has the privilege to run the stored procedure
"DWC_GLOBAL"."GRANT_PRIVILEGE_TO_SPACE".

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Database Users/Open SQL Schemas PUBLIC 149
Context

You can grant write privileges by running an SAP Datasphere specific stored procedure in the SQL console in
the SAP HANA Database Explorer.

Procedure

1. In the side navigation area, click (Space Management), locate your space tile, and click Edit to open it.

2. In the space page, go to Database Access Database Users , then select the database user and click
Open Database Explorer.
3. In the SQL console in SAP HANA Database Explorer, call the stored procedure to grant the ‘INSERT’,
‘UPDATE’, or ‘DELETE’ privilege to a space using the following syntax:

CALL "DWC_GLOBAL"."GRANT_PRIVILEGE_TO_SPACE" (
OPERATION => <operation>,
PRIVILEGE => <privilege>,
SCHEMA_NAME => <schema name>,
OBJECT_NAME => <object name>,
SPACE_ID => <space ID>);

Parameters are set as follows:

Parameter Values Description

operation • 'GRANT' [required] Enter 'GRANT' to give the write privi-


leges, or 'REVOKE' to remove the write privileges
• 'REVOKE'
to the space.

privilege • 'INSERT' [required] Enter the write privilege that you want
to grant (or revoke) to the space.
• 'UPDATE'
• 'DELETE'
 Note
• 'EXECUTE'
You can grant one privilege at a time.

schema_name '[name of the Open SQL schema]' [required] Enter the name of the Open SQL
Schema you want the space to be able to write
into.

object_name • '' [required] You can grant the write privileges, ei-
ther at the schema level or at the object level.
• null
• '[name of the objet]'
• At the schema level (all objets in the
schema): enter null or ' '.
• At the object level: enter a valid table name.

space_id '[ID of the space]' [required] Enter the ID of the space you are
granting the write privileges to.

To grant update write access to all objects (tables) in the schema:

CALL "DWC_GLOBAL"."GRANT_PRIVILEGE_TO_SPACE" (

Integrating Data and Managing Spaces in SAP Datasphere


150 PUBLIC Integrating Data via Database Users/Open SQL Schemas
OPERATION => 'GRANT',
PRIVILEGE => 'UPDATE',
SCHEMA_NAME => 'SALE#ETL',
OBJECT_NAME => '',
SPACE_ID => 'SALES');

To grant update write access to the table MY_TABLE:

CALL "DWC_GLOBAL"."GRANT_PRIVILEGE_TO_SPACE" (
OPERATION => 'GRANT',
PRIVILEGE => 'UPDATE',
SCHEMA_NAME => 'SALE#ETL',
OBJECT_NAME => 'MY_TABLE',
SPACE_ID => 'SALES');

To grant a space the ability to run procedures in the Open SQL schema:

CALL "DWC_GLOBAL"."GRANT_PRIVILEGE_TO_SPACE" (
OPERATION => 'GRANT',
PRIVILEGE => 'EXECUTE',
SCHEMA_NAME => 'SALE#ETL',
OBJECT_NAME => '',
SPACE_ID => 'SALES');

For more information on the SAP HANA statement CREATE PROCEDURE, see CREATE PROCEDURE
Statement (Procedural) in the SAP HANA SQL Reference Guide for SAP HANA Platform.

4. Run the query by clicking  (Run) or press F8.

Results

If the run is successful, you receive a confirmation message in the Result pane. You can then open the Data
Builder, create a data flow, and select the tables as targets.

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data via Database Users/Open SQL Schemas PUBLIC 151
6 Exchanging Data with SAP SQL Data
Warehousing HDI Containers

Users with a space administrator role can use SAP SQL Data Warehousing to build calculation views and other
SAP HANA Cloud HDI objects directly in the run-time SAP HANA Cloud database and then exchange data
between HDI containers and SAP Datasphere spaces. SAP SQL Data Warehousing can be used to bring existing
HDI objects into your SAP Datasphere environment, and to allow users familiar with the HDI tools to leverage
advanced SAP HANA Cloud features.

Context

To enable SAP SQL Data Warehousing on your SAP Datasphere tenant, an S-user must create an SAP ticket to
connect your SAP BTP account (see Enable SAP SQL Data Warehousing on Your SAP Datasphere Tenant).

 Note

See the blog SAP SQL Data Warehousing with HANA (published in December 2017) for an introduction
to SAP SQL Data Warehousing.

See the blog SAP BTP Showcase – Access the SAP HANA Cloud Database Underneath SAP Datasphere
(published in January 2021) for an overview of the Integrated SAP SQL Data Warehousing approach.

See the blog SAP Datasphere – Hybrid Access to SAP HANA for SQL Data Warehousing (published in
June 2021) for information about accessing your HDI objects on another server through a standard SAP
HANA connection.

The integrated SAP SQL data warehousing approach allows you to add HDI containers to your space and
exchange data between them:

• Bring data from calculation views and other objects built in your HDI container into your SAP Datasphere
space via local tables and data flows (see Add an HDI Container and Access its Objects in Your Space [page
155]).

• Use data exposed from your SAP Datasphere space as a source for calculation views and flowgraphs built
in your HDI container (see Consume Space Objects in Your HDI Container [page 156]).

Integrating Data and Managing Spaces in SAP Datasphere


152 PUBLIC Exchanging Data with SAP SQL Data Warehousing HDI Containers
• Select tables in your HDI container as targets for your SAP Datasphere data flows (see Allow Your Space to
Write to Your HDI Container [page 157]).

6.1 Prepare Your HDI Project for Exchanging Data with Your
Space

To allow your SAP Datasphere space to read from and, if appropriate, write to the HDI container, you must
configure your HDI project to build on your SAP Datasphere tenant and define the appropriate roles.

Procedure

1. Create the required roles to allow SAP Datasphere to read from and, optionally, write to the container.
You must define the roles DWC_CONSUMPTION_ROLE and DWC_CONSUMPTION_ROLE# (with grant option) in
the container to allow you to add it to your space and allow you to exchange data between the container
and the space.

The contents of the roles define the HDI objects that can be read from or written to. You can allow access to
individual objects or to all the objects in the container.

These examples show roles that provide read access to individual objects or to all objects in the container:

Read Access to Individual Objects Read Access to the Schema

This example provides read access to one table and one This example provides read access to all the objects in the
view in the container: container:

{ {
"role": { "role": {
"name": "DWC_CONSUMPTION_ROLE", "name": "DWC_CONSUMPTION_ROLE",
"object_privileges": [ "schema_privileges": [
{ {
"name":"MyNamespace::Table1", "privileges":[ "SELECT" ]
"type":"TABLE", }
"privileges":[ "SELECT" ] ]
}, }
{ }

"name":"MyNamespace::CalcView1",
"type":"VIEW",
"privileges":[ "SELECT" ]
}
]
}
}

Integrating Data and Managing Spaces in SAP Datasphere


Exchanging Data with SAP SQL Data Warehousing HDI Containers PUBLIC 153
To allow data flows in your space to write to the container you must, in addition, define the roles
DWC_WRITE_ROLE and DWC_WRITE_ROLE#.

These examples show roles that provide write access to individual objects or to all objects in the container:

Write Access to Individual Objects Write Access to the Schema

This example provides write access to two tables in the This example provides write access to all the objects in the
container: container:

{ {
"role": { "role": {
"name": "DWC_WRITE_ROLE", "name": "DWC_WRITE_ROLE",
"object_privileges": [ "schema_privileges": [
{ {
"name":"MyNamespace::Table1", "privileges":[ "SELECT",
"type":"TABLE", "INSERT", "UPDATE", "DELETE" ]
"privileges":[ "SELECT", }
"INSERT", "UPDATE", "DELETE" ] ]
}, }
{ }
"name":"MyNamespace::Table2",
"type":"TABLE",
"privileges":[ "SELECT" ]
}
]
}
}

For detailed information about hdbroles, see Roles (.hdbrole and .hdbroleconfig) in the SAP HANA Platform
documentation.

2. Configure your Cloud Foundry Settings to build to your SAP Datasphere tenant.
3. Edit your mta.yaml file to reference the database id of your SAP Datasphere tenant, your HDI container,
and the name of your user-provided service instance (see Consume Space Objects in Your HDI Container
[page 156]).
4. Build your database module.

Once the build is complete, the HDI container is available to be added to your SAP Datasphere space (see
Add an HDI Container and Access its Objects in Your Space [page 155]).

Integrating Data and Managing Spaces in SAP Datasphere


154 PUBLIC Exchanging Data with SAP SQL Data Warehousing HDI Containers
6.2 Add an HDI Container and Access its Objects in Your
Space
To access calculation views and other HDI objects as sources for your views and data flows, you must add the
HDI container to your SAP Datasphere space.

Context

You must define the roles DWC_CONSUMPTION_ROLE and DWC_CONSUMPTION_ROLE# (with grant option) in the
container to allow you to add it to your space and allow you to exchange data between them (see Prepare Your
HDI Project for Exchanging Data with Your Space [page 153]).

 Note

See the blog SAP Datasphere Integrated with SAP SQL Data Warehousing Sharing a Single DB Tenant
(published in July 2021) for a walkthrough of this process.

Procedure

1. In the side navigation area, click (Space Management), locate your space tile, and click Edit to open it.
2. In the HDI Container section, click the + button to open the Add Unassigned HDI Containers dialog.

If the + button is not available, then you need to raise a ticket (see Exchanging Data with SAP SQL Data
Warehousing HDI Containers [page 152]) or no HDI containers have yet been built on the SAP Datasphere
tenant.
3. Select one or more HDI containers and then click Add to add them to your space.

If your HDI container is not available:


• It may already be assigned to another space.

 Note

Each HDI container can only be added to one SAP Datasphere space. A database analysis user can
review the assignment of HDI assignments using the following query:

select * from DWC_TENANT_OWNER.SPACE_SCHEMAS where SCHEMA_TYPE =


'hdi_owned';

See Create a Database Analysis User to Debug Database Issues.

• It may have been created before assignment was enabled or the database_id may need to be
specified (see Set Up an HDI Container in the SAP HANA Service documentation).
4. Deploy your space.
5. Create a view or data flow and select one or more objects from the container as sources (see Import an
Object from a Connection or Other Source).

Integrating Data and Managing Spaces in SAP Datasphere


Exchanging Data with SAP SQL Data Warehousing HDI Containers PUBLIC 155
6.3 Consume Space Objects in Your HDI Container

To consume SAP Datasphere views that are exposed for consumption as sources for your calculation views and
flowgraphs, you must create a database user with appropriate privileges, create a user-provided service in SAP
BTP Cockpit, and configure your HDI project.

Context

 Note

See the blog SAP Datasphere Integrated with SAP SQL Data Warehousing Sharing a Single DB Tenant
(published in July 2021) for a walkthrough of this process.

Procedure

1. Add the HDI container to your space (see Add an HDI Container and Access its Objects in Your Space [page
155]).
2. Create a database user with the following privileges:

• Enable Read Access (SQL)


• Enable Write Access (SQL, DDL, DML)
• Enable HDI Consumption
See Create a Database User [page 147].
3. Open the Database User Details dialog and copy the credentials from the HDI Consumption section.
4. Go to the SAP BTP Cockpit, open Service Instances, and use the copied credentials to create a user-
provided service instance, which will provide the connection between SAP BTP Cockpit and your SAP
Datasphere tenant.

See Creating User-Provided Service Instances in Cloud Foundry Environment in the SAP Service Manager
documentation.
5. In your SAP Business Application Studio project, do the following:

• Prepare your project for building, including specifying the name of your user-provided service instance
in your mta.yaml file (see Prepare Your HDI Project for Exchanging Data with Your Space [page 153]).
• Create an hdbgrants file to grant read access to the views that you want to consume from your SAP
Datasphere space.
You can, optionally, grant access to the Application Function Library by including the following code in
hdbgrants:

"global_roles" : [
{
"roles" :
[ "AFL__SYS_AFL_AFLPAL_EXECUTE_WITH_GRANT_OPTION"]
}

Integrating Data and Managing Spaces in SAP Datasphere


156 PUBLIC Exchanging Data with SAP SQL Data Warehousing HDI Containers
]

 Note

The SAP HANA Cloud Script Server must be enabled to allow this grant (see Enable the SAP HANA
Cloud Script Server on Your SAP Datasphere Tenant).

• Create hdbsynonym and hdbsynonymconfig files to define synonyms for the consumable views.
6. Build your project.

6.4 Allow Your Space to Write to Your HDI Container

To allow data flows in your SAP Datasphere space to use tables in your HDI container as targets, you must set
the appropriate roles and add the container to your space.

Procedure

1. Define the roles DWC_WRITE_ROLE and DWC_WRITE_ROLE# (with grant option) in your HDI container (see
Prepare Your HDI Project for Exchanging Data with Your Space [page 153]).
2. Add the container to your space (see Add an HDI Container and Access its Objects in Your Space [page
155]).
3. Create a data flow and select a table in the container as the target for your data flow (see Add or Create a
Target Table in a Data Flow).

Integrating Data and Managing Spaces in SAP Datasphere


Exchanging Data with SAP SQL Data Warehousing HDI Containers PUBLIC 157
7 Integrating Data to and From SAP HANA
Cloud Data Lake

Users with an administrator role can connect an SAP Datasphere space with SAP HANA Cloud, data lake to
store and access large amounts of data.

What is SAP HANA Cloud, data lake?

Data lake is a fully managed cloud service for securely storing and analyzing large amounts of data that
is infrequently updated. It leverages inexpensive storage options to lower costs, while maintaining excellent
performance and full SQL access to data.

Data lake includes elastically scalable compute to provide high-performance analysis on-demand, and enabling
cost control when it isn’t in use. It’s fully integrated into SAP HANA Cloud, sharing common security, tenancy,
and tooling.

For more information on data lake, please visit the SAP HANA Cloud, Data Lake product page.

Working with data lake in SAP Datasphere

To benefit from data lake, you need to assign a SAP Datasphere space that gets access to the data lake.
Currently, only one space can be assigned to access and use the data lake.

Tables in the data lake are accessed via virtual tables in the open SQL schema. These tables can be consumed
in the Data Builder. SAP Datasphere provides two stored procedures to ease this task. Take a look at Data Lake
API [page 161] for more details and examples on how to use the procedures.

For a detailed step-by-step guide to connect and work with your data lake see Working with Data Lake [page
159].

Integrating Data and Managing Spaces in SAP Datasphere


158 PUBLIC Integrating Data to and From SAP HANA Cloud Data Lake
7.1 Working with Data Lake

Assign a SAP Datasphere space to access and work with SAP HANA Cloud, data lake.

Prerequisites

Before you can work with data lake, you need to enable it by selecting a storage size in System
Configuration Tenant Configuration (see Configure the Size of Your SAP Datasphere Tenant.

Context

You can assign a space that connects to data lake. Tables in the data lake can then be accessed via virtual
tables in the open SQL schema. These tables can also be consumed in the Data Builder.

 Note

• Currently, DDL statements in your data lake are not audited.


• For data flows, only APPEND (UPSERT option not selected) and TRUNCATE modes are supported.
APPEND (with UPSERT option selected) and DELETE modes are not supported.

Procedure

1. Go to Space Management and select the space that you want to give access to data lake. Alternatively, you
could create a dedicated space for the data lake.
2. Under Storage Assignment select the check box Use this space to access the data lake. If another space
already has access to the data lake, you won't be able to assign your space.
3. Click Save.

You can now use your preferred SQL tool to create tables in data lake and access these tables via SAP
HANA virtual tables in your open SQL schema. SAP Datasphere offers two stored procedures that you can
use to easily create and access the tables. For more information and examples on the stored procedures
see Data Lake API [page 161].
4. In your SQL tool, use the "DWC_GLOBAL"."DATA_LAKE_EXECUTE" ( IN STMT NCLOB ) stored
procedure in the schema DWC_GLOBAL to create tables in data lake. Please note, that the statements
issued via this procedure are not audited.
5. Then create SAP HANA virtual tables in your open SQL schema that refer to the tables in data lake. The
virtual tables are used to query the tables in data lake.

Use the following procedure to create a virtual table in your open SQL schema:

"DWC_GLOBAL"."DATA_LAKE_CREATE_VIRTUAL_TABLE"

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data to and From SAP HANA Cloud Data Lake PUBLIC 159
(
IN VIRTUAL_TABLE_NAME NVARCHAR(256),
IN DATA_LAKE_TABLE_NAME NVARCHAR(256),
IN TARGET_SCHEMA_SYS BOOLEAN DEFAULT false
).

Results

You can then work with your virtual tables in the Data Builder by selecting the open SQL schema used for the
data lake.

Example

Creating a table in data lake:

CALL "DWC_GLOBAL"."DATA_LAKE_EXECUTE"('CREATE TABLE TABLE1 ( MY_ID INTEGER ) ');

Creating a virtual table in open SQL schema:

CALL "DWC_GLOBAL"."DATA_LAKE_CREATE_VIRTUAL_TABLE"
( VIRTUAL_TABLE_NAME => 'TABLE1_VT',
DATA_LAKE_TABLE_NAME => 'TABLE1'
) ;

Inserting a record in the table:

INSERT INTO table1_vt VALUES (2);

Selecting data from the virtual table:

SELECT * FROM table1_vt;

Dropping the virtual table:

DROP TABLE table1_vt;

Dropping the table in data lake:

CALL "DWC_GLOBAL"."DATA_LAKE_EXECUTE"('DROP TABLE TABLE1');

Accessing the system views via the virtual table:

CALL "DWC_GLOBAL"."DATA_LAKE_CREATE_VIRTUAL_TABLE"
( VIRTUAL_TABLE_NAME => 'SYSTAB',
DATA_LAKE_TABLE_NAME => 'SYSTAB',
TARGET_SCHEMA_SYS => true
) ;

For more information see System Views in Data Lake Relational Engine (SAP HANA DB-Managed).

Integrating Data and Managing Spaces in SAP Datasphere


160 PUBLIC Integrating Data to and From SAP HANA Cloud Data Lake
7.2 Data Lake API
SAP Datasphere provides two procedures to access SAP HANA Cloud, data lake.

The following procedures are available in the schema DWC_GLOBAL:

• "DWC_GLOBAL"."DATA_LAKE_EXECUTE" ( IN STMT NCLOB )

This procedure is used to execute statements in data lake. It's a wrapper procedure around the data lake
procedure REMOTE_EXECUTE. This enables you to create or drop tables in data lake, or to load data from
files.

• "DWC_GLOBAL"."DATA_LAKE_CREATE_VIRTUAL_TABLE"
(
IN VIRTUAL_TABLE_NAME NVARCHAR(256),
IN DATA_LAKE_TABLE_NAME NVARCHAR(256),
IN TARGET_SCHEMA_SYS BOOLEAN DEFAULT false
).

• This procedure provides the option to create SAP HANA virtual tables in the open SQL schema that refer to
objects in data lake. These virtual tables are used to query data in the data lake. For more information see
What's the Difference Between Data Lake Relational Engine and SAP HANA Database SQL Statements?.
The procedure can also be used to access system views of data lake in schema SYS. The optional IN
parameter IN_TARGET_SCHEMA_SYS needs to be to true to create a virtual table on a data lake system
view. For more information see System Views in Data Lake Relational Engine (SAP HANA DB-Managed).
Virtual tables can be dropped directly using DROP TABLE <virtual_table_name> syntax.
For more information see Query Data in Data Lake Relational Engine (SAP HANA DB-Managed).

Example

Creating a table in data lake:

CALL "DWC_GLOBAL"."DATA_LAKE_EXECUTE"('CREATE TABLE TABLE1 ( MY_ID INTEGER ) ');

Creating a virtual table in open SQL schema:

CALL "DWC_GLOBAL"."DATA_LAKE_CREATE_VIRTUAL_TABLE"
( VIRTUAL_TABLE_NAME => 'TABLE1_VT',
DATA_LAKE_TABLE_NAME => 'TABLE1'
) ;

Inserting a record in the table:

INSERT INTO table1_vt VALUES (2);

Selecting data from the virtual table:

SELECT * FROM table1_vt;

Dropping the virtual table:

DROP TABLE table1_vt;

Integrating Data and Managing Spaces in SAP Datasphere


Integrating Data to and From SAP HANA Cloud Data Lake PUBLIC 161
Dropping the table in data lake:

CALL "DWC_GLOBAL"."DATA_LAKE_EXECUTE"('DROP TABLE TABLE1');

Accessing the system views via the virtual table:

CALL "DWC_GLOBAL"."DATA_LAKE_CREATE_VIRTUAL_TABLE"
( VIRTUAL_TABLE_NAME => 'SYSTAB',
DATA_LAKE_TABLE_NAME => 'SYSTAB',
TARGET_SCHEMA_SYS => true
) ;

Integrating Data and Managing Spaces in SAP Datasphere


162 PUBLIC Integrating Data to and From SAP HANA Cloud Data Lake
8 Managing and Monitoring Data
Integration

Users with a space administrator or integrator role can use the Data Integration Monitor app to schedule, run,
and monitor data replication and persistence tasks for remote tables and views, track queries sent to remote
source systems, and manage other tasks through flows and task chains.

 Note

For optimal performance, it is recommended that you consider staggering the scheduled run time of tasks
such as data flows and task chains that may contain these tasks. There is a limit on how many tasks can be
started at the same time. If you come close to this limit, scheduled task runs may be delayed and, if you go
beyond the limit, some scheduled task runs might even be skipped.

The Data Integration Monitor is composed of several monitors that help you track how data is managed within
your space. Note that the available monitors depend on space capabilities.

• Remote Tables (Not available if you are in space of type SAP HANA Cloud, data lake files): Copy data for
remote tables that have been deployed in your space and monitor the replication of the data. Copy or
schedule copying the full set of data from the source, or set up replication of data changes in real-time via
change data capturing (CDC). For more information, see Replicating Data and Monitoring Remote Tables
[page 168].

 Note

Remote tables from a source that is configured as data access Remote Only are not displayed in this
monitor. For more information, see Managing and Monitoring Data Integration [page 163]

• Local Tables (Not available if you are in space of type SAP HANA Cloud, data lake files): Monitor all local
tables which have been created in a space and display their metrics. From the detail screen, you can also
delete table records on-demand, using filter conditions or using a schedule.
• Local Tables (Files) (Not available if you are in space of type SAP HANA Cloud, SAP HANA database):
Check how and when your local tables (files) were last updated and if new data has still to be merged. For
more information, see Monitoring Local Tables (File) [page 195]
• Views (Not available if you are in space of type SAP HANA Cloud data lake files): Add, monitor and
schedule persisted views. Turn your views into persisted views to improve performance while accessing
your data. For more information, see Persisting and Monitoring Views [page 196].
• Flows: View and monitor the execution details of the executed data flows, replication flows or
transformation flows. For more information, see Monitoring Flows [page 239].

 Note

Data flows are not available for spaces of type SAP HANA Cloud data lake files

• Remote Queries:
• Remote Queries: Check and analyze how the communication is running between the federation layer
of SAP HANA Cloud and the connected remote source systems. For more information, see Monitoring
Remote Queries [page 249].

Integrating Data and Managing Spaces in SAP Datasphere


Managing and Monitoring Data Integration PUBLIC 163
• Remote Table Statistics (Not available if you are in space of type SAP HANA Cloud, data lake files):
Create statistics for your remote tables to improve federated query execution. For more information,
see Creating Statistics for Your Remote Tables [page 251].
• Task Chains: Monitor the status and progress of running and past task chains. For more information, see
Monitoring Task Chains [page 252].

Depending on your role and the tasks you need to perform in the Data Integration Monitor, you may need
specific privileges. See Privileges and Permissions.

Here are some examples of privileges and permissions you need to perform some tasks:

• To create or update connections, you need the Data Warehouse Connection privilege.
• To perform one-time actions in one of the monitors, such as replicating full set of data, redeploying
persisted views, loading new snapshot, or removing persisted data, you need the privilege Data Warehouse
Data Integration Update.
• To schedule tasks in one of the monitors and perform task execution and log retrieval in general, you need
the privilege Data Warehouse Data Integration Execute.
• To add a new view in the Views monitor, or to set up or change partitioned data loading in the Remote
Tables monitor, you need the privilege Data Warehouse Data Builder Read.

8.1 Navigating in the Monitor Tabs

Selecting a specific monitor tab from Data Integration Monitor ( ) displays a tabular list of the scheduled
tasks associated with the selected monitor type. It shows the details of runs executed in last 60 days. You can
personalize the columns you want to display clicking on  Select Columns. From the tabular display, you can
click > at the end of a particular task row to display details of that monitored item’s execution history including
the completion date and status of each run. After selecting a specific monitor type, for Remote Tables, Views,
Flows, or Task Chains, each of the monitor displays provides a summary listing of currently scheduled and
previously executed tasks. For example, here is a sample of the monitor display for the Remote Tables monitor

The display provides a summary of all currently scheduled and previously executed remote table replication
tasks, their schedule and refresh frequency, various memory and disk usage statistics, as well as the last task
update or execution. Similar displays are provided for each of the other monitor types.

To view more information on any specific task in the summary list, click > in the right-most column of the row
to access the Run Details page for the task you selected. Depending on the monitor type you selected, the Run
Details page provides a flexible two or three panel display in which you can resize ( and  ), expand/collapse
( ), or close (  ) the display of each panel to focus on specific detail information or log messages. In addition,
you can sort ( ) and filter ( ) information displayed in the different panels, as well as search log messages.

The following is an example of the Run Details display for a remote table replication task.

Integrating Data and Managing Spaces in SAP Datasphere


164 PUBLIC Managing and Monitoring Data Integration
In the left-most column or panel, the monitor displays a list of all previous runs of the selected task as well as
the status of the task, completed successfully or failed. In the right-most column or panel, the monitor displays
log messages detailing the steps executed for the selected task. For failed task runs or execution, you can click
on the highlighted Failed status to show more information.

By default, the failed tasks appear on the top. For more information on statuses and substatuses, see
Understanding Statuses and Substatuses [page 165].

 Note

From each monitor, you can navigate back to the Data Builder to change your objects by clicking Open in
Data Builder.

You also get information about who runs the task (Run By) and how it has run (Run Type): Direct (the task
started directly while clicking the Run button) or Scheduled (a schedule is created)

8.2 Understanding Statuses and Substatuses

When you run an activity (replicate a remote table, persist a view, etc..), the progress of the task is monitored
using statuses and substatuses. Statuses and substatuses are available in the relevant editors as well as in the
System Monitor.

In SAP Datasphere , the activities can take a minimum of 3 main statuses:

• Running or Loading: The task is currently running.


• Completed or Available: The task has completed successfully.
• Failed or Error: Something goes wrong during the task run and it could not be completed. Go to the details
screen of your run and check the logs to identify to issue.

Depending on your activity, there might exist additional statuses. For more information, see Replicating Data
and Monitoring Remote Tables [page 168], Persisting and Monitoring Views [page 196], Monitoring Flows
[page 239], Monitoring Remote Queries [page 249] and Monitoring Task Chains [page 252].

Integrating Data and Managing Spaces in SAP Datasphere


Managing and Monitoring Data Integration PUBLIC 165
Statuses are sometimes combined with substatuses:

Substatus Description

CANCELLED This task run was canceled, after it had started. In this case,
the data was rolled back and restored to the state that
existed before the task run was initially triggered.

FAIL_CONSENT_EXPIRED The authorization that allows SAP to run task chains


or schedule data integration tasks on your behalf has
expired. Select the option provided to renew your consent
(Changing SAP Datasphere Settings).

FAIL_CONSENT_INVALIDATED This task or task chain could not be executed, typically


due to a change in the Identity Provider configuration of
the tenant. In that case, no new task jobs or task chains
can be run or scheduled in the name of the affected user.
If the assigned user still exists in the new IdP, revoke
the scheduling consent and then grant it again. If the
assigned user no longer exists, assign a new task job or
task chain owner and provide the required task scheduling
consent. See the following SAP note 3089828 for more
information.

FAIL_CONSENT_NOT_AVAILABLE You have not authorized SAP to run task chains or schedule
data integration tasks on your behalf. Select the option
provided to give your consent (Changing SAP Datasphere
Settings).

FAIL_NOT_TRIGGERED This task job could not be executed due to a system outage
or some part of the database system not being available
at the time of the planned execution. Wait for the next
scheduled job execution time or reschedule the job.

FAIL_OWNER_MISSING This task job could not be executed because it does not
have an assigned system user. Assign an owner user to the
job.

FORBIDDEN The assigned user does not have the privileges necessary to
execute this task.

LOCKED The same task was already running, so this task job cannot
be run in parallel with an existing task’s execution.

PARTIALLY_COMPLETED (PARTIALLY_EXECUTED) This task failed but, for some sub-objects, the execution
ended successfully.

SCHEDULE_CANCELLED This task job could not be executed due to an internal error.
Contact SAP Support and provide them with the correlation
id and timestamp from this task job’s log detail information.

SET_TO_FAILED This task was running but the user set this task’s status to
FAILED.

STOPPED This task was stopped, but no rollback was performed.

SUCCESS_SKIPPED This task’s execution has not been triggered because a


previous run of the same task is still in progress.

TASK_EXCEPTION This task encountered an unspecified error during


execution.

Integrating Data and Managing Spaces in SAP Datasphere


166 PUBLIC Managing and Monitoring Data Integration
Substatus Description

TASK_EXECUTOR_ERROR This task encountered an internal error, likely during


preparation steps for execution, and the task could not be
started.

UNAUTHORIZED The user could not be authenticated, has been locked, or


deleted.

RESOURCE_LIMIT_ERROR The task can't be performed because sufficient resources


are not available. Navigate to the details screen of the
relevant log and check the message detail for more
information.

You can also monitor the statuses and substatuses in the  (System Monitor). For more information, see
Monitoring SAP Datasphere

8.3 Remote Tables in Data Access Remote Only Vs Data


Access Remote and Replication

An administrator or a data integrator has configured the source connection as Remote Only to prevent remote
tables from being replicated in SAP Datasphere.

 Note

If your connections have been set up in a previous version, they have been updated with data access .

The possibility to replicate the entire source object using remote table replication (snapshot or real-time) is
useful for some use cases such as performance optimization.

However, remote table replication offers limited support for use cases that require more flexibility when dealing
with larger amounts of data. Many of these use cases require ETL (Extract, Transform, Load) capabilities that
can be easily achieved with Replication Flow or Data Flow capabilities.

There are also cases where you don't want the data to be persisted on the Cloud, for example for data
protection or security concerns.

To support use cases where data does not need to be replicated using remote tables, the option data access:
Remote Only can be set to a connection when supported (currently, for SAP HANA Cloud and SAP HANA
on-premise). For more information, see SAP HANA Connections [page 92]. You can use the data access:
Remote Only for connections that both support remote tables and replication flows. This way, you will replicate
data only via replication flows (SAP recommendation) and not via remote tables, and thus avoid replicating the
same data twice.

The table below compares the actions you can do when your remote table is in data access Remote Only
compared to a remote table in data access Remote and Replication or with no particular option selected:

Integrating Data and Managing Spaces in SAP Datasphere


Managing and Monitoring Data Integration PUBLIC 167
Data Access Remote and Replication (or
Actions Data Access Remote Only no option)

Replication of Data Not possible directly. However, Possible. You can do it via Data
selected data can be replicated via
Replication Start Data Replication
Flows monitor.
or Data Replication Enable
For more information, see Creating a
Real-Time Data Replication when
Data Flow and Creating a Replication
supported. Data can be replicated
Flow
either from the table editor (in the
.
Data Builder) or from the Remote
Tables monitor (in the Data Integration
Monitor)

For more information, see Replicating


Data and Monitoring Remote Tables
[page 168]

Scheduling of data load Not possible Possible. For more information, see
Scheduling Data Integration Tasks
[page 256]

Partition of data load Not possible Possible. See Partitioning Remote


Table Data Loads [page 184]

Import/Edit/Delete/Refresh/Display Possible. For more information, see Possible. or more information, see
Importing Tables and Views from Importing Tables and Views from
Sources. Sources.

Monitoring of remote tables No specific monitor is available. Possible. You can monitor remote
However, you can monitor the tables via the Remote Tables monitor.
exchange of data between SAP For more information, see Monitoring
Datasphere and the remote source Remote Queries [page 249].
system in the Remote Queries Monitor.
For more information, see Monitoring
Remote Queries [page 249]

Create/Edit/Delete Statistics Yes. For more information, see Creating Not possible
Statistics for Your Remote Tables [page
251]

8.4 Replicating Data and Monitoring Remote Tables


In the Remote Tables monitor, you can find a remote table monitor per space. Here, you can copy data from
remote tables that have been deployed in your space into SAP Datasphere, and you can monitor the replication
of the data. You can copy or schedule copying the full set of data from the source, or you can set up replication
of data changes in real-time via change data capturing (CDC).

The monitor shows the remote tables that have already been deployed in your space. Also, it only shows
remote tables for connections that have been created within your space.

SAP Datasphere is using two types of adaptors to connect to remote tables:

Integrating Data and Managing Spaces in SAP Datasphere


168 PUBLIC Managing and Monitoring Data Integration
• SAP HANA smart data integration (used in connections with Data Provisioning option = Data Provisioning
Agent).
• SAP HANA smart data access adaptors (used in connections with no Data Provisioning option or Data
Provisioning option = Cloud Connector or Direct).

 Note

If your source data comes from an SAP HANA On-Premise system, select the adaptor following your use
case:

• You want to access the data remotely: SAP HANA smart data access (Data Provisioning Option: Direct)
would be the recommended adaptor to read the data. It allows higher degree of query pushdown to the
remote database, leading to better response times and less resource consumption.
• You want to replicate the data into SAP Datasphere: The preferred option for this is to use Replication
Flows, see Creating a Replication Flow. In case you require replication for remote tables, SAP HANA
smart data integration (Data Provisioning Option: Data Provisioning Agent) is the recommended
adaptor to push the data. It offers more options when loading the data, such as applying filter
conditions or data partitioning.

For more information on these adaptors, see Connecting SAP HANA Cloud, SAP HANA Database to
Remote Data Sources.

Initially, by default, when a remote table is deployed, the data is directly accessed live from the source and no
data is stored in SAP Datasphere.

For connection types that support replication, you can set up data replication to improve performance when
accessing the data.

The Data Access property of a remote table shows you whether data is accessed directly from the source or
whether data is replicated.

When replicating data, you can select:

• Start Data Replication


It copies the full set of data from your source object (usually a database table or view) into SAP
Datasphere.
• Enable Real-Time Data Replication
Real-time replication is only availabe for connection types and source objects that support copying data
changes in real-time into SAP Datasphere via change data capturing (CDC). Examples are database
connections such as SAP HANA, Microsoft SQL Server, or Oracle, supporting real-time replication for
tables after the data has been updated in the source. With schedules, you can copy the full set of data from
your source recurrently and asynchronously in the background.

 Note

Parallel execution is not supported. You can't perform two similar tasks at the same time and therefore
you need to wait until the currently running task is completed before launching a new identical task. For
example, if you want to start a data replication, but another task is already replicating data from the same
table, then you will get a notification and your task will fail.

The Frequency property of a remote table shows you whether or not data for a remote table is replicated and if
yes, at what frequency the replication takes place. Replicated data is stored in a replica table.

Integrating Data and Managing Spaces in SAP Datasphere


Managing and Monitoring Data Integration PUBLIC 169
For more information on which connection types support data replication (direct or scheduled replication) and
real-time replication, see Integrating Data via Connections [page 8].

As long as the prerequisites are met, you can switch between directly accessing the data from the source
(remote), or accessing to replicated data (data replication in real-time or in direct or schedule mode),
according to your needs.

The interactive graphic below summarizes how data is replicated:

• #unique_15/unique_15_Connect_42_subsection-im3 [page 171]


• #unique_15/unique_15_Connect_42_subsection-im2 [page 171]
• #unique_15/unique_15_Connect_42_subsection-im4 [page 171]

Integrating Data and Managing Spaces in SAP Datasphere


170 PUBLIC Managing and Monitoring Data Integration
• #unique_15/unique_15_Connect_42_subsection-im6 [page 171]
• #unique_15/unique_15_Connect_42_subsection-im5 [page 171]
• #unique_15/unique_15_Connect_42_subsection-im7 [page 172]
• #unique_15/unique_15_Connect_42_subsection-im8 [page 172]
• #unique_15/unique_15_Connect_42_subsection-im9 [page 172]
• #unique_15/unique_15_Connect_42_subsection-im10 [page 172]
• #unique_15/unique_15_Connect_42_subsection-im11 [page 172]
• #unique_15/unique_15_Connect_42_subsection-im12 [page 172]

Remote Table (Replicated):

Data is copied into SAP Datasphere and is read from the replica table:

• For data replication, replicated means that data is read from the replica table but not expected to be
updated in real-time. To get your data updated, you must start a direct data replication or create a schedule
to plan regular replication.
• For real-time replication, replicated means that data is read from the replica table and expected to be
updated in real-time.

Remote table (Remote)

Data is not accessed locally in SAP Datasphere. It’s read from a virtual table, which accesses the data directly
from the source via the adaptors. Remote tables are created from the Data Builder.

For more information, see Import Remote Tables.

Replica table

Remote data is read from these replica tables once the data access is replicated. For remote tables connected
via SAP HANA smart data access, only one remote table per source entity of a connection can be in status
"Replicated" (direct or scheduled replication, or real-time replication).

Virtual Table

Virtual tables enable you to access objects in other databases, without having to replicate in SAP HANA.

For more information, see Create Virtual tables

Replication

Start a data replication to improve performance when accessing the data. For connection types that support
replication, you can run a data replication to improve performance when accessing the data. You can replicate
data via a direct data replication, real-time replication, or you can schedule your replication tasks. You can also
partitions your data loads.

For more information, see Integrating Data via Connections [page 8] and Scheduling Data Integration Tasks
[page 256].

From the Data Builder, you can also restrict data loads to reduce the volume of data loaded in your remote
table.

For more information, see Restrict Remote Table Data Loads.

Integrating Data and Managing Spaces in SAP Datasphere


Managing and Monitoring Data Integration PUBLIC 171
SAP HANA smart data integration/SAP HANA smart data access

SAP HANA smart data access and SAP HANA smart data integration allow you to access remote data through
virtual tables without copying the data into SAP HANA.

For more information, see Connecting SAP HANA Cloud, SAP HANA database to Remote Data Sources.

DP Agent

The Data Provisioning Agent is a lightweight component running outside the SAP Datasphere environment. It
hosts data provisioning adapters for connectivity to remote sources, enabling data federation and replication
scenarios. The Data Provisioning Agent acts as a gateway to SAP Datasphere, providing secure connectivity
between the database of your SAP Datasphere tenant and the adapter-based remote sources.

For more information, see Preparing Data Provisioning Agent Connectivity

Odata

SAP OData is a standard Web protocol used for querying and updating data present in SAP using ABAP,
applying and building on Web technologies such as HTTP to provide access to information from a variety of
external applications, platforms, and devices. To be able to successfully validate and use a connection to an
OData service for remote tables, you need to configure some properties.

For more information, see Generic OData Connections [page 42].

Cloud Connector

The SAP Cloud Connector provides a secure tunnel between SAP Datasphere and on-premise systems. You
need to set up and configure Cloud Connector before creating a connection to an on-premise source that you
want to use for remote tables via SAP HANA smart data integration.

For more information, see Configure Cloud Connector.

Direct Connection

Some connections, don’t need extra connector (for example Amazon Athena, Google Big Query, etc.).

Remote Sources

You can connect several data sources to SAP Datasphere that support remote table functionality.

For more information, see Integrating Data via Connections [page 8]

Monitoring Remote Tables

In the monitor, for all the remote tables that have been previously deployed in your space, you can see how
the data is accessed, remotely from the source or replicated. For replicated data, the frequency shows how you
are replicating your data: in real-time, using a schedule (scheduled) or on-demand (none) to capture your data
changes. You monitor the size and replication status of the tables and you can see when the latest replication
for a remote table took place.

Integrating Data and Managing Spaces in SAP Datasphere


172 PUBLIC Managing and Monitoring Data Integration
Column Information

Connection (Business Name) or Connection (Technical Name Displays the name of the connection the remote table be-
longs to. Technical or Business Name is displayed, depend-

ing on how you have configured your UI settings in Your

Profile Settings

Technical Name (or Business Name) Displays the name of the remote table. Technical or Business
Name is displayed, depending on how you have configured

your UI settings in Your Profile Settings

Data Access Displays how data is currently accessed.

• Remote: Data is accessed directly from the source (fed-


eration) and read from the virtual table.
• Replicated: Data is copied in SAP Datasphere and is
read from the replica table.
• Replicated (Snapshot): The data is read from the
replica table but is not expected to be updated in
real-time. You can create a schedule to refresh the
data regularly.
• Replicated (Real-Time): The data is read from the
replica table and expected to be updated in real-
time.

Frequency Displays how often data is replicated. The value here shows
the intended frequency that might have not been reached
yet depending on the status of the remote table.

• –-: Refresh frequency doesn't apply.


• None: There is no schedule defined for this table.
• Scheduled: A schedule is defined for this table.
• Real-Time: Real-time replication has been set up for the
remote table.
• Paused: Data refresh is paused.

Integrating Data and Managing Spaces in SAP Datasphere


Managing and Monitoring Data Integration PUBLIC 173
Column Information

Status Displays the current status of the remote table.

• ---: Initial state


• Loading: Started to load a new snapshot
• Available: Snapshot has been loaded and is available in
the replica table
• Initializing: Started to enable real-time data replication
• Active: Data is replicated and updated in real-time
• Error: Failed to load a new snapshot, or failed to enable
real-time data replication, or data is replicated in real-
time but remote exceptions occurred
• Paused: Real-time replication is paused
• Disconnected: SAP HANA smart data integration Data
Provisioning Agent got disconnected
For information about how to connect the agent, see
Connect and Configure the Data Provisioning Agent.

Latest Update Displays when the data was last successfully updated in SAP
Datasphere.

Latest Change (Source) (For remote tables connected with SAP HANA smart data
integration in real-time replication only). Displays the date
and time when the data has been last updated in the source
system. It can take up to 5 minutes to show the information
in SAP Datasphere after the source has been updated.

 Note
If the Latest Change (Source) column shows a signifi-
cantly lower value than the Latest Update column, it is
very likely that there is an overload situation in transfer-
ring changes from the source to SAP Datasphere. In
that case, refer to the documentationBest Practices and
Sizing Guide for Smart Data Integration when used in
SAP Data Warehouse Cloud

Number of Records Displays the number of records replicated into SAP


Datasphere. You can then compare this number with the
number of records stored in the source system, and identify
errors in your data replication.

New Run Displays the next scheduled run (if a schedule is set for the
remote table).

Memory Used for Storage (MiB) Displays the size occupied by the remote table data in mem-
ory.

Integrating Data and Managing Spaces in SAP Datasphere


174 PUBLIC Managing and Monitoring Data Integration
Column Information

Disk Used for Storage (MiB) Displays the size occupied by the remote table data on disk.

Schedule Owner Displays the name of the current schedule owners.

You can personalize the columns you want to display clicking on  Select Columns

Actions in the Remote Tables

Depending on the combination of values in the Data Access, Frequency and Status columns, from the Remote
Tables monitor, you can perform the following actions on remote tables:

Action Description

Start Data Replication Directly start a copy of the full set of data from the source in
the background.

 Note
If you start data replication on a remote table whose
data access is Replicated (Real-time), you’ll remove the
real-time replication for this table. Your data will no lon-
ger be updated in real-time and you’ll have to start a new
data replication or create a schedule to get your data
updated.

For more information, see Replicate Full Set of Data [page


179].

Remove Replicated Data Stop replication and delete data from replica table.

 Note
If you need to remove the replicated data for a remote
table whose data access is Replicated (Real-time), you
must ensure that the following requirements are met to
allow a proper deletion:

• The data provisioning agent is not disconnected.


• The real-time replication is not paused at connec-
tion level and is working properly.

If one of the requirement is not met, some objects won't


be deleted and you'll have to delete them manually. See
3307334 for more information.

Integrating Data and Managing Spaces in SAP Datasphere


Managing and Monitoring Data Integration PUBLIC 175
Action Description

Enable Real-Time Data Replication Start replication of data changes in the source in real-time.

 Note
When enabling the replication in real-time, the data is
updated in real-time. You don’t need to start a new data
replication or to create a schedule to update your data

For more information, see Replicate Data Changes in Real-


Time [page 181].

Disable Real-Time Data Replication Disable real-time data replication even if your source system
is unreachable and switch the data access from Replicated
(Real-time) to Replicated (snapshot). It preserves the replica
table and its replicated data to allow further actions on the
remote table such as deleting or changing the properties
of the connection. For more information, see Disabling Real-
Time Data Replication [page 184].

 Note
The remote table must be connected via SAP HANA
smart data integration, and must not be connected via
SAP HANA smart data access.

Create Schedule Create a schedule to start a data replication in the back-


ground according to the settings defined in the schedule.

 Note
If you create a schedule for a remote table whose data
access is Replicated (Real-time), the replication type will
change from real-time replication to batch replication at
the next run of the schedule. The data will no longer be
updated in real-time

For more information, see Scheduling Data Integration Tasks


[page 256].

Edit Schedule Change how the schedule is specified, or change the owner
of the schedule.

For more information, see Take Over the Ownership of a


Schedule [page 265].

Delete Schedule Delete the schedule if required.

Integrating Data and Managing Spaces in SAP Datasphere


176 PUBLIC Managing and Monitoring Data Integration
Action Description

Assign Schedule To Me Take the ownership of the schedule.

For more information, see .Take Over the Ownership of a


Schedule [page 265].

Pause Schedule Pause the scheduled task.

For more information, see Pause or Resume a Scheduled


Task [page 266].

Resume Schedule Resume the scheduled task that was previously paused.

 Note

You can select several remote tables to group actions whenever it's applicable:

• Schedule data replication: You can delete, pause or resume schedules for several remote tables that
have schedules defined. You can also assign several schedules to your name and become the owner of
the schedules.
• Remove Replicated Data: You can select several remote tables that have replicated data available and
remove the replicated data for all of them.
• Many actions like Start Data Replication, Remove Replicated Data, Create Schedule, etc. cannot be
grouped. In such cases, they are disabled.

After triggering a group of actions, the status of each action is displayed in the notification area. As these
operations run in an asynchronous mode, the status will not be updated automatically, and you need to
manually Refresh the page to get the updated status.

Other Actions on Remote Tables (connected via SAP HANA Smart Data
Integration/SAP HANA Smart Data Access)

Other actions are available on remote tables, but behavior depends on how your remote table is connected (via
SAP HANA smart data access or SAP HANA smart data integration):

Remote Tables via SAP HANA Remote Tables via SAP HANA
Actions Smart Data Access Smart Data Integration Comments

Filtering data loads Supported but with Supported but with Action to be done from the
restrictions depending on restrictions depending on  (Data Builder). For more
data type and adapter used. data type and adapter used. information, see Restrict
Remote Table Data Loads

Exclude column when Supported Supported Action to be done from the


loading data  (Data Builder). For more
information, see Restrict
Remote Table Data Loads

Integrating Data and Managing Spaces in SAP Datasphere


Managing and Monitoring Data Integration PUBLIC 177
Remote Tables via SAP HANA Remote Tables via SAP HANA
Actions Smart Data Access Smart Data Integration Comments

Refresh Remote Table Supported with some Supported. Action to be done from the
Definition connections.  (Data Builder). For more
information, see Process
Source Changes in the Table
Editor

Partitioning data loads Partitioning on key columns Partitioning on key and non- Action to be done from
only and for snapshot key columns with certain Remote Tables monitor
replication only (no real- restrictions depending on details screen. For more
time) data type and adapter used. information, see Partitioning
Remote Table Data Loads
[page 184]

Statistics Supported Supported except for remote Action to be done


tables connected via an SAP from Remote Table
HANA smart data integration Statistics monitor.For more
CDI or an SAP HANA information, see Creating
smart data integration ABAP Statistics for Your Remote
adapters Tables [page 251].

Replicating Remote Tables Connected via SAP HANA Smart Data Access
For remote tables connected via SAP HANA smart data access, only one remote table per remote object can
have data access Replicated.

 Example

Example 1: Several remote tables (A, B, and C) are pointing to the same source object but none of the
remote tables is replicated:

• You are allowed to replicate only one remote table in this group. For example, you replicate remote table
B.
• After the replication (or the scheduling) is triggered on this remote table B, the replication and the
scheduling of the other remote tables (A and C) are disabled. Remote tables A and C will have to use
the replicated data of the replicated remote table (B), or you must remove the schedule and replicated
data of the remote table B to allow scheduling and replication of the remote table A or C.

 Example

Example 2: Several remote tables (A, B, and C) are pointing to the same source object and one remote
table (A) is replicated (or is scheduled to do so):

• On remote table A, normal replication and scheduling actions are enabled, depending if it is replicated
and/or scheduled.
• The replication and the scheduling of the other remote tables (B and C) are disabled. Remote tables B
and C will have to use the replicated data of the replicated remote table (A), or you must remove the
schedule and replicated data of the remote table A to allow scheduling and replication of the remote
table B or C.

Integrating Data and Managing Spaces in SAP Datasphere


178 PUBLIC Managing and Monitoring Data Integration
Repairing a Load

If the replication status shows an error for snapshot or real-time replication, you can repair the load.

1. Select the remote table and choose Data Replication Remove Replicated Data .
2. Select the remote table and choose Data Replication Start Data Replication or Data Replication
Enable Real-Time Data Replication .

Canceling a Run

If you need to cancel a running snapshot replication or an initialization task (for example if the replication or
initialization takes long) you can click the Cancel Run button available in the Remote Tables Details screen:

1. From the Remote Tables , Click  of the remote table you need to cancel the run.
2. Click Cancel Run.

 Note

A replication task consists of several steps that include exchanges with SAP HANA. Therefore, you
might get different status while clicking Cancel Run, depending where you are in the processing steps.

As a consequence, the replication status of your remote table turns into Failed, and a cancel replication task
starts.

 Note

If replica data already exists from a previous run, the remote table continues to use it after the run is
canceled.

8.4.1 Replicate Full Set of Data

Start a data replication to copy the full set of data from your source object (usually a database table or view)
into SAP Datasphere.

Context

Availability of data replication (snapshot) depends on the connection type. For more information on which
connection types support snapshot replication, see Integrating Data via Connections [page 8].

 Note

By default, the full set of data is copied and imported in a single background task. For remote tables
with larger data volumes, this can lead to out-of-memory errors or high memory peaks. To avoid such
situations, you can enable partitioned data transfer of large datasets for supported connection types

Integrating Data and Managing Spaces in SAP Datasphere


Managing and Monitoring Data Integration PUBLIC 179
(which is all remote table-enabled connections that use a Data Provisioning Agent). For more information,
see Partitioning Remote Table Data Loads [page 184].

Procedure

1. Go to Data Integration Monitor and select the space in which the relevant remote table resides.
2. Go to the Remote Tables.
3. Select the remote table that you want to replicate data for.

4. When real-time replication has been used before, select Data Replication Remove Replicated Data .
5. You have two options:

Option Description

Directly start the replication Select Data Replication Start Data Replication .

Schedule the replication Select Schedule Create Schedule . For more infor-
mation, see Schedule a Data Integration Task (Simple
Schedule) [page 256].

Results

When you directly start the replication, it starts in the background and you can see the progress of the
replication in the replication status. Once the replication has finished, you can see it in the Status column. For a
successful load, you will see the status Available and the date and time of the latest update. The Frequency for a
replication you started directly is None. The changed size of the remote table will also be shown in the monitor.

When you scheduled the replication, the columns Frequency and Next Run are updated with the scheduling
information. You can display the scheduling settings by clicking Scheduled in the Frequency column.

You can access the detailed logs for the remote table: Select the relevant remote table and click (Remote
Table Logs).

Once, you start a new data replication (either directly or via a schedule) while replicated data already exists, the
replicated data in SAP Datasphere is deleted and fully replaced by the new one. .

 Note

You can still access the current data until the replication ends. Once replication has successfully run, the
data is updated. Please note that this data refresh behavior requires temporarily additional in-memory and
disk-storage to be reserved during the processing of a new data replication.

Integrating Data and Managing Spaces in SAP Datasphere


180 PUBLIC Managing and Monitoring Data Integration
8.4.2 Replicate Data Changes in Real-Time
Use real-time replication to copy the data changes from your source object in real-time into SAP Datasphere.

Prerequisites

 Note

Availability of real-time / trigger based replication depends on the connection type. For more information
on which connection types support real-time replication, see Integrating Data via Connections [page 8].

• The object in the source has to be enabled for change data capture (CDC).
• If the database table in the source is CDC enabled but the remote table has been deployed before real-
time replication was possible with SAP Datasphere, you have to re-deploy the table to allow real-time
replication. You can re-deploy the table in the remote table editor in the Data Builder.
• For source views, real-time replication ist not supported.
• For ABAP Dictionary tables, real-time replication is not supported.
• For ABAP ODP sources, real-time replication is supported for in the following cases:
• ODP-BW: InfoProviders with a change log. This can be:
• DataStore objects (advanced) with data activation and change log (object type ADSO)
• Standard DataStore objects (classic) (object type ODSO)
• InfoObjects

 Note

To properly display real-time capability for ODP-BW InfoProviders in SAP Datasphere, please
update your SAP BW to:
• SAP BW 7.4: Support Package Stack 23 or higher (see SAPKW74023 )
• SAP BW 7.5: Support Package Stack 17 or higher (see SAPK-75017INSAPBW )

• ODP-CDS: For all ABAP CDS views with primary key and that are delta-enabled

 Note

If you apply filters to replicate your remote table, you must apply them on primary key fields
only. Indeed, delete operations on ABAP CDS Views are only recorded with their key fields in the
Operational Delta Queue (ODQ), but no content is recorded for other field types. In case of filtering
on non-key fields, there is strong risk that deletions are not propagated consistently to the target
remote table.

For more information, see:


• [CDS based data extraction – Part II Delta Handling. See SAP Blogs ]: "In case of a DELETE
operation the job generates an empty record with just the key field(s) filled, sets the deletion
indicator and hands them over to ODQ".
• 2930269 "Question1: How the deleted record looks like during delta load?"

• ODP-SAPI: For all extractors that are delta enabled (delta method ADD* not allowed) and have a
primary key. For more information, see CDS Views Enabled for Data Extraction.

Integrating Data and Managing Spaces in SAP Datasphere


Managing and Monitoring Data Integration PUBLIC 181
See the blog Finding the right CDS Extractor in SAP S/4HANA (published in January 2022).

 Note

For an overview of all extractors released for ODP replication (ODP SAPI), see SAP note
2232584 .

• For remote tables using an SAP HANA smart data access connection (SAP HANA on-premise via
Cloud Connector or SAP HANA Cloud), you can replicate your data changes in real-time but with some
restrictions as some remote object types are not supported:
• Remote object needs to be a table of type "COLUMN TABLE" (other table types like Row Tables or
Virtual Tables not supported)
• Some data types and SAP HANA table features are not supported. For more information, see
Unsupported Data Types and Other Limitations.

 Note

You can’t pause the replication for connections if SAP HANA smart data access is used.

• For remote tables connected via an SAP HANA smart data integration Cloud Data Integrator (CDI) adapter,
we recommend that you use a DP Agent version equal or higher than 2.6.1.

Procedure

1. Go to Data Integration Monitor and select the space in which the relevant remote table resides.
2. Go to the Remote Tables.
3. Select the remote table that you want to replicate data for.

4. When changing from scheduled data replication to real-time replication, you must select Data
Replication Remove Replicated Data to delete the current replicated data from the replica table and
delete the schedule.

5. Select Data Replication Enable Real-Time Data Replication .

Results

Real-time replication gets initialized by copying the data from the source. You will see the status Active in the
Status column once real-time replication has been successfully initialized. Whenever data now changes in the
source, the changes will be copied into SAP Datasphere. You won't be notified on individual replications. The
Latest Update column shows date and time when data has been lastly changed in SAP Datasphere.

 Note

No logs are generated when data is replicated in real-time mode. The displayed logs relate to previous
actions.

Integrating Data and Managing Spaces in SAP Datasphere


182 PUBLIC Managing and Monitoring Data Integration
8.4.3 Pausing and Restarting Real-Time Replication
To avoid any impacts on real-time replication while performing maintenance activities in the underlying source
system, or while applying changes to the SAP HANA smart data integration Data Provisioning agent, you can
pause and restart real-time replication.

Whenever you need to perform maintenance operations in one or more underlying source systems, with the
DW Space Administrator role you can pause real-time replication for selected connections that connect to
the sources through a Data Provisioning agent in the Connections application. For more information, see Pause
Real-Time Replication for a Connection [page 126].

Whenever you need to upgrade or patch your Data Provisioning agent, or you need to apply any changes to its
configuration, with the DW Administrator role you can pause real-time replication for all connections that
use a specific Data Provisioning agent in the Configuration Data Integration On-Premise Agents . For
more information, see Pause Real-Time Replication for an Agent.

 Note

From remote tables using an SAP HANA Cloud connection, you can’t pause the replication in the
Connections application if SAP HANA smart data access is used.

8.4.4 Resuming Real-Time Replication After a Fail


You want to resume a real-time replication after it has failed, once you have corrected the error.

Context

It can occur that your real-time replication has been paused with errors. To resume your real-time replication,
you need to check the relevant log in the details screen of the relevant remote table to identify what is the
error and how to solve it. Once the error has been corrected, use the Retry button to resume the real-time
replication:

Procedure

1. Go to the Details screen of the remote table using the button  Remote Table Logs.
2. Check the log details and correct the error.
3. Click retry to resume the real-time replication.
4. Refresh your screen.

The status of your real-time replication is now back to active.

 Note

The Retry button must be used once you’ve corrected the errors. It does not serve to correct the error
itself but it serves to resume the real-time replication only. It corresponds to the RETRY statement

Integrating Data and Managing Spaces in SAP Datasphere


Managing and Monitoring Data Integration PUBLIC 183
as available in SAP HANA Smart Data Integration. For more information, see PROCESS REMOTE
SUBSCRIPTION EXCEPTION Statement.

 Tip

You might need to consider monitoring the delta queues in the ABAP backend in case of real-time
replication exceptions raised in ABAP adapter-based connections (ABAP, SAP BW, SAP ECC, SAP S/
4HANA on premise).

8.4.5 Disabling Real-Time Data Replication

Disable real-time data replication even if your source system is unreachable, and switch the data access from
Replicated (Real-time) to Replicated (Snapshot).

If the source system is not reachable or if you simply don' t need to get your data updated in real-time, you can
disable the real-time data replication.

From the Remote Tables monitor, select one or more remote tables and click Data Replication Disable
Real-Time Data Replication .

 Note

The selected remote tables must fulfill the following requirements to support disabling real-time data
replication:

• The remote table must be in real-time data replication.


• The remote table must be connected via SAP HANA smart data integration (in the connection, Data
Provisioning option is set to Data Provisioning Agent) and not via SAP HANA smart data access.

When you disable the real-time data replication, it changes the data access from Replicated (Real-Time) to
Replicated (Snapshot). It preserves the replica table and its replicated data to allow further data consumption.

Switching the data access from Replicated (Real-Time) to Replicated (Snapshot) allows you to:

• Delete the remote table as it is not possible to delete a remote table, which is in data access Replicated
(Real-time).
• Change the properties of the connection. For more information, see Modify or Duplicate Remote Tables.

 Note

If your remote table is connected via SAP HANA smart data access, you can delete it. This will stop the data
replication in real-time and you’ll be able to modify the connection details and restart a data replication.

8.4.6 Partitioning Remote Table Data Loads

Replication capabilities depend on the data type and size. Larger data volumes can cause replication runs
to terminate because of memory shortage. From the Data Integration Monitor Remote Tables - Details

Integrating Data and Managing Spaces in SAP Datasphere


184 PUBLIC Managing and Monitoring Data Integration
screen, you can create partitions on columns from your dataset and break your data down into smaller and
more manageable parts.

Replicating remote tables with larger data volumes without having out-of-memory errors or high memory
peaks can sometimes be a challenge for Data modelers. To avoid such situations, you can create partitions
per remote tables and thus enable partitioned data transfer of large datasets for supported connection types
(which are all remote table-enabled connections that use a Data Provisioning agent).

 Note

The partition is effective for both snapshot and real-time enabling replication if your remote table is
connected to the source via SAP HANA smart data integration. However, it is not working for real-time
replication if your remote table is connected via SAP HANA smart data access.

Creating Partitions for Your Remote Tables

You can create partitions from the Data Integration Monitor Remote Tables – Details screen:

1. Click  of the remote table you need to create partitions for.


2. Go to the Partitions tab and click Define Partitions.
3. In the Define Partitions window, define your partitions settings. Note that the available options can differ
depending if your remote table is connected via SAP HANA smart data integration or SAP HANA smart
data access:

Property Description Comment

Column Select the column from which the It might happen that you can’t see all
partition will be defined. the columns available in your dataset
because:
• Not all data types are supported
for partitioning depending on
your data sources:
• SAP HANA data: Supported
data types include:
• String
• Integer, Integer
64, Decimal,
hana.SMALLINT,
hana.TINYINT
• Date, DateTime, Time,
Timestamp
• Binary
For more information,
see Partitioning Limits in
the SAP HANA Cloud,
SAP HANA Database
documentation.

Integrating Data and Managing Spaces in SAP Datasphere


Managing and Monitoring Data Integration PUBLIC 185
Property Description Comment

• (SAP HANA smart data


integration only) If you select a
non-key column to define your
partitions, you must ensure that
the column value of a record in
the source data is not changed
during replication. Otherwise,
data loading may fail because
a record with the same key
is replicated in two different
partitions.
• (SAP HANA smart data access
only) You can create partitions
only on key columns.
• (SAP HANA smart data
integration only) You can't define
partitions on columns, which
contain amounts associated
with a currency or quantities
associated with a unit.
• (SAP HANA smart data
integration only) For ODP data
sources, only the following
columns can be selected:
• Columns which support
where clause of the SELECT
statement.
• Columns equal to or less
than 45 characters.

 Restriction
For ODP data sources, you
can create partitions on non-
key columns only if your
remote tables have been
deployed using a DP agent
connection from release
2.5.4.1.

Moreover, if the repository


entity comprises filter
capabilities, you first need to
refresh the remote table in
the Data Builder, before you
can create partitions.

Parallel Processes (SAP HANA smart You can define up to 5 parallel By default, the partitions are loaded
data integration only) processes. one after the other to control
resource consumption.

Integrating Data and Managing Spaces in SAP Datasphere


186 PUBLIC Managing and Monitoring Data Integration
Property Description Comment

 Note
No parallel processes are
possible for remote tables
connected with an SAP HANA
smart data access adapter.

Partitions Create the number of desired  Note


partitions by entering a range for
each of them.
• Partitioning requires an in-
depth knowledge of the
values that are used or
 Caution
are valid for the chosen
Ranges must not overlap partitioning column.
• Negative values are not
. valid for partition bounds of
numeric columns and will be
converted to 0 at runtime.
• The low bound must be
lower than the high bound.
The order is checked
according to the type of the
partition column. Example:
for string-like columns, '90'
<= DOC_ID < '100' is not
valid, whereas '090' <=
DOC_ID < '100' is valid.
• (SAP HANA smart data
integration only) Values
entered in the bounds for
ABAP ODP remote tables
must match the ABAP
internal format: for example,
enter "20210714" if the date
is "July 14, 2021".
• (SAP HANA smart data
access only): Only single-
level range partitions are
supported by SAP HANA .
For more information, see
Range Partitioning.

4. Load a data replication to start the partitioning immediately, or alternatively create a schedule to run the
data replication later (or wait until your next schedule run will occur).

 Note

Partitions can be updated or deleted using the corresponding menu. But for remote tables connected
via SAP HANA smart data access, you must first remove the replicated data before you can delete the
partitions.

Integrating Data and Managing Spaces in SAP Datasphere


Managing and Monitoring Data Integration PUBLIC 187
Editing Existing Partitions for Your Remote Tables

Whenever it’s necessary, you can update the existing partitions using the menu Partitions Edit .

 Example

For example, you want to include a new year in your partitions, or you need to fine granular the existing
partitions.

 Note

Changes will be reflected only after a new data replication (snapshot or scheduled), or a real-time
replication.

 Restriction

For remote tables connected via SAP HANA smart data access, you must first remove the replicated data
before you can change the partitions.

Deleting Existing Partitions for Your Remote Tables

You can delete an existing partition using the menu Partitions Delete .

 Restriction

For remote tables connected via SAP HANA smart data access, you must first remove the replicated data
before you can change the partitions.

Monitoring Your Partitions

Once the partitioning task is completed, messages with information on partitions is created. You have different
places to monitor the partitioned data loads:

• In the task logs that are created: one log is created per partition and contains detailed information.
• In the Data Integration Monitor Remote Queries by displaying the SQL Statement: The SELECT
statements sent by SAP HANA smart data integration to the remote source are listed. The partition
information is part of the WHERE clause of the particular statement. For more information, see Monitoring
Remote Queries [page 249].

 Note

If your defined ranges don’t cover all the data, additional partitions are created in the background . For
example, if rows have been inserted or modified since your last data replication and they now don’t match
any of the defined ranges. An additional partition called 'Others' is created to cover this data.

Integrating Data and Managing Spaces in SAP Datasphere


188 PUBLIC Managing and Monitoring Data Integration
Examples of Partitioning

Let's take an example to understand how the partitioning is done. The following partitions for column Calyear, a
4-digit numeric field which can include NULL values, are defined in the UI:

1. Partition 1: 2000 <= Calyear < 2005


2. Partition 2: 2005 <= Calyear < 2010
3. Partition 3: 2010 <= Calyear < 2011

The replica table is partitioned accordingly. An "Others" partition could be added for data, which is not covered
by the defined partitions:

Partitions for replica table:

1. Partition 1: 2000 <= Calyear < 2005


2. Partition 2: 2005 <= Calyear < 2010
3. Partition 3: 2010 = Calyear
4. Partition 4: Others

The partition ranges are also used by SAP HANA smart data integration data load runtime as filters in the
SELECT statements sent to the remote source. Usually, the filters are exactly the same as the partition ranges,
with the exception of ABAP ODP Extractors. In this case, the filters are as follows:

SAP HANA smart data integration data load runtime for ABAP ODP Extractors:

1. Partition 1: 0000 <= Calyear <= 1999


2. Partition 2: 2000 <= Calyear <= 2004
3. Partition 3: 2005 <= Calyear <= 2009
4. Partition 4: 2010 = Calyear
5. Partition 5: 2011 <= Calyear <= 9999
6. Partition 6: Calyear = NULL

8.5 Monitoring Local Tables

Monitor all the local tables created for a space in the Data Builder and check their metrics.

Local tables can be used as source data or target data by SAP Datasphere apps. As a Data Integrator you might
want to monitor all local tables which have been created in the current space to check data size, updates in
these tables, etc.

Go to Data Integration Monitor Local Tables . All local tables that have been created in the Data Builder
are listed.

 Note

Tables created in Open SQL schemas are not listed. For more information, see Integrating Data via
Database Users/Open SQL Schemas [page 145].

The monitor displays the following properties:

Integrating Data and Managing Spaces in SAP Datasphere


Managing and Monitoring Data Integration PUBLIC 189
Properties Description

Technical Name or Business Name Displays the technical name or the business name of the
table, depending on how you have configured your UI
settings in Your Profile Settings .

Number of records Indicates the number of records contained in the table.

Growth in 30 Days Displays the growth in % of the number of records since the
last month. It is computed by taking the difference between
the number of records in the table today and the number
of records in the table 30 days ago, and it calculates the
growth in percentage.

 Note
Figures are only available for large tables (superior of 1
megabyte). It takes the value N/A if the table is smaller
than 1 megabyte and O if no changes have been made
during the month.

Used Disk (MiB) Displays the disk storage currently used by the local table in
MiB.

 Note
Value is set to Not Applicable for local tables that store
data in-memory.

Used In-Memory (MiB) Displays the memory currently used by the local table in
MiB.

 Note
Value is set to Not Applicable for local tables that store
data on disk.

Last updated Displays the timestamp when the local table was last
updated.

 Restriction
If the local table is updated through a replication flow,
the information is not available if the flow is running via
the load type Initial and Delta.

Last updated by Indicates how the table was last updated. For example,
name of the flow if it was updated by a flow.

 Note
If you are authorized to navigate to the relevant object,
a clickable link is provided and you can navigate to the
object details. If you are not authorized to navigate to
the object, you’ll get the value unauthorized.

Integrating Data and Managing Spaces in SAP Datasphere


190 PUBLIC Managing and Monitoring Data Integration
Properties Description

 Restriction
If the local table is updated through a replication flow,
the information is not available if the flow is running via
the load type Initial and Delta.

Number of partitions Displays the number of partitions defined on the local table.

 Note
Displays 0 if no partitions have been defined on the
local table. For more information, see Partitioning Local
Tables.

Delta capture Indicates if the local table allows delta capture. For more
information, see Capturing Delta Changes in Your Local
Table.

8.5.1 Performing Data Deletion for Local Tables

Delete records for local tables, on-demand, using filter conditions or using a schedule.

Table records can consume a lot of memory and you might need to do a cleanup of older records regularly.
Especially for tables where delta capture is enabled, as when you delete records, they are not physically deleted
but marked as deleted to allow other apps to propagate the changes to the different objects that consume
it in delta mode. For more information, see Capturing Delta Changes in Your Local Table. As these records
continue to consume memory until they are physically deleted from SAP Datasphere, you must clean up your
data regularly.

From the Data Integration Monitor ( ), go to the Local Tables monitor. Navigate to the details screen of the
local table you want to perform the data deletion.

The header of the screen displays metrics on the local table:

Metric Description

Used Disk Displays the size used by the local table if it is stored on the
disk.

Used In-Memory (MiB) Displays the size used by the table, if it is stored in-memory.
For more information, see Accelerate Table Data Access
with In-Memory Storage

Number of Records Displays the current number of records in the table.

Integrating Data and Managing Spaces in SAP Datasphere


Managing and Monitoring Data Integration PUBLIC 191
Metric Description

Delta Capture Indicates if the table tracks the changes made on your local
table after you have deployed it. For more information, see
Capturing Delta Changes in Your Local Table

In-Memory Storage Indicates if the table is stored in-memory.

The Data Deletion Tab

From the Data Deletion tab, you define criteria to delete the records:

Type of Deletion Description

Delete All Records You want to delete all records contained in the local table.
Note that in case of tables with delta capture enabled,
all records will be marked as deleted but will not be
deleted physically. For more information, see Capturing
Delta Changes in Your Local Table

Delete Filtered Records You define a filter and based on this filter, the records will be
deleted. Note that for tables with delta capture enabled the
records will be marked as deleted but will not be physically
deleted.

 Restriction
SQL based selection criteria are not supported.

 Note
Click Show Preview to preview records that match the
filter criteria and will be deleted. You must have the role
DW Integrator or DW Modeler to preview the data.

Delete Records Marked as "Deleted" This option will appear only if your table is a delta capture
table. With this option, you can physically delete records
marked as deleted and that have already been successfully
processed by other apps that consume them. You can
define a period of deletion. For example, records older than
30 days.

 Note
Data deletion preview is not supported for records
marked as "deleted", as the data editor does not

Integrating Data and Managing Spaces in SAP Datasphere


192 PUBLIC Managing and Monitoring Data Integration
Type of Deletion Description

support the switch between active records and delta


capture records.

In the Preview section, in addition to previewing the data to be deleted when supported, you can see 2
additional metrics:

Metric Description

Number of Deletable records Displays the number of records that will be deleted based
on your previous selection.

Size of Deletable Records Displays the size of the records that will be deleted based
on your previous selection.

Click Delete Records to delete records based on your selection.

 Note

Ensure that flows that consume the table are not affected.

Data Deletion Schedules

From the Data Deletion Schedules tab, you can automate data deletion with a schedule.

To create a schedule, click Create.

The Schedule Data Deletion window opens:

1. Define the settings of the schedule: Give a business name and a technical name to your schedule, and
select the type of deletion you would like to apply.
2. Define the Frequency and Time Range for your schedule. For more information, see Scheduling Data
Integration Tasks [page 256].
3. Click Next Step.
4. Review your settings and click Create Schedule.

From this tab, you can also change or delete a schedule, and monitor the existing schedules with the available
metrics:

Metric Description

Type Displays the type of deletion you applied on your table.

Name Displays the business name or technical name of the


schedule. Technical or Business Name is displayed,
depending on how you have configured your UI settings in

Profile Settings .

Integrating Data and Managing Spaces in SAP Datasphere


Managing and Monitoring Data Integration PUBLIC 193
Metric Description

Frequency Displays the frequency of the schedule.

 Note
Paused is displayed in the case the schedule has been
paused.

Last Run Start Displays when the last scheduled run has started (date and
time).

Last Run End Displays when the last scheduled run has ended (date and
time).

Status Displays the status of your scheduled run.

Records (Last Run) Number of records deleted with the last scheduled run.

Data Deletion Logs

From the Data Deletion Logs tab, you can monitor how the deletion task runs:

Metric Description

Start Displays the date and time when the deletion task started.

Duration Displays how long the task ran to delete all records.

Object Type Displays the type of object that contains the records to be
deleted.

Activity Displays what type of activity the task runs.

User Displays who initiates the deletion task.

Number of records Displays the number of records that have been deleted, or
marked for deletion in case of local table with delta capture
enabled.

Status Displays the status of the deletion task.

Integrating Data and Managing Spaces in SAP Datasphere


194 PUBLIC Managing and Monitoring Data Integration
8.6 Monitoring Local Tables (File)

Monitor your local tables (file). Check how and when they were last updated and if new data has still to be
merged.

 Note

The object store is not enabled by default in SAP Datasphere tenants. To enable it in your tenant, see SAP
note 3525760 .

For additional information on working with data in the object store, see SAP Note 3538038 .

The object store cannot be enabled in SAP Datasphere tenants provisioned prior to version 2021.03. To
request the migration of your tenant, see SAP note 3268282 .

Local tables can be used as source data or target data by SAP Datasphere apps. As a Data Integrator you
might want to monitor all local tables (file) deployed in your space and check how and when data has been last
updated and if some data must still be merged.

Go to Data Integration Monitor Local Tables (Files) . All local tables (file) that have been created in the
Data Builder are listed. For more information on local tables (file), see Creating a Local Table (File).

The monitor displays the following properties:

Property Description

Technical Name or Business Name Displays the technical name or the business name of the
table, depending on how you have configured your UI
settings in Your Profile Settings .

Buffer Merge Status Status of the data merge.

Data updates are waiting in the Inbound Buffer until they


are merged.

They can have the following status:

• Running: a merge task is still running.


• Failed: Last merge run task has failed. You might need
to check the log for more details.
• Merged: All data have been merged with the last merge
task.
• New Data: Data are waiting to be merged.
• Empty: There is no new data waiting in the inbound
buffer.

Last updated Displays the timestamp when the local table (file) was last
updated.

Last updated by Indicates how the local table (file) was last updated. For
example, name of the task chain that updated it.

Buffer Last Updated Displays the timestamp when the inbound buffer was last
updated by a replication flow.

Integrating Data and Managing Spaces in SAP Datasphere


Managing and Monitoring Data Integration PUBLIC 195
Property Description

Buffer Last Updated by Indicates the name of the replication flow that last updated
the inbound buffer.

Delta capture Indicates if the local table (file) allows delta capture. For
more information, see Capturing Delta Changes in Your
Local Table.

Partitions Indicates if the local table (file) contains partitions.

8.7 Persisting and Monitoring Views

From Data Integration Monitor Views , you can monitor views that have been created in the Data Builder.
You can persist these views (direct run or via a schedule) to make them available locally to improve the
performance when accessing your data. You can monitor the existing persisted views to keep control of your
data sizing and free up memory space.

Prerequisites: Your view must be deployed.

 Note

While persisting complex views, see Persisted Views and Memory Consumption [page 214].

The interactive graphic below summarizes how views are persisted:

Integrating Data and Managing Spaces in SAP Datasphere


196 PUBLIC Managing and Monitoring Data Integration
• #unique_150/unique_150_Connect_42_subsection-im1 [page 198]
• #unique_150/unique_150_Connect_42_subsection-im2 [page 198]
• #unique_150/unique_150_Connect_42_subsection-im3 [page 198]
• #unique_150/unique_150_Connect_42_subsection-im4 [page 198]
• #unique_150/unique_150_Connect_42_subsection-im5 [page 198]
• #unique_150/unique_150_Connect_42_subsection-im6 [page 198]

Integrating Data and Managing Spaces in SAP Datasphere


Managing and Monitoring Data Integration PUBLIC 197
• #unique_150/unique_150_Connect_42_subsection-im7 [page 198]
• #unique_150/unique_150_Connect_42_subsection-im8 [page 198]
• #unique_150/unique_150_Connect_42_subsection-im9 [page 199]
• #unique_150/unique_150_Connect_42_subsection-im10 [page 199]
• #unique_150/unique_150_Connect_42_subsection-im11 [page 199]

Persisted Views
The definition of the view is stored in SAP Datasphere and data is read from the persistency table to improve
performance when accessing the data.

Persistency Tables
Tables created in SAP HANA Cloud that stored the required data. When a view is persisted, data is read from
this table to improve performance. When you refresh the persisted data of a view, a new persistency table is
created. Data from the technical view are inserted into the new persistency table. If everything runs properly,
the old persisted data is replaced by the new inserted data. In case of issues, previously persisted data are still
available, until you solve the issue and refresh the data with success.

Technical Views
Copy of the virtual views to be persisted into SAP HANA Cloud. The definition of the view is stored (including
associations, hints, etc). It serves to create the persistency table.

Persist Data
Views can be persisted by running a data persistence or scheduling a data persistence task. You can also
partition your data load into smaller parts to improve the performance.

For more information, see Creating Partitions for Your Persisted Views [page 210] and Scheduling Data
Integration Tasks [page 256].

Views (SQL, Graphical) in virtual access


SQL or Graphical views are created in the data builder. The views are accessed directly from the source data.
Once they are deployed in a space, they can be persisted.

For more information, see Creating a Graphical View and Creating an SQL View.

Remote table
Table created from a connection to allow source data to be accessed in SAP Datasphere. Views can be created
using these tables as source. By default, data access of a remote table is "remote", but it can be "replicated".

For more info, see Replicating Data and Monitoring Remote Tables [page 168].

Local table
Table created from sources that have already been integrated into SAP Datasphere. It serves of source to
create a view.

For more information, see Creating a Local Table.

SAP HANA smart data integration/SAP HANA smart data access


SAP HANA smart data integration and SAP HANA smart data access allow you to access remote data through
virtual tables without copying the data into SAP HANA.

Integrating Data and Managing Spaces in SAP Datasphere


198 PUBLIC Managing and Monitoring Data Integration
For more information, see Connecting SAP HANA Cloud, SAP HANA Database to Remote Data Sources.

Flows
Define the flow of your data, starting by reading data from sources and writing data into a target after
performing transformations on the data.

For more information, see Creating a Data Flow and Creating a Replication Flow.

Remote Sources
You can connect several data sources to SAP Datasphere that support remote table functionalities. Views can
then be created from these remote tables.

For more information, see Integrating Data via Connections [page 8].

Local files
Local files can be imported and deployed to your space. They are then used to create a local table.

For more information, see Creating a Local Table from a CSV File.

Action in Views

The monitor displays the views in 3 tabs:

• All Views: It displays all the deployed views that have been created in the Data Builder for the current space.
• Scheduled (default): It displays the views for which a schedule is defined to persist the data.
• Persisted: It displays the views already persisted.

You can perform the following actions (might depend on your role and privilege, see Privileges and Permissions
and Roles and Privileges by App and Feature:

Integrating Data and Managing Spaces in SAP Datasphere


Managing and Monitoring Data Integration PUBLIC 199
Action Information

Data Persistence menu Start Data Persistence: Start a new data persistence to up-
date or create the persisted view.

 Note
You can set up a view as persisted even if it has been
created on top of remote data.

Remove Persisted Data: Remove the data that have been


persisted in the view and switch the access back to virtual.

Start View Analyzer: Analyzes each view definition, the con-


sumed views and the data sources used by local and remote
tables that compose your data model. You can then use
the provided information and warnings to decide on some
improvements for your data model. You can also generate an
SQL Analyzer Plan File that you can download (if you have
the required authorizations) for detailed analysis. For more
infromation, see Exploring Views with View Analyzer [page
221].

Schedule Data Persistence From this menu, you can :

• Create Schedule: Select the relevant persisted view and


create a simple or recurring schedule for your view. You
define your scheduling options and thus ensure that
you always have an up-to-date persisted view.
Edit Schedule: Your scheduling options need to be up-
dated? You can adapt them to your needs at any time
from this menu.
• Delete Schedule: You don't need to schedule a Data Per-
sistence task anymore? Then you can simply delete it
from this menu.
• Assign Schedule To Me: Become the owner of the
schedule.
• Pause Schedule: Pause the scheduled task
• Resume Schedule: Resume the pause scheduled task

 Note
You need to authorize SAP to run the recurring sched-
uled tasks on your behalf. You can do so via the message
that is displayed at the top of the monitor, or in your
profile settings under Schedule Consent Settings.

For more information, see Schedule a Data Integration Task


(Simple Schedule) [page 256].

Integrating Data and Managing Spaces in SAP Datasphere


200 PUBLIC Managing and Monitoring Data Integration
Action Information

Access the detailed logs to monitor what's happened with Select the relevant view and click:  (View Details).
your persisted view.

 Note

You can select several views to group actions whenever it's applicable:

• Schedule: You can delete, pause or resume schedules for several views that have schedules defined.
You can also assign several schedules to your name and become the owner of the schedules.
• Remove Persisted Data: You can select several views that have persisted data and remove the persisted
data for all of them.
• Many actions like Start Data Persistence, Run, Create Schedule, etc. cannot be grouped. In such cases,
they are disabled.

After triggering a group of actions, the status of each action is displayed in the notification area. As these
operations run in an asynchronous mode, the status will not be updated automatically, and you need to
manually Refresh the page to get the updated status.

 Note

You can still access the current data until the load ends. Once the data persistence has successfully run, the
data is updated. Please note that this data refresh behavior requires temporarily additional in-memory and
disk-storage to be reserved during the processing of a new data persistence.

Monitoring the Persisted Views

The list of available persisted views is displayed in a table containing several information:

Column Information

Technical Name (or Business Name) Name of the persisted view.Technical or Business Name is
displayed, depending on how you have configured your UI

settings in Your Profile Settings

Integrating Data and Managing Spaces in SAP Datasphere


Managing and Monitoring Data Integration PUBLIC 201
Column Information

Data Access Shows how you currently access your view.

• Persisted: The view is persisted can be used immedi-


ately.
• Partially Persisted: Not all view data has been persisted.
If your view contains an input parameter, only records
that match the input parameter default value are per-
sisted.
• Virtual: The view is accessed directly, no intermediate
persistency is used. Or the view was persisted and has
now been turned into virtual to free up memory space,
for example.

Frequency Shows if a schedule is defined for the view.

• None: There is no schedule task defined for this

view. You can define one from Schedule Create

Schedule .
• Scheduled: A schedule task is defined for this view. If
you click on Scheduled, you will get detailed informa-
tion on the schedule. You can update the schedule op-

tions at any time from Schedule Edit Schedule ,

or delete the schedule from Schedule Delete

Schedule .
• Paused: Schedule is paused and the data is not re-
freshed.

Last Updated Shows when the persisted view was last updated.

New Run If a schedule is set for the view, see by when the next run is
scheduled.

Number Of Records Shows the records of the persisted views.

Memory Used for Storage (MiB) Tracks how much size the view is using in your memory.

Disk Used for Storage (MiB) Tracks how much size the view is taking on your disk.

Integrating Data and Managing Spaces in SAP Datasphere


202 PUBLIC Managing and Monitoring Data Integration
Column Information

Status Shows the status of the persisted view.

• Available: The persisted view is available and can be


used.
• Loading: The persisted view is currently creating or up-
dating. You might need to refresh your screen until the
loading is completed to get the final status. Until then
the virtual access or the old persisted data is used if the
view is accessed.
• Error: Something goes wrong during the load of the data
to the persisted table. The old persisted data is used or
if the view was not successfully loaded before, the data
is still accessed via virtual access (status is virtual).
You need to fix the error to be able to complete the
persisted view creation or update.

Schedule Owner Displays the name of the schedule owner.

You can personalize the columns you want to display clicking on  Select Columns

Persisted Views and Input Parameters

Data persistence can be run in case a view contains a parameter only if the following prerequisites are met:

• The view contains one single input parameter,


• The input parameter has a default value maintained.

If a view contains several input parameters, or if the input parameter has no default value, the view can’t be
persisted.

Once persisted, the view takes the data access value Partially Persisted because only records that match the
input parameter default value are persisted.

Persisted Views and Deployment

When you deploy a persisted view, you need to consider the following cases:

• If you update or redeploy a persisted view, you need the right permission level (Data Integration - Update).
For more information, see Privileges and Permissions. Note that after a redeployment, the data persistence
might be deleted and the data access could be changed back to virtual for views with structural changes.
This will happen as soon as the underlying database object is recreated as part of the deployment. This can
be seen in the Data Access status in the View Builder or in the Data Integration Monitor Views . The
update of the data persistence has to be triggered again.
• If the view uses other views that were changed but not yet deployed, they are deployed as well. If these
views are persisted, the persistence is deleted as well.

Integrating Data and Managing Spaces in SAP Datasphere


Managing and Monitoring Data Integration PUBLIC 203
• If the view uses other views that were not changed, these views are not touched by the deployment and
therefore the persistence is still available
• If you update or redeploy a view while you are persisting data, the persistence will fail. In this case, try again
to persist the view or wait until the next scheduled run.
• If the persisted view is consuming a view for which a data access control has changed (a data access
control is added or removed, or its assignment has changed), then the persistence of your parent view is
removed when the underlying view is redeployed.

Canceling a Running Execution of a Data Persistence Task

If you need to cancel a running data persistence task, go to the Details screen of the relevant view and click
Cancel Run . This will turn the status of the task into failed and you will be able to restart the task to persist the
view later if needed.

 Note

A task to persist a view consists of several steps that includes exchanges with SAP HANA. Therefore, it
might take some time until the task is stopped and the status turns into Failed.

8.7.1 Data Persistence and Run Mode

From the Views Details screen, you can change the settings to persist your data to ensure that you choose the
best run mode according to your business case.

When persisting a view, an SQL procedure is called in SAP HANA. In case of high data volume to persist,
processing the SQL procedure can impact the memory consumption and the system resources. In the details
screen of the view under Settings, you can optimize the run mode depending on your needs.

Run Mode

You can change the run mode of your data persistence and choose between performance and memory
consumption:

• Performance-Optimized (Recommended): This is the recommended mode to run the task. SQL queries are
processed directly in-memory. No intermediate result is stored on the disk to ensure a quicker data access
and data consumption. However, this mode consumes memory resources and if your data model is very
complex, it can result in an out of memory and the task failed.
• Memory-Optimized: The mode uses an enhancement of the SAP HANA Execution Engine, which stores
intermediate results of SQL statements on the disk. Memory consumption of SQL statements is then lower
because intermediate results are not stored in-memory but are saved on the disk. It results in slower
runtime, though. For example, you can use it when you get an out of memory error while running the task,
as alternative to manual partitioning.

Integrating Data and Managing Spaces in SAP Datasphere


204 PUBLIC Managing and Monitoring Data Integration
 Note

You must ensure that you have sufficient disk space available. If not, it can result with an error of type
"Disk Full". You can monitor your disk space using the SAP HANA Cockpit. For more information, see
Disk Usage: Monitor Disk Volume

8.7.2 Persisted Views and Data Access Control

When Data Access Control is defined, it can impact the data persistence. Depending on where the Data Access
Control is set, it might prevent you from persisting a view.

Without Data Access Control (DAC), all users assigned to a space in SAP Datasphere can see and consume all
data available in this space. When DAC is applied to a data layer view or a business layer object, the rows of
data contained in the object are filtered based on the specified criteria and can impact the data persistence.

Persisted Views and Data Access Control on View Level

While defining Data Access Control in your view, you need to consider the impact on the data persistence:

• You can persist a view with Data Access Control set up.
• You can't persist a view consuming a view with Data Access Control set up.

 Example

A view A is created as join of view B and C. As Data Access Control is set on view B, all users can see
data from view C, but only authorized users can see data from view A and B. View A can't be persisted.

In such cases, we recommend persisting the data on views B and C. If Data Access Control is set to view
A instead of view B, then also view A can be made persisted. And it may not be necessary to persist
view B and C.

 Example

A view A is created on top of view B and view B is created on top of view C. Data Access Control is set on
view C. View B and view A can’t be persisted.

• If you change Data Access Control on a view, the following happens:


• If the view is already set as persisted, the persistence is removed when the view is deployed. You need
to persist the view again.
• If views on top are persisted, then the persistence of the top view is removed:
• For top views within the same space as the deployed view, the persistence is removed when you
deploy your changes.

 Example

View A is persisted, and it consumes view B. You add a DAC on view B. When redeploying view
B, persistence of view A is removed.

Integrating Data and Managing Spaces in SAP Datasphere


Managing and Monitoring Data Integration PUBLIC 205
• For other top views, the persistence is removed during the next data persistence run of the top
view.

Persisted Views and Data Access Control on Object Level

Example 1: Data Access Control is set on one column.

Data Access Control is set on "Region" column:

User A has authorizations to see data from "NA" region and User B is authorized to see data from "EMEA"
region.

The "Salary" view can be persisted.

Example 2: DAC is defined on one column, and inner join is set.

Taking back the previous example, we now create the "Salary_Bonus" view on top of the "Sales" and "Salary"
views. Data from the "Sales" view is not protected whereas data from the "Salary" view is protected with DAC
on the "Region" column. An inner join is set between the "Product" column issued in the "Sales" view and the
"Product Responsible" column used in the "Salary" view:

Integrating Data and Managing Spaces in SAP Datasphere


206 PUBLIC Managing and Monitoring Data Integration
Due to the inner join, only the sales records matching the authorized records from the "Salary" view are
selected. The "Sales" and "Salary" views can be persisted but the "Salary_Bonus" view can’t be persisted due
to DAC assigned to the "Salary" view.

Example 3: DAC is defined on one column and left outer join is set.
Again the "Salary_Bonus" view is created on top of the "Sales" and "Salary" views. Data from the "Sales" view is
not protected whereas data from the "Salary" view is protected with DAC on the "Region" column. A left outer
join is set between the "Product" column from the "Sales" view and the "Product Responsible" column from the
"Salary" view:

Due to the left outer join, the "Salary_Bonus" view contains all records of the "Sales" view. Even the
records from "Sales" view, where no authorized join partner exists in the "Salary" view are considered in the
"Salary_Bonus" view. The "Sales" and "Salary" views can be persisted but the Salary_Bonus" view can’t be
persisted due to DAC assigned to the "Salary" view.

Integrating Data and Managing Spaces in SAP Datasphere


Managing and Monitoring Data Integration PUBLIC 207
Example 4: DAC is defined on two columns and left outer join is set:

The "Salary_Bonus" view is created on top of the "Sales" and "Salary" views. Data from the "Sales" view is
protected with DAC on the "Product" column whereas data from the "Salary" view is protected with DAC on the
"Region" column. A left outer join is set between the "Product" column from the "Sales" view and the "Product
Responsible" column from the "Salary" view:

Due to left outer join, the salary records are selected as long as they are in the authorized region (EMEA),
even when no dedicated "product" authorizations (responsibility) are granted. You can persist the "Sales" and
"Salary" views but you can't persist the "Salary_Bonus" view.

Example 5: DAC is defined on two columns and inner join is set.

The "Salary_Bonus" view is created on top of "Sales" and "Salary" views. Data from the "Sales" view is
protected with DAC on the "Product" column whereas data from the "Salary" view is protected with DAC
on the "Region" column. An inner join is set between "Product" column from the "Sales" view and "Product
Responsible" column from the "Salary" view:

Integrating Data and Managing Spaces in SAP Datasphere


208 PUBLIC Managing and Monitoring Data Integration
Due to the inner join, only the authorized "Sales" records matching the authorized records from the "Salary"
view are selected. In the given example, this means that the "Salary_Bonus" view contains only data for region
EMEA and Product 10. You can persist the "Sales" and "Salary" views but you can't persist the "Salary_Bonus"
view.

Example 6: DAC is defined on 2 columns and a union is set.


Now, the "Sales_Bonus" view is created on top of "Sales External" and "Sales Internal" views. Data from both
"Sales External" and "Sales Internal" views are protected with DAC on "Product" column . A union is set:

With a Union, data records are selected from the underlying sources and the respective authorizations are
applied . The result set is the sum of the selected records. You can persist both the "Sales External" and "Sales
Internal" views but you can't persist the "Sales_Bonus" view.

Example 7: Permissions entity for DAC is a view containing run-time dependent logic.
A DAC can use a view or a table as permissions entity. A view used as permissions entity for a DAC can contain
run-time dependent logic. If such a DAC is assigned to a view and this view is persisted, wrong results may
occur when reading data from the persisted view. This behavior occurs if the runtime dependent logic inside
the view used as permissions entity returns different results when loading a snapshot and when reading later
on from the snapshot. In this case, only a subset of the expected entries of the view will be considered. The
subset may even be an empty set.

 Example

Let's assume we have the "EMEA_Sales" SQL View in which the "Current day" is fetched allowing users to
see the sales amount per day. A DAC is created using the "EMEA_Sales" view as permissions entity. This
DAC is assigned to a "Global_Sales" view.

Now, let's say that we are on April 1st, 2021 and you decide to persist the "Global_Sales" view. Imagine
that a user would like to consume data of the "Global_Sales" view on April 4, 2021. As a result, no data is
displayed as the view was persisted with entries of data from April 1st and the DAC is configured to always
fetch the entries from the current day (April 4).

Integrating Data and Managing Spaces in SAP Datasphere


Managing and Monitoring Data Integration PUBLIC 209
8.7.3 Creating Partitions for Your Persisted Views
From the Views- Details screen, you can create partitions on columns from your dataset and break your data
down into smaller and more manageable parts to enable persistency on views with large data volume.

Persisting views with larger data volumes without having out-of-memory errors or high memory peaks can
sometimes be a challenge for Data modelers. To avoid such situations, you can create partitions when
persisting a view and thus enable partitioned data transfer of large datasets for supported connection types.

 Note

For this action, Data Builder (Read) privilege is required which is not included in the DW Integrator role. To
perform this action, ask your tenant administrator to assign your user to a scoped role that is based either
on the DW Space Administrator role or on a custom role that includes the following required privileges: Data
Warehouse Data Integration (Read, Update, Execute) and Data Builder (Read).

 Caution

Some adapters used by the connected sources are not able to push down the filters for the partitioning.
Thus, performance can be impacted by the number of partitions created.

You can create partitions from the Views – Details screen:

1. Click  for the view in virtual access mode you would like to persist and for which you need to create
partitions for.
2. Go to the Partitions tab and click Define Partitions.

Selecting a Column

Select a column from the list, which will then become the partitioning column.

The columns available for selection are a subset of all view output columns. Some columns can be excluded
because:

• Supported data types include:


• String
• Integer, Integer 64, Decimal, hana.SMALLINT, hana.TINYINT
• Date, DateTime, Time, Timestamp
• Binary
For more information, see Partitioning Limits in the SAP HANA Cloud, SAP HANA Database documentation.
• You can't select a column containing analytical measures as partitioning-columns, only columns containing
dimensions can be used as partitioning columns.
• If you select a non-key column to define your partitions, you must ensure that the column value of a record
in the source data is not changed while the view is persisted. Otherwise, data loading may fail because a
record with the same key is persisted in two different partitions.

 Note

Partitioning is not recommended on columns that stem from remote tables using ABAP adapter with
ODP, as related filter conditions are not pushed to the source. Instead, use remote table partitioning. The

Integrating Data and Managing Spaces in SAP Datasphere


210 PUBLIC Managing and Monitoring Data Integration
changes will be reflected only after a new data replication (snapshot or scheduled), or a real-time data
replication. For more details, see Partitioning Remote Table Data Loads [page 184]. For connections using
OData adapter, Camel JDBC adapter, or Cloud Data Integration adapter, partitioning may not have the
desired effect as the related adapters have only limited capabilities and do may not push selections on
partitioning intervals to the source system.

You can either select on a column that suits your business case, or switch on the option Rate Column
Suitability:

It runs a technical analysis of your view definition to calculate the level of suitability for partitioning for each
column. For example, calculated columns or columns with moving values (such as last changed date) are not
suitable as it might be difficult to do a reliable partitioning on such data.

In a first step, general findings are displayed:

In a 2nd step, the columns are sorted by their partitioning suitability, and additional information is displayed:

• Rating: It shows the suitability column rating. Possible values can be good, intermediate or bad.
• Details: It shows additional details about important factors that influence the suitability for each column to
help you with your choice. It can be one message or multiple messages.

Integrating Data and Managing Spaces in SAP Datasphere


Managing and Monitoring Data Integration PUBLIC 211
Defining Intervals for Your Partitions

Create the number of desired partitions by clicking  (Add Partitions)

Enter an interval for each of them.

 Caution

Intervals must not overlap

Check the Locked option if you don't want to update a partition in the next run.

 Caution

If locked partitions contain data filtered by authorizations provided by data access controls, you need to
unlock partitions and load a new snaphot of data, whenever these authorizations are changed.

An OTHERS partition is set by default. It contains data that is not covered by your defined intervals . For
example, if rows have been inserted or modified since your last data persistence and they now don’t match any
of the defined intervals. This OTHERS partitions can't be locked.

 Note

• Partitioning requires an in-depth knowledge of the values that are used or are valid for the chosen
partitioning column.
• Negative values are not valid for partition bounds of numeric columns and will be converted to 0 at
runtime.
• The low bound must be lower than the high bound. The order is checked according to the type of the
partition column. Example: for string-like columns, '90' <= DOC_ID < '100' is not valid, whereas '090'
<= DOC_ID < '100' is valid.
• Values entered in the bounds for ABAP ODP remote tables must match the ABAP internal format: for
example, enter "20210714" if the date is "July 14, 2021".

Integrating Data and Managing Spaces in SAP Datasphere


212 PUBLIC Managing and Monitoring Data Integration
 Caution

Once the partitions are created and the view is persisted, don't change data type of the column you have
used to create the partition: if the data type is changed, it deletes the partitioning data.

Locking Partitions to Avoid Unnecessary Data Loads

When you think the data will no longer change, or no change is expected for a while, you can check this option
to avoid unnecessary data load.

In the example below, I defined 2 partitions as Locked (partitions 2 and 3):

When I load a new snapshot, only partitions 1, 4 and OTHERS are refreshed:

 Note

If you change the partition definition, you can lock unchanged partitions to avoid unnecessary data loads.
In the case of adding partitions, if you lock the unchanged partitions, only the new partitions will be loaded
(Locking new partitions is ineffective, as they need to be updated once after they are introduced. They will
only be locked in the subsequent runs.)

A full data load will happen for the following cases:

• You have no locked partitions.


• You have changed the partitioning column. A full snapshot will happen at next data load.

Integrating Data and Managing Spaces in SAP Datasphere


Managing and Monitoring Data Integration PUBLIC 213
Refreshing Single or Multiple Partitions

In some cases, you might want to refresh data for a single or for multiple partitions only, independently if they
are locked or not. To do so, select the desired partitions and click Load Selected Partitions.

 Example

In the example below, I have selected 2 partitions. Partition 1 is not locked and partition 2 is locked:

Only the selected partitions are updated:

 Note

If you change the partition definition, then the next load will be a full load. No manual load will be possible,
and the Load Selected Partitions button will be disabled.

8.7.4 Persisted Views and Memory Consumption

You want to persist a complex view and consider how it affects the memory consumption.

When you persist a view, all the data that is provided by the view is stored in SAP Datasphere. When the view
is complex and consume a lot of data, some intermediate results have also to be stored in the memory so
that they can be used as source for further calculations and subsequent views. The amount of memory that is
needed for the runtime of the overall scenario can exceed the memory resources, and thus an out of memory
exception is raised or the memory consumption is very high.

Integrating Data and Managing Spaces in SAP Datasphere


214 PUBLIC Managing and Monitoring Data Integration
Reducing Memory Consumption while Persisting Complex Views

If your view is very complex, using other views or remote tables, you might get an out of memory or increase
the memory consumption in SAP HANA Cloud while persisting it.

You might need to think about the following steps before persisting the view:

• Persist intermediate views: Identify views that combine a lot of data and are perhaps also used in
other scenarios. These views might read data from different sources and combine them. The memory
consumption could already be quite high, and the result of the view may have to be stored as intermediate
result in the memory. If this view is persisted, then the top views that do also heavy calculations can be
optimized: They reach out directly to the same SAP HANA instance and no additional memory for the
pre-calculations is needed. It could also be a good idea to split the large view into two or several smaller
views, and to persist these smaller views first. In most cases, the data persistence on a deeper level in
the view graph is more effective than just persisting the top view, because the data retrieval and the
calculations can be optimized if internal SAP HANA tables can be used. You might need to consider the
View Analyzer to check the memory consumptions and to identify views that can be optimized. For more
information, see Exploring Views with View Analyzer [page 221].
• Partition your data: While partitioning your data, not all the data is retrieved and stored at once in the
persisted views: The data is divided into several batches (partitions). Each partition is retrieved and
stored on its own, and thus consume less memory. For more information, see Creating Partitions for Your
Persisted Views [page 210].
• Handling of remote tables: The memory consumption can increase with usage of remote tables, especially
when filters used in the view definition can't be pushed down to remote tables. Make sure that the statistics
for the remote tables are up to date. If the remote tables are part of joins or aggregations are existing, then
statistics could optimize the data retrieval. The statistics can be run from the Remote Queries Remote
Tables Statistics . For more information, see Creating Statistics for Your Remote Tables [page 251].
You can check the SQL statement that is used to retrieve the data for a remote source in Remote Queries.
For more information, see Monitoring Remote Queries [page 249]. If the needed filter condition is not
part of the SELECT statement, then the filter is applied on all the data. This leads to a higher memory
consumption and a performance decrease. It might be a solution to replicate the data for some remote
tables where the data does not change regularly, for example master data tables. You can also try to persist
views that just do basic joins/unions on some remote tables, and are used in more than one view. If these
remote tables and views are persisted, then the data retrieval for other scenarios is highly improved in
terms of memory consumption and performance, because the top views can use tables in the same SAP
HANA instance.
• You can change the run mode of your data persistence and choose between performance and memory
consumptions. For more information, see Data Persistence and Run Mode [page 204].

Monitoring Memory Consumption While Persisting Views

Persisting views can consume a lot of memory and it might be useful to keep an eye on memory figures that are
stored with task logs.

You can monitor memory consumption in the task logs if you switch on the option Expensive Statement Tracing
(under  (Configuration) → Monitoring).

Integrating Data and Managing Spaces in SAP Datasphere


Managing and Monitoring Data Integration PUBLIC 215
 Note

• The memory consumption is measured at runtime only.


• The memory and disk size consumed by the data persistence table are not available on partition-level,
but only for the entire table.

You can also use SAP HANA Views:

• M_EXPENSIVE_STATEMENT: This view provides all statements, for which memory consumption, CPU
time or duration exceed the corresponding specified thresholds. For more information, see Configure
Monitoring
• Memory consumed during insert statements can be tracked using the view: M_ACTIVE_PROCEDURES .

For more information on how to use system view, see Working with SAP HANA Monitoring Views.

8.7.5 Data Persistence, Detailed Logs and Remote


Statements

Access the detailed log of a data persistence task to get statistics information and run details, and navigate to
the remote statements for more information.

When your views are based on remote tables in Remote access, the data is read from a remote source system,
and it can take times to persist the data.

In the details information of a data persistence run, you can view statistics information in the logs: number of
remote statements, volume of data fetched from the remote source system:

When you click on View Details you can have detailed information and you can navigate to the Remote Queries
monitor clicking on the space name.

 Note

You must have the relevant privileges to access the Remote Queries monitor (Data Warehouse Data
Integration privilege).

Integrating Data and Managing Spaces in SAP Datasphere


216 PUBLIC Managing and Monitoring Data Integration
• If your view consumes data from several spaces, the View Details provide you with information on the
different spaces the data is coming from.

 Note

If the data is coming from a space for which you don't have authorization to access, it will be indicated
as "unauthorized spaces".

Clicking on the space name will bring you to the Remote Queries monitor.

Integrating Data and Managing Spaces in SAP Datasphere


Managing and Monitoring Data Integration PUBLIC 217
 Note

Even if you are a member of the space, you must have the relevant privileges to access the Remote
Queries monitor: Data Warehouse Data Integration privilege.

• If your view contains partitions, the log also shows how many partitions and how many remote statements
are involved during data persistence:

You can analyze the remote statements per partitions because each partition has its own statement ID.
Click on View Details on the partition row level of the log for more information or to navigate to the Remote
Queries monitor if you have the relevant permissions.

8.7.6 Understanding the Data Persistence Metrics

Check the data persistence metrics to decide on design change to prevent performance issues

 Note

The data persistence task has run successfully.

Administrators or data integrators want to analyze metrics information for a persisted view. The information
can be used to manually schedule data persistence tasks so that peaks memory can be avoided, or quicky
identify overload memory situations and decide on design change to prevent such situations, for example.

When your data persistence task has completed successfully, in the run details of the data persistence task,
you can see additional tabs: Metrics and Partitions Metrics (if you have defined partitions to persist your view
data).

Integrating Data and Managing Spaces in SAP Datasphere


218 PUBLIC Managing and Monitoring Data Integration
The Metrics Tab

The Metrics tab displays information for the persisted view:

Metrics Name Description

Number of Persisted Records Displays the number of records persisted for the view.

Number of Records Read From Remote Sources Displays the number of records read from the remote
source.

Remote Sources Processing Time Displays the time (in seconds) spent reading data from the
remote sources.

SAP HANA Peak Memory Displays the maximum amount of memory (in MiB) the task
to persist view data has used during the runtime in SAP
HANA.

 Note
You can see this information if the option Enable
Expensive Statement Tracing is enabled, and if the task
exceeds the thresholds specified in  (Configuration)
→ Monitoring. Otherwise, no number is displayed.

For more information, see Monitoring SAP Datasphere.

SAP HANA Total CPU Time Displays the amount of CPU time (in ms) the task to persist
view data has used in SAP HANA.

 Note
You can see this information if the option Enable
Expensive Statement Tracing is enabled, and if the task
exceeds the thresholds specified in  (Configuration)
→ Monitoring. Otherwise, no number is displayed.

For more information, see Monitoring SAP Datasphere.

Used Storage Displays how much size (in MiB) the view is using in your
memory.

Number of Partitions Displays how many partitions have been defined for this
view.

Integrating Data and Managing Spaces in SAP Datasphere


Managing and Monitoring Data Integration PUBLIC 219
Metrics Name Description

For more information, see Creating Partitions for Your


Persisted Views [page 210].

Number of Locked Partitions Displays how many partitions have been locked for this
persisted view.

 Note
Locked partitions are not updated and metrics related
to these partitions will take the value N/A in the
Partition Metrics tab.

For more information, see Creating Partitions for Your


Persisted Views [page 210].

Run Mode Displays the run mode used to persist your view data:
performance-optimized or memory-optimized.

For more information, see Data Persistence and Run Mode


[page 204]

Data Access Controls Displays if view data is protected by data access controls.

For more information, see Persisted Views and Data Access


Control [page 205]

Elastic Compute Node Name Displays the technical name of the elastic compute node
used to persist the view. If no elastic compute node is used
the value is empty.

Input Parameter If your view contains an input parameter, it displays the


name of the parameter used to persist the view data.

The Partition Metrics Tab

The Partition Metrics tab displays the metrics for each defined partitions of the persisted view:

Partition Metrics Name Description

Name Displays the name of the partition.

Integrating Data and Managing Spaces in SAP Datasphere


220 PUBLIC Managing and Monitoring Data Integration
Partition Metrics Name Description

Partition Range Display the partition range for this partition as defined
during creation.

 Note
OTHERS partitions can contain NULL value or values
not included in the defined ranges.

For more information, see Creating Partitions for Your


Persisted Views [page 210].

Number of records Displays the number of records persisted with the partition.

SAP HANA Peak Memory Displays the maximum amount of memory (in MiB) the task
to persist view data for this partition range has used during
the runtime in SAP HANA.

 Note
You can see this information if the option Enable
Expensive Statement Tracing is enabled, and if the task
exceeds the thresholds specified in  (Configuration)
→ Monitoring. Otherwise, no number is displayed.

For more information, see Monitoring SAP Datasphere.

Data Access Controls Displays if view data is protected by data access controls.

For more information, see Persisted Views and Data Access


Control [page 205]

Elastic Compute Node Name Displays the technical name of the elastic compute node
used to persist the view. If no elastic compute node is used
the value is empty.

Input Parameter If your view contains an input parameter, it displays the


name of the parameter used to persist the view data.

8.7.7 Exploring Views with View Analyzer

Use the View Analyzer to explore graphical or SQL views and the entities they consume.

The View Analyzer provides you with statistics and useful information on each entity that composes your view,
no matter the complexity of the view. It analyzes each view definition, the consumed views, the local tables
and the data sources used by remote tables that compose your data model. You can then use this information
to optimize your data model and decide, for example, which view to persist when performance or technical
problems such as out-of-memory errors occur.

Integrating Data and Managing Spaces in SAP Datasphere


Managing and Monitoring Data Integration PUBLIC 221
8.7.7.1 Getting Started with View Analyzer

Use the View Analyzer to explore graphical or SQL views and the entities they consume.

The View Analyzer provides you with statistics and useful information on each entity that composes your view,
no matter the complexity of the view. It analyzes each view definition, the consumed views, the local tables
and the data sources used by remote tables that compose your data model. You can then use this information
to optimize your data model and decide, for example, which view to persist when performance or technical
problems such as out-of-memory errors occur.

To use View Analyzer,

1. Go to Data Integration Monitor Views . Select the view you need to analyze and navigate to the details
screen of this view. You need to select a space if you are assigned to several spaces.
2. Select Data Persistence Start View Analyzer .

 Note

For this action, Data Builder (Read) privilege is required which is not included in the DW Integrator role.
To perform this action, ask your tenant administrator to assign your user to a scoped role that is based
either on the DW Space Administrator role or on a custom role that includes the following required
privileges: Data Warehouse Data Integration (Read, Update, Execute) and Data Builder (Read). For more
information see, Create a Scoped Role to Assign Privileges to Users in Spaces.

You can choose between 4 options:

Integrating Data and Managing Spaces in SAP Datasphere


222 PUBLIC Managing and Monitoring Data Integration
1. Run the View Analyzer without memory consumption: The View Analyzer will not execute the data
persistence simulation. For views already persisted, the memory consumption of data persistence task
will be shown. The analyzer will analyze the entities that compose your data model and will provide you
with the information it has collected.

 Note

If your data persistence contains partitions, the number of partitions is displayed in the results, but
the data persistence simulation does not consider it.

2. Generate Explain Plan: In the View Analyzer click the option Generate Explain Plan to get a quick
overview of your model and analyze the performance of a view.
It is a lightweight tool that can be generated in case of data persistence simulation failure or out-of-
memory errors of SQL analyzer plan file fails.
Once the explain plan is generated:

Integrating Data and Managing Spaces in SAP Datasphere


Managing and Monitoring Data Integration PUBLIC 223
• You can directly display the details with View Details. You can expand or collapse rows to view
specific details or click the settings button to choose a specific column to view.
• You can also download the details directly as a CSV file. If there are more than 1000 rows, you are
required to download the results.

 Note

Views with input parameters with default value are supported.

You can start the View Analyzer from the Data Builder, but to view the results, you need to go to the
details page.

 Note

You can generate the Explain Plan with the same permission required to run View Analyzer without
memory consumption.

To download the Explain Plan, you must have either the scoped role DW Administrator DW
Administrator or a custom role that includes Data Warehouse Runtime (Read) privilege.

3. Run the View Analyzer with memory consumption: In addition to exploring the entities that compose
your data model, the analyzer will execute the data persistence simulation for non persisted views. It
will report the memory consumption that would be used to get an entity persisted or will provide you
with the actual memory consumption used when you have persisted your data. The simulations are
executed sequentially as they may add some workloads to the system. To reduce these workloads,
you can define a limit to the memory consumption that can be used to persist the view thanks to the
threshold. If one of the analyzed views exceeds the limit, then no further data persistence simulations
are run.

 Note

• To start the View Analyzer with memory consumption, you must have the space Read privilege.
If you are missing this privilege, the View Analyzer will run without memory consumption by
default.
• If your data persistence contains partitions, the number of partitions is displayed in the results,
but the data persistence simulation does not consider it.

If your data persistence contains partitions, the number of partitions is displayed in the results, but the
data persistence simulation does not consider it. You can run View Analyzer with memory consumption
for views with one input parameter with default value.
The maximum memory consumption is based on the statement memory limit of the space workload
configuration. For more information on statement limits, see Set Priorities and Statement Limits for
Spaces.

 Note

To use the View Analyzer with memory consumption during data persistence simulation, you must
enable Expensive Statement Tracing in  (Configuration). For more information, see Configure
Monitoring.

4. Generate SQL Analyzer Plan File: Before using this option, you must consider the following
requirements:

Integrating Data and Managing Spaces in SAP Datasphere


224 PUBLIC Managing and Monitoring Data Integration
 Note

• To download this file, you must have the DW Administrator role or a custom role that includes
the Data Warehouse Runtime (Read) privilege. For more information, see Privileges and
Permissions.
• To open this file, you must install a compatible SQL plan visualization tool, such as SQL
Analyzer Tool for SAP HANA.
• This option requires additional system resources.

With this option, the data persistence simulation is run for the main view only, and the analyzer
creates a file containing detailed information about your data model that you can download for further
analysis.

 Note

• If your data persistence contains partitions, the number of partitions is displayed in the results,
but the data persistence simulation does not consider it.
• If you want to analyze one of the underlying views, select the relevant view from the
dependency graph or the entity list, and launch the View Analyzer with SQL analyzer plan
file option.
• You can select only one view for analysis.

The Plan File is stored along with task logs and has a retention period that is the same as the task logs:

 Tip

While working with complex views, see Persisted Views and Memory Consumption [page 214].

Cross-space analysis can be analyzed in View Analyzer, but you can only view entities that you are
authorized to view. For more information see: Authorization and Permissions [page 226]

 Note

For views with one input parameter with default value, you can execute views with memory
consumption. For views with input parameters with default values, you can generate Explain Plan
and SQL analyzer plan file.

Integrating Data and Managing Spaces in SAP Datasphere


Managing and Monitoring Data Integration PUBLIC 225
8.7.7.2 Authorization and Permissions

Working in View Analyzer requires different privileges and permissions depending on the tasks you want to
perform. You will not be able to view entities that you are not authorized to view.

Unauthorized Notification

When a user does not have permission to access a space or spaces, a row titled unauthorized will be displayed.

You can only view entities that you are authorized to view. Entities from spaces that you are not allowed to view
will be labeled as unauthorized. Contact your administrator to get authorization to view the hidden entities.

 Note

To start the View Analyzer you will need the DW Integrator scoped role and the Data Builder (Read) privilege
which can be found in the DW Modeler scoped role. You will be able to see all entities with either the global
role DW Administrator or a custom role that includes Data Warehouse Runtime (Read) privilege.

Ask your tenant administrator to assign you a scoped or custom role that fits your required task.

This can also be seen in the lineage graph.

Integrating Data and Managing Spaces in SAP Datasphere


226 PUBLIC Managing and Monitoring Data Integration
Permissions

The View Analyzer provides you with statistics and useful information on each entity that composes your view.
There are a variety of ways to learn more about your view and make decisions to optimize your data model
and performance. Due to the potential sensitivity of information gathered in View Analyzer, either the DW
Administrator role or a custom role is required to run the view or to read the results of generated files.

Ask your tenant administrator to assign you a scoped or custom role that fits your required task.

Action Permission

Run View Analyzer without mem- To run View Analyzer without memory consumption, you must have a role that is either
ory consumption based on the scoped role DW Administrator or on a custom role that includes the Data
Warehouse Data Integration (Read, Update, Execute) and Data Builder (Read) privileges.

Integrating Data and Managing Spaces in SAP Datasphere


Managing and Monitoring Data Integration PUBLIC 227
Action Permission

Generate Explain Plan You can generate the Explain Plan with the same permission required to run View
Analyzer without memory consumption.

To generate Explain Plan, you must have either the scoped role DW Administrator or an
additional custom role that includes Data Warehouse Runtime (Read) privilege.

Run View Analyzer with memory To run the View Analyzer with memory consumption you must have a role that is ei-
consumption ther based on the scoped role DW Administrator or a custom role that includes Data
Warehouse Data Integration (Read, Update, Execute) and Data Builder (Read) and the
additional Space (Read) privilege.

Generate SQL Analyzer Plan File You can generate a SQL analyzer plan file with the same permission required to run View
Analyzer without memory consumption.

To download the SQL analyzer plan file, you must have either the scoped role DW
Administrator or an additional custom role that includes Data Warehouse Runtime
(Read) privilege.

 Note

For more information see: Privileges and Permissions

8.7.7.3 Analyze Results

1. Analyze your results

Integrating Data and Managing Spaces in SAP Datasphere


228 PUBLIC Managing and Monitoring Data Integration
To understand how you can interpret the analyzer's findings, let's take an example with
the following View_001_001. It consumes 4 views (in red) and 4 remote tables (in purple):

Now, let's start the View Analyzer with memory consumption. On the total of 8 GiB that I have allowed in
my statement memory limit, let's decide to limit the memory consumption to 4 GiB. This means that if a
data persistence task exceeds this limit, the next data persistence simulations will be canceled:

Once the View Analyzer has completed the task, 2 tabs are available in the details screen:

 Note

The Analyzed Views tab was previously called the Entities tab.

• Task Log: It contains all information related to the View Analyzer task itself. You can see that my 5
views have been analyzed, and an overview of some key indicators is displayed: memory consumption,
number of records, etc.

• Analyzed Views: It contains the details of the View Analyzer findings per analyzed entities. Information
is displayed in a table format:

Integrating Data and Managing Spaces in SAP Datasphere


Managing and Monitoring Data Integration PUBLIC 229
Column Description

Entity Name Name of the analyzed entity. It can be a local or a


remote table, a view, a shared entity, an external entity,
an intelligent lookup and or a data access control.

Space Name of the space the entity resides in.

Persisted Indicates if the view is already persisted, or the remote


table replicated.

Partitions Indicates if partitions have been defined for the


analyzed views.

 Note
If partitions have been defined for remote tables,
it's not displayed. Only partitions defined for views
are indicated by the View Analyzer.

Peak Memory (MiB) Indicates the memory used during the simulation, or in
case of the view already persisted, the memory used to
persist the view.

Row Indicates the number of rows.

Duration Indicates the duration in second to persist the view


(simulation or actual data persistence).

Adapter Indicates if an adapter is used to access the remote


table source.

If you click on an entity row or click the >, you'll access the details for the selected entity. For example,
if I click on the table SCARR:

Integrating Data and Managing Spaces in SAP Datasphere


230 PUBLIC Managing and Monitoring Data Integration
In the message's details, you can see that the remote table is using an adapter that has limited
capacities. This is why the entity was marked with a warning in the results.
From the Analyzed Views tab, you can do many actions to explore your results:

Action Description

 (Sort) Set sort order and sort columns:

 (Filter) Filter your results:

 (Inspect) Display the whole list of results that you can also export
in CSV format:

Integrating Data and Managing Spaces in SAP Datasphere


Managing and Monitoring Data Integration PUBLIC 231
Action Description

Start Analyzer You can select one or more views and start a new view
analyzer.

 Note
The analyzer will always start from the main
view. Selecting the views here in the entity list,
or in the lineage graph has an effect only on
the data persistence simulation. In case memory
consumption is chosen, the data persistence
simulation will be executed for the selected views,
regardless of their data persistence status.

 (Lineage View) Display the Lineage graph:

Integrating Data and Managing Spaces in SAP Datasphere


232 PUBLIC Managing and Monitoring Data Integration
Action Description

If you click on one entity from the graph, you can


see the view analyzer results and the indicators. For
example, I click on SpfiJoinScarr view:

 Note
It is also possible to select multiple views here and
start the View Analyzer with memory consumption.

Integrating Data and Managing Spaces in SAP Datasphere


Managing and Monitoring Data Integration PUBLIC 233
8.7.7.4 Warning Notifications

The entities results list show the name of the entity and the space that it resides in. The orange bar on the left
side of the table signifies any entities that have warnings. The warnings icon in the task bar opens a popover
that will provide more information about entities with warnings.

Expand a single text with the navigation icon ">"

Integrating Data and Managing Spaces in SAP Datasphere


234 PUBLIC Managing and Monitoring Data Integration
Click on this and it will take you to the relevant entity in the list.

Integrating Data and Managing Spaces in SAP Datasphere


Managing and Monitoring Data Integration PUBLIC 235
8.7.7.5 Data Persistence Guidance

Data persistence guidance provides recommendation on what views to persist in View Analyzer. The entire
lineage graph is reviewed and relative metrics are analyzed to provide guidance for the optimal views to persist
to improve performance, run-time, and memory consumption.

About Data Persistence Guidance

The View Analyzer provides statistics and useful information on each entity that composes your view regardless
of the the complexity of the model. A graphical view can look simple, but the underlying lineage graph can be
complex and built with many stacked views. This can lead to technical challenges, performance problems, or
time-out errors.

The best candidate to be persisted can be the main view or a view of the lineage graph.

 Note

Data persistence guidance is most useful with models that have complex lineage graphs.

The Data Persistence Candidate Score column gives you recommendations by identifying views that can help
improve your model if they are persisted.

The information tab provides more details of the View Analyzer Results and the metrics of the view.

Integrating Data and Managing Spaces in SAP Datasphere


236 PUBLIC Managing and Monitoring Data Integration
 Tip

Persist only one view at a time. Check the performance of the main view after the persistence run
completed. One possibility to check the performance is to simulate the data persistence in the View
Analyzer with the option, 'With memory consumption'. Continue the process in an iterative way until the
optimal performance is reached. The goal is to persist the minimal number of views required to optimize
performance.

 Note

To get the best results from data persistence guidance, it is recommended that the user have permission
to see all views in the model from all spaces. If a user is unauthorized to see some views, the view with
the highest rating value may remain hidden (a warning will be displayed). For more information about
authorization and permission in View Analyzer see Authorization and Permissions [page 226].

Data Persistence Candidate Score

The Data Persistence Candidate Score provides a recommendation for the best views to persist to increase
performance.

The Data Persistence Candidate Score is not a specific inherent value but a ranked score based on weighted
values from a list of metrics that most affect the performance of a view or a model. The top scores show the
best options to persist with the top five noted. The first is the 'Best Option'. The next are the 'Best Alternative'.
Persist one view at a time, check the performance, and iterate the process if needed.

Integrating Data and Managing Spaces in SAP Datasphere


Managing and Monitoring Data Integration PUBLIC 237
The primary metrics are: the number of sources, the number of targets, the overall number of sources, the
number of federated remote tables, the number of federated remote tables with limited adapter capabilities,
number of local tables (file), and if the view is a SQL script view.

 Note

The values of the score are relative weighted values determined by the primary metrics in the lineage
graph.

 Note

Branches of the views that are already persisted are not considered in the metric evaluation of the Data
Persistence Candidate Score. It is best to persist one view at a time and then check the performance.

A model can have a Data Persistence Candidate Score with values ranging from 0 - 10. Any change in the model,
for example, persisting one view, will change the values due to the change in the metrics used to calculate the
data persistence score.

Views using Shared Entities from Other Spaces

Analysis for data persistence guidance will be run for all entities including shared entities from other spaces.

The best candidate score will be computed regardless of the entities the views belong to. Results of the analysis
will be able to be viewed by those that have the authorization. If the best candidate is from a space the
user does not have the authorization to see the result from, a message will be shown. See Authorization and
Permissions [page 226] for more information.

Further Recommendations

Persisting data is one option to optimize a model or improve performance. Utilizing Transformation Flows is
also a good option to consider instead of persisting data depending on the scenario.

For more information see Creating a Transformation Flow.

Consider replacing remote tables with local tables by using Replication Flow for replicating the data.

For more information see Creating a Replication Flow.

Integrating Data and Managing Spaces in SAP Datasphere


238 PUBLIC Managing and Monitoring Data Integration
8.8 Monitoring Flows

In the Flows monitor, you can find all the deployed flows per space.

Prerequisites

• To use the Flows monitor, you need the DW Integrator role along with the DW Viewer or the DW
Modeler role.
• To run the flows, you need the DWC_DATAINTEGRATION privilege and the Update permission. For more
information, see Managing Roles and Privileges.
• To schedule the flows, you need the DWC_DATAINTEGRATION privilege and the Execute permission. For
more information, see Managing Roles and Privileges.

Context

In the Flows monitor, you can find all the deployed flows (data flows, replication flows, and transformation
flows) per space. Here, you can run the deployed flows, and you can view and monitor the execution details of
the flows.

 Note

For optimal performance, it is recommended that you consider staggering the scheduled run time of tasks
such as data flows and task chains that may contain these tasks. There is a limit on how many tasks can be
started at the same time. If you come close to this limit, scheduled task runs may be delayed and, if you go
beyond the limit, some scheduled task runs might even be skipped.

For more information and points to consider when using replication flows, see also SAP Note 3297105 .

Procedure

1. In the SAP Datasphere, navigate to Data Integration Monitor Flows .

You can find all the deployed flows per space. You can run the deployed flows under Run, and you can
view and monitor the execution details of the flows. For some flows, you can create, edit, delete, pause
or resume a schedule, or assign a schedule to your name under Schedule, and you can select several
data flows to group the scheduling actions. For more information about scheduling, see Schedule a Data
Integration Task (Simple Schedule) [page 256].

The list of deployed flows shows the following information:

Integrating Data and Managing Spaces in SAP Datasphere


Managing and Monitoring Data Integration PUBLIC 239
Column Information

Technical Name or Business Name Name of the flow. Technical or Business Name is dis-
played, depending on how you have configured your UI
settings in Your Profile Settings

Status Indicates the status of the last run of the flow:


• Running: The flow is currently running.
• Active: [Replication Flow only] A replication flow that
contains objects with load type Initial and Delta does
not have an end date. Once started, it remains in
status Active until it is stopped or paused or an issue
occurs.
• Completed: The flow is completed successfully.
• Failed: Something goes wrong during the flow run and
it could not be completed. Go to the details screen of
your flow and check the logs to identify to issue.
• Stopped: The flow has been stopped by a user.

All these statuses can be combined with a sub-status.


For more information on sub-statuses, see Understanding
Statuses and Substatuses [page 165].

 Note
If the status of a flow is empty, it means that the
flow has not run yet. The status is updated every 10
minutes for replication flow (except when it is run via
a task chain: It updates every 2 minutes), and every
1 minute for a transformation flow or a data flow. To
get an immediate updated status, navigate to the run
details page.

For more information on specifics for replication flow runs,


see Working With Existing Replication Flow Runs [page
242].

Type Indicates the type of flow: data flow, transformation flow


or replication flow.

Frequency Indicates if a schedule is defined for a data flow or a trans-


formation flow. :
• Scheduled: A schedule task is defined for this flow. If
you click on Scheduled, you get detailed information
on the schedule. You can update the schedule options
at any time from Schedule Edit Schedule ,
or delete the schedule from Schedule Delete
Schedule
• Paused: A schedule task is defined for this flow
but is currently paused. To resume the task, select
Schedule Resume

Integrating Data and Managing Spaces in SAP Datasphere


240 PUBLIC Managing and Monitoring Data Integration
Column Information

Duration Provides information on the total execution time of the


last flow run.

 Note
The initialization time for executing a flow takes an
average of 20 seconds even with smaller data loads,
causing longer runtime for the flow. It is calculated
using the task start time and task end time. For rep-
lication flow, it does not depend on the initial load
duration.

Last Run Start Provides information on when the last run of the flow
started.

Last Run End Provides information on when the last run of the flow
ended.

Next Run If a schedule is set for a flow, you can see here by when the
next run is scheduled.

Schedule Owner Displays the name of the current schedule owners when a
schedule is defined for a flow.

You can personalize the columns you want to display clicking on  Select Columns

 Note

Parallel execution is not supported: You can't run two identical flows at the same time. You need to wait
until the first run is completed before you can launch another one. Otherwise your second run gets the
status Failed (Run Already in Progress).

2. To see the run details of a flow, click  (Details) for the relevant flow.

The run details page opens. It shows the runs of the flow in the left panel and their corresponding
messages in the right panel. Here, you can use the Stop Run (for Replication Flow), Cancel Run (for
Transformation Flow) or Set to failed (for Data Flow) buttons to stop a run that is currently running. The
buttons are available until the run ends.
3. Select a flow run in the left panel to view its details. The run details contain messages for all runs and
metrics for the current run.
• Messages:
For failed runs, click View Details besides the error message to see more information about the error.
• Metrics:
For successful runs, the metrics provide the record count for source and target tables used in the flow.

 Note

• It takes a while to load the metrics once the flow run is completed.
• If the flow run has been optimized for better performance, no record count information is
shown.
• For historical runs of a flow, metrics can’t be retrieved.

Integrating Data and Managing Spaces in SAP Datasphere


Managing and Monitoring Data Integration PUBLIC 241
 Restriction

For data flows, the logs messages and metrics are only displayed in English.

For more information, see Metrics for Transformation Flows [page 247].
4. If your flow failed, you can use the following tools from the details page of the flow run to analyze and solve
your issue:
• [Data Flow only] You can analyze the errors that have occurred during your data flow runs,
downloading the run details as a JSON file you can provide to the SAP Support for troubleshooting:
Click  Download Run Details.
• [Data Flow only] You can also perform a complete health analysis and generate detailed logs if you
choose Run Start Run with Checkup . Send this health analysis log to your support team to get
help in case of issues.
• [Transformation Flow only]: You can run your transformation flow with Support Settings. For more
information, see Explore Transformation Flows [page 248].

5. To view and modify a flow, click Open in Data Builder.

8.8.1 Working With Existing Replication Flow Runs

You can pause a replication flow run and resume it at a later point in time, or you can stop it completely.

 Tip

This text explains the options related to replication flow runs (in the context of monitoring). For information
about how to make changes to an existing replication flow in the Data Builder, see Modify a Replication
Flow.

This topic contains the following sections:

• Statuses for Replication Flow Runs [page 242]


• Scheduling a Replication Flow [page 243]
• Pausing and Resuming a Replication Flow Run [page 243]
• Monitoring Premium Outbound Volume [page 243]
• Stopping a Replication Flow Run [page 244]
• Space Deletion [page 244]
• Adding or Removing Replication Objects [page 244]
• Changing the Delta Interval for a Replication Flow [page 244]

Statuses for Replication Flow Runs

On the left of the monitoring screen, you find the aggregated status information for your replication flow. If the
status for one or more objects in a run differs from the overall status, this is indicated by an additional status
value in brackets.

Integrating Data and Managing Spaces in SAP Datasphere


242 PUBLIC Managing and Monitoring Data Integration
As long as the status is Running or Active, you cannot start another run for the same replication flow.

Scheduling a Replication Flow

You can create a schedule for your replication flow and include your replication flow in a task chain if all objects
in it have load type Initial Only. For more information, see Schedule a Data Integration Task (Simple Schedule)
[page 256] and Creating a Task Chain.

Pausing and Resuming a Replication Flow Run

You may want to pause a replication flow run, for example, while a system update is running.

 Caution

You must always stop or pause a running replication flow before a source system downtime.

When you do so, the flow is stopped in SAP Datasphere, but not in the source. Consequently, the system still
keeps track of changes in the source, but does not copy them to the target.

 Note

If you have many data changes in your source, and if you pause the replication flow for a longer time (such
as several hours), the logging tables can become quite big.

To pause a replication flow that has a schedule, you need to pause the schedule. (Pausing the replication flow
run itself is not possible in this case.)

When you resume the flow, the system replicates all source data changes that happened while the replication
flow was paused, which means that the first replication flow run after a pause can take considerably longer
than usual.

Monitoring Premium Outbound Volume

When you open the Flows monitor for a replcation flow with premium outbound integration, a value called Used
Premium Outbound Data Volume is shown in the overview panel on the left. This is the total volume for all
replication flow runs with premium outbound integration (replication to a non-SAP target) in this tenant during
the last 360 days.

 Note

• The value may change significantly from one day to the other, for example when you run an initial load
for a replication flow, or when an initial load drops out of the statistics because it happened more than
360 days ago.
• The retention period has been increased from 90 days to 360 days starting as of July 11, 2024, so
depending on when you access this feature you may see the data for less than 360 days.

Integrating Data and Managing Spaces in SAP Datasphere


Managing and Monitoring Data Integration PUBLIC 243
The value is updated once per hour.

Stopping a Replication Flow Run

If you do so, the flow run is stopped permanently in SAP Datasphere as well as in the source. You can still run
it again, but it will then start from scratch (rather than from where it left off when you stopped it). If you stop a
replication flow run because you don't need it anymore, you should also delete it so that it does not clutter your
system. For more information, see Delete a Replication Flow.

Space Deletion

If you have replication flows in a space that is about to be deleted, make sure to stop your replication flows
before space deletion starts. This helps to avoid issues during space deletion and makes it possible for you to
start the replication flows again if the space gets restored at a later point in time.

When a space gets permanently deleted, all replication flows in it are deleted as well.

For more information about space deletion, see Delete Your Space [page 325].

Adding or Removing Replication Objects

This can be done in the Data Builder, see Modify a Replication Flow.

 Note

Adding or removing objects without stopping the replication flow first is only possible for replication flows in
status Active.

Changing the Delta Interval for a Replication Flow

This can be done in the Data Builder, see Modify a Replication Flow.

Integrating Data and Managing Spaces in SAP Datasphere


244 PUBLIC Managing and Monitoring Data Integration
8.8.2 Cancel a Transformation Flow Run

You want to cancel a transformation flow that is running.

Context

You can use the Cancel Run button to stop a transformation flow run that is currently running. This button is
available until the run ends.

Procedure

1. Navigate to Data Integration Monitor Flows .


2. Navigate to the run details of your transformation flow by clicking  (Details) for the relevant flow.

The Run Details page opens. It shows the runs of the flow in the left panel and their corresponding
messages in the right panel. Under Runs, you can view the status of the transformation flow runs (the Run
activity) and also the status of any attempts to cancel a transformation flow run (the Cancel activity).
3. To stop a run that is currently running, select the relevant run and click the Cancel Run button.

If the run has been canceled successfully, the status of the transformation flow run will have the value
Failed (Canceled).

Canceling a transformation flow involves rolling back database transactions. Therefore, it might take some
time until the transformation flow is canceled and the status changes to Failed (Canceled).

 Caution

For a transformation flow in a file space, when you cancel a run, it stops where it is, and does not roll
back what has already been processed. If you restart the run later, it will restart from the beginning, not
where it has stopped.

8.8.3 Watermarks

When you run a transformation flow that loads delta changes to a target table, the system uses a watermark (a
timestamp) to track the data that has been transferred.

On the Delta Capture Settings tab of the Data Integration Monitor, you can view the watermark for a source
table. Note that if a transformation flow run does not load delta data to a target table, no source table
information is displayed.

Integrating Data and Managing Spaces in SAP Datasphere


Managing and Monitoring Data Integration PUBLIC 245
How the System Calculates the Watermark

Local Tables

When you run a transformation flow that loads delta changes to a target table, the system only transfers data
that has a change timestamp greater than or equal to the watermark. Once the transformation flow run has
completed successfully, the system sets the watermark to the start time of the transformation flow run. Note
that if open transactions exist in the source table at the time the transformation flow run started, then the
earliest start timestamp of these transactions is used as the watermark.

Remote Tables Located in SAP BW Bridge Spaces

When you run a transformation flow that loads delta changes to a target table, the system sets the watermark
to the timestamp of the first request that has a timestamp greater than the value of the watermark. If no
request exists that has a timestamp that is greater than the value of the watermark, the system sets the value
of the watermark to the maximum timestamp (12/31/9999 25:59:59).

If the transformation flow has not yet been run, or if you reset the watermark, then the value of the watermark
is null. If you then run the transformation flow, the system uses the minimum timestamp (0001-01-01
00:00:00) as the watermark.

The system only transfers data that has a change timestamp greater than or equal to the watermark. Once the
transformation flow run has completed successfully, the system sets the watermark to the timestamp of latest
request of the source table that the system could ensure was loaded to the target table.

Resetting the Watermark

If you reset a watermark, the system will transfer all data to the target table the next time the transformation
flow runs (using the load type Initial and Delta). This means that you do not need to redeploy the
transformation flow and use the load type Initial Only.

Resetting the watermark can make sense in the following situations:

• If a table that can capture delta changes is joined with a second table, and columns in the second table
have been updated, you can reset the watermark to ensure that these changes are reflected in records that
have already been transferred.
• If corrupt data is present in the target table, you can reset the watermark for the transformation flow to
ensure that the latest data from the source table is loaded to the target table.

You can reset the watermark for a transformation flow in the Data Integration Monitor. Open your
transformation flow and navigate to the Delta Capture Settings tab. To reset the watermark, choose the Reset
Watermark button.

Integrating Data and Managing Spaces in SAP Datasphere


246 PUBLIC Managing and Monitoring Data Integration
8.8.4 Change the Run Mode for a Transformation Flow

You can change the run mode for a transformation flow.

Procedure

In the Data Integration Monitor, open your transformation flow. On the Settings tab, you can change the run
mode for a transformation flow and choose between performance and memory consumption:
• Performance-Optimized (Recommended): This is the recommended mode to run the transformation flow.
SQL queries are processed directly in-memory. No intermediate result is stored on the disk to ensure a
quicker data access and data consumption. However, this mode consumes memory resources and if your
data model is very complex, it can result in an out of memory and the task failed.
• Memory-Optimized: The mode uses an enhancement of the SAP HANA Execution Engine, which stores
intermediate results of SQL statements on the disk. Memory consumption of SQL statements is then lower
because intermediate results are not stored in-memory but are saved on the disk. It results in slower
runtime, though. For example, you can use it when you get an out of memory error while running the task,
as alternative to manual partitioning.

 Note

You must ensure that you have sufficient disk space available. If not, it can result with an error of type
"Disk Full". You can monitor your disk space using the SAP HANA Cockpit. For more information, see
Disk Usage: Monitor Disk Volume

8.8.5 Metrics for Transformation Flows

View metrics for completed transformation flow runs.

Metrics provide the record count for source and target tables used in the flow. You can view the following
metrics for a completed transformation flow run:

• LOAD_TYPE
The load type for a transformation flow run. For more information about load types, see Creating a
Transformation Flow.
• TRUNCATE
Indicates whether the Delete All Before Loading option is used for the transformation flow run. For more
information, see Create or Add a Target Table to Your Transformation Flow.
• REMOTE_ACCESS
Shows "true" if a remote access has been used when executing the transformation flow. This is possible
only when the transformation flow consumes a BW Bridge object imported as remote table.
• NUMBER_OF_RECORDS
The number of records written to the target table.
• MEMORY_CONSUMPTION_MIB
The peak memory consumption (in mebibytes) for the SAP HANA database while running the
transformation flow.

Integrating Data and Managing Spaces in SAP Datasphere


Managing and Monitoring Data Integration PUBLIC 247
• RUNTIME
Runtime is used to run the transformation flow. It can be HANA (for a transformation flow in a space with
Storage Type DAP HANA Database (Dish and In-Memory)) or SPARK (for transformation flow in a space
with Storage Type SAP HANA Data Lake Files).
• LOAD_UNIT_TARGET
Target tables created in spaces with SAP HANA Cloud, SAP HANA database storage have the value PAGE if
the selected storage is “Disk”, or COLUMN if the selected storage is “In-Memory”.
• EXECUTION_MODE
The “Run Mode” is defined in the settings for a transformation flow run: 0 means “Performance-Optimized
(Recommended)” and 1 means “Memory-Optimized”. For more information about the "Run Mode", see
Change the Run Mode for a Transformation Flow [page 247].
• VIEW TRANSFORM STACK SIZE ACCOUNT
The number of view transform nodes inside the transformation flow.
• SPARK_RESOURCE_ID
ID of the job used to run the transformation flow on an Apache Spark runtime. This ID must be provided to
the SAP support to retrieve and analyze the jobs, and identify the cause of errors.

8.8.6 Explore Transformation Flows

Use Run with Settings to explore graphical or SQL views and the entities they consume in a transformation flow.

The Run with Settings option provides you with the ability to simulate a run, but also to download a plan file
containing statistics and useful information on each entity that composes your view, no matter the complexity
of the view. You can analyze each view definition, the consumed views, the local tables and the data sources
used by remote tables that compose your data model. You can then use this information to optimize your
data model and decide, for example, which view to persist when performance or technical problems such as
out-of-memory errors occur.

1. Go to Data Integration Monitor Flows . Select the view you need to analyze and navigate to the details
screen of this view. You need to select a space if you are assigned to several spaces.
2. Select Run Run with Settings . You have two options:
• Simulate Run: Simulating a flow allows you to test a transformation flow and see if you get the desired
outcome. Based on the result, you can decide to resolve errors or to optimize the flow to improve
performances. No changes are saved in the target table.

 Note

A simulated run will not take the truncate function into account.

• Generate SQL Analyzer Plan File: Before using this option, you must consider the following
requirements:

 Note

• This option is only available to file spaces.


• To download this file, you must have either the roles of:
• the scoped role DW Administrator with the privileges DWC_DATABUILDER (Read) and
DWC_DATAINTEGRATION (Read).

Integrating Data and Managing Spaces in SAP Datasphere


248 PUBLIC Managing and Monitoring Data Integration
• an additional custom role with the privileges of Data Warehouse Runtime (Read),
DWC_DATABUILDER (Read), and DWC_DATAINTEGRATION (Read).
For more information, see Privileges and Permissions.
• To open this file, you must install a compatible SQL plan visualization tool, such as SQL
Analyzer Tool for SAP HANA.
• This option requires additional system resources.

This file allows you to analyze your transformation flow to resolve errors and enhance its
performances. With this option, you create a plan file (also called .plv file) containing detailed
information about your data model that you can download for further analysis. This file provides more
details than the Simulate Run option.

 Note

You can select only one flow per download.

The plan file is stored along with task logs and has a retention period that is the same as the task logs.

8.9 Monitoring Remote Queries

In the Remote Queries monitor, you track the queries sent to your remote connected source systems for
your space. You can monitor the communication between the federation layer of SAP HANA Cloud and the
connected remote source systems, and analyze them.

SAP Datasphere is connected to remote systems via SAP HANA Cloud using data federation. SAP HANA Cloud
already provides a corresponding monitoring view of actual SAP HANA runtime data, including statistics and
status information related to the execution of Data Manipulation Language (DML) statements. The Remote
Queries monitor retrieves this information, so that you can keep an eye on how the communication between
SAP Datasphere and the connected remote source systems is running and better analyze possibilities for
performance bottlenecks while running queries in SAP Datasphere.

The Remote Queries monitor allows you to monitor the queries sent to your remote connected source systems
for your space. This information is helpful when analyzing executions of story (or on views or previews), since it
provides details of single read requests, like runtimes or result set size.

 Note

Data Flows executions are not contained in this monitoring view as Data Flows don't use SAP HANA
Federation.

 Example

You build a story on top on a view. In this story, you create charts and set filters on those charts. Let’s say
you remove one of those filters. A query is then triggered to the connected remote source system. A new
entry appears in the Remote Queries monitor and you can see the last statement sent by the story.

The Remote Queries monitor only shows the last 100 statements. It provides you with the following
information:

Integrating Data and Managing Spaces in SAP Datasphere


Managing and Monitoring Data Integration PUBLIC 249
Column Description

Connection (Business Name or Technical Displays the name of the connection.


Name)

Start Time Displays the statement start time.

End Time Displays the statement end time.

Remote Statement Duration Displays the duration in milliseconds of the total remote statement (open/
fetch/close).

Rows Displays the number of rows that have been processed by the statement.

Status Displays the status of the statement.

The status can be one of the following:

• Closed: The statement has been completed without error.


• Executing: The statement is currently running.
• Error: The statement can't be completed.

Remote Statement Displays the outgoing statement that is sent to the remote connected
source system.

As the statement is most often too long and appears as truncated, you
need to click on More to get the full statement information.

Statement ID Displays the ID of the statement.

This ID is provided by SAP HANA and can be used to do deep dive analysis
in the SAP HANA Cockpit. For more information, see SAP HANA Cockpit.

Executing SQL Statement Displays the statement that is currently running in SAP HANA Cloud, and
which includes a remote SQL statement.

You might have a look at this statement if something goes wrong to identify
the root cause. As the statement is most often too long and appears as
truncated, you need to click on More to get the full statement information.

Transaction Displays the remote source transaction.

Adapter Displays the name of the adapter

Connection Type Displays the type of the connection.

Data Provisioning Agent Displays the name of the data provisioning agent. For more information, see
Preparing Data Provisioning Agent Connectivity.

Statement Details Displays the statement details.

Remote Source Displays the name of the remote source.

You can personalize the columns you want to display clicking on  Select Columns

Integrating Data and Managing Spaces in SAP Datasphere


250 PUBLIC Managing and Monitoring Data Integration
8.9.1 Creating Statistics for Your Remote Tables

Create statistics for your remote tables to improve federated query execution.

SAP Datasphere is using SAP HANA smart data integration and SAP HANA smart data access to connect to
remote data sources. Objects from remote data sources are made available with so-called SAP HANA virtual
tables, which are created when deploying remote tables in SAP Datasphere.

To ensure the best possible performance when accessing your remote tables and provide efficient query
execution plans, you can create statistics for your remote tables. SAP Datasphere uses the SAP HANA SQL
Optimizer that relies on accurate data statistics for virtual tables in SAP HANA and help you take decisions. For
more info, see SAP HANA SQL Optimizer.

 Restriction

• Your remote table must be in "remote" data access only.


• You can’t create statistics if your remote table is connected via an SAP HANA smart data integration
Cloud Data Integration (CDI) or an SAP HANA smart data integration ABAP adapters to the following
sources:
• SAP HANA smart data integration Cloud Data Integration (CDI) adapter
• Cloud Data Integration (CDI)
• SAP Marketing Cloud
• SAP S/4HANA Cloud
• SAP HANA smart data integration ABAP adapter:
• SAP ABAP
• SAP BW
• SAP ECC
• SAP S/4HANA On-Premise
• If your remote table is based on Google BigQuery source, and if it contains one or more columns with
data type STRUCT, then you can only create statistics of type RECORD COUNT.

Statistics can be created from the Remote Queries monitor.

1. From the Data Integration Monitor ( ), select Remote Queries Remote Table Statistics .

 Note

With the different tabs, you can navigate between the remote tables in Data Access: Remote or Data
Access: Replicated. However, statistics can be created only on remote tables with Data Access: Remote.
However, you can still navigate in existing statistics from previous actions.

2. Select the remote table for which you need to create statistics for, and choose Create Statistics. Or select
the remote table, click  to navigate to the details screen, and click the button Create Statistics.

You can create three types of data statistics object:

Statistics type Description

RECORD COUNT Creates a data statistics object that helps the query
optimizer calculate the number of records (rows) in a table

Integrating Data and Managing Spaces in SAP Datasphere


Managing and Monitoring Data Integration PUBLIC 251
Statistics type Description

data source. The RECORD COUNT type is a table-wide


statistic.

SIMPLE Creates a data statistics object that helps the query


optimizer calculate basic statistics, such as min, max, null
count, count, and distinct count.

HISTOGRAM Creates a data statistics object that helps the query


optimizer estimate the data distribution.

Statistics are created for each column of the table and can only be created if the data access is "Remote". Only
one data statistics object at a time is allowed per remote table.

 Note

Creating statistics can cause a significant workload on the source system, especially if the source object is
a view, like for example a CDS View in ABAP or an SQL View in a Database. RECORD COUNT is the simplest
statistics object, and complexity increases for SIMPLE and HISTOGRAM, which causes heavier workload.

You can update data statistics objects: when you choose the menu Create Statistics on a remote table for which
data statistics are already available, the windows Create Statistics opens, and the existing type of statistics line
is marked as active. You can then choose to create another type of statistics or update the existing ones.

You can delete the data statistics objects for one or several tables with the menu Delete Statistics.

8.10 Monitoring Task Chains

Monitor the status and progress of running and previously run task chains.

To view all task chains that are scheduled to run, are currently running, or have already been run in your space,
go to Data Integration Monitor Task Chains :

The monitor displays information on your task chains:

Properties Comments

Business Name or Technical Name Name of the task chain. Technical or Business Name is
displayed, depending on how you have configured your UI
settings in Your Profile Settings

Status Status of the task chain after last run.

• Running: the task chain is currently running.


• Completed: the task chain is completed successfully.

Integrating Data and Managing Spaces in SAP Datasphere


252 PUBLIC Managing and Monitoring Data Integration
Properties Comments

• Failed: Something goes wrong during the task chain run


and it could not be completed. Go to the details screen
of your task chain and check the logs to identify to
issue.

All these statuses can be combined with a sub-status.


For more information on sub-statuses, see Understanding
Statuses and Substatuses [page 165].

Frequency See if a schedule is defined for the task chain:

• None (--): There is no schedule defined for this

task chain. You can define one from Schedule

Create .
• Scheduled: A schedule is defined for this task chain. If
you click Scheduled, you will get detailed information
on the schedule. You can click Schedule to create, edit,
delete, pause or resume a schedule for a selected task
chain or to assign a schedule to your name.
• Paused: A schedule is defined for this task chain
but it's currently paused. You can resume it if you
click Schedule Resume

Duration How long it took to run the task chain.

Last Run Start Date and Time when the last run started.

Last Run End Date and Time when the last run ended.

Next Run When the next run is scheduled.

Schedule Owner Displays the name of the current schedule owners.

You can personalize the columns you want to display clicking on  Select Columns

 Note

A task chain consists of multiple tasks that run one after another. As it takes some time for the task chain
monitor to collect and update all a task’s status information, the task runtime information displayed here
might not exactly match the task runtime information shown in a respective detail monitor, for example, in
the Views or Flows monitors.

Click Run Start Run to manually start a new run of a selected task chain. If a task chain’s last run failed,
you can also choose the Run Retry Latest Run option to retry the failed task chain’s execution from the
point at which it previously failed.

 Note

If you add a remote table whose data access is Replicated (Real-time) in a task chain, the replication type
will change from real-time replication to batch replication at the next run of the task chain. The data will no
longer be updated in real-time.

Integrating Data and Managing Spaces in SAP Datasphere


Managing and Monitoring Data Integration PUBLIC 253
You can click Schedule to create, edit, delete, pause or resume a schedule for a selected task chain or to assign
a schedule to your name. Whenever it’s applicable, you can select several task chains to group the scheduling
actions. For more information, see Scheduling Data Integration Tasks [page 256].

 Note

For optimal performance, it is recommended that you consider staggering the scheduled run time of tasks
such as data flows and task chains that may contain these tasks. There is a limit on how many tasks can be
started at the same time. If you come close to this limit, scheduled task runs may be delayed and, if you go
beyond the limit, some scheduled task runs might even be skipped.

Click  in the right-most column to access the Task Chain Log Details page for the selected task chain. The
Details page provides a flexible three-panel display to access all the relevant information about the selected
task chain’s run history. It also provides details about individual child tasks or subtasks in the task chain, and
log messages about a selected task chain’s execution.

 Note

The task chain Details page also lets you resize ( and  ), expand/collapse ( ), or close (  ) the display of
each panel to focus on specific detail information or log messages. In addition, you can sort ( ) and filter
( ) information displayed in the different panels, as well as search log messages.

To view details about a specific task run, select it in the Runs list in the left-side panel. The Runs list provides a
summary of all the selected task chain’s previous runs as well as those scheduled in the future. It also indicates
the status of each task chain run, whether previous task chains completed successfully, failed, are currently
running, or are scheduled to be run.

 Note

For failed task chain runs, you can click on Failed in the Status field to show the likely cause of the failure,
when that information is available.

The center Run Details panel provides a list of task chain objects and child tasks (if any are defined) for a
selected task chain and displays their individual type and execution status.

The Run Details panel also provides a Cancel Run link in the top right-hand corner of the display for a specific
task chain. If you click this link, SAP Datasphere initiates the Cancel operation and immediately remove all
child tasks not already triggered. For any child task that has already started, the effect of the Cancel operation
depends on the individual object. For example, replication flows and Open SQL procedures cannot effectively

Integrating Data and Managing Spaces in SAP Datasphere


254 PUBLIC Managing and Monitoring Data Integration
be cancelled, so successful completion of the task chain cancellation will wait until those child tasks are
completed. Finally, when the task chain cancellation is complete, the task chain cancel activity will then display
a Run Status of Completed.

In the right-side panel, the task chain monitor displays log messages for a selected task chain. When you click
on an object in the Run Details task list, detailed log messages for the selected task or subtask are displayed in
the right-most panel.

 Note

From each monitor, you can navigate back to the Data Builder to change your objects by clicking Open in
Data Builder.

To return to the original Run Details display, click on the task chain name under the Parent Task Chain field in
the upper right corner of the page.

Retrying Failed Task Chain Runs

If a task chain’s last run failed, the Run Details panel provides a Retry Run option to retry the failed task
chain’s execution. You can also choose the Run Retry Latest Run option to retry the failed task chain's
execution. Note that the Retry Latest Run option is only active when the last taskchain runs failed.

Any user with authorization to run the task chain can retry a task chain’s execution. Retries of a task chain are
run on behalf of the original user and the retried task chain will have the same log ID as the original task chain
run. The monitor will create a new log ID for retries of failed subtasks and both the original and retried subtasks
appear in the hierarchy of tasks displayed in the Run Details panel.

Retries of a failed task chain will begin at the first failed child or subtask in the task chain, skipping any tasks
already completed successfully. The monitor updates the status of a retried task as the new run is triggered.

Integrating Data and Managing Spaces in SAP Datasphere


Managing and Monitoring Data Integration PUBLIC 255
8.11 Scheduling Data Integration Tasks

Schedule data integration tasks to run periodically at a specified date or time.

You can schedule or unschedule data integration tasks such as remote table replication, persisting views, and
data flow, replication flow, or task chain execution. You may also pause and then later resume execution of
scheduled tasks.

 Note

For optimal performance, it is recommended that you consider staggering the scheduled run time of tasks
such as data flows and task chains that may contain these tasks. There is a limit on how many tasks can be
started at the same time. If you come close to this limit, scheduled task runs may be delayed and, if you go
beyond the limit, some scheduled task runs might even be skipped.

Using a dedicated dialog box, you can specify the frequency and time range of the schedule by using a simple
form or by directly entering a cron expression.

 Note

When you click Create, the definition of the schedule that is created and saved is the one that is currently
displayed (either in the Simple Schedule area or the Cron Expression area).

 Note

If you create a schedule for a remote table whose data access is Replicated (Real-time), the replication type
will change from real-time replication to batch replication at the next run of the schedule. The data will no
longer be updated in real-time.

8.11.1 Schedule a Data Integration Task (Simple Schedule)

You can schedule or unschedule data integration tasks such as remote data replication, data persistence, data
flow, replication flow, or task chains runs. You may also pause and then later resume the run of scheduled
tasks.

Prerequisites

• With the DW Integrator role you can create, change, or delete a schedule and let the scheduling
component of SAP Datasphere run tasks on your behalf. For data flows, you need the DW Integrator role
along with the DW Viewer or the DW Modeler role.
• For replication flows, scheduling is possible if all objects in the flow have load type Initial Only.
• To run recurring scheduled tasks on your behalf, you need to authorize the job scheduling component of
SAP Datasphere. In your profile settings under Authorized Consent Settings, you can give and revoke your
consent to SAP Datasphere to run your scheduled tasks in the future. Note that when you don't give your
consent or revoke your consent, tasks that you own won't be executed but will fail.

Integrating Data and Managing Spaces in SAP Datasphere


256 PUBLIC Managing and Monitoring Data Integration
 Note

Your consent is valid for 365 days. If your consent will expire within the next four weeks, when you
attempt to schedule new tasks, SAP Datasphere displays a message warning that your consent is
approaching its expiration date. After the consent has expired, a log message informs you that the
tasks for which you own the schedule won’t be executed anymore. Renew your consent to resume
task execution according to the schedules. Additionally, in Data Integration Monitor, a warning message
appears four weeks before the expiry of consent.

For more information, see Changing SAP Datasphere Settings.

Context

The job scheduling component of SAP Datasphere runs scheduled tasks on behalf of the owner of the
schedule. Per default, the user who creates a schedule is the owner of the schedule. You can takeover the
ownership, however, if required. Scheduled tasks run in the background according to the settings defined in the
schedule.

Procedure

1. Go to Data Integration Monitor.


2. Select your space.
3. Open the monitor for the object for which you want to schedule a task.
4. If you haven't authorized SAP Datasphere yet to run your scheduled tasks for you, you will see a message
at the top of the monitor asking for your consent. Give your consent.
5. Select the object for which you want to schedule a task.

6. Click Schedule Create Schedule

You can also use schedules to automate actions on objects created from the Data Builder

The Create Schedule window opens.

 Note

Tasks for which you own a schedule without having authorized SAP Datasphere to run scheduled tasks
on your behalf before won't be executed but will fail.

7. Define your scheduling options:

Integrating Data and Managing Spaces in SAP Datasphere


Managing and Monitoring Data Integration PUBLIC 257
Categories Possible Values Examples

Frequency • Minutes: Define how often per Let's take a few examples. Imagine
hour that you want to trigger the that we are in April 14, 2021, it's
scheduled run during the desig- 9:00 am and you define your schedule
nated time or date range. You can without a start and end date:
specify scheduled runs every 10, • Minutes: You set frequency of
15, 20, or 30 minutes. scheduled task runs to 15 mi-
• Hourly: Define how often you nutes. Tasks will now be triggered
want to trigger the schedule per to run every 15 minutes within
day (every 1 to 23 hours of the designated time/date range.
day) and the start time. • Hourly: You select every "5" hours
Optionally, you can specify an off- "00" minutes. You need to con-
set from which the tasks will run. sider that a cycle run reset will
For example, so that a task runs occur every day at 0:00. Next
runs will be:
every 2 hours starting from 1:30
AM, specify 2 in Every and 1:30 in • Apr 14, 2021 10:00

Starting at (UTC). • Apr 14, 2021 15:00

• Daily: Define how often you want • Apr 14, 2021 20:00
to trigger the schedule per month • Apr 15, 2021 0:00
(every 1 to 28 days per month) • Apr 15, 2021 5:00
and the start time. If you define that the schedule
By default, the option Reset Every
must run at 0:15 mn, then the
Month is not selected, which
next runs will be
means that you schedule daily
tasks with a fixed frequency (for
• Apr 14, 2021 10:15

example, every two days) that is • Apr 14, 2021 15:15

not reset on the first day of each • Apr 14, 2021 20:15
month. You can select the option • Apr 15, 2021 0:15
so that the run cycle is reset ev-
• Apr 15, 2021 5:15
ery month.
• Daily: You select every "4" days at
• Weekly: Select which day(s) of "0:00" hours.
the week the schedule must be
Next runs will be:
triggered (multiple selection is al-
lowed) and the start time. • April 17, 2021

• Monthly: Define how often you • April 21, 2021


want to trigger the schedule per • April 25, 2021
year (every 1 to 11 months of the
year), which day of the month
• April 29, 2021
where the schedule must start, • May 1st, 2021 (if the op-
and define the start time. tion Reset Every Month is se-
lected) or May 3 (if the op-
tion is not selected)
• May 5, 2021 (if the option
Reset Every Month is se-
lected) or May 7 (if the op-
tion is not selected)
• Monthly: You select every "5"
months on "1"st day of month.
You will have a cycle run reset at

Integrating Data and Managing Spaces in SAP Datasphere


258 PUBLIC Managing and Monitoring Data Integration
Categories Possible Values Examples

1st day of the year. Next runs will


be:
• Jun 1, 2021 0:00
• Nov 1, 2021 0:00
• Jan 1, 2022 0:00 (Cycle run
reset)
• Jun 1, 2022 0:00

Time Zone • UTC: This is the default schedule UTC is the default setting. You can
setting. also choose a regional time zone
• Region/Location: You can choose in the format of Region/Location,
a regional time zone. This will for example, Europe/Berlin, Amer-
automatically adjust for daylight ica/New_York, or Asia/Kolkata.
saving time.
 Note

Make sure your location reflects


local daylight saving changes
when choosing a time zone.
Schedules will be automatically
updated to reflect daylight saving
time changes locally.

Duration Elastic compute nodes only.


See Run an Elastic Compute Node.

Time Range • Start Date: Specify a date if you Now imagine we are still in April 14,
want to the change the default 2021, but you set as start date "May
start date, which is the current 2, 2020" and as end date "March 31,
date. 2021". If we take the same examples
• End Date: Optionally you can as above the next runs will be:
specify an end date. Otherwise, • Daily:
the schedule runs indefinitely. • May 6, 2021
• May 11, 2021
• May 16, 2021
• May 21, 2021
• May 26, 2021
• Monthly:
• Jun 1, 2021 0:00
• Nov 1, 2021 0:00
• Jan 1, 2022 0:00 (Cycle run
reset)

Integrating Data and Managing Spaces in SAP Datasphere


Managing and Monitoring Data Integration PUBLIC 259
Categories Possible Values Examples

Preview Show Runs In: Preview your next five runs in ei-
• UTC ther UTC time, local time, or the
Region/Location time selected. The
• Local Time
preview will take into consideration
• Region/Location selected in the changes in daylight saving time.
Time Zone drop down.
Let's look at an example. You schedule
a task to run in the time zone Amer-
ica/Los_Angeles every week on Mon-
day at 20:00. Your start date is March
5, 2025. You want to preview your next
runs.

Show Runs In:


• UTC: You have scheduled your
task to run in the Amer-
ica/Los_Angeles time zone. You
check your next scheduled runs
in UTC time.
• Mar 11, 2025 3:00:00
• Mar 18, 2025 3:00:00
• Mar 25, 2025 3:00:00
• Apr 1, 2025 3:00:00
• Apr 8, 2025 3:00:00
• Local Time: You have scheduled
your task to run in the Amer-
ica/Los_Angeles time zone. You
are working in Europe/Berlin
time zone and check your sched-
uled run times in your local time
zone.
• Mar 11, 2025 4:00:00
• Mar 18, 2025 4:00:00
• Mar 25, 2025 4:00:00
• Apr 1, 2025 5:00:00
• Apr 8, 2025 5:00:00

Daylight saving time in the Berlin lo-


cation of Europe takes place at the
end of March 2025. UTC time does
not take daylight saving into consider-
ation. You can see that daylight sav-
ing time has automatically been ad-
justed for in your local time zone (Eu-
rope/Berlin).

8. Click Create to create the schedule.

Integrating Data and Managing Spaces in SAP Datasphere


260 PUBLIC Managing and Monitoring Data Integration
Results

The column Frequency shows the status Scheduled. By clicking the status you can display the scheduling
settings.

The column Schedule Owners displays the name of the current schedule owner.

The Next Run column shows the start date and time of the next run according to the scheduling settings.

As long as the consent from the owner of the schedule hasn't expired the task will run in the background.
You can access the log by selecting the relevant object and clicking (Remote Table Logs), (Views Logs), or
(Details).

Next Steps

Once a schedule is defined, you can adjust the scheduling settings at any time selecting Schedule Edit
Schedule . The next run of the task will use the adjusted scheduling settings.

The schedule can be removed at any time via the menu Schedule Delete Schedule with the result that no
new run will start anymore for the task. The Next Run column gets cleared.

You may also pause and then later resume execution of scheduled tasks via the menu Schedule Pause
Schedule option. See Pause or Resume a Scheduled Task [page 266] for more information.

Related Information

Take Over the Ownership of a Schedule [page 265]

8.11.2 Schedule a Data Integration Task (with Cron


Expression)

Schedule data integration tasks such as data replication, data persistence, or data flow runs or task chains
runs, by entering directly a cron expression.

Prerequisites

• With the DW Integrator role you can create, change, or delete a schedule and let the scheduling
component of SAP Datasphere run tasks on your behalf. For data flows, you need the DW Integrator role
along with the DW Viewer or the DW Modeler role.

Integrating Data and Managing Spaces in SAP Datasphere


Managing and Monitoring Data Integration PUBLIC 261
• SAP Datasphere. In your profile settings under Schedule Consent Settings, you can give and revoke your
consent to SAP Datasphere to run your scheduled tasks in the future. Note that when you don't give your
consent or revoke your consent, tasks that you own won't be executed but will fail.

 Note

Your consent is valid for 365 days. After the consent has expired, a log informs you that the tasks
for which you own the schedule won’t be executed anymore. Renew your consent to resume task
execution according to the schedules. Additionally, in To run recurring scheduled tasks on your behalf,
you need to authorize the job scheduling component of Data Integration Monitor, a warning message
appears 10 days before the expiry of consent.

For more information, see Changing SAP Datasphere Settings.

Context

The job scheduling component of SAP Datasphere runs scheduled tasks on behalf of the owner of the
schedule. By default, the user who creates a schedule is the owner of the schedule. You can takeover the
ownership, however, if required. Scheduled tasks run in the background according to the settings defined in the
schedule.

Procedure

1. Go to Data Integration Monitor.


2. Select your space.
3. Open the monitor for the object for which you want to schedule a task.
4. If you haven't authorized SAP Datasphereto run recurring scheduled tasks on your behalf, you need to
authorize it to run your scheduled tasks for you. You will see a message at the top of the monitor asking for
your consent. Give your consent.
5. Select the object for which you want to schedule a task.

6. Click Schedule Create Schedule

You can also use schedules to automate actions on objects created from the Data Builder

The Create Schedule window opens.

 Note

Tasks for which you own a schedule without having authorized SAP Datasphere to run scheduled tasks
on your behalf before won't be executed but will fail.

7. In the Frequency area, select Cron Expression from the Enter As drop down list.
8. In the Frequency area, select your preferred Region//Location from the Time Zone drop down list. Cron
expression will reflect the time in the Region/Location you have chosen.
9. Define when the task will be run by using the unix-cron string format (* * * * *) in the five following fields:
Minute, Hour, Day (Month), Month and Day (Week).

Integrating Data and Managing Spaces in SAP Datasphere


262 PUBLIC Managing and Monitoring Data Integration
Either enter the cron expression directly in the fields or copy it from another source to the cron expression
area by selecting one of the five fields and pasting it.

Syntax Description

x Where x is any number in the following range:


• Minute – 0-59
• Hour – 0-23
• Day (Month) – 1-31
• Month – 1-12
• Day (Week) – 0-6

x,y Where x and y are two or more values of a list separated by a comma.

Example: 4,5,6 in the Month field is equivalent to "April, May and June".

Valid in: All fields except Minute

x-y Where x and y are the first and last values of a range separated by a dash.

Example: 1-5 in the Day (Week) field is equivalent to "from Monday to Friday".

Valid in: All fields except Minute

*/x Where * is all possible values and x is the step value, separated by a slash.

Example: */3 in the Hour field is equivalent to 0,3,6,9,12,15,18,21. The asterisk (*) specifies
"every hour" but the /3 means "every 3 hours".

Valid in: All fields except Minute

a-b/x Where a and b are the first and last values of a range and where x is the step value.

Example: 1-15/2 in the Day (Month) field is equivalent to 1,3,5,7,9,11,13,15. 1-15 specifies "from
day 1 to day 15" of the month but /2 means "every 2 days".

Valid in: All fields except Minute

* Where the asterisk is all possible values for the field.

Example: an asterisk in the Hour field is equivalent to "every hour."

Valid in: All fields except Minute

Examples of cron expressions:

Expression Description

0 0 1-3 * * The task runs the first 3 days of each month at 00:00
(UTC).

0 18 1,15,25 * * The task runs every 1st, 15th and 25th of each month at
6:00 PM (UTC).

0 8 */1 * 1-5 The task runs from Monday to Friday at 8:00 AM (UTC).

Integrating Data and Managing Spaces in SAP Datasphere


Managing and Monitoring Data Integration PUBLIC 263
Expression Description

0 2-8/2 */1 * 1-5 The task runs every 2 hours between 2:00 AM (UTC) and
8:00 AM (UTC), from Monday to Friday.

 Note

You can copy to the clipboard the cron expression you've entered by clicking the copy icon button next
to the cron expression.

The Duration area is available only for elastic compute notes. See Run an Elastic Compute Node.

10. In the Start Date area, specify a date if you want to the change the default start date, which is the current
date.
11. Optionally, in the End Date area, specify a date. If you do not specify an end date, the schedule runs
indefinitely.
12. In the right area of the dialog box, you can see a preview of the schedule and when the five next runs are
planned. Schedules are viewed in Coordinated Universal Time (UTC) by default. You can also view the next
runs in local time or the Region/Location time that you have set as your preferrence.
13. Click Create to create the schedule.

Results

The column Frequency shows the status Scheduled. By clicking the status you can display the scheduling
settings.

The column Schedule Owners displays the name of the current schedule owner.

The Next Run colums shows the start date and time of the next run according to the scheduling settings.

As long as the consent from the owner of the schedule hasn't expired the task will run in the background. You
can access the log by selecting the relevant object and clicking .

Next Steps

Once a schedule is defined, you can adjust the scheduling settings at any time selecting Schedule Edit
Schedule . The next run of the task will use the adjusted scheduling settings.

The schedule can be removed at any time via the menu Schedule Delete Schedule with the result that no
new run will start anymore for the task. The Next Run colum gets cleared.

You may also pause and then later resume execution of scheduled tasks via the menu Schedule Pause
Schedule option. See Pause or Resume a Scheduled Task [page 266] for more information.

Integrating Data and Managing Spaces in SAP Datasphere


264 PUBLIC Managing and Monitoring Data Integration
8.11.3 Take Over the Ownership of a Schedule

Per default, the user who creates a task schedule owns the schedule which means that the job scheduling
component runs the task on the owner's behalf according to the defined schedule. You can assign the
ownership of the schedule to yourself.

Prerequisites

• With the DW Integrator role you can change the schedule and let the scheduling component run tasks
on your behalf. For data flows, you need the DW Integrator role along with the DW Viewer or the DW
Modeler role.
• To let the job scheduling component of SAP Datasphere run scheduled tasks on your behalf, you need to
give your consent. For more information, see Changing SAP Datasphere Settings.

Context

Changing the owner of the schedule is required when the user owning the schedule until now left the
organization and the user has been deleted, for example. In such a case the schedule isn't assigned to any
user and can't be executed successfully.

Procedure

1. Go to Data Integration Monitor.


2. Select your space.
3. Open the monitor for the relevant object.
4. If you haven't authorized SAP Datasphere yet to run your scheduled tasks for you, you will see a message
at the top of the monitor asking for your consent. Give your consent.
5. Select the object for which you want to schedule a task.

6. Select Schedule Assign Schedule to me


7. Click OK to save the schedule.

Results

The Owner field now shows your user ID as the new owner of the schedule. The scheduling component from
now own will execute the task on your behalf.

Integrating Data and Managing Spaces in SAP Datasphere


Managing and Monitoring Data Integration PUBLIC 265
8.11.4 Pause or Resume a Scheduled Task

Using options available for creating, editing, and removing task scheduling, you can pause or resumed
previously scheduled data integration tasks such as remote table replication, persisting views, data flow or
task chain execution.

After previously scheduling a task, you can select Schedule Pause Schedule to pause the next schedule
run of a task, or select Schedule Resume Schedule to resume a schedule previously paused.

If you pause a schedule, when you return to the task display, the Frequency column displays Paused for the
selected task, and the Next Run column is cleared of any future scheduled runs. The defined schedule for
future scheduled runs is still saved, however, so you can simply go back later and resume the schdule to run
future scheduled task again.

If you pause one or more objects within a replication flow that has a schedule, the schedule continues
unchanged for all other objects in the flow, wheras no replication takes place for the paused objects until
you resume them.

8.12 Deleting Task Logs to Reduce Storage Consumption

In the Configuration area, you can check how much spaces the task logs are using on your tenant, and decide to
delete the obsolete ones to reduce storage consumption.

Context

Each time an activity is running in SAP Datasphere (for example, replicate a remote table), task logs are
created to allow you to check if the activity is running smoothly or if there is an issue to solve. You access
these detailed task logs by navigating to the Data Integration Monitor - Details screen of the relevant object. For
example, clicking the button of the relevant remote table.

However, task logs can consume a lot of spaces in a tenant. Deleting old task logs that are no longer needed
can be useful to release storage space. This is why SAP Datasphere has a log deletion schedule activated by
default. You can change the schedule defining your own criteria or decide to take immediate deletion actions.

Procedure

1. In the side navigation area, click  (Configuration) → Tasks.


2. Check how much size the task logs consume on your tenant. If needed, decide how you want to delete the
task logs.

• Manually Delete Task Log: SAP Datasphere automatically triggers logs deletion using the following
default criteria:

Integrating Data and Managing Spaces in SAP Datasphere


266 PUBLIC Managing and Monitoring Data Integration
• Deletion tasks will be run every 4 months
• Task logs older than 200 days will be deleted
Go to the section Schedule Task Log Deletion, update the deletion criteria following your needs and
click Save.
• Manual Deletion: You want to manually delete task logs to take immediate action. Go to the Manually
Delete Task Log section and determine how long you want to keep the logs. For example, delete the
logs that are older than 100 days.
Click Delete.

Integrating Data and Managing Spaces in SAP Datasphere


Managing and Monitoring Data Integration PUBLIC 267
9 Securing Data with Data Access Controls

Users with a space administrator role can create data access controls to allow modelers to apply row-level
security to Data Builder and Business Builder objects. Once a data access control is applied to an object, any
user viewing its data either directly or via an object using it as a source, will see only those records they are
authorized to view, based on the specified criteria.

This topic contains the following sections:

• Permissions Entities [page 268]


• Data Access Control Best Practices [page 268]
• Applying Data Access Controls [page 269]
• Data Access Control Example [page 270]

Permissions Entities

Your criteria are defined in a table or view that lists SAP Datasphere user IDs (in the form required by your
identity provider) and assigns them to one or more criteria. You can define the following types of criteria:

• Single Values - Each user can only see the records that match any of the single values she is authorized for
in the permissions entity. See Create a "Single Values" Data Access Control [page 271].
• Operator and Values - Each user can only see the records that fulfill the operator-value pairs she is
authorized for in the permissions entity, including support for complex AND and OR combinations. See
Create an "Operator and Values" Data Access Control [page 274].
• Hierarchy - Each user can only see the records that match the hierarchy values she is authorized for in
the permissions entity, along with any of their descendents. See Create a "Hierarchy" Data Access Control
[page 277]

You can create one or more data access controls that consume each permissions entity, and select one or
more columns in each data access control to specify the criteria that it will enforce. You can apply a single
data access control to multiple views. Each view protected in this way will filter the results available in its data
preview to only those rows meeting the criteria for the current user.

 Note

Permissions entities must not, themselves, be protected by a data access control, or have any protected
view among their sources.

Data Access Control Best Practices

We recommend that you develop clear policies for securing data, and that you:

• Focus in particular on securing transactional data and sensitive master data.

Integrating Data and Managing Spaces in SAP Datasphere


268 PUBLIC Securing Data with Data Access Controls
• Avoid applying multiple data access controls to a view if possible.
• Secure data as soon as possible once it is ingested into SAP Datasphere and then only use the protected
view going forward, keeping the following factors in mind as you plan your modeling strategy for security
and performance:
• Encapsulating individual source tables in views and applying data access controls to them:
• Improves performance by filtering records immediately, as fewer go forward for subsequent
preparation.
• Can impact performance when any view that combines multiple sources is run, by increasing the
number of data access controls that must be applied.
• Can impact performance by excluding the possibility of persisting subsequent views, since views
cannot be persisted if any of their sources is protected by a data access control.
• Doing initial processing and joining separate data sources before applying a single, more targeted data
access control to the result:
• Improves performance by applying only a single data access control to a combined view.
• Improves performance by allowing the use of persistence at any stage during the preparation.
• Can impact performance as more records will pass through each stage of the data preparation.
• When using Operator and Values data access controls, try to control the size of the permissions entity and,
in particular, avoid having more than 5,000 permissions records for a single user:
• Consider applying the filter against a different column that could simplify calculation.
• You can use the * operator to provide access to all records.

 Note

The row-level security provided by the data access control can be circumvented while the view remains in
its space. It is enforced only when the view is:

• Shared to another space, or


• Consumed outside the space in SAP Analytics Cloud.

Applying Data Access Controls

For information about:

• Applying a data access control to a data layer view, see Apply a Data Access Control to a Graphical or SQL
View.
• Using a data access control to create authorization scenarios in the business layer, see Authorization
Scenario.
• Persisting data in a view that has a data access control applied to it, see Persisted Views and Data Access
Control [page 205].

 Note

If you experience performance issues with a view protected by a data access control, we recommend
enabling replication for source tables, particularly if any source contains more than 500,000 rows.

Integrating Data and Managing Spaces in SAP Datasphere


Securing Data with Data Access Controls PUBLIC 269
Data Access Control Example

This diagram shows a typical environment where a permissions entity is maintained by business users in one
space and shared to a second space, where technical users create data access controls from it and apply them
to views. These views are then shared to other spaces where they can be securely accessed by modelers and
analysts:

In this environment:

• The Permissions space users are a select group of business users who:
• Maintain the Permissions table, assigning users (which, in this case, are identified by their email
address) to the appropriate country, department, and any other relevant criteria:

User ID Country Department

anne.archer@acme.com US Sales

jennifer.jones@acme.com FR Sales

• Share the Permissions table to the IT space to use as a data source for data access controls.

Integrating Data and Managing Spaces in SAP Datasphere


270 PUBLIC Securing Data with Data Access Controls
 Note

Tables and views shared to a space cannot be directly used as the permissions entity for a data
access control. Modelers in the space receiving the shared object must encapsulate it in a view,
which can then serve as the permissions entity for one or more data access controls.

• The IT space users are technical users who:


• Use the shared Permissions table as the source for a Permissions view that they will use as a
permissions entity in their space.
• Create a Country data access control, which uses the Permissions view as its permissions entity.
• Maintain a connection to a source system, from which they import the Sales table.
• Create a Sales view for use by business analysts, to which they apply the Country data access
control.

 Note

The row-level security provided by the Country data access control can still be circumvented while
the view remains in the IT space.

• Share the Sales view to the Sales space where it can be used securely.
• The Sales space users use the protected Sales view to do analytics on sales data:
• Anne is a data modeler in the US sales department. She connects to SAP Datasphere and combines
the Sales view with other data. Whenever she previews data, she can only see US data.
• Jennifer is a business analyst in the French sales department. She connects to SAP Analytics Cloud,
builds a story on the Sales view, and can only see French sales data.

9.1 Create a "Single Values" Data Access Control

Users with the DW Space Administrator role (or equivalent privileges) can create data access controls in which
criteria are defined as single values. Each user can only see the records that match any of the single values she
is authorized for in the permissions entity.

Context

Before creating your data access control, you must have prepared a permissions entity with the following
columns:

• User ID column - Containing user ids in the format required by your identity provider (email addresses,
logon names, or other identifiers). If you are using SAML authentication, this column must contain values
in the form defined as your User Attribute / IdpUserID (see Enabling a Custom SAML Identity Provider).
This column must be selected as the Identifier Column in your data access control.

Integrating Data and Managing Spaces in SAP Datasphere


Securing Data with Data Access Controls PUBLIC 271
 Note

If a user has no entries in the permissions entity, then they will not have access to any records in the
protected view.

• One or more columns containing filter criteria (country or region names or ids, department names or ids,
or any other criteria to control how your data is exposed to users). These columns must be selected as
Criteria columns in your data access control.

Procedure

1. In the side navigation area, click  (Data Builder), select a space if necessary, and click New Data Access
Control to open the editor.
2. Complete the following properties:

Property Description

Business Name Enter a descriptive name to help users identify the object. This name can be changed at any
time.

Technical Name Displays the name used in scripts and code, synchronized by default with the Business Name.

To override the default technical name, enter a new one in the field. Technical names can
contain only alphanumeric characters and underscores.

 Note
Once the object is saved, the technical name can no longer be modified.

Package Select the package to which the object belongs.

Packages are used to group related objects in order to facilitate their transport between
tenants.

 Note
Once a package is selected, it cannot be changed here. Only a user with the DW Space
Administrator role (or equivalent privileges) can modify a package assignment in the
Packages editor.

For more information, see Creating Packages to Export [page 292].

Owner Enter the name of the person responsible for your data access control.

Structure Select Single Values.

Permissions Entity Select the table or view containing your user ids and criteria.

The permissions entity must match the structure selected in the Structure field. Click the Open
in New Tab button to the right of the field to open the entity in its own editor.

Integrating Data and Managing Spaces in SAP Datasphere


272 PUBLIC Securing Data with Data Access Controls
Property Description

Identifier Column Select a column containing user ids in the format required by your identity provider (email
addresses, logon names, or other identifiers).

3. Select your permissions entity:


a. Click the Permissions Entity field to open the Select Permissions Entity dialog.
b. In the Data Objects list, select the table or view containing your user logins and columns with filter
criteria.
c. In the Available Columns list, select one or more columns containing the criteria on which you want to
control the display of data.

These columns will be listed in the Criteria section. When you assign a data access control to a view,
these columns must be mapped to columns in the view to enforce filtering by these criteria.
d. Click OK to confirm the selection.
4. Click Save and then Deploy to deploy your data access control and make it available for use.

For information about attaching a data access control to a view, see Apply a Data Access Control to a
Graphical or SQL View.

 Note

You can use the View as User tool in the Data Viewer panel to review the effects of the data access
controls you apply by checking the records that another user will be allowed to see (see Viewing Object
Data).

5. The tools in the editor toolbar help you work with your object throughout its lifecycle:

Tool Description

 (Save) Save your changes to the design-time repository. You can use Save As to create a copy of the
object.

See Saving and Deploying Objects.

 (Deploy) Deploy your changes to make them available in the run-time environment.

See Saving and Deploying Objects.

 (Export) Export the object to a CSN/JSON file.

See Exporting Objects to a CSN/JSON File.

 (Impact and Open the Impact and Lineage Analysis graph for the object.
Lineage Analysis)
See Impact and Lineage Analysis.

Status [read-only] Displays the status of the object.

See Saving and Deploying Objects.

Versions Open the Version History dialog for the object.

See Reviewing and Restoring Object Versions.

Details Toggles the display of the Properties panel.

Integrating Data and Managing Spaces in SAP Datasphere


Securing Data with Data Access Controls PUBLIC 273
9.2 Create an "Operator and Values" Data Access Control

Users with the DW Space Administrator role (or equivalent privileges) can create data access controls in
which criteria are defined as operator and value pairs. Each user can only see the records that fulfill the
operator-value pairs she is authorized for in the permissions entity, including support for complex AND and OR
combinations.

Context

Before creating your data access control, you must have prepared a permissions entity with the following
columns:

• Permission ID column - Which should be marked as a key.


• User ID column - Containing user ids in the format required by your identity provider (email addresses,
logon names, or other identifiers). If you are using SAML authentication, this column must contain values
in the form defined as your User Attribute / IdpUserID (see Enabling a Custom SAML Identity Provider).

 Note

If a user has no entries in the permissions entity, then they will not have access to any records in the
protected view.

• Restriction Column - Each restriction per user must have a unique name, and one or more criteria can
belong to each restriction. All criteria belonging to a restriction act together as an AND condition. If you
have more than one restriction per user then all restrictions act together as an OR condition.
• Criterion Column - The criteria name must contain only alphanumeric characters and underscores and
appears in the Mappings area when a user applies the data access control to their view. It does not need to
exactly match a column name, but should guide the user to the column to which it should be mapped.
• Operator Column - The following operators are supported:
• ALL (or *) - Provides access to all records. No values required.
• EQ (or =) - Equal to First Value.
• NE (or <> or !=) - Not equal to First Value.
• GT (or >) - Greater than First Value.
• GE (or >=) - Greater than or equal to First Value.
• LT (or <) - Less than First Value.
• LE (or <=) - Less than or equal to First Value.
• CP (or LIKE) - Contains pattern First Value. Optionally Second Value can be set to an escape character.
See LIKE Predicate in the SAP HANA Platform documentation.
• BT (or BETWEEN) - Between First Value and Second Value.
• First Value Column - This value acts as the first argument or lower bound of the operator.
• Second Value Column - This value acts as the second argument or upper bound of the operator.

For example:

Integrating Data and Managing Spaces in SAP Datasphere


274 PUBLIC Securing Data with Data Access Controls
Permission ID User ID Restriction Criterion Operator First Value Second Value

1 bob@acme.c 0 Code BT CA CZ
om

2 bob@acme.c 0 Type EQ 1
om

3 bob@acme.c 1 Class CP ERR%


om

4 ann@acme.c 0 Admin ALL


om

Based on these three records:

• bob will have access to records with a:


• (restriction 0) - Code between CA and CZ, and having a Type equal to 1, or
• (restriction 1) - Class beginning with ERR.
• ann (restriction 0) - will have access to all records thanks to the ALL operator.

 Note

A criterion with an ALL operator can have any name and can be mapped to any column in a view.

Procedure

1. In the side navigation area, click  (Data Builder), select a space if necessary, and click New Data Access
Control to open the editor.
2. Complete the properties in the General section:

Property Description

Business Name Enter a descriptive name to help users identify the object. This name can be changed at any
time.

Technical Name Displays the name used in scripts and code, synchronized by default with the Business Name.

To override the default technical name, enter a new one in the field. Technical names can
contain only alphanumeric characters and underscores.

 Note
Once the object is saved, the technical name can no longer be modified.

Integrating Data and Managing Spaces in SAP Datasphere


Securing Data with Data Access Controls PUBLIC 275
Property Description

Package Select the package to which the object belongs.

Packages are used to group related objects in order to facilitate their transport between
tenants.

 Note
Once a package is selected, it cannot be changed here. Only a user with the DW Space
Administrator role (or equivalent privileges) can modify a package assignment in the
Packages editor.

For more information, see Creating Packages to Export [page 292].

Structure Select Operator and Values.

Permissions Entity Select the table or view containing your user ids and criteria.

The permissions entity must match the structure selected in the Structure field. Click the Open
in New Tab button to the right of the field to open the entity in its own editor.

Identifier Column Select a column containing user ids in the format required by your identity provider (email
addresses, logon names, or other identifiers).

3. Complete the properties in the Criteria section:

Property Description

Restriction Column Select the permissions entity column that contains the restriction names.

Each restriction per user must have a unique name, and one or more criteria can belong to
each restriction. All criteria belonging to a restriction act together as an AND condition. If you
have more than one restriction per user then all restrictions act together as an OR condition.

Criterion Column Select the permissions entity column that contains the criteria names.

The criteria name must contain only alphanumeric characters and underscores and appears in
the Mappings area when a user applies the data access control to their view. It does not need
to exactly match a column name, but should guide the user to the column to which it should be
mapped.

Operator Column Select the permissions entity column that contains operators.

First Value Column Select the permissions entity column that contains the lower bound values to apply to the
operators.

This value acts as the first argument or lower bound of the operator.

Second Value Column Select the permissions entity column that contains the upper bound values to apply to the
operators.

This value acts as the second argument or upper bound of the operator.

4. Use the Available Criteria table to control which criteria should be applied.
All available criteria are listed in the table and are all selected to be applied by default. To remove a
criterion, deselect it in the list.
5. Click Save and then Deploy to deploy your data access control and make it available for use.

Integrating Data and Managing Spaces in SAP Datasphere


276 PUBLIC Securing Data with Data Access Controls
For information about attaching a data access control to a view, see Apply a Data Access Control to a
Graphical or SQL View.

 Note

You can use the View as User tool in the Data Viewer panel to review the effects of the data access
controls you apply by checking the records that another user will be allowed to see (see Viewing Object
Data).

6. The tools in the editor toolbar help you work with your object throughout its lifecycle:

Tool Description

 (Save) Save your changes to the design-time repository. You can use Save As to create a copy of the
object.

See Saving and Deploying Objects.

 (Deploy) Deploy your changes to make them available in the run-time environment.

See Saving and Deploying Objects.

 (Export) Export the object to a CSN/JSON file.

See Exporting Objects to a CSN/JSON File.

 (Impact and Open the Impact and Lineage Analysis graph for the object.
Lineage Analysis)
See Impact and Lineage Analysis.

Status [read-only] Displays the status of the object.

See Saving and Deploying Objects.

Versions Open the Version History dialog for the object.

See Reviewing and Restoring Object Versions.

Details Toggles the display of the Properties panel.

9.3 Create a "Hierarchy" Data Access Control

Users with the DW Space Administrator role (or equivalent privileges) can create data access controls in which
criteria are defined as hierarchy values. Each user can only see the records that match the hierarchy values she
is authorized for in the permissions entity, along with any of their descendents.

Context

Before creating your data access control, you must have identified the following entities:

• A permissions entity containing the following columns:

Integrating Data and Managing Spaces in SAP Datasphere


Securing Data with Data Access Controls PUBLIC 277
• User ID column - Containing user ids in the format required by your identity provider (email addresses,
logon names, or other identifiers). If you are using SAML authentication, this column must contain
values in the form defined as your User Attribute / IdpUserID (see Enabling a Custom SAML Identity
Provider).

 Note

If a user has no entries in the permissions entity, then they will not have access to any records in
the protected view.

• Criterion column - A column containing the criterion data. This criterion data is organized into a
hierarchy via an external entity with a semantic usage of Hierarchy.
• An entity with a semantic usage of Hierarchy containing parent-child relationships for the records in the
permissions entity.

 Note

Only external hierarchies with a single pair of parent-child columns are supported. Level-based
hierarchies, dimensions with internal hierarchies, and entities with a semantic usage of Hierarchy with
Directory cannot be used, and the data in your hierarchy must respect the following rules:
• A single root node with a parent value of null
• No nodes with multiple parents
• No circular relationships

For example:

• The Geo Permissions table is used as a permissions entity, and contains the following records:

User ID Geo Criteria

bob@acme.com Europe

jim@acme.com France

ann@acme.com Paris

• The Geo Hierarchy table has a semantic usage of Hierarchy and contains the following records:

Child Geo Location Parent Geo Location

World

Europe World

France Europe

Paris France

• The Geo Hierarchy data access control uses the Geo Permissions table as a permissions entity and,
when it is applied to a view, ensures that the following users will only have access to records with the
appropriate geo location values:
• bob - Europe, France, or Paris.
• jim - France or Paris.
• ann - Paris.

Integrating Data and Managing Spaces in SAP Datasphere


278 PUBLIC Securing Data with Data Access Controls
Procedure

1. In the side navigation area, click  (Data Builder), select a space if necessary, and click New Data Access
Control to open the editor.
2. Complete the properties in the General Section:

Property Description

Business Name Enter a descriptive name to help users identify the object. This name can be changed at any
time.

Technical Name Displays the name used in scripts and code, synchronized by default with the Business Name.

To override the default technical name, enter a new one in the field. Technical names can
contain only alphanumeric characters and underscores.

 Note
Once the object is saved, the technical name can no longer be modified.

Package Select the package to which the object belongs.

Packages are used to group related objects in order to facilitate their transport between
tenants.

 Note
Once a package is selected, it cannot be changed here. Only a user with the DW Space
Administrator role (or equivalent privileges) can modify a package assignment in the
Packages editor.

For more information, see Creating Packages to Export [page 292].

Structure Select Hierarchy.

Permissions Entity Select the table or view containing your user ids and criteria.

The permissions entity must match the structure selected in the Structure field. Click the Open
in New Tab button to the right of the field to open the entity in its own editor.

Identifier Column Select a column containing user ids in the format required by your identity provider (email
addresses, logon names, or other identifiers).

3. Complete the properties in the Criteria section:

Property Description

Business Name Enter a descriptive name to help users identify the object. This name can be changed at any
time.

Integrating Data and Managing Spaces in SAP Datasphere


Securing Data with Data Access Controls PUBLIC 279
Property Description

Technical Name Displays the name used in scripts and code, synchronized by default with the Business Name.

To override the default technical name, enter a new one in the field. Technical names can
contain only alphanumeric characters and underscores.

 Note
Once the object is saved, the technical name can no longer be modified.

Criterion Column Select the column containing the authorized nodes from the permissions entity.

Hierarchy Entity Select the entity with a semantic usage of Hierarchy that defines the parent-child hierarchy
relationships between the records in the permissions entity.

4. Click Save and then Deploy to deploy your data access control and make it available for use.

For information about attaching a data access control to a view, see Apply a Data Access Control to a
Graphical or SQL View.

 Note

You can use the View as User tool in the Data Viewer panel to review the effects of the data access
controls you apply by checking the records that another user will be allowed to see (see Viewing Object
Data).

5. The tools in the editor toolbar help you work with your object throughout its lifecycle:

Tool Description

 (Save) Save your changes to the design-time repository. You can use Save As to create a copy of the
object.

See Saving and Deploying Objects.

 (Deploy) Deploy your changes to make them available in the run-time environment.

See Saving and Deploying Objects.

 (Export) Export the object to a CSN/JSON file.

See Exporting Objects to a CSN/JSON File.

 (Impact and Open the Impact and Lineage Analysis graph for the object.
Lineage Analysis)
See Impact and Lineage Analysis.

Status [read-only] Displays the status of the object.

See Saving and Deploying Objects.

Versions Open the Version History dialog for the object.

See Reviewing and Restoring Object Versions.

Details Toggles the display of the Properties panel.

Integrating Data and Managing Spaces in SAP Datasphere


280 PUBLIC Securing Data with Data Access Controls
9.4 Import SAP BW and SAP BW∕4HANA Analysis
Authorizations

You can import analysis authorizations defined in SAP BW and SAP BW∕4HANA systems into SAP Datasphere
to provide row-level protection for data imported from these systems.

This topic contains the following sections:

• Introduction [page 281]


• Importing Permissions [page 282]

Introduction

This feature is supported only for SAP BW v7.5 SP16 and higher and SAP BW∕4HANA v2.0 and higher systems
accessed via one of the following connection types:

• SAP ABAP (see SAP ABAP Connections [page 67])


• SAP BW (see SAP BW Connections [page 77])
• SAP BW∕4HANA Model Transfer connection (see Importing SAP BW∕4HANA Models)

An SAP BW or SAP BW∕4HANA administrator must have created the report that generates the
RSDWC_RSEC_DAC permissions table on the source system (see SAP Note 3062381 ). This SAP note also
includes instruction on how to implement BAdI RSDWC_DAC_RSEC_USER_UPDATE for situations where your
SAP BW or SAP BW∕4HANA instance does not use email addresses as user IDs.

 Note

This SAP note has been updated to support importing permissions via an SAP BW∕4HANA Model Transfer
connection. If you have previously used this feature, you must follow the updated steps to ensure you have
the latest version of the report.

The objects to be protected must have already been imported into your space. For objects imported from:

• SAP ABAP or SAP BW connections - Only analysis authorizations applied to characteristics used on
InfoProviders and CompositeProviders can be imported with this wizard.

 Note

Analysis authorizations applied to queries used as InfoProviders cannot be imported.

• SAP BW∕4HANA Model Transfer connections - Analysis authorizations applied to navigation attributes can
also be imported.

See the following blogs for additional information about this feature:

• For a general introduction, see Introducing Remote Authorizations from SAP BW/4HANA for SAP
Datasphere (updated February 2022).
• For information about leveraging SAP HANA-native analytical privileges in an open SQL schema in case of
slow-running views or out of memory issues, see HANA Analytical Privileges in SAP Datasphere – Part 1
and Part 2 (published January 2023).

Integrating Data and Managing Spaces in SAP Datasphere


Securing Data with Data Access Controls PUBLIC 281
Importing Permissions

The Import Permissions wizard will import the permissions table from the source system, create additional
objects to calculate the permissions, and apply a data access control to each selected object.

1. In the side navigation area, click  (Data Builder), select a space if necessary, and click Import Import
Permissions to open the Import Permissions wizard.
2. Select the connection from which you want to import permissions and click Next.
3. Import or accept the permissions table for this connection. If the permissions table is:
• Not imported - Enter a Business Name and Technical Name and click Import and Deploy. Once the
table has been imported and deployed, click Next.

 Note

You cannot choose the name of the permissions table for a SAP BW∕4HANA Model Transfer
connection.

• Imported - Click Next.


• Not generated on the server - Click Cancel. You cannot continue without importing the table.

 Note

For SAP BW∕4HANA Model Transfer connections, if SAP Note 3062381 has been correctly
applied and the wizard cannot find the permissions table, verify that the user specified in
your connection has permission to access the permissions table RSDWC_RSEC_DAC. Your SAP
BW∕4HANA administrator may need to add this permission to the user or provide an alternative
user with the necessary permission.

 Note

We recommend that the report generating the permissions table in SAP BW∕4HANA is run at least
once a day and the remote table in SAP Datasphere is kept in remote (federated) access to ensure
that it is always up-to-date. If you decide to replicate the permissions table, you should schedule at
minimum a daily refresh. Real-time replication is not supported for this table.

4. Select the objects for which you want to import permissions, and then click Next.
Only those objects that have previously been imported from this connection, and which can be protected,
are listed.
5. Review the summary of the objects that will be created in SAP Datasphere.
For each object to be protected, new objects are created:

Integrating Data and Managing Spaces in SAP Datasphere


282 PUBLIC Securing Data with Data Access Controls
For each selected remote table, the following objects are For each selected view, the following objects are created
created: or modified:
• <table> - Permissions Script • <view> - Base View (<view>_B) - A copy of
(<table>_P) - An SQL view to calculate the permis- the view to protect, used for generating the permis-
sions to apply to the object. sions script view.

• <table> - Permissions (<table>_D) - A • <view> - Permissions Script


data access control to consume the SQL view. (<view>_P) - An SQL view to calculate the permis-
• <table> - Protected (<table>_V) - An SQL sions to apply to the object.
view to wrap the table selected for protection and • <view> - Permissions (<table>_D) - A data
apply the data access control. Once the import is access control to consume the SQL view, which will
complete, you should only access the table via this be applied to the view selected for protection.
view. • <view> (<view>) - The original view is protected
by applying the data access control to it and can con-
tinue to be used.

 Note

If you have previously protected some or all of the selected objects, they will appear in the wizard with a
Status/Action of Overwrite and will be overwritten.

6. Click Import and Deploy to create, import, and deploy all these objects.
7. Click Finish to exit the wizard and perform all the actions necessary to protect the data.
All the objects are listed in the Repository Explorer and, where appropriate, also in the Data Access Controls
and Data Builder start pages.

Integrating Data and Managing Spaces in SAP Datasphere


Securing Data with Data Access Controls PUBLIC 283
9.5 Process Source Changes in the Data Access Control
Editor

If the permissions entity that is consumed by your data access control is modified, then the next time you open
the data access control, you will be asked to process the changes. If the source change has generated warnings
or errors in your data access control, its status will be updated and you will receive a notification inviting you to
review the changes.

Context

Your object's status may be updated when changes to one or more of its sources are saved or deployed:

• If no warnings or errors were generated by the source changes, then the status will not be changed.
• If warnings or errors were generated by the source changes and:
• Source changes were saved but not deployed, your object's status will be set to Design-Time Error.
• Source changes were saved and deployed, your object's status will be set to Run-Time Error.

Procedure

1. If a source change has generated warnings or errors in your view, you will receive a notification inviting you
to review them, and you can click the notification to open it directly.

The Source Updates dialog opens, listing the sources that have been modified.
2. Click OK to dismiss the dialog. The following validation messages may be displayed:

Change Impact in Dependent Data Access Controls

Add Column Adding columns to the permissions entity has no impact on the data access control.

Change Business Name Modifying the business name of the permissions entity or any of its columns has no
impact on the data access control

Change Column Data Type Source column data type changes are not allowed for columns that are used by a data
access control.

Integrating Data and Managing Spaces in SAP Datasphere


284 PUBLIC Securing Data with Data Access Controls
Change Impact in Dependent Data Access Controls

Delete Column Deleting columns from the permissions entity generates errors in dependent data ac-
cess controls if the columns are used:
• If the column was selected as the Identifier Column, an error message is displayed.
You must specify a new column in this field.
• If the column was selected as a column in the Criteria list, an error message is
displayed. You must remove this column from the Criteria list.

 Note
You cannot add new criteria columns to a data access control that is used by
views. If this change is necessary, we recommend that you create a new data
access control with the additional columns and use it to replace the existing
data access control in the views to which it is attached.

3. Once you have processed the modifications to the Identifier Column or Criteria list, and if your data access
control is applied to any views, warning messages are displayed urging you to review the impacted views to
ensure that the changes have not compromised row level security.
4. Click  (Save) to save the changes to your object and dismiss any info and warning messages.

If the source changes were saved but not deployed, your object's status should change from Design-Time
Error to Changes to Deploy.

 Note

As the new version of your object depends on the source updates, you should coordinate with the
person who has modified the source about deployment. The new version of your object will work
correctly in the run-time only when both it and its source are deployed.

5. Click  (Deploy) to deploy your object.

Your object's status should change from Changes to Deploy (or Run-Time Error) to Deployed.

Integrating Data and Managing Spaces in SAP Datasphere


Securing Data with Data Access Controls PUBLIC 285
10 Importing SAP and Partner Business
Content from the Content Network

Users with an administrator role or space administrator role, can use the Semantic Onboarding app to import
business content and sample content from SAP and partners published to the Content Network.

Context

To import business content and sample content from SAP and partners via the Semantic Onboarding app, you
must have either:

A global role that allows you to update any space and to A combination of a global role and a scoped role:
create new spaces if necessary, by granting you the following
• A global role that grants you the following privilege:
privileges:
• Data Warehouse General (-R------) - To access
• Data Warehouse General (-R------) - To access SAP SAP Datasphere.
Datasphere.
• Lifecycle (-----M--) - To use the Content
• Lifecycle (-----M--) - To use the Content Network Network tiles in the Semantic Onboarding app.
tiles in the Semantic Onboarding app.
• A scoped role that grants you access to the space or
• Spaces (-------M) - To update all spaces and space spaces to import into with the following privileges:
properties.
• Data Warehouse General (-R------) - To access
• Spaces (C-------) - To create spaces. SAP Datasphere.
• Spaces Files (-------M) - To create, read, update, and • Spaces (--U-----) - To update your spaces and
delete all objects in all spaces. their properties.
The DW Administrator role, for example, grants these privi- • Spaces Files (CRUD----) - To create, read, up-
leges. date, and delete objects in your spaces.
• Data Warehouse Data Builder (CRUD--S-) - To
create, read, update, delete and share data builder
objects (and any other relevant object privileges
to allow you to create and update other types of
objects contained in the package).

The DW Space Administrator global and scoped role tem-


plates, for example, grant this combination of privileges.

For more information, see Privileges and Permissions and Standard Roles Delivered with SAP Datasphere.

 Note

Use of sample content and business content is optional and not part of the business functionality of the
product. Content packages may be discontinued or replaced at any time.

Detailed documentation for all business content is available at SAP Datasphere Business Content.

Integrating Data and Managing Spaces in SAP Datasphere


286 PUBLIC Importing SAP and Partner Business Content from the Content Network
 Note

The Content Network is not available for trial users.

We also provide content for download at our Community Content site.

Procedure

1. In the side navigation area, click  (Semantic Onboarding), and then click one of the following tiles:

• Business Content: End-to-end business scenarios created by SAP for various industries and lines of
business.
• 3rd Party Content: End-to-end business scenarios created by SAP partners. Some third-party content
has to be purchased in the SAP Store.
• Samples: Not used in SAP Datasphere.
2. Click a package to open it and see an overview of its content.
3. Review the How to install this package section to understand the requirements for installing the content.
You may be instructed to:
• Create a space with a specific business and technical name before importing the content.
• Create a connection in the new space.
• Visit the SAP Store and, in some cases, make a payment to a partner.
• Contact a partner to complete the installation or obtain data.
4. Review the Import Overview section. For more details, and to modify any of the options, click the Import
Options tab:

Section Description

Import Schedule Specifies what happens when the package is updated:


• None - No updates are made on your system. This option is automatically selected
for packages that cannot be updated.
• Notify for new content - You receive a notification and can choose whether to import
the new version of the package to your system.

Integrating Data and Managing Spaces in SAP Datasphere


Importing SAP and Partner Business Content from the Content Network PUBLIC 287
Section Description

Overwrite Preferences Controls whether objects that are already present in your space can be overwritten
during an import.

 Note
Only object definitions can be transported. Data cannot be transported between
SAP Datasphere tenants via the  (Transport) app.

You can choose between:


• Don’t overwrite objects or data - [default] Imports only objects that are not already
present in your space. Objects that are already present are not updated, and are
listed with warnings in the Import window.
• Overwrite data only - Not supported for SAP Datasphere.
• Overwrite objects and data - Imports all objects from the package, even if certain
objects are already present in your space. These objects may be updated, and these
updates may impact data (for example, if a column is deleted).

 Note
• The Including permissions option is not supported for SAP Datasphere.
• Connections are never overwritten by an import.
• When overwriting objects with a release state, the objects you are import-
ing must have the same release state or another release state that is con-
sistent with their lifecycle (see Releasing Stable Views for Consumption).

Deployment Preference Check the Deploy after import checkbox to automatically deploy objects after import. If
you have selected:
• Don't overwrite objects or data - Only newly imported objects will be deployed.
• Overwrite objects and data - All imported objects in the space will be deployed.

 Note
This option does not support packages that contain objects from one or more
spaces where objects are shared from one space to another (see Sharing Entities
and Task Chains to Other Spaces. In this case, you should manually deploy the
objects from the source space and then deploy the objects from the target space.

Integrating Data and Managing Spaces in SAP Datasphere


288 PUBLIC Importing SAP and Partner Business Content from the Content Network
Section Description

Content Select the space or spaces to import your content into in the Space Mapping field. You
must choose one space for each space listed in the package:
• You can select any space for which you have the Spaces.Update permission to
import the space's content into.
If you have the Spaces.Manage permission, then you can choose to import into any
space in your tenant.
• If the space does not exist in your tenant and you have the Spaces.Create permis-
sion, then you are invited to create it by default.
• You cannot "merge" content from two or more spaces included in your package. You
must select a separate space for each space listed.

 Note
Choosing space mappings is only possible if the source objects for any shared
objects are included in the package. For example if your package contains shared
objects in Space A that have been shared from Space B, then the source objects
from Space B must be included in the package. If the sources of any shared
objects cannot be resolved in the package, then the package must be imported into
spaces with the same technical names as the spaces listed in the package.

Review the objects to be imported. All objects are imported by default.

You can select and deselect individual objects, or types of objects. The Impacted Object
column shows warnings for objects that will overwrite content on your system.

5. Click Import to begin importing the content.

You will receive a notification when the process is complete, or if an error occurs.

If you selected Don’t overwrite objects or data, a notification will let you know if any of the content wasn’t
imported to avoid overwriting existing content.

To view the log of updates and imports for a package, select the package in the content network and
choose  History.

Integrating Data and Managing Spaces in SAP Datasphere


Importing SAP and Partner Business Content from the Content Network PUBLIC 289
11 Transporting Content Between Tenants

Users with an administrator or space administrator role can use the Transport app to transfer content between
tenants via a private cloud storage area.

The following object types can be exported and imported via the  (Transport) app:

 Note

Only object definitions can be transported. Data cannot be transported between SAP Datasphere tenants
via the  (Transport) app.

Object Type Details

Connections The definition of a connection does not have dependencies on any other objects.

See Create a Connec-


tion [page 16].

Remote Tables The definition of a remote table contains information about its connection.

See Import Remote Ta-


bles.

Local Tables The definition of a local table contains the structure of the table only, and does not have depend-
encies on any other objects.
See Creating a Local
Table.

Flows The definition of a data flow, replication flow, or transformation flow contains the definitions of all
its sources and its target table. When you export a flow, these objects are exported too.
See Creating a Data
Flow, Creating a Rep-
lication Flow, and Cre-
ating a Transformation
Flow.

Views The definition of a view contains the definitions of all its sources and any used data access
controls. When you export a view, these objects are exported too.
See Creating a Graphi-
cal View and Creating
an SQL View.

Intelligent Lookups The definition of an intelligent lookup contains the definitions of its input and lookup entities.
When you export an intelligent lookup, these entities are exported too.
See Creating an Intelli-
gent Lookup.

Analytic Models The definition of an analytic model contains the definitions of its fact and dimension sources.
When you export an analytic model, these entities are exported too.
See Creating an Ana-
lytic Model.

Integrating Data and Managing Spaces in SAP Datasphere


290 PUBLIC Transporting Content Between Tenants
Object Type Details

E/R Models The definition of an E/R model does not include the objects that it visualizes as dependencies.
These objects must be selected manually.
See Creating an Entity-
Relationship Model.

Data Access Controls The definition of a data access control contains the definition of its permissions entity. When you
export a data access control, the permissions entity is exported too.
See Securing Data with
Data Access Controls
[page 268].

Task Chains The definition of a task chain does not include the objects that it automates. These objects must
be selected manually.
See Creating a Task
Chain.

Business Entities / The definition of a business entity contains all its versions (which are also listed separately), along
Business Entity Ver- with the definition of its source data entity and any authorization scenarios. When you export a
sions business entity (or one of its versions), these objects are exported too.

See Creating a Busi-


ness Entity.

Fact Models The definition of a fact model contains all its versions (which are also listed separately), along with
the definition of all its source fact models and business entities. When you export a fact model (or
See Creating a Fact
one of its versions), these objects are exported too.
Model.

Consumption Models The definition of a consumption model contains all its perspectives (which are also listed sepa-
rately), along with the definition of all its source fact models and business entities. When you
See Creating a Con-
export a consumption model (or one of its perspectives), these objects are exported too.
sumption Model.

Authorization Scenar- The definition of an authorization scenario contains the definition of its data access control. When
ios you export an authorization scenario, the data access control is exported too.

See Creating an Au-


thorization Scenario.

 Note

Only local and remote tables, views, data flows, and E/R models can be selected for exporting in SAP
Datasphere tenants provisioned prior to version 2021.03. While objects created in SAP Datasphere tenants
provisioned prior to version 2021.03 can be transported to tenants provisioned after that date, transport of
objects in the other direction is not possible.

 Note

You can also export content from and import content to your space via:

• Export to CSN/JSON File buttons in selected Data Builder editors (see Importing and Exporting Objects
in CSN/JSON Files).
• The datasphere command line interface objects commands (see Manage Modeling Objects via the
Command Line).

Integrating Data and Managing Spaces in SAP Datasphere


Transporting Content Between Tenants PUBLIC 291
For information about importing business content in the form of end-to-end business scenarios for specific
industries and lines of business provided by SAP and its partners via the  (Semantic Onboarding) app, see
Importing SAP and Partner Business Content from the Content Network [page 286].

11.1 Creating Packages to Export

Users with the DW Space Administrator role can create packages to model groups of related objects for
transport between tenants. Modelers can add objects to packages via the Package field, which appears in
editors when a package is created in their space. Once a package is complete and validated, the space
administrator can export it to the Content Network. The structure of your package is preserved and, as the
objects it contains evolve, you can easily export updated versions of it.

This topic contains the following sections:

• Introduction to Packages [page 292]


• Create a Package [page 293]
• Export a Package [page 295]
• Update a Package [page 295]
• Delete a Package [page 296]

Introduction to Packages

Each package must contain a complete and coherent set of objects:

• A package can contain one or more objects from one space.


• Each object can belong to only one package.
• A package must:
• Contain the complete lineage of all the objects that belong to it, or
• Include other packages containing all or part of this lineage as required packages.
• Required packages can be selected from other spaces.
If an object in a package has, as part of its lineage, an object shared from another space, then the shared
object must be included in a package defined in its origin space, which is then added as a required package.

Integrating Data and Managing Spaces in SAP Datasphere


292 PUBLIC Transporting Content Between Tenants
In our example:

• Package A1 - Contains T1 and V1, where V1 has a single source, T1.


• Package B1 - Contains T2 and V2, where V2 has two sources, T2 and V1.
• Package B1 adds Package A1 as a required package.
• Package B2 - Contains T3, and V3, where V3 has two sources, T3 and V2.
• Package B2 adds Package B1 as a required package.

Create a Package

Users with the DW Space Administrator role can create packages:

1. In the side navigation area, click  (Transport)  (Packages) to open the list of packages.
Use the Space filter field to show packages from all spaces or only one space.
2. Click Add and specify a space, if necessary, to create a new package and open it in the editor.
3. Enter the following properties on the General tab:

Integrating Data and Managing Spaces in SAP Datasphere


Transporting Content Between Tenants PUBLIC 293
Property Description

Business Name Enter a descriptive name to help users identify the object. This name can be changed at
any time.

Technical Name Displays the name used in scripts and code, synchronized by default with the Business
Name.

To override the default technical name, enter a new one in the field. Technical names can
contain only alphanumeric characters and underscores.

 Note
Once the object is saved, the technical name can no longer be modified.

Business Purpose Provide a description, purpose, contacts, and tags to help other users understand the
package.

Category The only available category is My Content.

Target Location Select the location within My Content to export to.

Version Enter the version number of the package in the format 1.0.0 (major version, minor
version, patch number).

You can re-export your package with the same version number to overwrite the current
exported version or change the version to export a package update. The version number
must always go up and never down.

Status Displays the current status of the package.

Possible statuses are:


• Not Exported - Initial state.
• Exporting - Export is in progress. The package and its contents are locked.
• Exported - The latest version of the package is exported.
• Changes to Export - One or more of the objects in the package is new or has been
updated. You can re-export at any time.
• Design Time Error - For one or more of the objects in the package, the dependen-
cies of these objects cannot be resolved in the package. Add missing dependencies
to the package or contact a space administrator to ask them to review and resolve
the situation (see Add an Object to a Package).

Current Exported Version Displays the last exported version of the package.

4. [optional] Add any packages that contain objects that the objects you add to your package depend on as
part of their lineage.
1. Click the Required Packages tab.
2. Click Add to open the Select Packages dialog, which lists all packages you have permission to see.
The list includes packages created in your space and packages shared to your space.
3. Select one or more packages and click Select to add them as required packages.
5. Add objects from your space to the package.
1. Click the Objects tab.
2. Click Add to open the Add Objects dialog, which lists all objects in your space.

Integrating Data and Managing Spaces in SAP Datasphere


294 PUBLIC Transporting Content Between Tenants
3. Select the objects you want to add to your package.
As objects can only be added to one package, any objects that have been added to a different package
cannot be selected.
4. Click Next and wait for the dialog to analyze your objects and extend your selection by adding any
sources and other required dependencies.
Each object must be able to resolve its entire lineage either by adding the objects it depends on to the
package or by referencing them via a required package.
5. Click Add Objects to add the extended object selection to your package.
You can add your objects even if errors are shown. In general errors can be resolved by adding the
package containing the objects that cannot be added as a required package.
6. Click Save to save your package and validate all dependencies.
• If no errors are found, your package is ready to be exported.
• If errors are found, you should try to resolve them by adding required packages or objects or by
removing objects whose dependencies cannot be resolved, and then re-saving.
Once your package is created, it becomes available to modelers in your space who can add further objects
to it (see Packages).

 Note

Though modelers can add objects to a package in the object editors, they cannot remove objects from
a package.

Export a Package

Users with the DW Space Administrator role can export packages:

1. Verify that your package is ready to export:


• Your package must be validated.
• All deployable objects in your package must have a status of Deployed (see Deploy Objects).
• Any required packages must have a status of Exported.
2. Click Export to export your package to the Content Network with the filename
<TechnicalName_Version>.
When the export is complete, the package status changes to Exported and you will receive a notification
confirming the export.
3. [optional] Review your exported package in the Export app by navigating to the target location (see
Exporting Content for Sharing with Other Tenants [page 296]).

Update a Package

Once your package is exported, you can continue to evolve it by adding, removing, or updating objects.
Modelers in your space may also deploy changes to the objects it contains and add new objects to it.

When any updates are made to a package that has been exported, its status changes to Changes to Export. You
can choose at any time to export these changes. If you:

Integrating Data and Managing Spaces in SAP Datasphere


Transporting Content Between Tenants PUBLIC 295
• Leave the version number unchanged, your new export will overwrite the previous exported version.
• Set a new, higher version number, a new package version is exported.

Delete a Package

Select a package and click Delete to delete it.

You cannot delete a package if it is listed among the Required Packages for any other package.

11.2 Exporting Content for Sharing with Other Tenants

You can use the Transport app to export content from one or more spaces for sharing with other tenants.

Prerequisites

To export content from a space via the Export app, you must have either:

A global role that allows you to access and export content A combination of a global role and a scoped role:
from any space, by granting you the following privileges:
• A global role that grants you the following privilege:
• Data Warehouse General (-R------) - To access SAP • Data Warehouse General (-R------) - To access
Datasphere. SAP Datasphere.
• Lifecycle (-----M--) - To use the Transport apps. • Lifecycle (-----M--) - To use the Transport apps.
• Spaces (-------M) - To update all spaces and space • A scoped role that grants you access to the space or
properties. spaces to export from with the following privileges:
• Spaces Files (-------M) - To create, read, update, and • Data Warehouse General (-R------) - To access
delete all objects in all spaces. SAP Datasphere.
The DW Administrator role, for example, grants these privi- • Spaces Files (-R------) - To view specific spaces
leges. and their contents.
• Data Warehouse Data Builder (-R------) - To
view Data Builder objects (and any other relevant
object privileges to allow you to create and update
other types of objects contained in the package).

The DW Space Administrator global and scoped role tem-


plates, for example, grant this combination of privileges.

For more information, see Privileges and Permissions and Standard Roles Delivered with SAP Datasphere.

Integrating Data and Managing Spaces in SAP Datasphere


296 PUBLIC Transporting Content Between Tenants
Context

You can share content across tenants from one or more landscapes and set access rights for each sharing
destination, for example, to allow other tenants to update a package and export it again.

Each tenant can store up to 2 GB of exported content for free in the My Content area.

The following object types can be exported and imported via the  (Transport) app:

 Note

Only object definitions can be transported. Data cannot be transported between SAP Datasphere tenants
via the  (Transport) app.

Object Type Details

Connections The definition of a connection does not have dependencies on any other objects.

See Create a Connec-


tion [page 16].

Remote Tables The definition of a remote table contains information about its connection.

See Import Remote Ta-


bles.

Local Tables The definition of a local table contains the structure of the table only, and does not have depend-
encies on any other objects.
See Creating a Local
Table.

Flows The definition of a data flow, replication flow, or transformation flow contains the definitions of all
its sources and its target table. When you export a flow, these objects are exported too.
See Creating a Data
Flow, Creating a Rep-
lication Flow, and Cre-
ating a Transformation
Flow.

Views The definition of a view contains the definitions of all its sources and any used data access
controls. When you export a view, these objects are exported too.
See Creating a Graphi-
cal View and Creating
an SQL View.

Intelligent Lookups The definition of an intelligent lookup contains the definitions of its input and lookup entities.
When you export an intelligent lookup, these entities are exported too.
See Creating an Intelli-
gent Lookup.

Analytic Models The definition of an analytic model contains the definitions of its fact and dimension sources.
When you export an analytic model, these entities are exported too.
See Creating an Ana-
lytic Model.

Integrating Data and Managing Spaces in SAP Datasphere


Transporting Content Between Tenants PUBLIC 297
Object Type Details

E/R Models The definition of an E/R model does not include the objects that it visualizes as dependencies.
These objects must be selected manually.
See Creating an Entity-
Relationship Model.

Data Access Controls The definition of a data access control contains the definition of its permissions entity. When you
export a data access control, the permissions entity is exported too.
See Securing Data with
Data Access Controls
[page 268].

Task Chains The definition of a task chain does not include the objects that it automates. These objects must
be selected manually.
See Creating a Task
Chain.

Business Entities / The definition of a business entity contains all its versions (which are also listed separately), along
Business Entity Ver- with the definition of its source data entity and any authorization scenarios. When you export a
sions business entity (or one of its versions), these objects are exported too.

See Creating a Busi-


ness Entity.

Fact Models The definition of a fact model contains all its versions (which are also listed separately), along with
the definition of all its source fact models and business entities. When you export a fact model (or
See Creating a Fact
one of its versions), these objects are exported too.
Model.

Consumption Models The definition of a consumption model contains all its perspectives (which are also listed sepa-
rately), along with the definition of all its source fact models and business entities. When you
See Creating a Con-
export a consumption model (or one of its perspectives), these objects are exported too.
sumption Model.

Authorization Scenar- The definition of an authorization scenario contains the definition of its data access control. When
ios you export an authorization scenario, the data access control is exported too.

See Creating an Au-


thorization Scenario.

 Note

Only local and remote tables, views, data flows, and E/R models can be selected for exporting in SAP
Datasphere tenants provisioned prior to version 2021.03. While objects created in SAP Datasphere tenants
provisioned prior to version 2021.03 can be transported to tenants provisioned after that date, transport of
objects in the other direction is not possible.

Procedure

1. In the side navigation area, click  (Transport)  (Export) and then click  New Export.

Integrating Data and Managing Spaces in SAP Datasphere


298 PUBLIC Transporting Content Between Tenants
 Note

You can, alternatively, use thePackages app to prepare and export packages (see Creating Packages to
Export [page 292]).

2. On the Choose content page, expand the SPACE folder to show your spaces, select the objects you want to
export, and then click Next.

You can select a space to include all its objects, or expand it to select individual objects within it.

You can select objects from any space of which you are a member, and select objects from more than one
space.

If an object depends on other objects, then they will also be selected. For example, when you select a view,
all its sources and any used data access controls will be selected.

 Note

If an object selected for export depends on one or more objects that have been shared from other
spaces (see Sharing Entities and Task Chains to Other Spaces), then you must be a member of each of
those spaces in order to include all the necessary objects in your package.

If you are not a member of one or more sharing spaces, then you can still export your package (no error
will be displayed), but it will fail to import correctly unless all the missing objects are already present in
the target tenant.

You can review all the dependencies of your object by using the Impact and Lineage Analysis tool (see
Impact and Lineage Analysis).

3. On the Set package properties page, complete the properties, and then click Next.

Property Description

Name/ Description Enter a clear name to identify the package.

Details Enter more detailed information about the contents of the package, including terms and
conditions, if appropriate.

Category Select My Content.

Location By default, the package is exported to the My Content folder. To export to a subfolder,
click Browse and select or create a new folder.

Sharing Select the sharing destinations for the export and their permissions on the content (see
Adding Sharing Destinations [page 300]).

Include Data Audit Not used in SAP Datasphere.

Terms and Conditions Enable this switch to require users to accept terms and conditions before importing the
content.

4. On the Review and export page, review the overview of your package.

If there are any problems, select the warning icon  to see how to fix them.
5. Click Export to export your package.

You can check the progress of your export in the Notifications list. You'll also receive a message when it is
complete, and you can check the Export Summary to see the status of all the objects in the package.

Integrating Data and Managing Spaces in SAP Datasphere


Transporting Content Between Tenants PUBLIC 299
When the export completes successfully, the package becomes available in the Transport / Import / My
Content section of the Import app for the tenants you added.

You can edit the content, share it with other tenants, or delete it from the Transport / Export / My Content
page (see Managing Exported Content [page 301]).

11.2.1 Adding Sharing Destinations

You can add sharing destinations to your export and specify their level of permissions for the content in the
Package Properties page of the Export app.

Procedure

1. Click the  (Add destination) button above the Destinations list to open the Destinations dialog.
2. Complete the dialog as follows:

Property Description

Destination Type Select the category of the destination that you want to add:
• System: Add an individual tenant by URL. Select a Landscape, and copy the URLs to
the Systems list.

 Note
SAP Datasphere is only available on the Cloud Foundry platform. You can share
to tenants that use the same type of Cloud Foundry data center as yours. These
host systems use two digits, such as eu10 or us30.

• Customer: Add all tenants that belong to one customer based on their ERP ID. Type
the ID in the Customer field.

Access Select one or more permissions that you want to grant to the sharing destination:
• View: Your package will appear in the My Content page for the destination tenants,
and their users will be able to view the name and summary.
• Import: The tenants' users can import the package from the My Content list.
• Edit: Your package will appear in the Manage Packages page for the destination
tenants, where users can make changes to the package and export it again.
• Delete: From the Manage Packages page, the tenants' users can delete the package.

3. Click Add to add the destination to the Added Destinations list.

You can add multiple destinations and edit access or remove destinations from the list.
4. Select OK to finish adding destinations and return to the wizard.

Integrating Data and Managing Spaces in SAP Datasphere


300 PUBLIC Transporting Content Between Tenants
11.2.2 Managing Exported Content

If you've exported packages for sharing with other tenants, or have edit or delete permissions for another
tenant's exports, you can work on them in the Transport / Export / My Content page.

In the side navigation area, click  (Transport)  (Export) and then click the My Content folder to
display the list of packages that you have exported and those from other systems that you have permission to
edit or delete.

For packages:

• To get an overview of a package, select its name.


• To edit the contents or details of a package and export it again, select the checkbox next to it and select
 Edit. If you need to change the content, set the Modify Content switch to ON from the Choose Content
page. Otherwise, leave this switch off. Then, follow the steps in Exporting Content for Sharing with Other
Tenants [page 296].
• To change the sharing destinations for a package, select the checkbox next to it and select  Share.
Follow the steps for setting access rights in Exporting Content for Sharing with Other Tenants [page 296].
• To delete packages, select the checkboxes next to them and choose  Delete. The packages are removed
and can't be imported or edited from any tenant. Each tenant that imported the package will still have
access to the content itself, though.

For folders:

• To create a new folder, select  Create New Folder and type a name and description.
• To share a folder and all its contents, select the checkbox next to a folder and select  Share. See the
access steps in Exporting Content for Sharing with Other Tenants [page 296] for more information.
• To delete a folder and all its contents, select the checkbox next to a folder and select  Delete. The folder
and its contents will be removed for any tenants they were shared with.

11.3 Importing Content from Another Tenant

You can use the Transport app to import content that has been shared from another tenant.

Context

To import content into a space via the Import app, you must have either:

Integrating Data and Managing Spaces in SAP Datasphere


Transporting Content Between Tenants PUBLIC 301
A global role that allows you to update any space and to A combination of a global role and a scoped role:
create new spaces if necessary, by granting you the following
• Data Warehouse General (-R------) - To access SAP
privileges:
Datasphere.
• Data Warehouse General (-R------) - To access SAP • A global role that grants you the following privilege:
Datasphere.
• Data Warehouse General (-R------) - To access
• Lifecycle (-----M--) - To use the Transport apps. SAP Datasphere.
• Spaces (-------M) - To update all spaces and space • Lifecycle (-----M--) - To use the Transport apps.
properties.
• A scoped role that grants you access to the space or
• Spaces (C-------) - To create spaces. spaces to import into with the following privileges:
• Spaces Files (-------M) - To create, read, update, and • Spaces (--U-----) - To update your spaces and
delete all objects in all spaces. their properties.
The DW Administrator role, for example, grants these privi- • Spaces Files (CRUD----) - To create, read, up-
leges. date, and delete objects in your spaces.
• Data Warehouse Data Builder (CRUD--S-) - To
create, read, update, delete and share data builder
objects (and any other relevant object privileges
to allow you to create and update other types of
objects contained in the package).

The DW Space Administrator global and scoped role tem-


plates, for example, grant this combination of privileges.

For more information, see Privileges and Permissions and Standard Roles Delivered with SAP Datasphere.

Procedure

1. In the side navigation area, click  (Transport)  (Import) to open the list of content available for
import.

If you don’t see expected packages to import, click the Settings button and review the import settings. You
can:
• Enable the Show content shared by all systems option and optionally exclude content from specific
tenants by adding them to the Add systems to block content sharing list.
• Disable the Show content shared by all systems option and choose specific tenants to import from by
adding them to the Add systems to allow content sharing list.
If, after checking these settings, an expected package is still not listed, contact an administrator for help.
2. Click a package to open it and see an overview of its content.
3. Review the Import Overview section. For more details, and to modify any of the options, click the Import
Options tab:

Integrating Data and Managing Spaces in SAP Datasphere


302 PUBLIC Transporting Content Between Tenants
Section Description

Import Schedule Specifies what happens when the package is updated:


• None - No updates are made on your system. This option is automatically selected
for packages that cannot be updated.
• Notify for new content - You receive a notification and can choose whether to import
the new version of the package to your system.

Overwrite Preferences Controls whether objects that are already present in your space can be overwritten
during an import.

 Note
Only object definitions can be transported. Data cannot be transported between
SAP Datasphere tenants via the  (Transport) app.

You can choose between:


• Don’t overwrite objects or data - [default] Imports only objects that are not already
present in your space. Objects that are already present are not updated, and are
listed with warnings in the Import window.
• Overwrite data only - Not supported for SAP Datasphere.
• Overwrite objects and data - Imports all objects from the package, even if certain
objects are already present in your space. These objects may be updated, and these
updates may impact data (for example, if a column is deleted).

 Note
• The Including permissions option is not supported for SAP Datasphere.
• Connections are never overwritten by an import.
• When overwriting objects with a release state, the objects you are import-
ing must have the same release state or another release state that is con-
sistent with their lifecycle (see Releasing Stable Views for Consumption).

Deployment Preference Check the Deploy after import checkbox to automatically deploy objects after import. If
you have selected:
• Don't overwrite objects or data - Only newly imported objects will be deployed.
• Overwrite objects and data - All imported objects in the space will be deployed.

 Note
This option does not support packages that contain objects from one or more
spaces where objects are shared from one space to another (see Sharing Entities
and Task Chains to Other Spaces. In this case, you should manually deploy the
objects from the source space and then deploy the objects from the target space.

Integrating Data and Managing Spaces in SAP Datasphere


Transporting Content Between Tenants PUBLIC 303
Section Description

Content Select the space or spaces to import your content into in the Space Mapping field. You
must choose one space for each space listed in the package:
• You can select any space for which you have the Spaces.Update permission to
import the space's content into.
If you have the Spaces.Manage permission, then you can choose to import into any
space in your tenant.
• If the space does not exist in your tenant and you have the Spaces.Create permis-
sion, then you are invited to create it by default.
• You cannot "merge" content from two or more spaces included in your package. You
must select a separate space for each space listed.

 Note
Choosing space mappings is only possible if the source objects for any shared
objects are included in the package. For example if your package contains shared
objects in Space A that have been shared from Space B, then the source objects
from Space B must be included in the package. If the sources of any shared
objects cannot be resolved in the package, then the package must be imported into
spaces with the same technical names as the spaces listed in the package.

Review the objects to be imported. All objects are imported by default.

You can select and deselect individual objects, or types of objects. The Impacted Object
column shows warnings for objects that will overwrite content on your system.

4. Click Import to begin importing the content.

You will receive a notification when the process is complete, or if an error occurs.

If you selected Don’t overwrite objects or data, a notification will let you know if any of the content wasn’t
imported to avoid overwriting existing content.

To view the log of updates and imports for a package, select the package in the content network and
choose  History.
5. If your content included one or more connections that didn't exist in your space before the import, then
you should:
a. In the Connections app, open each connection, complete the configuration and enter appropriate
credentials to connect to your system (see Edit a Connection [page 123]).
b. In the Repository Explorer or in the object editors, select the imported objects that depend on an
imported connection (and which could not, therefore, be automatically deployed) and click  (Deploy)
to deploy them.

 Note

If you checked the Deploy after import checkbox in the Import Options, you can find a notification
about the failed deployment of objects in the Notifications panel. Clicking the notification directly
leads you to the list of imported objects in the Repository Explorer.

Integrating Data and Managing Spaces in SAP Datasphere


304 PUBLIC Transporting Content Between Tenants
6. If you are a DW Administrator and have created a space as part of the import, you are assigned to the space
as a DW Space Administrator. You can assign other users to the space (including as space administrators)
as appropriate.

11.4 Transporting Your Content through SAP Cloud


Transport Management

Integrate SAP Datasphere with SAP Cloud Transport Management service to transport content packages
across different landscapes.

SAP Cloud Transport Management service lets you manage software deliverables between accounts of
different environments (such as Cloud Foundry, ABAP, and Neo), by transporting them across various
runtimes.

To transport your SAP Datasphere content through SAP Cloud Transport Management, you first must
configure your SAP Cloud Transport Management system.

Once you’ve configured the system, you must connect it to your SAP Datasphere tenant.

Then you can use the SAP Datasphere Transport app to transfer your content to SAP Cloud Transport
Management.

 Note

Users with the DW Administrator global role (or users with both a scoped DW Space Administrator role and
a global role providing the Lifecycle privilege), can use the Transport app to export content.

Finally, you can return to SAP Cloud Transport Management to ensure the transferred content is available and
ready to be imported to another SAP Datasphere tenant.

Configuring SAP Cloud Transport Management

1. As an administrator of SAP Cloud Transport Management, set up an environment to transport content


objects directly in SAP Datasphere, see Initial Setup to Transport Content Objects Directly in an
Application.
2. Configure the landscape for your transports, by creating transport destinations, routes, and nodes.

 Tip

• Go to System→Administration→App Integration to find the token URL and add the extension
<Token URL>?grant_type=client_credentials.
• You must create the OAuth Client (client ID and secret) with the purpose API Access and
the access Analytics Content Network Interaction. See Create OAuth2.0 Clients to Authenticate
Against SAP Datasphere for more information.

1. Create a destination in the SAP Cloud Transport Management subaccount for each target environment
of SAP Datasphere. See Create Transport Destinations .

Integrating Data and Managing Spaces in SAP Datasphere


Transporting Content Between Tenants PUBLIC 305
To integrate with SAP Datasphere, the destination must have an endpoint URL in this format:
https://<tenant_URL>/api/v1/content/deploy/

2. Create nodes and routes. See Create Transport Nodes and Create Transport Routes . Make sure to
select the checkbox Allow Upload to Node when creating a source node. When defining a target node,
the content type must be Application Content to be able to select the destination name you've created
in the previous step.
3. Go to your SAP Business Technology Platform subaccount. Choose the SAP Cloud Transport Management
instance of your choice. In the instance details page, choose … → Download to download the SAP Cloud
Transport Management instance key.

Connecting to SAP Cloud Transport Management from SAP Datasphere

1. In the side navigation area, click Transport Export .


2. Navigate to My Content and click Settings from the menu bar.
3. In the Settings dialog box, click the Upload JSON button to upload the instance key you previously
downloaded from SAP Cloud Transport Management (See step 3 in the previous section).
4. The connection of your SAP Datasphere tenant with SAP Cloud Transport Management is now live: click
OK.

Using the SAP Datasphere Transport App to Transfer Content to SAP Cloud
Transport Management

1. In the side navigation area, click Transport Export and select the package you want to export.
2. Click  Upload to SAP Cloud Transport Management Node from the menu bar: in the dialog box, you can
see the list of selected packages.

Integrating Data and Managing Spaces in SAP Datasphere


306 PUBLIC Transporting Content Between Tenants
3. In the Export Node dropdown, choose the source node of SAP Cloud Transport Management to which you
want to export your packages. For more information on source and target nodes, see Create Transport
Nodes.
4. Select the checkbox to agree to use your user rights (based on user's email ID) to import the package in
the target tenant.
5. Click Upload.

Using SAP Cloud Transport Management to Transfer Content to SAP


Datasphere

1. Log into your SAP Cloud Transport Management tenant: in the Transport Nodes panel, select the target
node where you have uploaded the SAP Datasphere package.
2. In the Target Node Details page, you can see your packages that are ready for import: select the package
and click Import Selected.

Once the import is completed, you see the content of the package available in its corresponding destination in
your SAP Datasphere account.

 Note

The user who uploaded the package must be a member of the target tenant.

The authentication method (defined by the tenant owner) must be the same on the source and on the
target tenant. If the selected method is single sign-on (SSO), make sure to define the same user attribute
for the source and the target tenant. For more information on authentication methods, see Configuring
Identity Provider Settings.

Troubleshooting

In case of import failure, you can access the logs to investigate the error by adding this endpoint to the target
tenant URL in your browser:
{TenantURL}/contentmanager/v4/jobs/{ContentReferenceID}?brief=false

 Note

The Content Reference ID is retrieved from the SAP Cloud Transport Management import logs.

Integrating Data and Managing Spaces in SAP Datasphere


Transporting Content Between Tenants PUBLIC 307
12 Managing Your Space

Users with a space administrator role are responsible for controlling user access to their space, and monitoring
and otherwise managing the space.

All data acquisition, preparation, and modeling in SAP Datasphere happens inside spaces. A space is a secure
area - space data cannot be accessed outside the space unless it is shared to another space or exposed for
consumption.

An administrator must create one or more spaces. They allocate disk and memory storage to the space, set
its priority, and can limit how much memory and how many threads its statements can consume. See Creating
Spaces and Allocating Storage.

If an administrator has assigned you the role of space administrator for a certain space via a scoped role,
you can assign other users, create connections to source systems, secure data with data access controls, and
manage other aspects of the space.

To view information on the properties of a space that have been set by an administrator, see:

• Create a Space
• Allocate Storage to a Space
• Set Priorities and Statement Limits for Spaces

12.1 Control User Access to Your Space

You can assign users to your space and manage them.

Prerequisites

To assign users to your space and manage them from the Users area in your space details page, you must have
a scoped role that grants you access to your space with the following privileges:

• Data Warehouse General (-R------) - To access SAP Datasphere.


• Spaces (-RU-----) - To open and update your space in the Space Management tool.
• Space Files (-R------) - To view objects in your space.
• Scoped Role User Assignment (-------M) - To manage the users who can access your space.

The DW Space Administrator role template, for example, grants these privileges. For more information, see
Privileges and Permissions and Standard Roles Delivered with SAP Datasphere.

Integrating Data and Managing Spaces in SAP Datasphere


308 PUBLIC Managing Your Space
Introduction to Space User Access

If a user with an administrator role has assigned you to a space with space administration privileges via
a scoped role, you can assign and manage users in your space. For more information about how a DW
Administrator assigns users to roles in particular spaces via scoped roles, see Create a Scoped Role to Assign
Privileges to Users in Spaces.

In this example, based on scoped roles that the DW Administrator has created, users are assigned to spaces
with scoped roles as follows:

Spaces Users - Scoped Roles

Sales US Sally - Sales Modeler

Bob - Sales Modeler

Joan - Sales Spaces Admin

Sales Europe Sally - Sales Modeler

Bob - Sales Modeler

Jim - Senior Sales Modeler

Sales Asia Joan - Sales Spaces Admin

Joan has the role of space administrator for the spaces Sales US and Sales Asia. She can change the user
assignment. She can assign any user to her 2 spaces with the scoped roles to which her spaces are assigned:
Sales Modeler and Sales Spaces Admin.

She does the following changes in the Sales US space:

• As Bob no longer needs to work in the space Sales US, Joan unassigns Bob from the space.
• As Jim needs to work in space Sales US with the modeling privileges, Joan assigns Jim to the space with
the Sales Modeler scoped role.

She does the following change in the Sales Asia space:

• As Bob then needs to manage the space Sales Asia, Joan assigns Bob to the space with the Sales Space
Admin scoped role.

With Joan's changes, users are now assigned to spaces with scoped roles as follows:

Spaces Users - Scoped Roles

Sales US Sally - Sales Modeler

Jim - Sales Modeler

Joan - Sales Spaces Admin

Sales Europe Sally - Sales Modeler

Bob - Sales Modeler

Jim - Senior Sales Modeler

Integrating Data and Managing Spaces in SAP Datasphere


Managing Your Space PUBLIC 309
Spaces Users - Scoped Roles

Sales Asia Joan - Sales Spaces Admin

Bob - Sales Spaces Admin

Access the Users Area in Your Space

In the side navigation area, click (Space Management), locate your space tile, and click Edit to open it.

In the Users area of the space page, you can see and edit which users are assigned to your space with which
scoped roles. You can assign users to your space among all the users in the tenant, with one or more scoped
roles to which your space is assigned.

 Note

• Space Administrator column - A green tick appears for each user who has space administration
privileges for the space, which includes having the privilege (permission) Scoped Role User Assignment
(Manage).
• Scoped Role column - The one or more scoped roles that are assigned with each user to the space are
displayed. Only a DW Administrator can click a scoped role, which opens the scoped role page.

All user assignment changes you do in the Users area of your space page are reflected in the scoped roles to
which your space is assigned.

Assign Users to Your Space

You can add a user who will have certain privileges in your space by selecting the user and one or more scoped
roles.

1. Click Add. All the users created for the tenant are displayed in the dialog that opens.
2. Select one or more users and click Next. All scoped roles to which your space is assigned are displayed.
3. Select one or more scoped roles and click Create.

Change Privileges for a User in Your Space

You can change the privileges that the user has in your space by assigning it to another scoped role. A user can
be assigned to one or more scoped roles in a space.

1. Select a user and click Edit. All scoped roles to which your space is assigned are displayed.
2. Select one or more scoped roles and click Select.

Integrating Data and Managing Spaces in SAP Datasphere


310 PUBLIC Managing Your Space
Unassign Users from Your Space

You can unassign a user from your space. The user will be unassigned from the space in the one or more
scoped roles which gave access to the space.

1. Select a user and click Remove.


2. In the confirmation dialog, click Remove.

You can also use the SAP Datasphere command line interface to assign users to a space. For more information,
see Manage Spaces via the Command Line.

12.2 Create Time Data and Dimensions

Create a time table and dimension views in your space to provide standardized time data for your analyses.
The time table contains a record for each day in the specified period (by default from 1900 to 2050), and the
dimension views allow you to work with this date data at a granularity of day, week, month, quarter, and year,
and to drill down and up in hierarchies.

Prerequisites

To work with the time table and dimensions in your space, you must have a scoped role that grants you access
to your space with the following privileges:

• Data Warehouse General (-R------) - To access SAP Datasphere.


• Spaces (-R------) - To open your space in the Space Management tool.
• Space Files (-R------) - To view objects in your space.
• Data Warehouse Data Builder (-RUD----) - To read, update and delete objects in the Data Builder.

The DW Space Administrator role template, for example, grants these privileges. For more information, see
Privileges and Permissions and Standard Roles Delivered with SAP Datasphere.

Procedure

1. In the side navigation area, click (Space Management), locate your space tile, and click Edit to open it.
2. In the Time Data section, click Create Time Tables and Dimensions to open the dialog.
The dialog lists all the objects that will be created:
• Time Table - A table containing one record for each day from, by default, the year 1900 to the year
2050.
• Four views with a semantic usage of Dimension:
• Time Dimension - Day - Suitable for an association from a column containing dates in the form Dec
31, 2020, and providing hierarchies with levels:

Integrating Data and Managing Spaces in SAP Datasphere


Managing Your Space PUBLIC 311
• Year, Quarter, Month, Day
• Year, Month, Day
• Year, Week, Day

 Note

Monday is identified as 00 in the Day of Week column.

• Time Dimension - Month - Suitable for an association from a column containing months in the form
202012 (for December 2020), and providing hierarchies with levels:
• Year, Quarter, Month
• Year, Month
• Time Dimension - Quarter - Suitable for an association from a column containing quarters in the
form 20204 (for the fourth quarter of 2020), and providing one hierarchy with levels:
• Year, Quarter
• Time Dimension - Year - Suitable for an association from a column containing years in the form
2020.
• Three translation tables - which contain the names of days, months, and quarters in English, French,
German, and Spanish.
3. [optional] Modify the values in any of the following fields:

• Enter different years in the From Year and To Year fields to expand or reduce the scope of the time table
data.

 Note

The To Year cannot be more than 100 years later than the current year.

• Edit any of the proposed business names for the tables and views to be created. The technical names
cannot be modified.
4. Click Create.

When creation is complete, the content of the section changes to show that the objects are now present
in the space. You can use the tools to edit the Time Table values, refresh the time data, or delete these
entities.

 Note

From v2023.08, we use the ISOWEEK() function to generate ISO-standard year and week numbers
(see ISOWEEK Function (Datetime) in the SAP HANA Platform documentation. To regenerate time data
generated before this version, click Edit, modify your From Year or To Year, and click Save.

The tables and views are listed with your other space objects and can be opened (in read-only form) from
the Data Builder and Repository Explorer.

Users assigned to your space can now use the views as dimensions by creating associations to them from
their entities containing time data.

Integrating Data and Managing Spaces in SAP Datasphere


312 PUBLIC Managing Your Space
12.3 Logging Read and Change Actions for Audit

You can enable audit logs for your space so that read and change actions (policies) are recorded.
Administrators can then analyze who performed which action at which point in time.

This topic contains the following sections:

• Enable Audit Logs [page 313]


• Disable Audit Logs [page 314]
• Export Audit Logs [page 314]

Prerequisites

To enable audit logs for your space, you must have a scoped role that grants you access to your space with the
following privileges:

• Data Warehouse General (-R------) - To access SAP Datasphere.


• Spaces (-RU-----) - To open and update your space in the Space Management tool.
• Space Files (-R------) - To view objects in your space.

The DW Space Administrator role template, for example, grants these privileges. For more information, see
Privileges and Permissions and Standard Roles Delivered with SAP Datasphere.

Enable Audit Logs

• Enable audit logs of SAP Datasphere objects for read and change operations on the space level in Space
Management <Your Space> Edit Space Auditing . You can set the retention time in days for each
audit log policy (read or change). The default and minimum retention time is 7 days and the maximum
retention time is 10 000 days.
If audit logs have been enabled, entries of all SAP Datasphere related objects are saved in an SAP HANA
audit log. These logs don't include the objects of the database access schemas, like open SQL schemas,
for example.

 Note

If you choose to enable audit logs, be aware that they can consume a large quantity of GB of disk
in your SAP Datasphere tenant database, especially when combined with long retention periods.
Administrators can delete audit logs when needed, which will free up disk space. For more information,
see Delete Audit Logs.

• For individual database schemas (open SQL schemas), you can enable auditing for read and
change operations in Space Management <Your Space> Database Access Database Users Edit
Privileges . The retention time can be defined per schema.

Integrating Data and Managing Spaces in SAP Datasphere


Managing Your Space PUBLIC 313
 Note

Data Lake

Please note, that the statements issued via the execute-procedure for data lake are currently not
audited in SAP Datasphere.

As a result, SAP HANA policies are created for the schemas of the space.

The policy names of the SAP Datasphere administered objects are:

• DWC_DPP_<space name>_READ
• DWC_DPP_<space name>_CHANGE

Disable Audit Logs

When you disable an audit policy for a space or for a database user (open SQL schema), all related audit log
entries are deleted.

 Note

You can keep the audit log entries generated for your space by exporting them before the space is deleted.

Export Audit Logs

If you want to keep the audit log entries generated for your space, you can export them before they are
automatically deleted by these actions: deleting a space, deleting a database user (open SQL schema),
disabling an audit policy for a space, disabling an audit policy for a database user (open SQL schema) and
unassigning an HDI container from a space.

To export audit log entries using the SAP HANA Database Explorer, follow these steps:

1. Expose for consumption the view containing the audit log entries that you want to export. See Exposing
Data For Consumption.
2. Create a database user in your space. See Create a Database User [page 147]
3. Connect to your Open SQL schema with SAP HANA database explorer. See Connect to Your Open SQL
Schema [page 148]
4. Export the data. See Export Schemas, Tables, and Other Catalog Objects in the SAP HANA Database
Explorer guide.

Related Information

Monitor Database Operations with Audit Logs


Delete Your Space [page 325]

Integrating Data and Managing Spaces in SAP Datasphere


314 PUBLIC Managing Your Space
12.4 Translating Metadata for SAP Analytics Cloud

Users with a scoped role containing the Translation privilege can translate metadata such as business names
and column names for dimensions and analytic models, and hierarchy dimension labels for SAP Analytics
Cloud stories.

For example, you may have stories that have column names in a default language, but you'd like to have those
names displayed in other languages so that your SAP Analytics Cloud story can be viewed in the language of
your choice.

To translate objects' metadata, you can have it translated into multiple target languages using third-party
translation tools or you can have those descriptions translated in-line using the Translation tool. You can use
the tool to review translations, and edit or add translations from scratch for some or all languages.

To manage translation, you must be assigned to a scoped role that inherits a template (such as DW Space
Administrator), which grants the Translation privilege (see Create a Scoped Role to Assign Privileges to Users in
Spaces).

You can refer to the list below for the supported languages in SAP Datasphere.

This topic contains the following sections:

• Enable Translation [page 315]


• Select Objects to Translate [page 316]
• Translate Metadata via XLIFF Files [page 316]
• Translate Metadata Manually [page 317]
• Delete Translations [page 317]
• Manage Translation [page 317]

Prerequisites

To enable translation for your space, and to manage translation in the Translation tool, you must have a scoped
role that grants you access to your space with the following privileges:

• Data Warehouse General (-R------) - To access SAP Datasphere.


• Spaces (-R------) - To enable translation and select the source language for your space.
• Space Files (-R------) - To view objects in your space.
• Translation (CR-D----) - To access the Translation tool, select objects to translate, manage and delete
translations.

The DW Space Administrator role template, for example, grants these privileges. For more information, see
Privileges and Permissions and Standard Roles Delivered with SAP Datasphere.

Enable Translation

DW Space Administrators can turn on the toggle to enable content translation for a space:

Integrating Data and Managing Spaces in SAP Datasphere


Managing Your Space PUBLIC 315
1. In the Main Menu, click (Space Management).
2. Select a space.
3. In the Overview tab, in the General Settings Translation sections, select a language as Source
Language. For more information on available languages, see Learn About the Translation Process in SAP
Analytics Cloud.

Select Objects to Translate

You can add objects to Translation tool:

1. Click  (Add). The Add Objects dialog opens.


2. Select the objects you want to translate.

 Note

Only Analytic Models can be selected in the Add Objects dialog, but the dimensions and hierarchies
attached to added Analytic Models are automatically selected to the Translation tool.

3. Click Add Objects. The selected Analytic Models and their dependencies are added to the tool.

Translate Metadata via XLIFF Files

You can translate large amounts of metadata with the help of XLIFF files. The  (Download) button in the
Translation tool allows you to download the strings to be translated to a XLIFF file, which can be used to
translate the content. Then, the translated strings must be uploaded back to the Translation tool in the XLIFF
file format with the  (Upload) button.

1. In the Translation tool, select at least one entity that you want to translate.
2. Click  (Download) to download the XLIFF files. If one object is selected to be downloaded, the
corresponding XLIFF file can be downloaded as a .zip file.
3. You can choose to select All Strings or Outstanding Strings while exporting the XLIFF file(s). The
Outstanding Strings option is enabled only for partially translated objects.
The XLIFF file is downloaded to the default location in your system.

 Note

The downloaded XLIFF file contains the source locale and the list of text that needs translation. You
can use this XLIFF file in any external translation tool to generate the translated content as a separate
XLIFF file per locale.

4. Once the translation is done, click  (Upload) to open the Import Translations dialog.
5. Click Upload Files to upload the translated XLIFF files back into the Translation tool.
6. Click Import to upload the XLIFF files.

 Note

You can import multiple XLIFF files. Each XLIFF file contains the translated content in one language, which
is automatically fetched from the language code embedded in the name of the uploaded XLIFF files. These

Integrating Data and Managing Spaces in SAP Datasphere


316 PUBLIC Managing Your Space
XLIFF files can be uploaded together since they belong to one object. The Import option is disabled if more
than one object is selected.

Translate Metadata Manually

You can view the source and the translated text in the Translation tool, and add or edit the translated text
in-line.

1. Select the object you want to translate manually and click  (Edit). The list of source text appears.
2. Choose the Target Language and start adding the translated text in the empty field.
3. Select Save. The message Translation updated successfully appears, and the user who had marked the
model or dimension for translation is notified.

You have now successfully saved the newly translated text.

Delete Translations

To delete translations, select one or more objects in the Translation tool and click  (Delete).

 Note

Deleting translations removes the object and attached translated strings from the Translation tool. You'll be
able to add the object back to the tool, but no translation will be attached to it any longer.

Manage Translation

The Translation tool shows the spaces you are assigned to with enabled translation. Each space contains
entities that are selected for translation. The tool allows you to perform and manage these translations.

From the side navigation, choose  (Translation) and select a space. The spaces available here are the ones
enabled for translation. For more information, see Enable Translation [page 315].

 Note

• If the model has changed, new strings will automatically appear in the translation tool as Not
Translated. Expired strings will still be available in the Translation tool.
• You can upload the XLIFF files even if there is a mismatch in the HTML tags between the source and
target language strings.

The Translation tool displays the following buttons and fields:

Integrating Data and Managing Spaces in SAP Datasphere


Managing Your Space PUBLIC 317
Toolbar
Tool Description

 (Add) Adds objects to the Translation tool. Only Analytic Models can be selected in the Add
Objects dialog, but the dimensions and hierarchies attached to selected Analytic Models
are automatically added to the Translation tool. See Select Objects to Translate [page 316].

 (Refresh) Refreshes the list of objects.

 (Edit) Open the selected object for manual in-line translation. Users with the Create permission
can access this functionality. See Translate Metadata Manually [page 317].

 (Delete) Deletes translations for the selected objects. See Delete Translations [page 317].

 (Download)  (Upload) Downloads translatable (source) XLIFF files and uploads the translated XLIFF files. See
Translate Metadata via XLIFF Files [page 316].

 (Show) Specifies the fields to display for the tool objects:

• path of the object


• description of the object
• type of the object
• name of the object's creator
• source language from which the object has been created
• Requested On: date when the translation was requested
• Changed On: date when the translation strings have been updated
• Status: status of the translation. These status buttons are clickable and you can view
the status of translations uploaded for each language.

 (Filter) Filters translated, not translated, and expired objects.

 (Search) Searches objects based on Name, Description, Translated by, Translated Language, Re-
quested on, Changed on, Status, and Owner.

Tool Fields
Title Description

Technical Name/Business Name Names of the object requiring translation.

Type Object type requiring translation. It can be a:

• remote table
• local table
• view
• analytic model
• hierarchy

Creator Name of the person who seletected the object for translation

Requested On When the translation was requested

Changed On When the strings or translations for strings were last up-
dated

Integrating Data and Managing Spaces in SAP Datasphere


318 PUBLIC Managing Your Space
Title Description

Status Status of the translation per object:

• Translated: Once all strings are translated completely,


the status is shown as Translated.
• Not Translated: For a new object, or for objects whose
translation is expired or deleted completely, the status
is shown as Not Translated.
• Partially Translated: When the strings in an object
are not translated completely, the status is shown as
Partially Translated. Once it is translated again, the sta-
tus changes back to Translated.

To see the status of the translation of an object per lan-


guage, click on the object. A dialog appears with the list of
translated languages and corresponding status in which the
object was translated.

12.5 Save and Deploy Your Space

When you save a space, it is stored in the SAP Datasphere repository, which contains the design-time
definitions of all your spaces. When you deploy a space, you are creating a run-time version for use in the
SAP Datasphere database.

Prerequisites

To save and deploy the changes you've made in your space, you must have a scoped role that grants you
access to your space with the following privileges:

• Data Warehouse General (-R------) - To access SAP Datasphere.


• Spaces (-RU-----) - To open and update your space in the Space Management tool.
• Spaces Files (-RU-----) - To read and update objects in your space.

The DW Space Administrator role template, for example, grants these privileges. For more information, see
Privileges and Permissions and Standard Roles Delivered with SAP Datasphere.

Space Deployment

The current status of a space is shown in its Deployment Status property, which can have the following values:

• Not Deployed - The space has never been deployed and exists only as a design-time version.
• Deployed - The space is deployed to the run-time database and its design-time and run-time versions are
identical.

Integrating Data and Managing Spaces in SAP Datasphere


Managing Your Space PUBLIC 319
• Deploying - The space is currently being deployed.
• Changes to Deploy - The design-time version of the space contains changes that need to be deployed to
make them available in the run-time version.
• Design-Time Error - The design-time version of the space is invalid. You can save a save with design-time
errors, but you cannot deploy it. To correct design-time errors, review each of the validation messages in
the editor, make the necessary changes, and then save the space.
• Run-Time Error - The run-time version of the space is invalid and can no longer be used. To correct the
errors, review each of the validation messages in the editor, make the necessary changes, and then save
and deploy the space.

 Note

For spaces that have been created before the deploy feature exists, the Status area will display “No
Information”. Spaces are deployed anyway. If you click Deploy, the status will display Deployed.

You can also view the date of the last deployment in the Deployed On area.

 Note

• When you assign a user to the space, the user becomes active when you save the space. You can view
this in the Active column of the Users area in your space.
• Any change that you make in the Time Data area of your space is immediately deployed even if you
haven’t saved nor deployed the space. A message is displayed in this area, which indicates that the
changes are immediately applied
• The option Expose for Consumption by Default does not need a redeployment to be active.

 Note

On rare occasions when the run-time database is not available, an information message is displayed and
certain features that depend on the run-time are disabled:

• The Deploy, Delete and Monitor buttons or commands are greyed out.
• In the Space Management start page, dashes (-) appear instead of numbers in the bars (for example,
the Used Disk bar at the top of the page and in the space tiles), in the space status area at the top of the
page and space status colors are not displayed in the individual space tiles.
• In a space page, the Time Data area is disabled.

12.6 Review Your Space Status

Color codes and icons indicate the health of a space.

You can see the status of all spaces at the top of the Space Management page.

You also can see the status of an individual space at the top-left of the space.

Each status shows the used storage and may indicate whether a space needs extra attention. The status also
shows whether a space is active or not (locked).

Integrating Data and Managing Spaces in SAP Datasphere


320 PUBLIC Managing Your Space
Icon Description

 (Cold) The used storage is 5% or less.

 (Healthy) The used storage is between 6% and 90%

 (Critical) The used storage is greater than 90%.

This space is close to exceeding it's storage quota. It might make sense to delete unnecessary
data in your space or to extend the assigned storage.

If your space exceeds it's storage quota, it might change to a locked state.

 (Locked) A space is locked:

• if the space exceeds its allocations of memory or disk storage,


• if the audit logs consume too much disk storage,
• if a space administrator has manually locked the space.

For more information on how to unlock a space, see Unlock a Locked Space [page 321].

When a space is locked, its capabilities are limited. Users can continue to create and modify
objects and save their changes in the repository, but they cannot deploy their changes to the
run-time database.

Restrictions include:

• changes to data layer objects can be saved, but deployment is blocked


• uploading a CSV file is blocked
• wrangling is blocked
• open SQL users are blocked
• insert privilege is removed from the Space Manager
• assigning users and creating open SQL schemas in Space Management is blocked
• all direct execution or scheduling of tasks in the Data Integration Monitor or in the Data
Builder that allow to add data to the space.

12.7 Unlock a Locked Space

When a space exceeds its assigned storage or when the audit logs enabled in the space consume too much
disk storage, the space is automatically locked after 60 minutes if you do not free up space.

When a space is locked, users assigned to the space can continue to create and modify objects and save their
changes in the repository, but they cannot deploy their changes to the run-time database.

 Note

You cannot lock or unlock a file space.

Integrating Data and Managing Spaces in SAP Datasphere


Managing Your Space PUBLIC 321
Prerequisites

To manually lock or unlock your space, you must have a scoped role that grants you access to your space with
the following privileges:

• Data Warehouse General (-R------) - To access SAP Datasphere.


• Spaces (-RU-----) - To open and update your space in the Space Management tool.
• Spaces Files (-RU-----) - To read and update objects in your space.

The DW Space Administrator role template, for example, grants these privileges. For more information, see
Privileges and Permissions and Standard Roles Delivered with SAP Datasphere.

Space Locked as it Has Exceeded its Assigned Storage

If a space exceeds its allocations of memory or disk storage, it will be locked until a space user deletes the
excess data or an administrator assigns additional storage.

In this situation, these actions are possible:

• Users assigned to a space can delete data to bring the space back under the limit of its assigned storage.
• A space administrator can use the Unlock button on the Space Management page or on the space page to
unlock the space for a 24-hour grace period, in case urgent changes must be deployed.
• An administrator can assign more disk and/or memory storage to the space (see Allocate Storage to a
Space).

Space Locked as Audit Logs Consume Too Much Disk Storage

If the total size of all audit logs across all spaces has reached 40% of the tenant disk storage, the system
automatically locks any spaces - and deactivate any database analysis users - whose audit logs consume more
than 30% of the total audit log size.

In this situation, these actions are possible:

• A space administrator can use the Unlock button on the Space Management page or on the space page to
unlock the space for a 24-hour grace period, in case urgent changes must be deployed.
• An administrator can delete audit logs and free up disk space (see Delete Audit Logs).
• A space administrator can disable audit logs or decrease the number of days audit logs are kept for (see
Logging Read and Change Actions for Audit [page 313]).

Lock a Space

If you want to put a space on hold and you are a space administrator, you can manually lock it.

Integrating Data and Managing Spaces in SAP Datasphere


322 PUBLIC Managing Your Space
• To lock one or more spaces, in the side navigation area, click (Space Management), locate and select
one or more space tiles, and click the Lock button.
• To lock only one space, you can alternatively select the space tile, click Edit to open it and click the Lock
button.

 Note

You can unlock one or more spaces that have been manually locked by using the Unlock button.

12.8 Monitor Your Space Storage Consumption

See the storage amount assigned to and used by your space.

Prerequisites

To monitor the storage consumption of your space, you must have a scoped role that grants you access to your
space with the following privileges:

• Data Warehouse General (-R------) - To access SAP Datasphere.


• Spaces (-R------) - To open your space in the Space Management tool.
• Space Files (-R------) - To view objects in your space.

The DW Space Administrator role template, for example, grants these privileges. For more information, see
Privileges and Permissions and Standard Roles Delivered with SAP Datasphere.

Procedure

To open the monitoring page, select your space and click Monitor from the Space Management page or
alternatively open your space and click Monitor on the upper-right side of your space.

 Note

You cannot monitor a file space.

Action Example

See the amount of disk storage and memory storage used in


your space.

For more information about storage capacity, see Allocate


Storage to a Space.

Integrating Data and Managing Spaces in SAP Datasphere


Managing Your Space PUBLIC 323
Action Example

Filter by schema or by storage type. Filter values by adding


or removing them from the drop-down menu.

The values are displayed in a doughnut and bar chart.

The hidden replica tables of SAP HANA virtual


tables are stored in a separate schema named
_SYS_TABLE_REPLICA_DATA.

For more information about replica tables, see Replicating


Data and Monitoring Remote Tables.

See the Table Storage Consumption graph that displays all


the relevant tables according to your selected filter, which
makes it easy to get an overview over the consumed storage.

See Table Details that provides more information such as


the name, schema, storage type, record count, and the used
storage of each table.

Sort your list of tables in ascending, descending order or


group them together as well as search for certain values.

12.9 Manage Your Space via the Command Line

You can use the SAP Datasphere command line interface, datasphere, to read, update, and delete your
space. You can set space properties, assign (or remove) users, create database users, create or update objects
(tables, views, and data access controls), and associate HDI containers to your space.

To use datasphere to manage your space you must have an SAP Datasphere user with the DW Space
Administrator role or equivalent permissions (see Roles and Privileges by App and Feature).

For more information, see Manage Spaces via the Command Line.

Integrating Data and Managing Spaces in SAP Datasphere


324 PUBLIC Managing Your Space
12.10 Export Your Space Data

You can export the data contained in your space at any time. For example, you may want to export data before
deleting your space.

Prerequisites

To export the data contained in your space, you need to connect to an Open SQL schema or to allow the space
to access the open SQL schema. To do so, you must have a database user and a password (see Create a
Database User [page 147]).

Procedure

To export data using the SAP HANA Database Explorer, follow these steps:

1. Expose for consumption the views containing the data you want to export (see Exposing Data For
Consumption).
2. Create a database user in your space (see Create a Database User [page 147]).
3. Connect to your Open SQL schema with SAP HANA database explorer (see Connect to Your Open SQL
Schema [page 148]).
4. Export the data (see Export Schemas, Tables, and Other Catalog Objects in the SAP HANA Database
Explorer guide).

12.11 Delete Your Space

Delete a space if you are sure that you no longer need any of its content or data.

When you delete a space, the space is moved to the Recycle Bin area. From the Recycle Bin, a user with an
administrator role can either restore the space or permanently delete the space from the database to recover
the disk storage used by the data in the space (see Restore Spaces from, or Empty the Recycle Bin).

 Note

For spaces that have been deleted before version 2023.05, all related audit logs have been kept. A user with
an administrator role can decide to delete them (see Delete Audit Logs ).

 Note

Before deleting your space, you may want to:

• Export the data contained in your space (see Export Your Space Data [page 325]).

Integrating Data and Managing Spaces in SAP Datasphere


Managing Your Space PUBLIC 325
• Export the audit log entries generated for your space (see Logging Read and Change Actions for Audit
[page 313]).

Only users with an administrator role can create spaces. Users with an administrator or a space administrator
role can delete spaces.

Prerequisites

To move your space to the Recycle Bin area, you must have a scoped role that grants you access to your space
with the following privileges:

• Data Warehouse General (-R------) - To access SAP Datasphere.


• Spaces (-RUD----) - To open, update and delete your space in the Space Management tool.
• Space Files (-R------) - To view objects in your space.
• Scoped Role User Assignment (-------M) - To manage the users who can access your space.

The DW Space Administrator role template, for example, grants these privileges. For more information, see
Privileges and Permissions and Standard Roles Delivered with SAP Datasphere.

Procedure

 Note

If your space contains replication flows with objects of load type “Initial and Delta”, you should make sure
that these replication flows are stopped before you delete the space. If you restore the space at a later point
in time, the replication flows can then be started again. For more information about replication flows, see
Working With Existing Replication Flow Runs [page 242].

1. In the side navigation area, click (Space Management).


2. Locate and select your space, and click the Delete button.
3. In the confirmation message, enter DELETE if you are sure that you no longer need any of its content or
data, then click the Delete button.
The space is moved to the Recycle Bin area until it is restored or permanently deleted from the database
(see Restore Spaces from, or Empty the Recycle Bin).

 Note

The Recycle Bin is only accessible to users with an administrator role.

Once the space is in the Recycle Bin:

• You cannot edit the deleted space.


• As data is not deleted, the space still consumes disk storage.
• Data integration tasks and schedules (including those for elastic compute nodes) are paused.

Integrating Data and Managing Spaces in SAP Datasphere


326 PUBLIC Managing Your Space
• The deleted space is removed from the elastic compute nodes it was included in. If an elastic compute
node is in a running state, you can stop the run and restart it again so that replicated data is removed from
the node.
• The database users/Open SQL schemas and HDI containers of the deleted space are disabled.
• For remote tables connected via SAP HANA smart data access, with real-time replication, data replication
is stopped and data is removed.
• For remote tables connected via SAP HANA smart data integration, real-time replication is stopped.
• You cannot create a new space with the ID of a space that is in the recycle bin. If you want to delete a space
and recreate it with the same ID, you must first delete the space from the recycle bin (see Restore Spaces
from, or Empty the Recycle Bin).
• A database analysis user can still access the deleted space. For more information on a database analysis
user, see Create a Database Analysis User to Debug Database Issues.

Integrating Data and Managing Spaces in SAP Datasphere


Managing Your Space PUBLIC 327
Important Disclaimers and Legal Information

Hyperlinks
Some links are classified by an icon and/or a mouseover text. These links provide additional information.
About the icons:

• Links with the icon : You are entering a Web site that is not hosted by SAP. By using such links, you agree (unless expressly stated otherwise in your
agreements with SAP) to this:

• The content of the linked-to site is not SAP documentation. You may not infer any product claims against SAP based on this information.

• SAP does not agree or disagree with the content on the linked-to site, nor does SAP warrant the availability and correctness. SAP shall not be liable for any
damages caused by the use of such content unless damages have been caused by SAP's gross negligence or willful misconduct.

• Links with the icon : You are leaving the documentation for that particular SAP product or service and are entering an SAP-hosted Web site. By using
such links, you agree that (unless expressly stated otherwise in your agreements with SAP) you may not infer any product claims against SAP based on this
information.

Videos Hosted on External Platforms


Some videos may point to third-party video hosting platforms. SAP cannot guarantee the future availability of videos stored on these platforms. Furthermore, any
advertisements or other content hosted on these platforms (for example, suggested videos or by navigating to other videos hosted on the same site), are not within
the control or responsibility of SAP.

Beta and Other Experimental Features


Experimental features are not part of the officially delivered scope that SAP guarantees for future releases. This means that experimental features may be changed by
SAP at any time for any reason without notice. Experimental features are not for productive use. You may not demonstrate, test, examine, evaluate or otherwise use
the experimental features in a live operating environment or with data that has not been sufficiently backed up.
The purpose of experimental features is to get feedback early on, allowing customers and partners to influence the future product accordingly. By providing your
feedback (e.g. in the SAP Community), you accept that intellectual property rights of the contributions or derivative works shall remain the exclusive property of SAP.

Example Code
Any software coding and/or code snippets are examples. They are not for productive use. The example code is only intended to better explain and visualize the syntax
and phrasing rules. SAP does not warrant the correctness and completeness of the example code. SAP shall not be liable for errors or damages caused by the use of
example code unless damages have been caused by SAP's gross negligence or willful misconduct.

Bias-Free Language
SAP supports a culture of diversity and inclusion. Whenever possible, we use unbiased language in our documentation to refer to people of all cultures, ethnicities,
genders, and abilities.

Integrating Data and Managing Spaces in SAP Datasphere


328 PUBLIC Important Disclaimers and Legal Information
Integrating Data and Managing Spaces in SAP Datasphere
Important Disclaimers and Legal Information PUBLIC 329
www.sap.com/contactsap

© 2025 SAP SE or an SAP affiliate company. All rights reserved.

No part of this publication may be reproduced or transmitted in any form


or for any purpose without the express permission of SAP SE or an SAP
affiliate company. The information contained herein may be changed
without prior notice.

Some software products marketed by SAP SE and its distributors


contain proprietary software components of other software vendors.
National product specifications may vary.

These materials are provided by SAP SE or an SAP affiliate company for


informational purposes only, without representation or warranty of any
kind, and SAP or its affiliated companies shall not be liable for errors or
omissions with respect to the materials. The only warranties for SAP or
SAP affiliate company products and services are those that are set forth
in the express warranty statements accompanying such products and
services, if any. Nothing herein should be construed as constituting an
additional warranty.

SAP and other SAP products and services mentioned herein as well as
their respective logos are trademarks or registered trademarks of SAP
SE (or an SAP affiliate company) in Germany and other countries. All
other product and service names mentioned are the trademarks of their
respective companies.

Please see https://www.sap.com/about/legal/trademark.html for


additional trademark information and notices.

THE BEST RUN

You might also like