Integrating Data and Managing Spaces in SAP
Integrating Data and Managing Spaces in SAP
6 Exchanging Data with SAP SQL Data Warehousing HDI Containers. . . . . . . . . . . . . . . . . . . . . 152
6.1 Prepare Your HDI Project for Exchanging Data with Your Space. . . . . . . . . . . . . . . . . . . . . . . . . . . 153
6.2 Add an HDI Container and Access its Objects in Your Space. . . . . . . . . . . . . . . . . . . . . . . . . . . . . .155
6.3 Consume Space Objects in Your HDI Container. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 156
6.4 Allow Your Space to Write to Your HDI Container. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 157
7 Integrating Data to and From SAP HANA Cloud Data Lake. . . . . . . . . . . . . . . . . . . . . . . . . . . 158
7.1 Working with Data Lake. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 159
7.2 Data Lake API. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 161
10 Importing SAP and Partner Business Content from the Content Network. . . . . . . . . . . . . . . 286
Users with a space administrator or integrator role can create connections to source systems and databases
and can schedule and monitor data replication and other data integration tasks. Space administrators are,
additionally, responsible for controlling user access to their space, creating data access controls to secure
data, enabling other forms of data integration, transporting content between tenants, and monitoring and
otherwise managing the space.
Tip
The English version of this guide is open for contributions and feedback using GitHub. This allows you
to get in contact with responsible authors of SAP Help Portal pages and the development team to
discuss documentation-related issues. To contribute to this guide, or to provide feedback, choose the
corresponding option on SAP Help Portal:
• Feedback Edit page : Contribute to a documentation page. This option opens a pull request on
GitHub.
• Feedback Create issue : Provide feedback about a documentation page. This option opens an
issue on GitHub.
More information:
• Contribution Guidelines
• Introduction Video: Open Documentation Initiative
• Blog Post: Introducing the Open Documentation Initiative
An administrator will assign you the DW Space Administrator role, create your space, and assign you to it. Once
this is done, you can prepare your space as follows:
• Assign SAP Datasphere users to your space (see Control User Access to Your Space [page 308]).
• Transport objects securely to and from your space (see Transporting Content Between Tenants [page
290]).
Space administrators and integrators can create connections to source systems to allow space users to
acquire data from those systems (see Integrating Data via Connections [page 8]).
• Create database users to allow external tools to connect to the space and write data to Open SQL schemas
associated with the space (see Integrating Data via Database Users/Open SQL Schemas [page 145]).
• If your space has access to the SAP HANA Cloud, data lake, you can access it via an Open SQL schema
(see Integrating Data to and From SAP HANA Cloud Data Lake [page 158]).
• Add SAP HDI containers to your space (see Exchanging Data with SAP SQL Data Warehousing HDI
Containers [page 152])
• Generate a time table and associated time dimension views for use in the space (see Create Time Data and
Dimensions [page 311]).
SAP and our partners provide business content to support end-to-end business scenarios for various
industries and lines of business via the Content Network and our Community Content site (see Importing
SAP and Partner Business Content from the Content Network [page 286]).
We recommend that you create data access controls, which can be applied to views to provide row-level
filtering of your space data (see Securing Data with Data Access Controls [page 268]).
You can enable, run, schedule, and monitor data replication tasks in the (Data Integration Monitor) (see
Managing and Monitoring Data Integration [page 163]).
Users with a space administrator or integrator role can create connections to SAP and non-SAP source
systems, including cloud and on-premise systems and partner tools, and to target systems for outbound
replication flows. Users with modeler roles can import data via connections for preparation and modeling in
SAP Datasphere.
Each connection type supports a defined set of features. Depending on the connection type and the
connection configuration, a connection can be used for one or more of the following features:
Feature Description
Remote Tables Allows modelers to import tables from a source when creating graphical and SQL
views or entity-relationship models. During the import, the tables are deployed as
remote tables. Depending on the connection type, you can use remote tables to:
Note
If your source data comes from an SAP HANA On-Premise system, select the
adaptor following your use case:
• You want to access the data remotely: SAP HANA smart data access
(Data Provisioning Option: Direct) would be the recommended adaptor
to read the data. It allows higher degree of query pushdown to the
remote database, leading to better response times and less resource
consumption.
• You want to replicate the data into SAP Datasphere: The preferred option
for this is to use Replication Flows, see Creating a Replication Flow. In
case you require replication for remote tables, SAP HANA smart data
integration (Data Provisioning Option: Data Provisioning Agent) is the
recommended adaptor to push the data. It offers more options when
loading the data, such as applying filter conditions or data partitioning.
For more information on these adaptors, see Connecting SAP HANA Cloud,
SAP HANA Database to Remote Data Sources.
Data Flows Allows modelers to use objects from a connected source when creating data
flows to extract, transform and load data.
Model Import Allows modelers to import objects from any supporting SAP system with rich
metadata and, if supported, with their dependencies. This way, you can leverage
your existing investment in your customer system landscape without having to
rebuild the objects manually.
Replication Flows Allows modelers to replicate data from multiple source objects from a connected
source to a target including simple projection and mapping capabilities.
You can enable a single initial load or request initial and delta loads to load
changes at regular intervals. It depends on the connection type if you can use the
connection as source or target connection.
Note
Some connection types, such as SAP S/4HANA Cloud and on-premise or SAP ABAP, support replication
with both remote tables via Data Provisioning Agent (SAP HANA Smart Data Integration) and replication
flows (for on-premise systems via Cloud Connector).
Regardless of whether you use the same connection or two separate connections, for replicating data from
a dedicated source object we strongly recommend to only use remote tables or replication flows and not
both.
To allow space users using a feature, the feature must be enabled in the connection configuration. You can find
more information about which features are supported by which connection type and what prerequisites and
connection properties are required to enable (or disable) a feature from the links in the overview below.
To connect to remote systems, applications, databases, or storages, SAP Datasphere provides different
connection types. The following connection types are available with SAP Datasphere:
Remote Tables
(Remote Ac-
cess and Snap-
shot or Sched-
(Remote Ac-
cess and Snap-
shot or Sched-
Business Data not applicable not applicable not applicable not applicable not applicable not applicable
Product Con-
(replication
nections [page
30] flows are cre-
ated and de-
[created when
ployed via the
installing an
data product in-
SAP Business
stallation)
Data Cloud data
product from
the connected
system for the
first time and
only available
within an SAP-
managed inges-
tion space]
(Remote Ac-
cess and Snap-
shot or Sched-
Microsoft Azure yes (including yes (source) yes no Cloud Non SAP
SQL Database real-time repli-
Connections cation**)
[page 56]
(Remote Ac-
cess and Snap-
shot or Sched-
SAP ABAP Con- yes (including yes (source) yes no On-Premise, SAP
nections [page real-time repli- Cloud
67] cation**)
[generated and
only available
within an SAP
BW bridge
space if SAP
BW bridge has
been provi-
sioned]
SAP ECC Con- yes (including via connection yes no On-Premise SAP
nections [page real-time repli- type SAP ABAP
85] cation**) Connections
[page 67]
(source)
SAP HANA yes (including yes (source and yes no On-Premise, SAP
Connections real-time repli- target) Cloud
[page 92] cation**)
(Remote Ac-
cess and Snap-
shot or Sched-
SAP S/4HANA yes (including yes (source) yes yes Cloud SAP
Cloud Connec- real-time repli-
tions [page 107] cation**)
SAP S/4HANA yes (including yes (source) yes yes On-Premise SAP
On-Premise real-time repli-
Connections cation**)
[page 112]
*Data is pushed into the SAP Datasphere space by using the database user (Open SQL schema) SQL Interface.
Typically, the push action is triggered from the partner tool. When data is pushed by the source, you can import
a table from the Sources tab of the Source Browser panel in the Data Builder which is then deployed as local
table (see Connections to Partner Tools [page 122]).
** For more information about any constraints for using real-time replication, see Replicate Data Changes in
Real-Time [page 181].
***Connections marked as "On-Premise" can be used to connect to sources located in the cloud, too. Note
that in order to connect to a cloud-based SAP ABAP or Microsoft SQL Server, for example, you have to connect
through the on-premise installed SAP HANA smart data integration Data Provisioning Agent.
Managing Connections
In the (Connections) app, you get an overview of all connections created in your space.
Tool Description
(Add Connection) Create Create a connection to allow users assigned to the space to use the connected
remote system for data modeling and data access in SAP Datasphere.
Connection
For more information, see Create a Connection [page 16].
Edit Select and edit a connection to change its properties. Warnings might indicate
that you need to edit a connection.
Delete Select and delete one or more connections if they are not used anymore.
Validate Select and validate a connection to get detailed status information and make
sure that it can be used for data modeling and data access. Always validate a
connection after you have created or edited it.
Pause/Restart For connections that connect to a remote system through SAP HANA Smart Data
Integration and its Data Provisioning Agent, you can pause and restart real-time
replication for selected connections, if required.
For more information, see Pause Real-Time Replication for a Connection [page
126].
(Reload Connection List) Reload the connections list to include the latest updates into the list.
(Sort Connections) Open the Sort dialog to control the ordering of the connections list.
(Filter Connections) Select one or more filter values to restrict the connection list according to your
needs.
(Select Columns) Open the Columns dialog to control the display of columns in the results table.
Modify the column list in any of the following ways, and then click OK to apply
your changes:
• To select a column for display, select its checkbox. To hide a column deselect
its checkbox.
• Click on a column token to highlight it and use the arrow buttons to move it in
the list.
• Click Reset to go back to the default column display.
Search Enter one or more characters in the Search field to restrict the list to connections
containing the string.
When there is one or more messages for your connections, a button is displayed
on the top right corner of the Connections app specifying the number of warning
messages for all connections in the list.
(Warning Messages)
Click the button to open the list of messages. Clicking a message title selects the
corresponding connection in the list. Clicking (Navigation) for a message opens
a more detailed message containing guidance on how to solve the issue.
Create a connection to allow users assigned to a space to use the connected source or target for data modeling
and data access in SAP Datasphere.
Prerequisites
Procedure
1. In the side navigation area, click (Connections) and select a space if necessary.
2. On the Connections tab, click (Add Connection) Create Connection to open the connection
creation wizard.
3. Click an appropriate connection type tile for your source.
One or more of the following options help you to quickly find the connection type you need:
Option Description
Note
When hovering over a connection type, an Information button appears that you can click to get
information about the features supported by the connection type. This includes information about
whether the connected system can be used as source and/or target.
In the information, a warning icon indicates if a feature is not available, and clicking the icon tells you
what to do in that case. Also, if a connection type is inactive, the information tells you what to do to be
able to use it.
4. Complete the properties in the following sections and click Next Step.
Connection Details Enter or select properties that are needed to locate the
source system or database.
Note
If you connect to a partner tool, you will be redirected to an embedded partner UI to there enter
credentials and required connection properties, if applicable.
For an overview of all connection types including links to detailed property documentation, see Connection
Types Overview [page 10].
5. [optional] For some connection types that use a Data Provisioning Agent to connect to the source, for
example SAP ABAP, SAP HANA, or Microsoft SQL Server, you can define advanced properties which
Property Description
Business Name Enter a descriptive name to help users identify the object.
This name can be changed at any time.
Technical Name Displays the name used in scripts and code, synchronized
by default with the Business Name.
Note
Once the object is saved, the technical name can no
longer be modified.
Note
Once a package is selected, it cannot be changed
here. Only a user with the DW Space Administrator
role (or equivalent privileges) can modify a package
assignment in the Packages editor.
7. If you connect to a partner tool such as Precog, in the Additional Configuration section of this wizard step,
you'll find information about IP addresses that you can copy in order to add them to the IP allowlist in SAP
Datasphere.
For more information, see Manage IP Allowlist.
8. Click Create Connection to create the connection and add it to the overview of available connections.
When creating a connection to a partner tool such as Precog, the system checks whether the required
IP address or addresses are already available in the IP allowlist. Note that connection creation takes
longer because checking the allowlist in the database requires some time.
9. Select the connection and click (Validate Connection) to check if your connection is valid and can be
used for the enabled features.
Note
When creating a connection to SAP HANA on-premise using SAP HANA smart data access via Cloud
Connector, the system checks for a required internal service. In the connection overview, a warning
icon indicates that the service might not be ready yet. This happens for the first connection you create
or when you create a connection after the service has been disabled (after an automated weekly check
returning that there is no such connection anymore). Getting the service ready might take up to 45
minutes. Validate the connection and check the message details for more information.
Results
Depending on the connection type and the features that you have enabled for your connection:
• You can use the connection either to import tables when creating views or ER models, to create data flows,
or to create replication flows, or all three.
For more information, see:
• Importing Tables and Views from Sources
• Creating a Data Flow
• Creating a Replication Flow
Note
Generally, when importing a table, it is created as remote table. However, tables imported from partner
tool connections are created as local tables (see Connections to Partner Tools [page 122]).
• You can use remote tables imported from the connection either to access data directly live in the
connected source (federation) or to copy the data into SAP Datasphere (replication).
For more information, see Replicating Data and Monitoring Remote Tables [page 168].
• You can use the connection to import entities from your source.
For more information, see:
• SAP BW∕4HANA Model Transfer Connections [page 83]
• Importing SAP BW∕4HANA Models
Extend connectivity beyond SAP Datasphere standard remote connectivity and cover additional data sources
that are available with Adverity.
Adverity provides connectivity to a variety of data sources, including marketing data sources, databases and
others.
When creating a connection to a partner tool, an Open SQL schema is generated in SAP Datasphere. To be
able to send data to the Open SQL schema, the IP addresses of the partner tool need to be added to the
SAP Datasphere IP allowlist, and during connection creation the partner tool on their side needs to establish a
connection to the schema.
Prerequisites
Wizard step Configure Partner Connection shows an embedded Adverity UI. Here, you log on to the regional
Adverity server and perform the following steps:
Note
As a prerequisite to proceed to the next step in the wizard, your browser needs to be enabled to store
cookies. This is not the case if you run the browser in incognito mode.
Note
If there is no appropriate workspace available yet, you can create a new one.
5. Click Save to move to the next connection creation wizard step in SAP Datasphere.
You can now see the destination in the Adverity workspace. There, you can define a prefix for the tables created
for this destination to more easily find them later on in the SAP Datasphere Data Builder.
Use an Amazon Athena connection to access data from Amazon Athena, an interactive query service which can
be used to analyze data in Amazon S3 using standard SQL.
Prerequisites
• Remote
• Replication (snapshot)
For more information, see Replicating Data and Monitoring Remote Tables [page 168].
Connection Details
Property Description
Region Enter the AWS region in your Amazon Athena regional end-
point that you use to make your requests. For example, us-
west-2.
Property Description
Access Key Enter the access key ID of the user that the application must
use to authenticate.
Secret Key Enter the secret access key of the user that the application
must use to authenticate.
Features
Remote Tables are enabled without the need to set any additional connection properties.
Use an Amazon Redshift connection to access data from Amazon Redshift 8.x databases.
Prerequisites
The connection type supports the remote table as well as the data flow feature.
• Remote
• Replication (snapshot)
For more information, see Replicating Data and Monitoring Remote Tables [page 168].
Note
• In the view editor of the Data Builder, you can see technical artifacts from the source that you cannot
use for modeling. Affected artifacts are tables containing "pkey" or tables from technical schemas
pg_catalog, pg_internal, and information_schema, for example. Please ignore these artifacts.
Connection Details
Property Description
Port Displays the default port 5439. You can overwrite the port, if required.
Database Name Enter the name of the database to which you want to connect.
Security
Property Description
Use SSL Select the SSL certificate verification mode to use when connecting. The default is false.
[if Use SSL = true] Validate Select if the server certficate should be validated (true) or not (false). The default is false.
Server Certificate
If you select true, then the host name in the certificate must match the host name specified
in the Host field.
Credentials
Property Description
User Name Enter the name of the Redshift user that has privileges to connect to the database.
Features
Data Flows are enabled without the need to set any additional connection properties.
Use an Amazon Simple Storage Service connection to connect to and access data from objects in Amazon S3
buckets.
Data Flows You can use the connection to add source objects to a data flow.
Replication Flows You can use the connection to add source and target objects to a replication flow.
Note
You can only use a non-SAP target for a replication flow if your admin has assigned
capacity units to Premium Outbound Integration. For more information, see Premium
Outbound Integration and Configure the Size of Your SAP Datasphere Tenant.
Connection Details
Property Description
Endpoint Enter the endpoint URL of the Amazon S3 server, for example s3.amazonaws.com. The
protocol prefix is not required.
Note
When using Assume Role, you must enter the regional endpoint, for example s3.us-
west-2.amazonaws.com.
Protocol Select the protocol. The default value is HTTPS. The value that you provide overwrites the
value from the endpoint, if already set.
Root Path [optional] Enter the root path name for browsing objects. The value starts with the charac-
ter slash. For example, /My Folder/MySubfolder.
If you have specified the root path, then any path used with this connection is prefixed with
the root path.
Property Description
Use Server Side Encryption [optional] Select true (default) if you want to use S3 objects encrypted through server side
encryption.
• Encryption with Amazon S3 Managed Keys (SSE-S3) (default) if your objects are en-
crypted using the default encryption configuration for objects in your Amazon S3
buckets.
• Encryption with AWS Key Management Service Keys (SSE-KMS) if your Amazon S3
buckets are configured to use AWS KMS keys.
[if Encryption Option = Enter the KMS key Amazon Resource Name (ARN) for the customer managed key which
Encryption with AWS Key has been created to encrypt the objects in your Amazon S3 buckets.
Management Service Keys
(SSE-KMS)] KMS Key ARN
Assume Role
Property Description
Use Assume Role [optional] Select true if you want to use temporary security credentials and restrict access
to your Amazon S3 buckets based on an Identity and Access Management role (IAM role).
The default value is false.
[if Use Assume Role = true] Enter the role Amazon Resource Name (ARN) of the assumed IAM role.
Role ARN
[if Use Assume Role = true] [optional] Enter the name to uniquely identify the assumed role session.
Role Session Name
[if Use Assume Role = true] [optional] Enter the duration. The default duration is 3600 seconds.
Duration of Role Session (in
Seconds)
[if Use Assume Role = true] [optional] If an external ID is used to make sure that SAP Datasphere as a specified third
External ID party can assume the role, enter the ID.
[if Use Assume Role = true] [optional] If you want to use a session policy, enter the respective IAM policy in JSON
Role Policy format.
Credentials
Property Description
Access Key Enter the access key ID of the user that is used to authenticate to Amazon S3.
Secret Key Enter the secret access key of the user that is used to authenticate to Amazon S3.
Features
Feature Description
Data Flows Data Flows are enabled without the need to set any additional connection properties.
Replication Flows Replication Flows are enabled without the need to set any additional connection properties.
For information about the required prerequisites, see Prepare Connectivity to Apache Kafka.
Supported Features
Replication Flows You can use the connection to add target objects to a replication flow.
Note
You can only use a non-SAP target for a replication flow if your admin has assigned
capacity units to Premium Outbound Integration. For more information, see Premium
Outbound Integration and Configure the Size of Your SAP Datasphere Tenant.
Connection Details
Property Description
Cloud Connector
Property Description
Use Cloud Connector [optional] Set to true if your source is an on-premise source and you want to use the
connection for replication flows. The default is false.
[if Use Cloud Connector = [optional] Select Derive Virtual Host and Port from Connection Details or Enter Virtual Host
true] Virtual Destination and Port in Separate Fields.
If host and port entered in the connection details match the virtual host and port from the
Cloud Connector configuration, you can select Derive Virtual Host and Port from Connection
Details and don't need to enter the values manually.
[if Virtual Destination = Enter Enter the virtual host that you defined during Cloud Connector configuration.
Virtual Host and Port in
Separate Fields] Virtual Host
[if Virtual Destination = Enter Enter the virtual port that you defined during Cloud Connector configuration.
Virtual Host and Port in
Separate Fields] Virtual Port
Authentication
Property Description
Authentication Type Select the authentication type to use to connect to the Kafka brokers.
• No Authentication
• User Name And Password
Note
For basic authentication with user name and password, we recommend to use TLS
encryption to ensure secure communication.
SASL Authentication Type [read-only] Displays the Simple Authentication and Security Layer (SASL) authentication
mechanism to use depending on your selection for Authentication Type.
Kerberos
If Authentication Type = Kerberos with Username and Password or Kerberos with Keytab File:
Property Description
Kafka Kerberos Service Name Enter the name of the Kerberos service used by the Kafka broker.
Kafka Kerberos Realm Enter the realm defined for the Kafka Kerberos broker.
Kafka Kerberos Config Upload the content of the krb5.conf configuration file.
Choose (Browse) and select the file from your download location.
Once uploaded, you can check the configuration file by clicking the (inspection) button.
Credentials (SCRAM)
If Authentication Type = Salted Challenge Response Authentication Mechanism (512) or Salted Challenge
Response Authentication Mechanism (256):
Property Description
Kafka SASL User Name Enter the user name that is used for Kafka SASL SCRAM authentication.
Kafka SASL Password Enter the Kafka SASL SCRAM connection password.
Property Description
User Name Enter the user name that is used for Kafka SASL PLAIN authentication.
Property Description
User Name Enter the name of the user that is used to connect to the Kerberos service.
Choose (Browse) and select the file from your download location.
Once uploaded, you can check the keytab file by clicking the (inspection) button.
Property Description
User Name Enter the name of the user that is used to connect to the Kerberos service.
Security
Property Description
Validate Server Certificate Select true (default) to validate the TLS server certificate.
[if Use TLS = true and Select true to use mutual authentication with validating both client and server certificates.
Validate Server Certificate = The default is false.
true:]Use mTLS
Note
The supported filename extensions for the key are .pem (privacy-enhanced mail)
and .key.
Note
The supported filename extensions for the certificate are .pem (privacy-enhanced mail)
and .crt.
Features
Feature Description
Replication Flows Replication Flows are enabled without the need to set any additional connection properties.
If your SAP Datasphere tenant is included in an SAP Business Data Cloud formation, Business Data Product
connections are used to connect to other SAP systems included in the same formation.
Business Data Product connections are managed differently compared to other connections:
When your SAP Datasphere tenant is added to an SAP Business Data Cloud formation, the connections to the
source systems of the formation become available in SAP Datasphere with predefined technichal and business
names.
An SAP Datasphere administrator can find the systems and corresponding connections on the (System)
(Configuration) Business Data Products page and can change the business name of a connection
if required. For a system and its corresponding connection, the administrator can authorize the spaces to
which data products provided by the connected system can be installed (see Authorize Spaces to Install SAP
Business Data Cloud Data Products).
When a data product is installed to one or more authorized target spaces, the Business Data Product
connection is created in an SAP-managed ingestion space (if it does not already exist). In the ingestion
space, limited functionality is available for the connection, such as validating the connection or changing its
business name (on the Edit dialog). The connection does not support the standard connection features but
only supports the replication flows that have been created and deployed during data product installation.
Note
Business Data Product connections do not appear in the target spaces, neither in the Connections app nor
in the Source Browser panel of the Data Builder.
Related Information
Use the connection to connect to Apache Kafka hosted on either the Confluent Platform or Confluent Cloud.
The connection type has two endpoints: the Kafka brokers and the Schema Registry.
For information about the required prerequisites, see Prepare Connectivity to Confluent.
Replication Flows You can use the connection to add source and target objects to a replication flow.
Note
You can only use a non-SAP target for a replication flow if your admin has assigned
capacity units to Premium Outbound Integration. For more information, see Premium
Outbound Integration and Configure the Size of Your SAP Datasphere Tenant.
With Schema Registry, replication flows use a schema reference when serializing a Kafka
topic message instead of providing the schema definition in every message. In the target
settings of the replication flow configuration, you can decide to use Schema Registry.
Connection Details
Property Description
Cloud Connector
Property Description
Use Cloud Connector Set the property to true if your platform is on-premise. The default is false.
[if Use Cloud Connector = Select the location ID for the Cloud Connector instance that is set up for connecting to the
true] Location Kafka brokers.
Note
To select another location ID than the default location, Connection.Read privilege is
required. The privilege is neither included in the DW Integrator nor in the DW Space
Administrator role. If you need to select a location ID, ask your tenant administrator to
either assign your user to a global role that is based on the DW Administrator role or
to assign your user to a custom global role (with license type SAP Datasphere) that
includes the required Connection.Read privilege.
[if Use Cloud Connector = [optional] Select Derive Virtual Host and Port from Connection Details or Enter Virtual Host
true] Virtual Destination and Port in Separate Fields.
If host and port entered in the connection details match the virtual host and port from the
Cloud Connector configuration, you can select Derive Virtual Host and Port from Connection
Details and don't need to enter the values manually.
[if Virtual Destination = Enter Enter the virtual host that you defined during Cloud Connector configuration.
Virtual Host and Port in
Separate Fields] Virtual Host
[if Virtual Destination = Enter Enter the virtual port that you defined during Cloud Connector configuration.
Virtual Host and Port in
Separate Fields] Virtual Port
Authentication
Property Description
Authentication Type Select the authentication type to use to connect to the Kafka brokers.
• No Authentication
• User Name And Password
Note
For basic authentication with user name and password, we recommend to use TLS
encryption to ensure secure communication.
SASL Authentication Type [read-only] Displays the Simple Authentication and Security Layer (SASL) authentication
mechanism to use depending on your selection for Authentication Type.
Property Description
Kafka Kerberos Service Name Enter the name of the Kerberos service used by the Kafka broker.
Kafka Kerberos Realm Enter the realm defined for the Kafka Kerberos broker.
Kafka Kerberos Config Upload the content of the krb5.conf configuration file.
Choose (Browse) and select the file from your download location.
Once uploaded, you can check the configuration file by clicking the (inspection) button.
Credentials (SCRAM)
If Authentication Type = Salted Challenge Response Authentication Mechanism (512) or Salted Challenge
Response Authentication Mechanism (256):
Property Description
Kafka SASL User Name Enter the user name that is used for Kafka SASL SCRAM authentication.
Kafka SASL Password Enter the Kafka SASL SCRAM connection password.
Property Description
User Name Enter the user name that is used for Kafka SASL PLAIN authentication.
Property Description
User Name Enter the name of the user that is used to connect to the Kerberos service.
Choose (Browse) and select the file from your download location.
Once uploaded, you can check the keytab file by clicking the (inspection) button.
Property Description
User Name Enter the name of the user that is used to connect to the Kerberos service.
Property Description
Validate Server Certificate Select true (default) to validate the TLS server certificate.
[if Use TLS = true and Select true to use mutual authentication with validating both client and server certificates.
Validate Server Certificate = The default is false.
true:]Use mTLS
Note
The supported filename extensions for the key are .pem (privacy-enhanced mail)
and .key.
Note
The supported filename extensions for the certificate are .pem (privacy-enhanced mail)
and .crt.
Schema Registry
Property Description
URL Enter the URL of the Schema Registry service. The required format is <protocol>://
<host>:<port>.
Authentication Type Select the authentication type to be used to connect to the Schema Registry.
Note
We recommended that you configure Schema Registry to use HTTPS for secure
communication, because the basic protocol passes user name and password in
plain text.
• No Authentication
User Name [if Authentication Type = User Name And Password] Enter the name of the user used to
connect to the Schema Registry.
Password [if Authentication Type = User Name And Password] Enter the password.
Property Description
Use Cloud Connector Select the location ID for the Cloud Connector instance that is set up for connecting to the
Schema Registry.
Note
Since Schema Registry might be used in another location than the Kafka brokers, you
have to enter the Cloud Connector properties for Schema Registry separately from the
properties for the Kafka brokers.
[if Use Cloud Connector = [optional] Select Derive Virtual Host and Port from Connection Details or Enter Virtual Host
true] Virtual Destination and Port in Separate Fields.
If host and port entered in the connection details match the virtual host and port from the
Cloud Connector configuration, you can select Derive Virtual Host and Port from Connection
Details and don't need to enter the values manually.
[if Virtual Destination = Enter Enter the virtual host that you defined during Cloud Connector configuration.
Virtual Host and Port in
Separate Fields] Virtual Host
[if Virtual Destination = Enter Enter the virtual port that you defined during Cloud Connector configuration.
Virtual Host and Port in
Separate Fields] Virtual Port
Property Description
Validate Server Certificate Select true (default) to validate the TLS server certificate.
Features
Feature Description
Replication Flows Replication Flows are enabled without the need to set any additional connection properties.
Connection Details
Property Description
Authentication
Property Description
Authentication Type [read-only] Displays API Key and Secrect for authentication based on API keys.
Property Description
API Key Enter the Confluent Cloud API key that is used to control access to Confluent Cloud.
Security
Property Description
Validate Server Certificate Select true (default) to validate the TLS server certificate.
Schema Registry
Property Description
URL Enter the URL of the Schema Registry service. The required format is <protocol>://
<host>:<port>.
Authentication Type Select the authentication type to be used to connect to the Schema Registry.
Note
We recommended that you configure Schema Registry to use HTTPS for secure
communication, because the basic protocol passes user name and password in
plain text.
User Name Enter the Confluent Cloud API key that is used to control access to the Schema Registry.
Property Description
Validate Server Certificate Select true (default) to validate the TLS server certificate.
Features
Feature Description
Replication Flows Replication Flows are enabled without the need to set any additional connection properties.
Use a Cloud Data Integration connection to access data from SAP cloud applications which provide OData-
based APIs for data integration and have a Cloud Data Integration (CDI) provider service implemented.
Prerequisites
The connection type supports the remote table as well as the data flow feature.
• Remote
• Replication (snapshot)
• Replication (real-time)
For more information, see Replicating Data and Monitoring Remote Tables [page 168].
Connection Details
Property Description
Root Path [optional] Enter the root path name to restrict browsing to a
certain CDI namespace or provider.
Cloud Connector
Property Description
Use Cloud Connector [optional] Set to true if your source is an on-premise source and you want to use the
connection for data flows. The default is false.
[if Use Cloud Connector = [optional] Select Derive Virtual Host and Port from Connection Details or Enter Virtual Host
true] Virtual Destination and Port in Separate Fields.
If host and port entered in the connection details match the virtual host and port from the
Cloud Connector configuration, you can select Derive Virtual Host and Port from Connection
Details and don't need to enter the values manually.
[if Virtual Destination = Enter Enter the virtual host that you defined during Cloud Connector configuration.
Virtual Host and Port in
Separate Fields] Virtual Host
[if Virtual Destination = Enter Enter the virtual port that you defined during Cloud Connector configuration.
Virtual Host and Port in
Separate Fields] Virtual Port
Property Description
Authentication Type Select the authentication type to use to connect to the Cloud
Data Integration service endpoint.
Property Description
X.509 Client Certificate To upload the certificate or certificate chain that is used to authenticate to the remote
system, click (Browse) and select the file from your download location.
X.509 Client Private Key To upload the private key, click (Browse) and select the file from your download location.
X.509 Client Private Key [optional] If the private key is encrypted, enter the password required for decryption.
Password
OAuth 2.0
If Authentication Type = OAuth 2.0:
Property Description
• Client Credentials
• User Name and Password
OAuth Token Endpoint Enter the token endpoint that the application must use.
OAuth Response Type Select the OAuth response type. The values are token or
none.
OAuth Token Request Content Type Select the value for the content-type HTTP header that the
application must use when requesting a token. The values
are URLEncoded or JSON.
Property Description
Property Description
Property Description
User Name Enter the user name that the application must use for authentication.
Features
Data Flows are enabled without the need to set any additional connection properties.
Use a Generic JDBC connection to access data from tables and views in any supported data source for which a
JDBC driver is available.
You can use this connection type to connect to most databases for which SAP Datasphere does not already
provide a connection type. In general, the connection type supports most databases that have SQL-based data
types and functions, and a JDBC driver. For latest information about supported data sources and versions, see
the SAP HANA smart data integration Product Availability Matrix (PAM) .
Note
For information about unsupported data sources, see SAP Note 3130999 .
Prerequisites
• Remote
• Replication (snapshot)
For more information, see Replicating Data and Monitoring Remote Tables [page 168].
Connection Details
Property Description
JDBC Driver Class Enter the JDBC driver class for the database you are using.
Note
The JDBC URL is stored as plain text and is not stored in
encrypted form. It is visible when editing the connection
or when accessing the remote source with a database
analysis user. Therefore, do not provide any sensitive
information like passwords in the URL.
Credentials
Property Description
Features
Prerequisites
The connection type supports the remote table as well as the data flow feature.
• Remote
• Replication (snapshot)
For more information, see Replicating Data and Monitoring Remote Tables [page 168].
Connection Details
Property Description
Note
The OData service URL needs to be publicly available.
Version Select the OData version used to implement the OData service (V2 or V4).
Cloud Connector
Property Description
Use Cloud Connector [optional] Set to true if your source is an on-premise source and you want to use the
connection for data flows. The default is false.
[if Use Cloud Connector = [optional] Select Derive Virtual Host and Port from Connection Details or Enter Virtual Host
true] Virtual Destination and Port in Separate Fields.
If host and port entered in the connection details match the virtual host and port from the
Cloud Connector configuration, you can select Derive Virtual Host and Port from Connection
Details and don't need to enter the values manually.
[if Virtual Destination = Enter Enter the virtual host that you defined during Cloud Connector configuration.
Virtual Host and Port in
Separate Fields] Virtual Host
[if Virtual Destination = Enter Enter the virtual port that you defined during Cloud Connector configuration.
Virtual Host and Port in
Separate Fields] Virtual Port
Authentication
Property Description
• No Authentication
• User Name And Password for basic authentication
• OAuth 2.0 (default value)
OAuth 2.0
If Authentication Type = OAuth 2.0:
Property Description
OAuth Grant Type Displays Client Credentials as grant type used to retrieve an
access token.
OAuth Token Endpoint Enter the API endpoint to use to request an access token.
OAuth Token Request Content Type [optional] Enter the content type of the OAuth2 token re-
quest.
Property Description
Property Description
Features
Remote Tables are enabled without the need to set any additional connection properties.
Data Flows are enabled without the need to set any additional connection properties. If your source is an
on-premise source, make sure you have maintained the properties in the Cloud Connector section.
Different OData services support different sets of custom headers to pass parameters to the service when
using the connection. You can specify one or more custom HTTP headers to add them to the OData request.
Click Add to add name and value for a custom HTTP header field.
Note
Use a Generic SFTP connection to access files on a Secure File Transfer Protocol (SFTP) server.
• JSON, JSONL
• Parquet
• ORC
• CSV
• XLS, XLSX
For information about the required prerequisites, see Prepare Connectivity for Generic SFTP.
Connection Details
Property Description
Category Select the category. Select Cloud if you want to connect to an SFTP server in the public
cloud, or select On-Premise if you want to connect to an SFTP server in your local network.
Host Key Enter the public SSH host key (public key of the SFTP server, not the key fingerprint).
Choose (Browse) and select the file from your download location.
If you don’t know your host key, ask your administrator or use command line tools like ssh-
keyscan to obtain the host key (only through a trustable channel). For more information, see
Prepare Connectivity for Generic SFTP.
Root Path [optional] Enter the root path name for browsing objects. The value starts with the charac-
ter slash. For example, /My Folder/MySubfolder.
If you have entered root path, then any path used with this connection is prefixed with the
root path.
Property Description
Use Cloud Connector [optional] Set to true if your source is an on-premise source and you want to use the
connection for data flows. The default is false.
[if Use Cloud Connector = [optional] Select Derive Virtual Host and Port from Connection Details or Enter Virtual Host
true] Virtual Destination and Port in Separate Fields.
If host and port entered in the connection details match the virtual host and port from the
Cloud Connector configuration, you can select Derive Virtual Host and Port from Connection
Details and don't need to enter the values manually.
[if Virtual Destination = Enter Enter the virtual host that you defined during Cloud Connector configuration.
Virtual Host and Port in
Separate Fields] Virtual Host
[if Virtual Destination = Enter Enter the virtual port that you defined during Cloud Connector configuration.
Virtual Host and Port in
Separate Fields] Virtual Port
Authentication
Property Description
Credentials (SSH)
If Authentication Type = SSH:
Property Description
User Name Enter the user who is accessing the SFTP server.
Private Key Enter the user SSH private key used for SSH key authentication. The server must know the
user SSH public key.
Choose (Browse) and select the file from your download location.
Passphrase Enter the passphrase needed to decrypt the private key in SSH key authentication.
Property Description
Features
Data Flows are enabled without the need to set any additional connection properties. If your source is an
on-premise source, make sure you have maintained the properties in the Cloud Connector section.
Use the connection to connect to and access data from Google BigQuery.
For information about the required prerequisites, see Prepare Connectivity to Google BigQuery.
Supported Features
Remote Tables You can use remote tables imported from the connection either to access data directly live
in the connected source (federation) or to copy the data into SAP Datasphere (snapshot
replication).
For more information, see Replicating Data and Monitoring Remote Tables [page 168].
Data Flows You can use the connection to add source objects to a data flow.
Replication Flows You can use the connection to add target objects to a replication flow.
Note
You can only use a non-SAP target for a replication flow if your admin has assigned
capacity units to Premium Outbound Integration. For more information, see Premium
Outbound Integration and Configure the Size of Your SAP Datasphere Tenant.
Connection Details
Property Description
Project Enter the ID of the Google Cloud project to which you want
to connect. You can find the project ID in the json key file that
you need for the credentials.
Note
If you provide an invalid location, the connection valida-
tion still passes, but failures will happen when you're
using the connection in the Data Builder.
Credentials
Property Description
Key Upload the json key file that is used for authentication.
Features
Feature Description
Remote Tables Remote Tables are enabled without the need to set any additional connection properties.
Data Flows Data Flows are enabled without the need to set any additional connection properties.
Replication Flows Replication Flows are enabled without the need to set any additional connection properties.
Use the connection to connect to and access objects from Google Cloud Storage.
Data Flows You can use the connection to add source objects to a data flow.
Replication Flows You can use the connection to add source and target objects to a replication flow.
Note
You can only use a non-SAP target for a replication flow if your admin has assigned
capacity units to Premium Outbound Integration. For more information, see Premium
Outbound Integration and Configure the Size of Your SAP Datasphere Tenant.
Connection Details
Property Description
Root Path [optional] Enter the root path name for browsing objects.
The value starts with the character slash. For example, /My
Folder/MySubfolder.
If you have specified the root path, then any path used with
this connection is prefixed with the root path.
Credentials
Property Description
Key Enter the content of the json key file that is used for authen-
tication.
Features
Feature Description
Data Flows Data Flows are enabled without the need to set any additional connection properties.
Replication Flows Replication Flows are enabled without the need to set any additional connection properties.
Use the connection to access objects from an Hadoop Distributed File System (HDFS) server.
Connection Details
Property Description
Note
Along with Remote Procedure Call (RPC), HDFS can also extend connections with
WebHDFS.
If you select RPC, make sure that both the NameNode and the DataNodes are exposed
and can be connected from SAP Datasphere. For more information, see Finding SAP
Datasphere IP addresses.
Port Enter the port of the HDFS namenode. If you do not provide any value, the default value for
the selected protocol is used.
Root Path [optional] Enter the root path name for browsing objects. The value starts with the charac-
ter slash. For example, /My Folder/MySubfolder.
If you have specified the root path, then any path used with this connection is prefixed with
the root path.
Custom Parameters
Property Description
Authentication
Property Description
Authentication Type Select the authentication type that the application must use.
To connect to HDFS the application supports User Name And Password, Simple, or Kerberos
authentication mechanisms.
Property Description
Credentials (Simple)
If Authentication Type = Simple:
Property Description
Credentials (Kerberos)
If Authentication Type = Kerberos:
Property Description
keytab Enter the content of the keytab file of the user from the local system.
Choose (Browse) and select the file from your download location.
Configuration
If Authentication Type = Kerberos:
Property Description
krb5.config Enter the content of the krb5.conf configuration file from your local system.
Choose (Browse) and select the file from your download location.
Data Flows are enabled without the need to set any additional connection properties.
Connection Details
Property Description
Protocol Select the protocol (wasb for unencrypted access or wasbs for TLS-encrypted access). The
default is wasbs.
Account Name Enter the name of the Microsoft Azure storage account through which Microsoft Azure Blob
Storage is provisioned.
Root Path [optional] Enter the root path name for browsing objects. The value starts with the charac-
ter slash. For example, /My Folder/MySubfolder.
If you have entered root path, then any path used with this connection is prefixed with the
root path.
Endpoint Suffix Displays the default value for the endpoint suffix which is "core.windows.net". You can
override the default value according to the domain assigned to the endpoint of your blob
service.
Property Description
Account Key Enter the storage account access key (shared key authorization).
Features
Data Flows are enabled without the need to set any additional connection properties.
Use the connection to access objects in Microsoft Azure Data Lake Gen1 (ADL Gen1).
Caution
After your Microsoft Azure Data Lake Storage Gen1 has been migrated to Microsoft Azure Data Lake
Storage Gen2, recreate existing Gen1 connections with the Gen2 connection type and adapt your models
and processes to the new connections.
Connection Details
Property Description
Storage Account Name Enter the name of the Azure Data Lake Storage Gen1 account.
Tenant ID Enter the ID of the Azure Data Lake Storage Gen1 tenant.
Root Path [optional] Enter the root path name for browsing. It starts with a slash and the file system
name. For example /MyFileSystem/MyFolder. The file system must be provided.
Credentials
Property Description
Client Key Enter the client key (also referred to as client secret or authentication key).
Features
Data Flows are enabled without the need to set any additional connection properties.
Use the connection to connect to and access objects in Microsoft Azure Data Lake Gen2 (ADL Gen2).
For information about the required prerequisites, see Prepare Connectivity to Microsoft Azure Data Lake Store
Gen2.
Supported Features
Data Flows You can use the connection to add source objects to a data flow.
Replication Flows You can use the connection to add source and target objects to a replication flow.
Note
You can only use a non-SAP target for a replication flow if your admin has assigned
capacity units to Premium Outbound Integration. For more information, see Premium
Outbound Integration and Configure the Size of Your SAP Datasphere Tenant.
Connection Details
Property Description
Storage Account Name Enter the name of the Azure Data Lake Storage Gen2 used for authentication.
Root Path [optional] Enter the root path name for browsing. It starts with a slash and the file system
name. For example /MyFileSystem/MyFolder. The file system must be provided. Any
path used with this connection will be prefixed with this root path.
Property Description
Shared key provides full access to your storage account while with shared access signature
you can provide secure delegate access to the storage account resources.
Property Description
Account Key Enter the account key used in the shared key authorization.
Property Description
SAS Token Enter the shared access signature token (SAS token) used in shared access signature
authentication.
OAuth 2.0
Property Description
[if OAuth Grant Type = Enter the token endpoint that the application must use to get the access token.
Client Credentials with X.509
Client Certificate or Client
Credentials] OAuth Token
Endpoint
[if OAuth Grant Type = User Enter the client endpoint to get the access token for authorization method User Name and
Name and Password] OAuth Password.
Client Endpoint
Property Description
X.509 Client Certificate To upload the certificate or certificate chain that is used to
authenticate to the remote system, click (Browse) and
select the file from your download location.
X.509 Client Private Key To upload the private key, click (Browse) and select the
file from your download location.
Property Description
Property Description
Features
Feature Description
Data Flows Data Flows are enabled without the need to set any additional connection properties.
Replication Flows Replication Flows are enabled without the need to set any additional connection properties.
Use a Microsoft Azure SQL Database connection to access table data from a Microsoft Azure SQL database.
For information about the required prerequisites, see Prepare Connectivity to Microsoft Azure SQL Database.
Remote Tables You can use remote tables imported from the connection either to access data directly live
in the connected source (federation) or to copy the data into SAP Datasphere (replication).
For remote tables, real-time replication is supported. For information about any constraints,
see Replicate Data Changes in Real-Time [page 181].
Data Flows You can use the connection to add source objects to a data flow.
Replication Flows You can use the connection to add source objects to a replication flow.
Note
The connection type supports replication with both remote tables via Data Provisioning Agent (SAP HANA
Smart Data Integration) and replication flows.
Whether you use the same connection or two separate connections, we strongly recommend using only
remote tables or only replication flows for replicating data from a dedicated source object, but not both.
Connection Details
Property Description
Port Enter the port number of the Azure server. The default is 1433.
Database Name Enter the name of the database to which you want to connect.
Security
Property Description
Credentials
Property Description
User Name Enter the name of the Azure user that has privileges to connect to the database.
Feature Description
Data Flows Data Flows are enabled without the need to set any additional connection properties.
Replication Flows Replication Flows are enabled without the need to set any additional connection properties.
Use a Microsoft SQL Server connection to access data from a Microsoft SQL Server database (on-premise).
Prerequisites
The connection type supports the remote table as well as the data flow feature.
• Remote
• Replication (snapshot)
• Replication (real-time)
Note
SAP Datasphere uses trigger-based replication. For more information, see Microsoft SQL Server
Trigger-Based Replication in the SAP HANA Smart Data Integration and SAP HANA Smart Data Quality
Installation and Configuration Guide.
For more information, see Replicating Data and Monitoring Remote Tables [page 168].
Connection Details
Property Description
Note
Connecting via instance name is not supported.
Port Enter the Microsoft SQL Server port number. The value range is 1–65535.
Note
Dynamic ports are not supported.
Version Select the Microsoft SQL Server version. Supported versions are Microsoft SQL Server
2012, 2014, 2016, 2017, 2019, and 2022 (default is 2022).
Note
Microsoft SQL Server 2012 is not supported for remote tables.
Cloud Connector
Note
Cloud Connector is not required if your Microsoft SQL Server database is available on the public internet.
Property Description
Use Cloud Connector [optional] Set to true if your source is an on-premise source and you want to use the
connection for data flows. The default is false.
[if Use Cloud Connector = [optional] Select Derive Virtual Host and Port from Connection Details or Enter Virtual Host
true] Virtual Destination and Port in Separate Fields.
If host and port entered in the connection details match the virtual host and port from the
Cloud Connector configuration, you can select Derive Virtual Host and Port from Connection
Details and don't need to enter the values manually.
[if Virtual Destination = Enter Enter the virtual host that you defined during Cloud Connector configuration.
Virtual Host and Port in
Separate Fields] Virtual Host
[if Virtual Destination = Enter Enter the virtual port that you defined during Cloud Connector configuration.
Virtual Host and Port in
Separate Fields] Virtual Port
Security
Property Description
Use SSL Select whether you’re using SSL. The default value is true.
[if Use SSL = true] Host Enter the host name that is in the SSL certificate.
Name in Certificate
Credentials
Property Description
Features
Data Flows are enabled without the need to set any additional connection properties. Make sure you have
maintained the properties in the Cloud Connector section.
When you've selected a Data Provisioning Agent, you can configure advanced connection properties to enable
specific customer scenarios. In the connection creation wizard, you configure advanced properties in an
additional wizard step. When editing the connection to configure the properties, scroll down and click the Show
Advanced Properties button. Advanced properties have default values that you can override according to your
customer scenario’s needs.
Available properties:
For more information, see Microsoft SQL Server Log Reader Remote Source Configuration in the SAP HANA
Smart Data Integration and SAP HANA Smart Data Quality Installation and Configuration Guide.
Use an Open Connectors connection to access data from sources that are connected to the SAP Open
Connectors account that is integrated with your space.
You can use the broad SAP Open Connectors connectivity available from SAP Business Technology Platform
directly within your SAP Datasphere space.
Note
Using SAP Open Connectors may incur additional charges. You can review your usage in your SAP BTP
cockpit.
With SAP Open Connectors, (API) hubs provide uniform APIs to access a collection of resources enabling
one-to-many API access to multiple API providers. With integrating SAP Open Connectors SAP Datasphere
gets access to third-party cloud applications that are available with SAP Open Connectors hub categories.
Each hub category contains connectors each representing a pre-built API integration that enables a connection
into a specific API Provider endpoint, for example Salesforce, Snowflake, or Microsoft SharePoint. For any
SAP Datasphere supports the following SAP Open Connectors hub categories:
• CRM
• DB
• General
• Marketing
• Social
• Documents
Note
• For information about SAP Open Connectors availability in data centers, see SAP Note 2903776 .
• Some SAP Open Connectors types might not work in case the API hasn't been fully implemented.
We recommend to check the connector on SAP Open Connectors side first before using it in
SAP Datasphere. For more information, see Working with Connectors in the SAP Open Connectors
documentation.
• For sources which SAP Datasphere natively supports with dedicated connection types, such as
Amazon S3 or Microsoft SQL Server, use the corresponding SAP Datasphere connection type. We
do not recommend using SAP Open Connectors connectivity in this case.
• Input parameters are not supported for data flows. For resources with mandatory parameters the
parameters have to be replaced in SAP Open Connectors, or the respective connector has to be
cloned and adjusted to work without any mandatory parameter. For more information, see Setup
Configuration and Parameters (for custom connectors) in the SAP Open Connectors documentation.
Prerequisites
Connection Details
Property Description
Open Connectors Instance Select the instance to which you want to connect.
Data Flows are enabled without the need to set any additional connection properties.
Use the connection to connect to and access data from an Oracle database (on-premise).
For information about the required prerequisites, see Prepare Connectivity to Oracle.
Supported Features
Remote Tables You can use remote tables imported from the connection either to access data directly live
in the connected source (federation) or to copy the data into SAP Datasphere (replication).
For remote tables, real-time replication is supported. For information about any constraints,
see Replicate Data Changes in Real-Time [page 181].
Note
SAP Datasphere uses trigger-based replication. For more information, see Oracle Trig-
ger-Based Replication in the SAP HANA Smart Data Integration and SAP HANA Smart
Data Quality Installation and Configuration Guide.
Data Flows You can use the connection to add source objects to a data flow.
Connection Details
Property Description
Host Enter the host name or IP address on which the remote Oracle database is running.
Note
The port number can vary depending on whether SSL is used or not.
Service Name Enter the service name of Oracle database. When creating a remote source, you must set
only one of the following parameters: Database Name and Service Name.
If you set both, the Data Provisioning Agent connects to Oracle by the service name as the
first choice.
Version Select a version. Supported versions are Oracle 12c, Oracle 18c, and Oracle 19c. Default
version is Oracle 19c.
Cloud Connector
Note
Cloud Connector is not required if your Oracle database is available on the public internet.
Property Description
Use Cloud Connector [optional] Set to true if your source is an on-premise source and you want to use the
connection for data flows. The default is false.
[if Use Cloud Connector = [optional] Select Derive Virtual Host and Port from Connection Details or Enter Virtual Host
true] Virtual Destination and Port in Separate Fields.
If host and port entered in the connection details match the virtual host and port from the
Cloud Connector configuration, you can select Derive Virtual Host and Port from Connection
Details and don't need to enter the values manually.
[if Virtual Destination = Enter Enter the virtual host that you defined during Cloud Connector configuration.
Virtual Host and Port in
Separate Fields] Virtual Host
[if Virtual Destination = Enter Enter the virtual port that you defined during Cloud Connector configuration.
Virtual Host and Port in
Separate Fields] Virtual Port
Property Description
Use SSL Select whether you’re using SSL. The default value is true.
Note
If you select SSL and you want to use the connection for data flows, in addition to
the ODBC driver file, additional files must have been uploaded to SAP Datasphere (see
Upload Third-Party ODBC Drivers (Required for Data Flows)).
[if Use SSL = true] Enter the distinguished name (DN) of the primary data server certificate.
Distinguished Name (DN) in
Certificate Note
If this parameter is set, the DN field in the server certificate is verified to match this
parameter. If it doesn’t match, the connection to the primary data server fails.
Credentials
Property Description
Features
Data Flows are enabled without the need to set any additional connection properties.
When you've selected a Data Provisioning Agent, you can configure advanced connection properties to enable
specific customer scenarios. In the connection creation wizard, you configure advanced properties in an
additional wizard step. When editing the connection to configure the properties, scroll down and click the Show
Advanced Properties button. Advanced properties have default values that you can override according to your
customer scenario’s needs.
Available properties:
For more information, see Oracle Log Reader Remote Source Configuration in the SAP HANA Smart Data
Integration and SAP HANA Smart Data Quality Installation and Configuration Guide.
Extend connectivity beyond SAP Datasphere standard remote connectivity and cover additional data sources
that are available with Precog.
Precog provides connectivity to a variety of data sources, including marketing data sources, databases and
others.
When creating a connection to a partner tool, an Open SQL schema is generated in SAP Datasphere. To be
able to send data to the Open SQL schema, the IP addresses of the partner tool need to be added to the
SAP Datasphere IP allowlist, and during connection creation the partner tool on their side needs to establish a
connection to the schema.
Wizard step Configure partner settings shows an embedded Precog UI. Here, you log on to the regional Precog
server and perform the following steps:
Note
As a prerequisite to proceed to the next step in the wizard, your browser needs to be enabled to store
cookies. This is not the case if you run the browser in incognito mode.
You can now see the destination with the name of the generated Open SQL schema in Precog and start loading
data for the tables available with Precog.
Use an SAP ABAP connection to access data from SAP ABAP on-premise systems through RFC or to access
data from cloud source systems such as SAP S/4HANA Cloud through Web Socket RFC.
For information about the required prerequisites, see Prepare Connectivity to SAP ABAP Systems.
Remote Tables You can use remote tables imported from the connection either to access data directly live
in the connected source (federation) or to copy the data into SAP Datasphere (replication).
• Data from virtual tables through RFC for ODP sources (extractors):
• Extraction context ABAP_CDS provides access to ABAP Core
Data Services (CDS) Views that include the annotation
@Analytics.dataextraction.enabled: true (in SAP ABAP on-prem-
ise systems) and do not contain the annotation @Analytics.Query: true.
For information about which ABAP CDS Views are available for extraction, see CDS
Views Enabled for Data Extraction in the SAP S/4HANA documentation.
Note
For SAP S/4HANA on-premise, ABAP CDS Views are the preferred method
over SAPI extractors when integrating data with SAP Datasphere.
Note
For accessing CompositeProviders and Queries in SAP BW∕4HANA we
strongly recommended to use the model import with the specific connec-
tion type SAP BW/4HANA Model Transfer. For more information, see SAP
BW∕4HANA Model Transfer Connections [page 83].
Note
To import BW Hierarchies and deploy them as remote tables in SAP
Datasphere, use the /BI*/H<InfoObject name> tables with data access via
SAP ABAP Dictionary tables instead of using ODP extraction context BW. In
the Data Builder, you can find the tables in the ABAPTABLES folder of the
connection.
Note
Importing SAPI DataSources into SAP Datasphere is not supported if the
DataSource contains mandatory selection fields. You will receive an error mes-
sage when trying to import tables for such DataSources.
• Data from SAP ABAP Dictionary tables in SAP ABAP on-premise systems
Real-time replication is supported for ODP sources. For information about any constraints,
see Replicate Data Changes in Real-Time [page 181].
Note
Certain SAPI DataSources (from FI-AA Business Content, for example) may send
duplicate records during snapshot replication or real-time initialization. Remote table
replication runs with optimized INSERT processing on DataSource key level which leads
to unique constraint violation errors. In such cases, we recommend to use a data flow
which uses the remote table as source and for which the target table has selected the
Append mode with the Update Records By Primary Key (UPSERT) option.
Data Flows You can use the connection to add source objects to a data flow.
• extraction-enabled ABAP CDS views that are C1-released, that is views with annotation
@Analytics.dataextraction.enabled: true and that are available in the
connected system (access via ABAP Pipeline Engine)
For more information, see:
• in the SAP S/4HANA Cloud documentation: Data Integration and CDS Views Ena-
bled for Data Extraction
• in the SAP S/4HANA documentation: CDS Views Enabled for Data Extraction
• Tables from SAP Landscape Transformation Replication Server (SAP LT Replication
Server) sources, for which an appropriate configuration has been created in SAP LT
Replication Server. In the Sources of the data flow editor, you then can find the source
objects in an SLT - SAP LT Replication Server folder of the connection.
Note
• The availability of the data flow feature depends on the used version and Support
Package level of SAP S/4HANA or the DMIS addon in the source. Make sure your
source systems meet the required minimum versions. We recommend to use the
latest available version of SAP S/4HANA and the DMIS add-on where possible and
have the latest SAP notes and TCI notes implemented in your systems.
For more information about required versions, recommended system landscape,
considerations for the supported source objects, and more, see SAP Note
2890171 .
• Data flows currently support one-time loading of data (also known as initial load)
only.
• The data preview in the data flow editor of the Data Builder is not available, except
for:
• CDS views if the version of the connected system is SAP S/4HANA on-prem-
ise 1909 or SAP S/4HANA 1909 or a higher version respectively.
• SAP Landscape Transformation Replication Server objects if ABAP Add-on
DMIS 2018 SP08 or DMIS 2020 SP04 is installed or a higher DMIS version.
Replication Flows You can use the connection to add source objects to a replication flow.
• extraction-enabled ABAP CDS views that are C1-released, that is views with annotation
@Analytics.dataextraction.enabled: true and that are available in the
connected system (access via ABAP Pipeline Engine)
For more information, see:
• in the SAP S/4HANA Cloud documentation: Data Integration and CDS Views Ena-
bled for Data Extraction
• in the SAP S/4HANA documentation: CDS Views Enabled for Data Extraction
Note
When adding source objects in replication flows, you can find the CDS views in the
CDS_EXTRACTION container.
• standard and custom CDS view entities that are exposed using the SQL service from
SAP BTP, ABAP environment, orSAP S/4HANA Cloud, respectively
For more information, see:
• in the SAP S/4HANA Cloud documentation: Data Consumption Using SAP
Datasphere
• in the SAP BTP documentation: Data Consumption Using SAP Datasphere
Note
When adding source objects in replication flows, you can find the CDS view entities
in the SQL_SERVICE container.
Note
• The availability of the replication flow feature depends on the used version and
Support Package level of SAP S/4HANA or the DMIS addon in the source. Make
sure your source systems meet the required minimum versions. We recommend to
use the latest available version of SAP S/4HANA and the DMIS add-on where pos-
sible and have the latest SAP notes and TCI notes implemented in your systems.
For more information about required versions, recommended system landscape,
considerations for the supported source objects, and more, see SAP Note
2890171 .
• The SAP BW and SAP ECC connection types don't support replication flows. In-
stead, you can use the SAP ABAP connection type to use replication flows for
replication from SAP BW, SAP BW/4HANA or SAP ECC systems.
The connection type supports replication with both remote tables via Data Provisioning Agent (SAP HANA
Smart Data Integration) and replication flows (for on-premise systems via Cloud Connector).
Regardless of whether you use the same connection or two separate connections, for replicating data from
a dedicated source object in the source we strongly recommend to only use remote tables or replication
flows and not both.
See SAP Data Warehouse Cloud - First Guidance: Data Integration for ABAP Source Systems (published
February 2022) for general best practices for integrating data from ABAP source systems with SAP
Datasphere.
Connection Details
Property Description
Protocol Select RFC (Use for on Premise ABAP Systems) to connect to an ABAP on-premise system.
If you want to use the on-premise system for building data flows, you additionally need to
set the Cloud Connector properties.
Select Web Socket RFC (Use for S/4HANA Cloud Systems) if you want to use the connection
for data flows only and your source is a cloud source such as SAP S/4HANA Cloud.
[if Protocol = RFC (Use for Select Application Server to connect to a single application server, or select Message Server
on Premise ABAP Systems)] to use load balancing and connect to a message server that distributes the load to different
SAP Logon Connection Type application servers.
Note
The SAP Logon connection type you select here must match the connection type in
the system mapping in the Cloud Connector configuration (load balancing logon or
connecting to a specific application server).
[if SAP Logon Connection Enter the name of the application server to which you want to connect to.
Type = Application Server]
Application Server
[if SAP Logon Connection Enter the SAP ABAP system instance number.
Type = Application Server]
System Number
[if SAP Logon Connection Enter the name of the message server to which you want to connect to.
Type = Message Server]
Message Server
[if SAP Logon Connection Enter the message server port (numerical).
Type = Message Server]
Message Server Port
[if Protocol = Web Socket Enter the name of the application server to which you want to connect to.
RFC (Use for S/4HANA Cloud
Systems)] Application Server
[if Protocol = Web Socket Enter the port number of the WebSocket RFC connection endpoint.
RFC (Use for S/4HANA Cloud
Systems)] Port
Language Enter the two-digit ISO language code, EN for English or DE for German, for example. In the
Data Builder, object and field descriptions are then shown in the specified language. If not
set, the application uses the default logon language of the ABAP system for data flows, and
EN is used for remote tables and views.
Cloud Connector
Property Description
Use Cloud Connector Set to true if your source is an on-premise source and you want to use the connection for
data flows or replication flows. The default is false.
[if Use Cloud Connector = Select Derive Virtual Host and Port from Connection Details or Enter Virtual Host and Port in
true] Virtual Destination Separate Fields.
If host and port entered in the connection details match the virtual host and port from the
Cloud Connector configuration, you can select Derive Virtual Host and Port from Connection
Details and don't need to enter the values manually.
Note
If the SAP Logon Connection Type for your connection is Application Server: When
you select Derive Virtual Host and Port from Connection Details for ABAP-based connec-
tions, virtual port is set to sapgw<system number> to make sure that it matches
the virtual port defined in the Cloud Connector configuration.
[if Virtual Destination = Enter Enter the virtual host that you defined during Cloud Connector configuration.
Virtual Host and Port in
Separate Fields] Virtual Host
[if Virtual Destination = Enter Enter the virtual port that you defined during Cloud Connector configuration.
Virtual Host and Port in
Separate Fields] Virtual Port Note
• If the SAP Logon Connection Type for your connection is Application Server: Enter
sapgw<system number> to make sure that it matches the virtual port defined
in the Cloud Connector configuration.
• If the SAP Logon Connection Type for your connection is Message Server: Enter
a numerical port and make sure it is the port defined in the Cloud Connector
configuration.
Authentication
Property Description
Authentication Type Select the authentication type to use to connect to the SAP ABAP system.
• User Name And Password for basic authentication (default value) - This option is read-
only if you set Cloud Connector to false.
• OAuth 2.0 - You can select this option only if you have selected the RFC protocol and
if you have set Cloud Connector to true to enable replication flows and data flows.
Remote tables currently are not supported with OAuth authentication.
OAuth 2.0
Property Description
OAuth Grant Type Displays Client Credentials as grant type used to retrieve an
access token.
OAuth Token Endpoint Enter the API endpoint to use to request an access token.
OAuth Token Request Content Type [optional] Enter the content type of the OAuth2 token re-
quest.
Property Description
Property Description
User Name Enter the user name that is used to connect to the SAP ABAP system.
Feature Description
Property Description
Data Provisioning Select an agent if you want to use the connection to access data via
Agent imported remote tables and to build views.
[if you selected an Set to On if you want to use ABAP RFC streaming for loading tables
agent] Streaming based on small batches. Set to Off to use non-RFC streaming. The
Read default is Off.
[if Streaming Read = Enter the gateway host where the ABAP adapter would register
On] Gateway Host an RFC server instance to receive callbacks with the table data
batches. Usually, the gateway host is the same as the target ABAP
system host.
[if Streaming Read = Enter the port number of the gateway server where the ABAP
On] Gateway Port adapter would register an RFC server instance to receive call-
backs with the table data batches. You can find the port
number for sapgw<ABAP_instance_number> in file /etc/
services on the ABAP host. The default port number is
33<ABAP_instance_number>. If you have a file /etc/
services with this mapping on the Data Provisioning Agent’s
host, you can also enter sapgw<ABAP_instance_number>
instead of the port number.
[if Streaming = On] Enter the name of the RFC destination that you have created in
RFC Destination the source. For more information, see Prerequisites for ABAP RFC
Streaming.
Data Flows Data Flows are enabled without the need to set any additional connection properties. If your
source is an on-premise source, make sure you have maintained the properties in the Cloud
Connector section.
Replication Flows Replication Flows are enabled without the need to set any additional connection properties.
If your source is an on-premise source, make sure you have maintained the properties in the
Cloud Connector section.
When you've selected a Data Provisioning Agent, you can configure advanced connection properties to enable
specific customer scenarios. In the connection creation wizard, you configure advanced properties in an
additional wizard step. When editing the connection to configure the properties, scroll down and click the Show
Advanced Properties button. Advanced properties have default values that you can override according to your
customer scenario’s needs.
Available properties:
For more information, see SAP ABAP Adapter Remote Source Configuration in the SAP HANA Smart Data
Integration and SAP HANA Smart Data Quality Installation and Configuration Guide.
You can only display the connection. Here, you can copy the SAP BW service key, which you will need to enter
when you create an SAP BW bridge project in the SAP BW Bridge Modeling Tools. For more information, see
Copy the SAP BW Service Key.
Restrictions
The connection type SAP BW bridge connection doesn’t show up in any other space, it cannot be modified, and
you cannot create such a connection yourself.
SAP BW bridge connections only support remote tables and model import. Data flows and replication flows are
not supported.
Prerequisites
The connection type supports the remote table as well as the data flow feature.
Note
If you want to provide SAP BW∕4HANA business semantics to SAP Datasphere, you can use connection
type SAP BW/4HANA Model Import and import SAP BW∕4HANA models based on analytic queries. During
the import, the semantics of the underlying SAP BW∕4HANA models are translated into native SAP
Datasphere entities that you can use to access the data stored in SAP BW∕4HANA, to enhance the data in
SAP Datasphere, or to consume them by analytic clients. For more information, see SAP BW∕4HANA Model
Transfer Connections [page 83].
Note
This connection type doesn't support replication flows. Instead, we recommend using the SAP ABAP
connection type for replication flows.
Remote Tables
• data from virtual tables through RFC for ODP sources (extractors):
• Extraction context ABAP_CDS provides access to ABAP Core Data Services (CDS) Views that include
the annotation @Analytics.dataextraction.enabled: true
For information about which ABAP CDS Views are available for extraction, see CDS Views Enabled for
Data Extraction in the SAP S/4HANA documentation.
• Extraction context BW provides access to InfoProviders in SAP BW or SAP BW∕4HANA systems:
• DataStore objects (advanced or classic)
• CompositeProviders
• Queries as InfoProviders (with a subset of OLAP capabilities) - can be enabled in the runtime
properties of queries in the BW Modeling tools
• InfoObjects (characteristics) - can be enabled in the general settings of the InfoObject editor in the
BW Modeling tools
• Extraction context SAPI provides access to Service API (SAPI) DataSources
Importing SAPI DataSources into SAP Datasphere is not supported if the DataSource contains
mandatory selection fields. You will receive an error message when trying to import tables for such
DataSources.
• Remote
• Replication (snapshot)
• For ODP sources: Replication (real-time)
For more information about the prerequisites, see Replicate Data Changes in Real-Time [page 181].
For more information, see Replicating Data and Monitoring Remote Tables [page 168].
Note
Certain SAPI DataSources (from FI-AA Business Content, for example) may send duplicate records during
snapshot replication or real-time initialization. Remote table replication runs with optimized INSERT
processing on DataSource key level which leads to unique constraint violation errors. In such cases we
recommend to use a data flow which uses the remote table as source and for which the target table has
selected the Append mode with the Update Records By Primary Key (UPSERT) option.
Data Flows
For legacy SAP BW systems that do not have the ABAP Pipeline Engine extension or DMIS Addon installed,
you can leverage Operational Data Provisioning (ODP) connectivity und use ODP extractors as sources in data
flows. Note that ABAP Dictionary tables are not supported as sources in data flows.
Note
The data preview in the data flow editor of the Data Builder is not available for SAP BW sources.
Connection Details
Property Description
SAP Logon Connection Type Select Application Server to connect to a single application server, or select Message Server
to use load balancing and connect to a message server that distributes the load to different
application servers.
Note
The SAP Logon connection type you select here must match the connection type in
the system mapping in the Cloud Connector configuration (load balancing logon or
connecting to a specific application server).
[if SAP Logon Connection Enter the name of the application server to which you want to connect to.
Type = Application Server]
Application Server
[if SAP Logon Connection Enter the SAP ABAP system instance number.
Type = Application Server]
System Number
[if SAP Logon Connection Enter the name of the message server to which you want to connect to.
Type = Message Server]
Message Server
[if SAP Logon Connection Enter the message server port (numerical).
Type = Message Server]
Message Server Port
Language Enter the two-digit ISO language code, EN for English or DE for German, for example. In the
Data Builder, object and field descriptions are then shown in the specified language. If not
set, the application uses the default logon language of the ABAP system for data flows, and
EN is used for remote tables and views.
Cloud Connector
Property Description
Use Cloud Connector Set to true if your source is an on-premise source and you want to use the connection for
data flows. The default is false.
[if Use Cloud Connector = Select Derive Virtual Host and Port from Connection Details or Enter Virtual Host and Port in
true] Virtual Destination Separate Fields.
If host and port entered in the connection details match the virtual host and port from the
Cloud Connector configuration, you can select Derive Virtual Host and Port from Connection
Details and don't need to enter the values manually.
Note
When you select Derive Virtual Host and Port from Connection Details for ABAP-based
connections, virtual port is set to sapgw<system number> to make sure that it
matches the virtual port defined in the Cloud Connector configuration.
[if Virtual Destination = Enter Enter the virtual host that you defined during Cloud Connector configuration.
Virtual Host and Port in
Separate Fields] Virtual Host
[if Virtual Destination = Enter Enter the virtual port that you defined during Cloud Connector configuration.
Virtual Host and Port in
Separate Fields] Virtual Port Note
• If the SAP Logon Connection Type for your connection is Application Server: Enter
sapgw<system number> to make sure that it matches the virtual port defined
in the Cloud Connector configuration.
• If the SAP Logon Connection Type for your connection is Message Server: Enter
a numerical port and make sure it is the port defined in the Cloud Connector
configuration.
Credentials
Property Description
User Name Enter the user name that is used to connect to the SAP ABAP system.
Features
Data Provisioning Agent Select an agent if you want to use the connection to access data via imported remote tables
and to build views.
[if you selected an agent] Set to On if you want to use ABAP RFC streaming for loading tables based on small batches.
Streaming Read Set to Off to use non-RFC streaming. The default is Off.
[if Streaming Read = On] Enter the gateway host where the ABAP adapter would register an RFC server instance to
Gateway Host receive callbacks with the table data batches. Usually, the gateway host is the same as the
target ABAP system host.
[if Streaming Read = On] Enter the port number of the gateway server where the ABAP adapter would register an
Gateway Port RFC server instance to receive callbacks with the table data batches. You can find the
port number for sapgw<ABAP_instance_number> in file /etc/services on the
ABAP host. The default port number is 33<ABAP_instance_number>. If you have a
file /etc/services with this mapping on the Data Provisioning Agent’s host, you can
also enter sapgw<ABAP_instance_number> instead of the port number.
[if Streaming = On] RFC Enter the name of the RFC destination that you have created in the source. For more
Destination information, see Prerequisites for ABAP RFC Streaming.
Data Flows are enabled without the need to set any additional connection properties. Make sure you have
maintained the properties in the Cloud Connector section.
When you've selected a Data Provisioning Agent, you can configure advanced connection properties to enable
specific customer scenarios. In the connection creation wizard, you configure advanced properties in an
additional wizard step. When editing the connection to configure the properties, scroll down and click the Show
Advanced Properties button. Advanced properties have default values that you can override according to your
customer scenario’s needs.
Available properties:
For more information, see SAP ABAP Adapter Remote Source Configuration in the SAP HANA Smart Data
Integration and SAP HANA Smart Data Quality Installation and Configuration Guide.
Use an SAP BW/4HANA Model Transfer connection to import analytic queries from SAP BW∕4HANA with their
Composite Providers and InfoObjects.
Prerequisites
Property Description
Property Description
Host Displays the SAP HANA host retrieved from SAP BW∕4HANA
via the selected tunnel connection.
Port Displays the SAP HANA port retrieved from SAP BW∕4HANA
via the selected tunnel connection.
Note
If required in exceptional cases, you can change the values in the Host, Port, and Schema fields.
In some use cases, for example when SAP BW∕4HANA is hosted by a third-party cloud provider, host, port,
and schema can't be retrieved and you must enter them manually.
In case of an error when retrieving host, port, and schema for example because of missing authorizations,
because of any of the services necessary for the tunnel connection not being active, or because of any
other issue with the tunnel connection, you can enter host, port, and schema manually to be able to
proceed with connection creation before validating and fixing the tunnel connection. We recommend,
however, to validate and fix the tunnel connection first.
You can reset the values for the Host, Port, and Schema fields to the values retrieved from SAP BW∕4HANA
by clicking Reset.
Property Description
Validate Server Certificate Specify whether to validate the certificate of the remote SAP
HANA server. The default value is true.
Note
A certificate signed by a certificate authority (CA) is
needed here. If the certificate is self-signed and not CA
signed, select false.
Property Description
User Name Enter the SAP HANA database user name (case sensitive).
Note
The SAP HANA user needs read privileges for the ABAP
schema in the SAP HANA database.
Features
On saving the connection, SAP Datasphere retrieves SAP HANA host and port via the SAP Analytics Cloud
tunnel connection.
You can now find your new connection in the list of the available connections for the space and can use it
in the Business Builder to import analytic queries from your connected SAP BW∕4HANA system into SAP
Datasphere. During the model import alongside other objects we will create and deploy remote tables in the
repository. The remote tables can be used in the following places:
Note
In the view or data flow editor in the data builder, SAP BW/4HANA Model Transfer connections with their
remote tables are not available as sources in the Sources tab of the source tree. Instead, as described
above, the remote tables are available from the Repository tab.
Use an SAP ECC connection to access data from virtual tables through RFC for ODP sources (extractors)
and ABAP Dictionary tables from SAP ERP Central Component (SAP ECC) systems (on-premise). For source
Prerequisites
The connection type supports the remote table as well as the data flow feature.For more information, see
Supported Connection Type Features [page 9].
Note
This connection type doesn't support replication flows. Instead, we recommend using the SAP ABAP
connection type for replication flows.
Remote Tables
You can use the connection type to access:
• data from virtual tables through RFC for ODP sources (extractors):
• Extraction context ABAP_CDS provides access to ABAP Core Data Services (CDS) Views that include
the annotation @Analytics.dataextraction.enabled: true
For information about which ABAP CDS Views are available for extraction, see CDS Views Enabled for
Data Extraction in the SAP S/4HANA documentation.
• Extraction context BW provides access to InfoProviders in SAP BW or SAP BW∕4HANA systems:
• DataStore objects (advanced or classic)
• CompositeProviders
• Queries as InfoProviders (with a subset of OLAP capabilities) - can be enabled in the runtime
properties of queries in the BW Modeling tools
• InfoObjects (characteristics) - can be enabled in the general settings of the InfoObject editor in the
BW Modeling tools
• Extraction context SAPI provides access to Service API (SAPI) DataSources
Note
Importing SAPI DataSources into SAP Datasphere is not supported if the DataSource contains
mandatory selection fields. You will receive an error message when trying to import tables for such
DataSources.
• Remote
• Replication (snapshot)
For more information, see Replicating Data and Monitoring Remote Tables [page 168].
Note
Certain SAPI DataSources (from FI-AA Business Content, for example) may send duplicate records during
snapshot replication or real-time initialization. Remote table replication runs with optimized INSERT
processing on DataSource key level which leads to unique constraint violation errors. In such cases we
recommend to use a data flow which uses the remote table as source and for which the target table has
selected the Append mode with the Update Records By Primary Key (UPSERT) option.
Data Flows
For legacy SAP ECC systems that do not have the ABAP Pipeline Engine extension or DMIS Addon installed,
you can leverage Operational Data Provisioning (ODP) connectivity und use ODP extractors as sources in data
flows. Also ODP extractors from SAP S/4HANA on-premise systems that have a version lower than 1909 are
supported. Note that ABAP Dictionary tables are not supported as sources in data flows.
Note
The data preview in the data flow editor of the Data Builder is not available for SAP ECC sources.
See SAP Data Warehouse Cloud - First Guidance: Data Integration for ABAP Source Systems (published
February 2022) for general best practices for integrating data from ABAP source systems with SAP
Datasphere.
Connection Details
Property Description
SAP Logon Connection Type Select Application Server to connect to a single application server, or select Message Server
to use load balancing and connect to a message server that distributes the load to different
application servers.
Note
The SAP Logon connection type you select here must match the connection type in
the system mapping in the Cloud Connector configuration (load balancing logon or
connecting to a specific application server).
[if SAP Logon Connection Enter the name of the application server to which you want to connect to.
Type = Application Server]
Application Server
[if SAP Logon Connection Enter the SAP ABAP system instance number.
Type = Application Server]
System Number
[if SAP Logon Connection Enter the name of the message server to which you want to connect to.
Type = Message Server]
Message Server
[if SAP Logon Connection Enter the message server port (numerical).
Type = Message Server]
Message Server Port
Language Enter the two-digit ISO language code, EN for English or DE for German, for example. In the
Data Builder, object and field descriptions are then shown in the specified language. If not
set, the application uses the default logon language of the ABAP system for data flows, and
EN is used for remote tables and views.
Cloud Connector
Property Description
Use Cloud Connector Set to true if your source is an on-premise source and you want to use the connection for
data flows. The default is false.
[if Use Cloud Connector = Select Derive Virtual Host and Port from Connection Details or Enter Virtual Host and Port in
true] Virtual Destination Separate Fields.
If host and port entered in the connection details match the virtual host and port from the
Cloud Connector configuration, you can select Derive Virtual Host and Port from Connection
Details and don't need to enter the values manually.
Note
When you select Derive Virtual Host and Port from Connection Details for ABAP-based
connections, virtual port is set to sapgw<system number> to make sure that it
matches the virtual port defined in the Cloud Connector configuration.
[if Virtual Destination = Enter Enter the virtual host that you defined during Cloud Connector configuration.
Virtual Host and Port in
Separate Fields] Virtual Host
[if Virtual Destination = Enter Enter the virtual port that you defined during Cloud Connector configuration.
Virtual Host and Port in
Separate Fields] Virtual Port Note
• If the SAP Logon Connection Type for your connection is Application Server: Enter
sapgw<system number> to make sure that it matches the virtual port defined
in the Cloud Connector configuration.
• If the SAP Logon Connection Type for your connection is Message Server: Enter
a numerical port and make sure it is the port defined in the Cloud Connector
configuration.
Credentials
Property Description
User Name Enter the user name that is used to connect to the SAP ABAP system.
Features
Property Description
Data Provisioning Agent Select an agent if you want to use the connection to access data via imported remote tables
and to build views.
[if you selected an agent] Set to On if you want to use ABAP RFC streaming for loading tables based on small batches.
Streaming Read Set to Off to use non-RFC streaming. The default is Off.
[if Streaming Read = On] Enter the gateway host where the ABAP adapter would register an RFC server instance to
Gateway Host receive callbacks with the table data batches. Usually, the gateway host is the same as the
target ABAP system host.
[if Streaming Read = On] Enter the port number of the gateway server where the ABAP adapter would register an
Gateway Port RFC server instance to receive callbacks with the table data batches. You can find the
port number for sapgw<ABAP_instance_number> in file /etc/services on the
ABAP host. The default port number is 33<ABAP_instance_number>. If you have a
file /etc/services with this mapping on the Data Provisioning Agent’s host, you can
also enter sapgw<ABAP_instance_number> instead of the port number.
[if Streaming = On] RFC Enter the name of the RFC destination that you have created in the source. For more
Destination information, see Prerequisites for ABAP RFC Streaming.
Data Flows are enabled without the need to set any additional connection properties. Make sure you have
maintained the properties in the Cloud Connector section.
When you've selected a Data Provisioning Agent, you can configure advanced connection properties to enable
specific customer scenarios. In the connection creation wizard, you configure advanced properties in an
additional wizard step. When editing the connection to configure the properties, scroll down and click the Show
Advanced Properties button. Advanced properties have default values that you can override according to your
customer scenario’s needs.
Available properties:
For more information, see SAP ABAP Adapter Remote Source Configuration in the SAP HANA Smart Data
Integration and SAP HANA Smart Data Quality Installation and Configuration Guide.
Prerequisites
The connection type supports the remote table as well as the data flow feature.
• Remote
• Replication (snapshot)
• Replication (real-time)
For more information, see Replicating Data and Monitoring Remote Tables [page 168].
Connection Details
Property Description
Root Path [optional] Enter the root path name to restrict browsing to a
certain CDI namespace or provider.
OAuth 2.0
Property Description
• Client Credentials
• User Name and Password
OAuth Token Endpoint Enter the token endpoint that the application must use.
OAuth Response Type Select the OAuth response type. The values are token or
none.
OAuth Token Request Content Type Select the value for the content-type HTTP header that the
application must use when requesting a token. The values
are URLEncoded or JSON.
Property Description
Property Description
Features
Data Flows are enabled without the need to set any additional connection properties.
Use an SAP HANA connection to access data from a remote SAP HANA database (on-premise or cloud).
For information about the required prerequisites, see Prepare Connectivity to SAP HANA.
Supported Features
Remote Tables You can use remote tables imported from the connection either to access data directly live
in the connected source (federation) or to copy the data into SAP Datasphere (replication).
For remote tables, real-time replication is supported. For information about any constraints,
see Replicate Data Changes in Real-Time [page 181].
Data Flows You can use the connection to add source objects to a data flow.
Replication Flows You can use the connection to add source and target objects to a replication flow.
The connection type supports replication with both remote tables via Data Provisioning Agent (SAP HANA
Smart Data Integration) and replication flows (for on-premise systems via Cloud Connector).
Regardless of whether you use the same connection or two separate connections, for replicating data from
a dedicated source object in the source we strongly recommend to only use remote tables or replication
flows and not both.
Connection Details
Property Description
Host Enter the fully qualified host name or IP address on which the remote SAP HANA server is
running.
Port Enter the SQL port number of the SAP HANA Cloud instance: 443.
Authentication
Property Description
Authentication Type Select the authentication type to use to connect to SAP HANA.
Property Description
X.509 Client Certificate To upload the certificate or certificate chain that is used to authenticate to the remote
system, click (Browse) and select the file from your download location.
X.509 Client Private Key To upload the private key, click (Browse) and select the file from your download location.
X.509 Client Private Key [optional] If the private key is encrypted, enter the password required for decryption.
Password
Property Description
Features
Feature Description
Remote Tables Remote Tables are enabled without the need to set any additional connection properties.
Property Description
Data Provisioning [read-only] Displays the midleware used to connect to the on-
Option premise system.
The option for SAP HANA Cloud is always set to Direct because no
middleware is required.
Data Access Select how you want to access data from source objects:
For more information about the use cases for the options, see Re-
mote Tables in Data Access Remote Only Vs Data Access Remote
and Replication [page 167].
Data Flows Data Flows are enabled without the need to set any additional connection properties.
Replication Flows Replication Flows are enabled without the need to set any additional connection properties.
Connection Details
Property Description
Host Enter the fully qualified host name or IP address on which the remote SAP HANA server is
running.
Port Enter the TCP SQL port number of the remote SAP HANA server.
For more information about port assignment for SAP HANA, see Port Assignment in Tenant
Databases in the SAP HANA Platform documentation.
Cloud Connector
Property Description
Use Cloud Connector Set to true if your source is an on-premise source and you want to use the connection for
one of the following features:
[if Use Cloud Connector = [optional] Select Derive Virtual Host and Port from Connection Details or Enter Virtual Host
true] Virtual Destination and Port in Separate Fields.
If host and port entered in the connection details match the virtual host and port from the
Cloud Connector configuration, you can select Derive Virtual Host and Port from Connection
Details and don't need to enter the values manually.
[if Virtual Destination = Enter Enter the virtual host that you defined during Cloud Connector configuration.
Virtual Host and Port in
Separate Fields] Virtual Host
[if Virtual Destination = Enter Enter the virtual port that you defined during Cloud Connector configuration.
Virtual Host and Port in
Separate Fields] Virtual Port
Property Description
Enable SSL encryption Select whether to enable SSL encryption on the connection to the remote SAP HANA
database. The default value is true.
Note
To use SSL encryption with a remote SAP HANA database connected via SAP HANA
smart data integration, the Data Provisioning Agent must already be correctly config-
ured for SSL support.
For more information, see Configure SSL for SAP HANA On-Premise [Manual Steps] in
the SAP HANA Smart Data Integration and SAP HANA Smart Data Quality documenta-
tion.
[if Enable SSL encryption Select whether to validate the certificate of the remote SAP HANA server. The default value
= true] Validate Server is true.
Certificate
If set to false, the host name used for the connection is used for verification.
Note
• SSL is name-based; connecting to an IP address, or to “localhost” is unlikely to
work.
• When using SAP HANA smart data access via Cloud Connector for remote tables:
To validate the server certificate, the certificate must have been uploaded to SAP
Datasphere.
For more information, see Manage Certificates for Connections.
[if Validate Server Certificate Verify the host name field of the server certificate:
= true] Host Name in Server
Certificate • If the string is “*”, any name matches.
• If the string starts with “CN=”, it is treated as a common name, and the textual repre-
sentation of the common name entry in the certificate must be exactly the same.
Authentication
Property Description
Authentication Type Select the authentication type to use to connect to SAP HANA.
Property Description
X.509 Client Certificate To upload the certificate or certificate chain that is used to authenticate to the remote
system, click (Browse) and select the file from your download location.
X.509 Client Private Key To upload the private key, click (Browse) and select the file from your download location.
X.509 Client Private Key [optional] If the private key is encrypted, enter the password required for decryption.
Password
Property Description
Feature Description
Property Description
Data Provisioning Select the midleware to use when connecting to and accessing the
Option on-premise system:
Note
When creating a connection to SAP HANA on-premise using
SAP HANA smart data access via Cloud Connector, the system
checks for a required internal service. In the connection over-
view, a warning icon indicates that the service might not be
ready yet. This happens for the first connection you create or
when you create a connection after the service has been disa-
bled (after an automated weekly check returning that there is
no such connection anymore). Getting the service ready might
take up to 45 minutes. Validate the connection and check the
message details for more information.
[if Data Provisioning Select an agent from the list of agents configured for SAP
Option = Data
Datasphere.
Provisioning Agent]
Data Provisioning
Agent
[if Data Provisioning Select how you want to access data from source objects:
Option = Cloud
Connector or Data • Select Remote Only if you want to federate data from the
Provisioning Agent] source objects of the connection and access them remotely
Data Access in the source system.
• Select Remote and Replication (default) if you want to allow
copying the data from the source object into SAP Datasphere.
For more information about the use cases for the options, see Re-
mote Tables in Data Access Remote Only Vs Data Access Remote
and Replication [page 167].
Data Flows Data Flows are enabled without the need to set any additional connection properties. Make
sure you have maintained the properties in the Cloud Connector section.
Replication Flows Replication Flows are enabled without the need to set any additional connection properties.
Make sure you have maintained the properties in the Cloud Connector section.
When you've selected a Data Provisioning Agent, you can configure advanced connection properties to enable
specific customer scenarios. In the connection creation wizard, you configure advanced properties in an
additional wizard step. When editing the connection to configure the properties, scroll down and click the Show
Advanced Properties button. Advanced properties have default values that you can override according to your
customer scenario’s needs.
Available properties:
For more information, see SAP HANA Remote Source Configuration in the SAP HANA Smart Data Integration
and SAP HANA Smart Data Quality Installation and Configuration Guide.
Use an SAP HANA Cloud, Data Lake Relational Engine connection to access table data from the Relational
Engine component of a standalone SAP HANA Cloud, data lake.
Standalone SAP HANA Cloud, data lake is a standalone component in the SAP Business Technology Platform
Cockpit. It is composed of the default data lake Files component and the data lake Relational Engine
component, and it is not integrated with SAP HANA database. For more information, see What is SAP HANA
Cloud, Data Lake in the SAP HANA Cloud, Data Lake documentation.
The connection type supports the remote table as well as the data flow feature.
• Remote
• Replication (snapshot)
Note
• With connection type SAP HANA Cloud, Data Lake Relational Engine you cannot connect to the SAP
HANA Cloud, data lake instance that is available with SAP Datasphere. To connect to this instance, use
your Open SQL schema in the space which you have selected to access the data lake.
For more information, see Integrating Data to and From SAP HANA Cloud Data Lake [page 158].
• SAP HANA Cloud, data lake Relational Engine connections are secure by default. They are encrypted
without the need of configuring any TLS settings. Also, uploading server certificates isn't required.
• To allow SAP Datasphere to access SAP HANA Cloud, data lake Relational Engine, it might be required
to add the SAP Datasphere outbound IP address and HANA IP addresses to the allowed IP addresses
for the data lake instance in SAP HANA Cloud Central.
For more information, see:
• Data Lake Connections in the SAP HANA Cloud documentation
• Finding SAP Datasphere IP addresses
Connection Details
Property Description
Host Enter the host of the SAP HANA Cloud, data lake Relational Engine server.
Credentials
Property Description
User Name Enter the name of the SAP HANA Cloud, data lake Relational
Engine user.
Note
This user needs to have corresponding privileges to
access the tables or views in SAP HANA Cloud, data
lake Relational Engine that you want to use in SAP
Datasphere. For more information, see SAP HANA
Cloud, Data Lake User Management for Data Lake IQ
in the SAP HANA Cloud, Data Lake documentation.
Features
Remote Tables are enabled without the need to set any additional connection properties.
Data Flows are enabled without the need to set any additional connection properties.
Use an SAP HANA Cloud, Data Lake Files connection to access data from the Files component of a standalone
SAP HANA Cloud, data lake.
Context
Standalone SAP HANA Cloud, data lake is a standalone component in the SAP Business Technology Platform
Cockpit. It is composed of the default data lake Files component and the data lake Relational Engine
Note
• With connection type SAP HANA Cloud, Data Lake Files you cannot connect to the SAP HANA Cloud,
data lake instance that is available with SAP Datasphere. To connect to this instance, use your Open
SQL schema in the space which you have selected to access the data lake.
For more information, see Integrating Data to and From SAP HANA Cloud Data Lake [page 158].
• SAP HANA Cloud, data lake Files connections are secure by default. They are encrypted without the
need of configuring any TLS settings. Also, uploading server certificates isn't required.
• To allow SAP Datasphere to access SAP HANA Cloud, data lake Files storage, it might be required to
add the SAP Datasphere outbound IP address and HANA IP addresses to the allowed IP addresses for
the data lake instance in SAP HANA Cloud Central.
For more information, see:
• Data Lake Connections in the SAP HANA Cloud documentation
• Finding SAP Datasphere IP addresses
Supported Features
Data Flows You can use the connection to add source objects to a data flow.
Replication Flows You can use the connection to add source and target objects to a replication flow.
Connection Details
Property Description
Host Enter the SAP HANA Cloud, data lake Files REST API endpoint, for example <instance
id>files.hdl.prod-eu10.hanacloud.ondemand.com.
Root Path [optional] Enter a root path for browsing files. The value starts with the character slash. For
example, /My Folder/MySubfolder.
If you have specified the root path, then any path used with this connection is prefixed with
the root path.
Property Description
Keystore File Enter the Base64-encoded value of the client keystore file. The supported format is
PCKS#12 (P12 and PFX files).
You can also choose Browse and select the file from your download location.
Features
Feature Description
Data Flows Data Flows are enabled without the need to set any additional connection properties.
Replication Flows Replication Flows are enabled without the need to set any additional connection properties.
Use an SAP Marketing Cloud connection to access data from SAP Marketing Cloud via its OData-based APIs for
data integration and SAP HANA Smart Data Integration.
Prerequisites
The connection type supports the remote table as well as the data flow feature.
• Remote
• Replication (snapshot)
• Replication (real-time)
For more information, see Replicating Data and Monitoring Remote Tables [page 168].
Connection Details
Property Description
Root Path [optional] Enter the root path name to restrict browsing to a
certain CDI namespace or provider.
Property Description
User Name Enter the user name that the application must use for authentication.
Features
Data Flows are enabled without the need to set any additional connection properties.
For information about the required prerequisites, see Prepare Connectivity to SAP SuccessFactors.
Remote Tables You can use remote tables imported from the connection either to access data directly live
in the connected source (federation) or to copy the data into SAP Datasphere (replication).
Note
• We strongly recommend to use remote tables for replication only and not to use it
for federated access because of the following reason:
When using remote tables to replicate data from SAP SuccessFactors, the SAP
SuccessFactors connection type uses snapshot-based pagination for SAP Success-
Factors entities that support this method and this way ensures data consistency
(see Snapshot-Based Pagination in the SAP SuccessFactors Platform documenta-
tion). For federated access, snapshot-based pagination is not supported because
of its service limits (see Service Limits for Snapshot-Based Pagination in the SAP
SuccessFactors Platform documentation). Instead, federated access uses client-
side pagination, which potentially can lead to data loss or data duplication (see
Client-Side Pagination in the SAP SuccessFactors Platform documentation).
Please note that since the Generic OData connection type doesn’t use snapshot-
based pagination at all, we do not recommend to use it to federate or replicate SAP
SuccessFactors data with remote tables.
• When replicating employee-related data, independently of their validity not only
current data is replicated but all data including historical data.
• Picklists used in SAP SuccessFactors are not available in remote tables because
they are exposed as navigation properties and not as exposed properties of the
entity. For more information on picklists, see SAP Note 2088679 .
Data Flows You can use the connection to add source objects to a data flow.
Connection Details
Property Description
URL Enter the OData service provider URL of the SAP SuccessFactors service that you want to
access.
Version [read-only] Displays the OData version used to implement the SAP SuccessFactors OData
service.
Property Description
Note
HTTP basic authentication in SAP SuccessFactors will
soon be retired. For more information, see Deprecation
of HTTP Basic Authentication in SAP SuccessFactors
What's New Viewer.
OAuth 2.0
If Authentication Type = OAuth 2.0:
Property Description
OAuth Grant Type [read-only] Displays SAML Bearer as the grant type used to
retrieve an access token.
OAuth Token Endpoint Enter the API endpoint to use to request an access token:
<SAP SuccessFactors API Server>/oauth/
token.
Note
When editing a connection that has been created before entering the SAML assertion in the credentials was
required, you can set the Provide SAML Assertion property in the OAuth 2.0 section to true to switch to the
new method and then enter a valid SAML assertion in the Credentials section.
Note that once you change the Provide SAML Assertion property to true, enter the SAML assertion and save
the connection, you cannot turn back to use the /oauth/idp API to generate the SAML assertion.
Property Description
Client ID Enter the API key received when registering SAP Datasphere
as OAuth2 client application in SAP SuccessFactors.
SAML Assertion Enter a valid SAML assertion that has been generated for
authentication.
Note
If the SAML assertion expires, the connection gets in-
valid until you update the connection with a new valid
SAML assertion.
Property Description
Features
Feature Description
Remote Tables Remote Tables are enabled without the need to set any additional connection properties.
Data Flows Data Flows are enabled without the need to set any additional connection properties.
Use an SAP S/4HANA Cloud connection to access or import extraction-enabled ABAP Core Data Services
views (ABAP CDS views) from SAP S/4HANA Cloud.
For information about the required prerequisites in the connected systems and SAP Datasphere, see Prepare
Connectivity to SAP S/4HANA Cloud.
Remote Tables You can use remote tables imported from the connection either to access data directly live
in the connected source (federation) or to copy the data into SAP Datasphere (replication).
• for federation: standard and custom CDS view entities that are exposed using the
ABAP SQL service from SAP S/4HANA Cloud (recommended for federation scenar-
ios)
For more information, see Using ABAP SQL Services for Accessing Data from SAP
S/4HANA Cloud.
Note
On the Sources tab of the remote-table-related Data Builder editors in SAP
Datasphere, the service binding name from the SQL_SCHEMA authorization field is
visible as (virtual) schema.
• for federation and replication: ABAP CDS Views exposed as OData services for data
extraction. Data is accessed via Cloud Data Integration (legacy)
Note
This legacy option is still supported, however we recommend using the ABAP SQL
service for federation (if available) and replication flows for replication.
For remote tables, real-time replication is supported. For information about any con-
straints, see Replicate Data Changes in Real-Time [page 181].
Model Import You can use the connection to import semantically-rich entities from SAP S/4HANA Cloud
and all their sources and dependencies. For more information, see Importing Entities with
Semantics from SAP S/4HANA.
Data Flows You can use the connection to add source objects to a data flow.
• extraction-enabled ABAP CDS views that are C1-released, that is views with annotation
@Analytics.dataextraction.enabled: true and that are available in the
connected system (access via ABAP Pipeline Engine)
For more information, see:
• Data Integration in the SAP S/4HANA Cloud documentation
• CDS Views Enabled for Data Extraction in the SAP S/4HANA Cloud documentation
Note
• Data flows currently support one-time loading of data (also known as initial load)
only.
• The data preview in the data flow editor of the Data Builder is only available if the
version of the connected system is SAP S/4HANA Cloud 2302 or higher.
Replication Flows You can use the connection to add source objects to a replication flow. (recommended for
replication scenarios)
• standard and custom CDS view entities that are exposed using the ABAP SQL service
from SAP S/4HANA Cloud
For more information, see Data Consumption Using SAP Datasphere.
Note
When adding source objects in replication flows, you can find the CDS view entities
in the SQL_SERVICE container.
• extraction-enabled ABAP CDS views that are C1-released, that is views with annotation
@Analytics.dataextraction.enabled: true and that are available in the
connected system (access via ABAP Pipeline Engine)
For more information, see:
• Data Integration in the SAP S/4HANA Cloud documentation
• CDS Views Enabled for Data Extraction in the SAP S/4HANA Cloud documentation
Note
When adding source objects in replication flows, you can find the CDS views in the
CDS_EXTRACTION container.
Note
The connection type supports replication with both remote tables via Data Provisioning Agent (SAP HANA
Smart Data Integration) and replication flows.
Regardless of whether you use the same connection or two separate connections, for replicating data from
a dedicated source object in the source we strongly recommend to only use remote tables or replication
flows and not both. Generally, for replication scenarios, we recommend to use replication flows.
Connection Details
Property Description
Application Server Enter the name of the API-URL from the communication arrangements that
you set up in SAP S/4HANA Cloud (without https://), for example myXXXXX-
api.s4hana.ondemand.com or myXXXXX-api.s4hana.cloud.sap.
For more information about the required communication arrangements, see Prepare Con-
nectivity to SAP S/4HANA Cloud.
System ID Enter the system ID of the system to which you want to connect.
Language Enter a two-digit ISO language code, EN for English or DE for German, for example. Object
and field descriptions in the data flow editor of the Data Builder are then shown in the
specified language. If you don't enter any language code, the application uses the default
logon language of the SAP S/4HANA system.
Port Enter the port number of the WebSocket RFC connection endpoint. The default port num-
ber is 443.
Authentication
Property Description
Authentication Type Select the authentication type to use to connect to SAP S/4HANA Cloud.
Note
For federated access with remote tables using the ABAP SQL service, authentica-
tion with User Name and Password is required.
Property Description
X.509 Client Certificate To upload the certificate or certificate chain that is used to authenticate to the remote
system, click (Browse) and select the file from your download location.
X.509 Client Private Key To upload the private key, click (Browse) and select the file from your download location.
X.509 Client Private Key [optional] If the private key is encrypted, enter the password required for decryption.
Password
Property Description
User Name Enter the user name that is used to connect to the SAP S/4HANA system.
Property Description
Data Provisioning Option Select the middleware to use when connecting to and accessing SAP S/4HANA Cloud:
Note
This legacy option is still supported, however we recommend using the ABAP SQL
service for federation (if available) and replication flows for replication.
• Direct: if you want to federate data from the source objects of the connection and ac-
cess them remotely in the source system using the ABAP SQL service (recommended
for federation scenarios).
Note
This option is supported for authentication type User Name And Password.
[if Data Provisioning Option [read-only] Displays how data from source objects can be accessed:
= Data Provisioning Agent or
Direct] Data Access • Remote and Replication: if you selected data provisioning option Data Provisioning
Agent (using the SAP HANA smart data integration CloudDataIntegrationAdapter).
• Remote Only: if you selected data provisioning option Direct (using the ABAP SQL
service).
[if Data Provisioning Option Select an agent from the list of agents configured for SAP Datasphere.
= Data Provisioning Agent]
Data Provisioning Agent
[if Data Provisioning Option = Specify how ABAP data types are mapped to SAP HANA data types:
Direct] Type Map
• native: no conversion of ABAP data types
• semantic: conversion of the ABAP data types to an ABAP-independent data format
• semanticDatsTimsAsWchar: like semantic, except that the ABAP types DATS and TIMS
are mapped to the ODBC type SQL_WCHAR to allow for lossless conversion of date
literals and time literals
Features
Feature Description
Remote Tables To enable Remote Tables, complete the connection properties in the Remote Tables section.
Note
For federation scenarios, we recommend using the ABAP SQL service (data provision-
ing option Direct).
Data Flows Data Flows are enabled without the need to set any additional connection properties.
Replication Flows Replication Flows are enabled without the need to set any additional connection properties.
Note
For replication scenarios, we recommend using replication flows.
Use an SAP S/4HANA On-Premise connection to access data from SAP S/4HANA on-premise systems.
For information about the required prerequisites in the connected systems and SAP Datasphere, and about
supported source system versions, see Prepare Connectivity to SAP S/4HANA On-Premise.
Remote Tables You can use remote tables imported from the connection either to access data directly live
in the connected source (federation) or to copy the data into SAP Datasphere (replication).
• for federation: standard and custom CDS view entities that are exposed using the
ABAP SQL service from SAP S/4HANA (recommended for federation scenarios)
For more information, see Using ABAP SQL Services for Accessing Data from SAP
S/4HANA.
Note
On the Sources tab of the remote-table-related Data Builder editors in SAP
Datasphere, the service binding name from the SQL_SCHEMA authorization field is
visible as (virtual) schema.
Note
ABAP CDS Views are the preferred method over SAPI extractors when
integrating SAP S/4HANA data with SAP Datasphere.
Note
Importing SAPI DataSources into SAP Datasphere is not supported if the
DataSource contains mandatory selection fields. You will receive an error
message when trying to import tables for such DataSources.
Note
Certain SAPI DataSources (from FI-AA Business Content, for example)
may send duplicate records during snapshot replication or real-time initi-
alization. Remote table replication runs with optimized INSERT process-
ing on DataSource key level which leads to unique constraint violation
errors. In such cases, we recommend to use a data flow which uses the
remote table as source and for which the target table has selected the
Append mode with the Update Records By Primary Key (UPSERT) option.
Note
This legacy option is still supported, however we recommend using the ABAP SQL
service for federation (if available) and replication flows for replication.
Model Import You can use the connection to import semantically-rich entities from an SAP S/4HANA
system which you have selected in the Import Entities wizard. For more information, see
Importing Entities with Semantics from SAP S/4HANA.
You can import extraction-enabled ABAP CDS views that are C1-released, that is views with
annotation @Analytics.dataextraction.enabled: true and that are available
in the connected system. For more information, see CDS Views Enabled for Data Extraction
in the SAP S/4HANA documentation.
The options to access the data depend on the version of the SAP S/4HANA system. You
can see the data access option when you have selected a connection in the Import Entities
wizard:
• Replication Flow to Local Tables - available if the connected system is SAP S/4HANA
2021 or higher (SAP_BASIS 756 and higher). It's the default and recommended option
for these system versions.
• Remote Tables - available if the connected system is SAP S/4HANA 1809 or higher
(SAP_BASIS 753 and higher)
In the Import Entities wizard, you can change from the default Replication Flow to Local
Tables to Remote Tables if your system is SAP S/4HANA 2021 or higher (SAP_BASIS
756 and higher).
Note
Note that associated entities can only be included in the import when data access
is Replication Flow to Local Tables. For more information, see Importing Entities with
Semantics from SAP S/4HANA.
Data Flows You can use the connection to add source objects to a data flow.
• extraction-enabled ABAP CDS views that are C1-released, that is views with annotation
@Analytics.dataextraction.enabled: true and that are available in the
connected system (access via ABAP Pipeline Engine)
For more information, see CDS Views Enabled for Data Extraction in the SAP S/4HANA
documentation.
• Tables from SAP Landscape Transformation Replication Server (SAP LT Replication
Server) sources, for which an appropriate configuration has been created in SAP LT
Replication Server.
Note
• The availability of the data flow feature depends on the used version and Support
Package level of SAP S/4HANA or the DMIS addon in the source. Make sure your
source systems meet the required minimum versions. We recommend to use the
latest available version of SAP S/4HANA and the DMIS add-on where possible and
have the latest SAP notes and TCI notes implemented in your systems.
For more information about required versions, recommended system landscape,
considerations for the supported source objects, and more, see SAP Note
2890171 .
• Data flows currently support one-time loading of data (also known as initial load)
only.
• The data preview in the data flow editor of the Data Builder is not available, except
for:
• CDS views if the version of the connected system is SAP S/4HANA 1909 or
higher.
• SAP Landscape Transformation Replication Server objects if ABAP Add-on
DMIS 2018 SP08 or DMIS 2020 SP04 is installed or a higher DMIS version.
Replication Flows You can use the connection to add source objects to a replication flow.
• extraction-enabled ABAP CDS views that are C1-released, that is views with annotation
@Analytics.dataextraction.enabled: true and that are available in the
connected system (access via ABAP Pipeline Engine)
For more information, see CDS Views Enabled for Data Extraction in the SAP S/4HANA
documentation.
• Data from ODP sources (extraction contexts SAPI and BW)
Note
The availability of the replication flow feature depends on the used version and Support
Package level of SAP S/4HANA or the DMIS addon in the source. Make sure your
source systems meet the required minimum versions. We recommend to use the latest
available version of SAP S/4HANA and the DMIS add-on where possible and have the
latest SAP notes and TCI notes implemented in your systems.
For more information about required versions, recommended system landscape, con-
siderations for the supported source objects, and more, see SAP Note 2890171 .
Note
The connection type supports replication with both remote tables via Data Provisioning Agent (SAP HANA
Smart Data Integration) and replication flows (for on-premise systems via Cloud Connector).
Regardless of whether you use the same connection or two separate connections, for replicating data from
a dedicated source object in the source we strongly recommend to only use remote tables or replication
flows and not both.
Connection Details
Property Description
SAP Logon Connection Type Select Application Server to connect to a single application server, or select Message Server
to use load balancing and connect to a message server that distributes the load to different
application servers.
Note
The SAP Logon connection type you select here must match the connection type in
the system mapping in the Cloud Connector configuration (load balancing logon or
connecting to a specific application server).
[if SAP Logon Connection Enter the name of the application server to which you want to connect to.
Type = Application Server]
Application Server
[if SAP Logon Connection Enter the SAP ABAP system instance number.
Type = Application Server]
System Number
[if SAP Logon Connection Enter the name of the message server to which you want to connect to.
Type = Message Server]
Message Server
[if SAP Logon Connection Enter the message server port (numerical).
Type = Message Server]
Message Server Port
Language Enter the two-digit ISO language code, EN for English or DE for German, for example. In the
Data Builder, object and field descriptions are then shown in the specified language. If not
set, the application uses the default logon language of the ABAP system for data flows, and
EN is used for remote tables and views.
Property Description
Use Cloud Connector Set to true if you want to use the connection for one of the following features:
• remote tables for data federation via the ABAP SQL service
• data flows
• replication flows
• model import
Note
When the connection is configured for using the ABAP SQL service for data federation
with remote tables, you can't use the same connection for model import.
[if Use Cloud Connector = [optional] Select Derive Virtual Host and Port from Connection Details or Enter Virtual Host
true] Virtual Destination and Port in Separate Fields.
If host and port entered in the connection details match the virtual host and port from the
Cloud Connector configuration, you can select Derive Virtual Host and Port from Connection
Details and don't need to enter the values manually.
Note
When you select Derive Virtual Host and Port from Connection Details for ABAP-based
connections, virtual port is set to sapgw<system number> to make sure that it
matches the virtual port defined in the Cloud Connector configuration.
Note
When the connection is configured for using the ABAP SQL service for data federation
with remote tables, you must enter virtual host and port manually. Deriving virtual host
and port from the connection details is not supported because the Cloud Connector
configuration requires two system mappings having different virtual ports (one for RFC
protocol for data flows and replication flows, and one for HTTP/HTTPS protocoll for
remote tables using the ABAP SQL service).
[if Virtual Destination = Enter Enter the virtual host that you defined during Cloud Connector configuration.
Virtual Host and Port in
Separate Fields] Virtual Host
[if Virtual Destination = Enter Enter the virtual port that you defined during Cloud Connector configuration.
Virtual Host and Port in
Separate Fields] Virtual Port Note
• If the SAP Logon Connection Type for your connection is Application Server: Enter
sapgw<system number> to make sure that it matches the virtual port defined
in the Cloud Connector configuration.
• If the SAP Logon Connection Type for your connection is Message Server: Enter
a numerical port and make sure it is the port defined in the Cloud Connector
configuration.
Authentication
Property Description
Authentication Type Select the authentication type to use to connect to SAP S/4HANA.
• User Name And Password for basic authentication (default value) - This option is read-
only if you set Cloud Connector to false.
• OAuth 2.0 - You can select this option only if you have set Cloud Connector to true to
enable replication flows and data flows. Remote tables and model import currently are
not supported with OAuth authentication.
OAuth 2.0
Property Description
OAuth Grant Type Displays Client Credentials as grant type used to retrieve an
access token.
OAuth Token Endpoint Enter the API endpoint to use to request an access token.
OAuth Token Request Content Type [optional] Enter the content type of the OAuth2 token re-
quest.
Property Description
User Name Enter the user name that is used to connect to the SAP ABAP system.
Property Description
Remote Tables
Property Description
Data Provisioning Option Select the middleware to use when connecting to and accessing SAP S/4HANA:
Note
This legacy option is still supported, however we recommend using the ABAP SQL
service for federation (if available) and replication flows for replication.
• Cloud Connector: if you want to federate data from the source objects of the connec-
tion and access them remotely in the source system using the ABAP SQL service
(recommended for federation scenarios).
[if Data Provisioning Option [read-only] Displays how data from source objects can be accessed:
= Data Provisioning Agent
or Cloud Connector] Data • Remote and Replication: if you selected data provisioning option Data Provisioning
Access Agent (using the SAP HANA smart data integration ABAPAdapter).
• Remote Only: if you selected data provisioning option Cloud Connector (using the ABAP
SQL service).
[if Data Provisioning Option Select an agent from the list of agents configured for SAP Datasphere.
= Data Provisioning Agent]
Data Provisioning Agent
[if you selected an agent] Set to On if you want to use ABAP RFC streaming for loading tables based on small batches.
Streaming Read Set to Off to use non-RFC streaming. The default is Off.
[if Streaming Read = On] Enter the gateway host where the ABAP adapter would register an RFC server instance to
Gateway Host receive callbacks with the table data batches. Usually, the gateway host is the same as the
target ABAP system host.
[if Streaming Read = On] Enter the port number of the gateway server where the ABAP adapter would register an
Gateway Port RFC server instance to receive callbacks with the table data batches. You can find the
port number for sapgw<ABAP_instance_number> in file /etc/services on the
ABAP host. The default port number is 33<ABAP_instance_number>. If you have a
file /etc/services with this mapping on the Data Provisioning Agent’s host, you can
also enter sapgw<ABAP_instance_number> instead of the port number.
[if Streaming = On] RFC Enter the name of the RFC destination that you have created in the source. For more
Destination information, see Prerequisites for ABAP RFC Streaming.
[if Data Provisioning Option = Specify how ABAP data types are mapped to SAP HANA data types:
Cloud Connector] Type Map
• native: no conversion of ABAP data types
• semantic: conversion of the ABAP data types to an ABAP-independent data format
• semanticDatsTimsAsWchar: like semantic, except that the ABAP types DATS and TIMS
are mapped to the ODBC type SQL_WCHAR to allow for lossless conversion of date
literals and time literals
Replication Flows
Property Description
Fast Serialization Turn on RFC fast serialization to help improve replication flow performance by avoiding
unnecessary, costly conversions in ABAP. This can be useful particularly in cases of large
data volumes. The default is On.
Note
• Fast serialization has no impacts on data flows.
• Fast Serialization for replication flows is not related to RFC Serialization in the
advanced properties of connections using a Data Provisioning Agent for remote
tables.
• For information about the prerequisites for using fast serialization in SAP
Datasphere, see Prepare Connectivity to SAP S/4HANA On-Premise.
Use the connection validation to check if the prerequisites in the source system for
using fast serialization are met.
Features
Feature Description
Remote Tables To enable Remote Tables, complete the connection properties in the Remote Tables section.
Data Flows Data Flows are enabled without the need to set any additional connection properties. Make
sure you have maintained the properties in the Cloud Connector section.
Replication Flows Replication Flows are enabled without the need to set any additional connection properties.
Make sure you have maintained the properties in the Cloud Connector section.
Model Import To enable Model Import, select the live data connection that connects to your SAP S/
4HANA system and select a Data Provisioning Agent.
Note
When the connection is configured for using the ABAP SQL service for data federation
with remote tables, you can't use the same connection for model import.
When you've selected a Data Provisioning Agent, you can configure advanced connection properties to enable
specific customer scenarios. In the connection creation wizard, you configure advanced properties in an
additional wizard step. When editing the connection to configure the properties, scroll down and click the Show
Advanced Properties button. Advanced properties have default values that you can override according to your
customer scenario’s needs.
Available properties:
For more information, see SAP ABAP Adapter Remote Source Configuration in the SAP HANA Smart Data
Integration and SAP HANA Smart Data Quality Installation and Configuration Guide.
Extend connectivity beyond SAP Datasphere standard remote connectivity and cover additional data sources
that are available with partner tools.
Connections to partner tools are created differently compared to other connections in SAP Datasphere.
When creating a connection to a partner tool, an Open SQL schema is generated in SAP Datasphere and
information about host, port and credentials of the Open SQL schema is transferred to the partner tool. To be
able to send data to the Open SQL schema, during the connection creation process the partner tool on their
side establishes a connection to the Open SQL schema.
To successfully validate a new connection and before data can be written to the Open SQL schema and you can
use it in the Data Builder, the IP address or addresses of the partner tool need to be added to the IP allowlist in
SAP Datasphere. To get the relevant IP addresses, depending on the partner tool you might need to contact the
Account Manager or the support team (for Adverity, for example), or you can find and copy the IP addresses in
the last step of the connection cration wizard (for Precog, for example).
Note
• The Open SQL schema (respectively the database user for which the schema has been created)
doesn't appear in the Connections app but only in the Database Users section in Space Management.
• In the Data Builder, unlike other Open SQL schemas, the Open SQL schema generated with the partner
connection doesn't appear in Source Browser Sources . Instead, you can find the corresponding
partner connection in Source Browser Sources Connections .
Once the partner tool fetches data from a source, in the Data Builder you can find the corresponding
table in Source Browser Sources Connections <partner connection name> . When importing
the table into the Data Builder, a local table is created and deployed. You can use it the same way as any
other table originating from an Open SQL schema.
You can access the Open SQL Schema that has been generated for the partner connection from a third-party
BI client via ODBC. For more information, see Consume Data in Power BI and Other Clients, Tools, and Apps via
ODBC/JDBC.
Related Information
Edit a connection, for example if the setup of the connected remote system changed, if the connection has
been imported from another tenant, or if a new feature is available for the connection type after you have
created your connection. A warning in the connections list might indicate the need to edit a connection.
• When the connection isn't used and its real-time replication status is Inactive, you can change any property
except for the technical name.
• When you use the connection and its real-time replication status is Active, you can only change the
business name and description. If no package has been assigned yet, you can select a package. Note that
credential changes are not considered when saving the connection.
If you want to change other connection properties, pause real-time replication in the Connections app, or if
required disable real-time replication for all affected remote tables in the Data Integration Monitor.
• When the connection has real-time replication status Active but the Data Provisioning Agent is
disconnected, you can only change the business name and description. If no package has been assigned
yet, you can select a package. Note that credential changes are not considered when saving the
connection.
If you want to change other connection properties, you must first reconnect the agent and pause real-time
replication in the Connections app, or if required disable real-time replication for all affected remote tables
in the Data Integration Monitor.
• When the connection has real-time replication status Paused, you can only change a subset of the
properties. It depends on the connection type if any properties can be changed at all and if yes which
properties can be changed. Note that the user name cannot be changed in this case.
When the Data Provisioning Agent is disconnected, you need to reconnect the agent or use a different
agent to enable editing specific connection properties.
1. In the side navigation area, click (Connections) and select a space if necessary.
2. Select the relevant connection and click Edit.
Tip
For imported connections, for example, and for other connections with warning messages, you can
click the warning button on the top right corner of the app to select the connection by then clicking the
corresponding message title.
Users with the DW Space Administrator role (or equivalent privileges) can create packages in the Packages
editor. Once a package is created in your space, you can select it when editing a connection.
Packages are used to group related objects in order to facilitate their transport between tenants.
Note
Once a package is selected, it cannot be changed here. Only a user with the DW Space Administrator role
(or equivalent privileges) can modify a package assignment in the Packages editor.
Select a connection and open the validation message to get detailed status information.
1. In the side navigation area, click (Connections) and select a space if necessary.
2. Select the relevant connection and click Validate.
The message informs you about the availability of supported features and provides details in case of errors
in the connection.
Select one or more connections and delete them if they are not used.
1. In the side navigation area, click (Connections) and select a space if necessary.
2. Select one or more connections and click Delete.
If the connection is still used for accessing data via remote tables, for importing models, or for data flows or
replication flows, you cannot delete it. A message informs you of any dependent objects.
Pause real-time replication for a connection while updating or maintaining the source. After you have finished
with your maintenance activities you can restart real-time replication.
Context
You can pause real-time replication for connections that connect to a source through SAP HANA Smart Data
Integration and its Data Provisioning Agent.
If you need to upgrade or patch your Data Provisioning Agent, or you need to apply any changes to its
configuration, you can pause all connections for the Data Provisioning Agent. For more information, see Pause
Real-Time Replication for an Agent.
Procedure
1. In the side navigation area, click (Connections) and select a space if necessary.
2. In the list of connections, you can see the Real-Time Replication Status for each connection. Depending on
the status you can perform different actions.
Shown when the Remote Tables feature for an SAP HANA connection is set to Data
Access = Remote Only.
Active Pause
Shown when the connection type supports real-time replication and for the connection
at least one table is replicated via real-time replication (even if the status in the Remote
Tables monitor is Error).
Paused Restart
Shown when the connection type supports real-time replication and for the connection
at least for one table real-time replication is paused.
Inactive Pause
Shown when the connection type supports real-time replication and for the connection
currently there is no table replicating via real-time replication.
Disconnected none
Shown when the connection uses a Data Provisioning Agent that currently has status
Disconnected.
3. From the list of connections, select the connection in question and click Pause.
The status in the list of the connections changes from Active or Inactive to Paused.
In the Remote Tables monitor, the status for affected tables changes to Paused and actions related to
real-time replication are not available for these tables. Also, you cannot start real-time replication for any
table of a paused connection.
4. You can now perform maintenance activities in the source.
5. Once your maintenance activities finished, restart real-time replication. Select the connection in question
and click Restart.
The status in the list of the connections as well as in the Remote Tables monitor changes accordingly and
you can again perform real-time related actions for the tables or start real-time replication.
Select one or more filter values to restrict the connection list according to your needs.
1. In the side navigation area, click (Connections) and select a space if necessary.
2. Click (Filter Connections) and select one or more filter values to get a better overview of the
connections in the list.
Model Import
Remote Tables
Replication Flows
On-Premise
Partner Tools
SAP
You may ask yourself, when to use which option. Here is a summary of the differences.
• For both the SAP BW connection and the SAP ABAP connection, only queries as InfoProviders are
supported. For this type of queries, there are some restrictions. For more information, see Query as
InfoProvider.
• The difference between the SAP BW connection and the SAP ABAP connection is, that the SAP BW
connection is a subset of the SAP ABAP connection, being restricted to SAP BW as a source whereas the
SAP ABAP connection is more comprehensive to other on-premise source systems.
• When you use the SAP BW/4HANA Model Transfer connection, you can import standard analytic queries,
and you don’t have to deal with the restrictions of queries as InfoProviders. But it is only available for SAP
BW/4HANA, not for SAP BW.
• For SAP BW, the SAP BW connection and the SAP ABAP connection can be used, but not the SAP BW/
4HANA Model Transfer connection. This means for SAP BW, you can only import queries as InfoProviders.
• For SAP BW/4HANA, you can use all three options: the SAP BW connection, the SAP ABAP connection,
and the SAP BW/4HANA Model Transfer connection.
You can manage TLS server certificates and connections via the Certificates and Connections REST APIs.
Creating and editing connections via the API is supported for SAP SuccessFactors connections only.
Introduction
SAP Datasphere exposes REST APIs that allow you to programmatically manage TLS server certificates and
connections using dedicated endpoints.
The API specifications are available at the SAP Business Accelerator Hub .
Using the Certificates API, you can perform the following actions:
To manage certificates via the REST API, you must have an administrator role or equivalent privileges:
Using the Connections API, you can perform the following actions:
To manage connections via the REST API, you must have a scoped role based on the DW Integrator role or
equivalent privileges:
You must, in addition, obtain the following parameters for an OAuth client created in your SAP Datasphere
tenant (with Purpose set to Interactive Usage and the Redirect URI set to the URI provided by the client, tool, or
app that you want to connect):
• Client ID
• Secret
• Authorization URL
• Token URL.
For more information about obtaining the OAuth client parameters, see Create OAuth2.0 Clients to
Authenticate Against SAP Datasphere.
With the OAuth Client created in your SAP Datasphere tenant, your client, tool, or application will be able to
authenticate against the system’s OAuth service and obtain authorization to authenticate against the API.
To obtain an OAuth access token, send a GET request to one of the API endpoints (see below) including
Authorization: Auth Type <OAuth 2.0> and passing the <client ID>, <secret>, <authorization
URL>, and <token URL> values.
You must pass the access token in the Authorization header of all requests that you make to the API.
Sample Code
https://<tenant_url>/api/v1/datasphere/configuration/security/certificates
You must have a valid x-csrf-token before creating a POST, PUT, or DELETE request to the API endpoints.
You can get a token by sending a GET request to one of the API endpoints (see below) including the x-csrf-
token: fetch header.
The CSRF token is returned in the x-csrf-token response header. You must pass the token in the x-csrf-
token:<token> header of all POST, PUT, or DELETE requests that you make to the API.
https://<tenant_url>/api/v1/datasphere/configuration/security/certificates
To retrieve a list of TLS server certificates available in the tenant, use the GET request with the/api/v1/
datasphere/configuration/security/certificates endpoint and enter:
https://<tenant_url>/api/v1/datasphere/configuration/security/certificates
To upload a TLS server certificate, use the POST request with the/api/v1/datasphere/configuration/
security/certificates endpoint and enter:
https://<tenant_url>/api/v1/datasphere/configuration/security/certificates
Note
The file that you upload must have a supported file extension: .pem (privacy-enhanced mail), .crt, or .cer.
To delete a TLS server certificate from the tenant, use the DELETE request with the/api/v1/datasphere/
configuration/security/certificates/<fingerprint> endpoint and enter:
https://<tenant_url>/api/v1/datasphere/configuration/security/certificates/
<fingerprint>
Note
To list all connections of a space, use the GET request with the/api/v1/datasphere/spaces/<space_id>/
connections endpoint and enter:
https://<tenant_url>/api/v1/datasphere/spaces/<space_id>/connections
To read the JSON definition of a connection (without its credentials) in a space , use the GET request
with the/api/v1/datasphere/spaces/<space_id>/connections/<connection_technical_name>
endpoint and enter:
https://<tenant_url>/api/v1/datasphere/spaces/<space_id>/connections/
<connection_technical_name>
To create an SAP SuccessFactors connection in a space by providing the connection definition in stringified
json format, use the POST request with the/api/v1/datasphere/spaces/<space_id>/connections
endpoint and enter:
https://<tenant_url>/api/v1/datasphere/spaces/<space_id>/connections
The following example shows how to create a connection to SAP SuccessFactors for OData V4 and basic
authentication:
Sample Code
{
"name": "<technical name>",
"businessName" : "<business name>",
"description":"<description>",
"authType": "Basic",
"url": "https://<SAP SuccessFactors API Server>/odatav4/<supported SAP
SuccessFactors service group>",
"version" :"V4",
"username": "<username>",
"password":"<password>"
}
The following example shows how to create a connection to SAP SuccessFactors for OData V2 and OAuth2
authentication:
Sample Code
{
"name":"<technical name>",
name Technical Name [required] Enter the technical name of the connection. The technical
name can only contain alphanumeric characters and underscores (_).
Underscore (_) must not be used at the start or end of the name. The
maximum length is 40 characters. The name must be unique within the
space.
Note
Once the object is saved, the technical name can no longer be modi-
fied.
businessName Business Name [optional] Enter a descriptive name to help users identify the object. This
name can be changed at any time.
description Description [optional] Provide more information to help users understand the object.
authType Authentication Type [required] Enter the authentication type to use to connect to the OData
endpoint.
Note
HTTP basic authentication in SAP SuccessFactors will soon be re-
tired. For more information, see Deprecation of HTTP Basic Authen-
tication in SAP SuccessFactors What's New Viewer.
url URL [required] Enter the OData service provider URL of the SAP SuccessFac-
tors service that you want to access.
version Version [required] Enter the OData version used to implement the SAP Success-
Factors OData service (V2 or V4).
oauth2GrantType OAuth Grant Type [required] Enter SAML Bearer as the grant type used to retrieve an
access token.
oauth2TokenEndp OAuth Token Endpoint [required] Enter the API endpoint to use to request an access token:
oint <SAP SuccessFactors API Server>/oauth/token.
oauth2CompanyId OAuth Company ID [required] Enter the SAP SuccessFactors company ID (identifying the
SAP SuccessFactors system on the SAP SuccessFactors API server) to
use to request an access token.
clientId Client ID [required] Enter the API key received when registering SAP Datasphere
as OAuth2 client application in SAP SuccessFactors.
clientSecret SAML Assertion [required] Enter a valid SAML assertion that has been generated for
authentication.
Note
If the SAML assertion expires, the connection gets invalid until you
update the connection with a new valid SAML assertion.
Validate Connections
https://<tenant_url>/api/v1/datasphere/spaces/<spaceId>/connections/
<connection_technical_name>/validation
To edit an SAP SuccessFactors connection in a space by providing a new definition, use the PUT request
with the/api/v1/datasphere/spaces/<space_id>/connections/<connection_technical_name>
endpoint and enter:
https://<tenant_url>/api/v1/datasphere/spaces/<space-id>/connections/
<connection_technical_name>
Delete Connections
To delete a connection from a space, use the DELETE request with the/api/v1/datasphere/spaces/
<spaceId>/connections/<connection_technical_name> endpoint and enter:
https://<tenant_url>/api/v1/datasphere/spaces/<spaceId>/connections/
<connection_technical_name>
Users with a modeler role and who have access to SAP BW bridge, can import SAP BW bridge objects into SAP
Datasphere as remote tables.
SAP BW bridge enables you to use SAP BW functionality in the public cloud and to import SAP BW bridge data
models into SAP Datasphere.
The SAP BW bridge connection differs from other connection types, as it can't be created or modified. It is
being generated when SAP BW bridge is provisioned.
To learn more about SAP BW bridge, see Acquiring and Combining Data.
SAP Business Data Cloud is a fully managed SaaS solution that unifies and governs all SAP data and
seamlessly connects with third-party data—giving line-of-business leaders context to make even more
impactful decisions.
If your SAP Datasphere tenant is part of an SAP Business Data Cloud formation, then the SAP Business Data
Cloud administrator can:
• Install insight apps to the SAP Datasphere and SAP Analytics Cloud tenants in the formation.
• Activate data packages to allow the contained data products to be installed in SAP Datasphere.
For detailed information, see the SAP Business Data Cloud documentation.
An SAP Business Data Cloud can install insight apps to the SAP Datasphere and SAP Analytics Cloud tenants
in the formation (see Installing Insight Apps in the SAP Business Data Cloud documentation).
• SAP-managed spaces are created in SAP Datasphere to contain the insight app content.
• Replication flows, tables, views, and analytic models are created in these spaces to ingest, prepare and
expose the required data to SAP Analytics Cloud.
SAP Datasphere users can work with insight app content in the following ways:
• Review the installed content (see Reviewing Installed Insight Apps [page 140]).
• Upload permissions records to control access to the data (see Applying Row-Level Security to Data
Delivered through Insight Apps [page 141]).
• Build on top of the delivered data products and content to extend the app (see Extending Insight Apps
[page 142])
An SAP Business Data Cloud can activate data packages to allow the contained data products to be installed in
SAP Datasphere (see Activating Data Packages in the SAP Business Data Cloud documentation).
• An SAP Datasphere administrator must choose the spaces to which data products from the originating
SAP system can be installed (see Authorize Spaces to Install SAP Business Data Cloud Data Products).
• SAP Datasphere modelers can install data products to their space for use in their modeling projects (see
Evaluating and Installing SAP Business Data Cloud Data Products.
SAP-Managed Spaces
SAP-managed spaces are automatically created to contain the data products and content installed from SAP
Business Data Cloud:
• By default, no users are granted access to these spaces, but a SAP Datasphere administrator can add
users to the spaces to monitor data integration and review the content.
• All the delivered content is SAP-managed, read-only, and cannot be modified as it is protected by a
namespace (see Namespaces).
• No user can create any objects in SAP-managed spaces.
When an insight app is installed, data is loaded from the source system into SAP Datasphere, combined and
prepared for analytics, and then exposed for consumption in SAP Analytics Cloud, where business users can
consume it as stories.
Data is loaded from the source system into SAP Datasphere, where the following spaces are created:
• Ingestion space - contains the data products as local tables, and the replication flows that load data to
them.
• Preparation space - contains views built on top of the data products to prepare them for consumption.
• App space - contain analytic models built on top of the views to expose the data for consumption in SAP
Analytics Cloud. This space has the name of the insight app.
Note
These spaces are SAP-managed. You cannot create objects in them, share objects to or from them or
otherwise import or export content to or from them.
The spaces created for the insight app and the objects they contain are SAP-managed, and cannot be edited.
However, users who are members of the relevant spaces can view these objects in the standard editors.
If the insight app contains one or more task chains, then a user with access to the space and a DW Integrator
role (or equivalent privileges) must run each task chain at least once, and should then create a schedule to run
the task chain regularly in the future. For more information, see:
All insight apps delivered through SAP Business Data Cloud are secured through data access controls. By
default, no user can view any data presented in the provide SAP Analytics Cloud stories. To make data available
to the users authorized to see it, you must upload your authorizations to the provided permissions table.
Note
When data is delivered to SAP Datasphere through SAP Business Data Cloud, any authorizations that
are applied to the data in the source system are not imported. For general information about providing
row-level security, see Securing Data with Data Access Controls.
Procedure
1. In your preparation space, identify the fact views and the permissions table.
2. Prepare your permissions records in the operator and values format to provide row-level access to the
data in your facts (see Create an "Operator and Values" Data Access Control).
Note
• By default, users cannot see any data in the protected views. To enable them to see any data, you
must have at least one permissions record in the permissions table.
• To show all records to a selected user, use the ALL or * operator.
• For performance reasons, avoid having more than 5,000 permissions records for a single user.
The data products installed via SAP Business Data Cloud as part of an insight app do not include any
extensions defined in your source system. However, you can modify the data products to add any required
custom columns, and adjust the delivered views and analytic models to consume them.
Context
If your organization has extended the SAP Business Data Cloud tables exposed via data products as part
of your insight app, you will need to copy the relevant data products and associated content from an SAP-
managed space to a new space to add these custom columns.
Note
If SAP updates the data products and content, these updates are not copied over automatically to your
copied space and therefore, you would need to repeat this process for any updated data products.
Procedure
1. Identify all the relevant spaces which contain your data product or depend on it.
In this example, two data products are consumed by views and eventually exposed via analytic models:
4. Update the data products in the new ingestion space to include the extension columns by re-importing
them using the Import Entities wizard. This should update the local entities and replication flow with the
new columns, and optionally start the replication flow.
For more information, see Importing Entities with Semantics from SAP S/4HANA.
5. Stop replication into the original ingestion space and start replication into the copied space, if not yet done.
For more information, see Running a Flow.
In our example, replication is now active only in the copied ingestion space:
6. Ensure that the data access controls applied to the fact views are still protecting data appropriately.
For more information, see Applying Row-Level Security to Data Delivered through Insight Apps [page 141].
7. Modify the objects in the preparation and app spaces, to take into account the new extension columns
added in the ingestion space.
For more information, see Process Source Changes in the Graphical View Editor or Process Source
Changes in the SQL View Editor.
Users with a space administrator role can create database users to read data exposed by the space and to
write data to Open SQL schemas attached to space, providing a secure method for exchanging data with the
space via ODBC access to the run-time SAP HANA Cloud database.
Depending on the privileges they are granted, a database user can use their Open SQL schema to:
• Create tables and views and write data to them (see Connect to Your Open SQL Schema [page 148]).
Users assigned to a space can use these objects as sources when creating views and data flows (see Using
the Source Browser).
Note
Views with parameters cannot be imported to SAP Datasphere through an Open SQL schema.
• Access the space schema and read data from the space.
Database users can only read data from views that are created in the space with which they are
associated if the view has the Expose for Consumption property enabled (see Exposing Data For
Consumption).
• Create a table to act as a target to receive data written from a data flow (see Allow the Space to Access the
Open SQL Schema [page 149]).
• Connect an analytics client or other tool to SAP Datasphere (see Consume Data in Power BI and Other
Clients, Tools, and Apps via an OData Service).
• Consume space data in an SAP SQL Data Warehousing HDI container (see Consume Space Objects in Your
HDI Container [page 156]).
• Work with SAP HANA Cloud, data lake (see Integrating Data to and From SAP HANA Cloud Data Lake [page
158]).
Note
Users with the DW Administrator role can create database user groups, which are similar to database
users, but which are not assigned to a single space and can allow multiple spaces to read from and write to
their schemas (see Creating a Database User Group).
Users with the DW Space Administrator role can create database users, granting them privileges to read from
and/or write to an Open SQL schema with restricted access to the space schema.
Procedure
1. In the side navigation area, click (Space Management), locate your space tile, and click Edit to open it.
2. In the Database Users section, click Create to open the Create Database User dialog.
3. Complete the properties as appropriate and click Create to create the user:
Property Description
Database User Name Suffix Enter the name of your database user, which will be appended to the space name to give
the name of the Open SQL schema.
Can contain a maximum of (40 minus the space name) uppercase letters or numbers
and must not contain spaces or special characters other than _ (underscore).
Enable Password Policy Require the database user to change their password with the frequency defined in the
password policy (see Set a Password Policy for Database Users).
Enable Automated Predic- Allow the user to access the SAP HANA Cloud machine learning libraries.
tive Library and Predictive
For information about enabling and using these libraries, see Enable the SAP HANA
Analysis Library
Cloud Script Server on Your SAP Datasphere Tenant
Enable Read Access (SQL) Allow the user to read all views that have Expose for Consumption enabled via their Open
SQL schema.
Enable Write Access (SQL, Allow the user to create objects and write data to their Open SQL schema. These objects
DDL, DML) are available to any modeler working in the space for use as sources for their views and
data flows (see Using the Source Browser).
Enable Audit Logs for Read Enable logging of all read and change operations performed by the user.
Operations / Enable Audit
For more information, see Logging Read and Change Actions for Audit [page 313].
Logs for Change Operations
4. In your space page, click Deploy to deploy your space and create the database user in the run-time
database.
You will receive a notification when the deployment of the space is complete.
5. Click the button for your user in the Database Users list to open the Database User Details dialog.
6. Request a password for your database user:
If you want to work with the SAP HANA database explorer, you will need to enter your password to
grant the explorer access to your Open SQL schema. When connecting to your Open SQL schema with
other tools, you should additionally note the following properties:
• Database User Name
• Host Name
• Port
7. If you are using your database user to consume space data in an SAP SQL Data Warehousing HDI
container, click the Copy Full Credentials button to copy the json code that you can use to initialize your
user-defined service (see Consume Space Objects in Your HDI Container [page 156]).
8. Click Close to return to your space page.
You can now use your database user (see Connect to Your Open SQL Schema [page 148]).
When you have created a database user, you can connect to your Open SQL schema with the SAP HANA
database explorer or with other tools.
To obtain the host name, port, and user name required to connect to your Open SQL schema, click the
button for your user in the Database Users list.
For security reasons, the password is not displayed in this dialog. If you have forgotten your password, you can
generate a new one by clicking Request New Password.
To open your Open SQL schema directly in the SAP HANA database explorer from your space page, select your
user in the Database Users list, and then click Open Database Explorer. You will need to enter your password in
order to add your schema to the list of databases.
To browse the objects in your schema, expand its node, expand Catalog, and click on an object category. To
browse the space schema, click Choose Schema and select the space schema in the list.
To import or export data, open the SQL console, or perform other actions on your schema, right-click it in the
list of databases and select the appropriate command.
For detailed information about working with the SAP HANA database explorer, see the SAP HANA Database
Explorer guide in the SAP HANA Cockpit documentation.
You can connect to your Open SQL schema with third-party ETL or analytics tools, via ODBC or JDBC. You will
need to download and install the SAP HANA client or the JDBC driver and provide your connection information.
For more information, see Connecting to SAP HANA Cloud in the SAP HANA Cloud, SAP HANA Database
documentation.
For information about importing data from cloud storage services, see Importing and Exporting Data in the SAP
HANA Cloud, SAP HANA Database documentation.
If the SAP HANA Cloud script server is enabled (see Enable the SAP HANA Cloud Script Server on Your SAP
Datasphere Tenant and your database user has the Enable Automated Predictive Library and Predictive Analysis
Library option enabled, you can work with the SAP HANA Automated Predictive Library (APL) and SAP HANA
Predictive Analysis Library (PAL) machine learning libraries.
For detailed information about using the machine learning libraries, see:
To list the privileges granted to your user, run the following statement in the SQL console:
To grant the space write privileges in the Open SQL schema and the ability to write data to target tables in the
schema, use the GRANT_PRIVILEGE_TO_SPACE stored procedure. Once this is done, data flows running in the
space can select tables in the Open SQL schema as targets and write data to them, and task chains can run
procedures in the schema.
Prerequisites
Only the database user for the open SQL schema has the privilege to run the stored procedure
"DWC_GLOBAL"."GRANT_PRIVILEGE_TO_SPACE".
You can grant write privileges by running an SAP Datasphere specific stored procedure in the SQL console in
the SAP HANA Database Explorer.
Procedure
1. In the side navigation area, click (Space Management), locate your space tile, and click Edit to open it.
2. In the space page, go to Database Access Database Users , then select the database user and click
Open Database Explorer.
3. In the SQL console in SAP HANA Database Explorer, call the stored procedure to grant the ‘INSERT’,
‘UPDATE’, or ‘DELETE’ privilege to a space using the following syntax:
CALL "DWC_GLOBAL"."GRANT_PRIVILEGE_TO_SPACE" (
OPERATION => <operation>,
PRIVILEGE => <privilege>,
SCHEMA_NAME => <schema name>,
OBJECT_NAME => <object name>,
SPACE_ID => <space ID>);
privilege • 'INSERT' [required] Enter the write privilege that you want
to grant (or revoke) to the space.
• 'UPDATE'
• 'DELETE'
Note
• 'EXECUTE'
You can grant one privilege at a time.
schema_name '[name of the Open SQL schema]' [required] Enter the name of the Open SQL
Schema you want the space to be able to write
into.
object_name • '' [required] You can grant the write privileges, ei-
ther at the schema level or at the object level.
• null
• '[name of the objet]'
• At the schema level (all objets in the
schema): enter null or ' '.
• At the object level: enter a valid table name.
space_id '[ID of the space]' [required] Enter the ID of the space you are
granting the write privileges to.
CALL "DWC_GLOBAL"."GRANT_PRIVILEGE_TO_SPACE" (
CALL "DWC_GLOBAL"."GRANT_PRIVILEGE_TO_SPACE" (
OPERATION => 'GRANT',
PRIVILEGE => 'UPDATE',
SCHEMA_NAME => 'SALE#ETL',
OBJECT_NAME => 'MY_TABLE',
SPACE_ID => 'SALES');
To grant a space the ability to run procedures in the Open SQL schema:
CALL "DWC_GLOBAL"."GRANT_PRIVILEGE_TO_SPACE" (
OPERATION => 'GRANT',
PRIVILEGE => 'EXECUTE',
SCHEMA_NAME => 'SALE#ETL',
OBJECT_NAME => '',
SPACE_ID => 'SALES');
For more information on the SAP HANA statement CREATE PROCEDURE, see CREATE PROCEDURE
Statement (Procedural) in the SAP HANA SQL Reference Guide for SAP HANA Platform.
Results
If the run is successful, you receive a confirmation message in the Result pane. You can then open the Data
Builder, create a data flow, and select the tables as targets.
Users with a space administrator role can use SAP SQL Data Warehousing to build calculation views and other
SAP HANA Cloud HDI objects directly in the run-time SAP HANA Cloud database and then exchange data
between HDI containers and SAP Datasphere spaces. SAP SQL Data Warehousing can be used to bring existing
HDI objects into your SAP Datasphere environment, and to allow users familiar with the HDI tools to leverage
advanced SAP HANA Cloud features.
Context
To enable SAP SQL Data Warehousing on your SAP Datasphere tenant, an S-user must create an SAP ticket to
connect your SAP BTP account (see Enable SAP SQL Data Warehousing on Your SAP Datasphere Tenant).
Note
See the blog SAP SQL Data Warehousing with HANA (published in December 2017) for an introduction
to SAP SQL Data Warehousing.
See the blog SAP BTP Showcase – Access the SAP HANA Cloud Database Underneath SAP Datasphere
(published in January 2021) for an overview of the Integrated SAP SQL Data Warehousing approach.
See the blog SAP Datasphere – Hybrid Access to SAP HANA for SQL Data Warehousing (published in
June 2021) for information about accessing your HDI objects on another server through a standard SAP
HANA connection.
The integrated SAP SQL data warehousing approach allows you to add HDI containers to your space and
exchange data between them:
• Bring data from calculation views and other objects built in your HDI container into your SAP Datasphere
space via local tables and data flows (see Add an HDI Container and Access its Objects in Your Space [page
155]).
• Use data exposed from your SAP Datasphere space as a source for calculation views and flowgraphs built
in your HDI container (see Consume Space Objects in Your HDI Container [page 156]).
6.1 Prepare Your HDI Project for Exchanging Data with Your
Space
To allow your SAP Datasphere space to read from and, if appropriate, write to the HDI container, you must
configure your HDI project to build on your SAP Datasphere tenant and define the appropriate roles.
Procedure
1. Create the required roles to allow SAP Datasphere to read from and, optionally, write to the container.
You must define the roles DWC_CONSUMPTION_ROLE and DWC_CONSUMPTION_ROLE# (with grant option) in
the container to allow you to add it to your space and allow you to exchange data between the container
and the space.
The contents of the roles define the HDI objects that can be read from or written to. You can allow access to
individual objects or to all the objects in the container.
These examples show roles that provide read access to individual objects or to all objects in the container:
This example provides read access to one table and one This example provides read access to all the objects in the
view in the container: container:
{ {
"role": { "role": {
"name": "DWC_CONSUMPTION_ROLE", "name": "DWC_CONSUMPTION_ROLE",
"object_privileges": [ "schema_privileges": [
{ {
"name":"MyNamespace::Table1", "privileges":[ "SELECT" ]
"type":"TABLE", }
"privileges":[ "SELECT" ] ]
}, }
{ }
"name":"MyNamespace::CalcView1",
"type":"VIEW",
"privileges":[ "SELECT" ]
}
]
}
}
These examples show roles that provide write access to individual objects or to all objects in the container:
This example provides write access to two tables in the This example provides write access to all the objects in the
container: container:
{ {
"role": { "role": {
"name": "DWC_WRITE_ROLE", "name": "DWC_WRITE_ROLE",
"object_privileges": [ "schema_privileges": [
{ {
"name":"MyNamespace::Table1", "privileges":[ "SELECT",
"type":"TABLE", "INSERT", "UPDATE", "DELETE" ]
"privileges":[ "SELECT", }
"INSERT", "UPDATE", "DELETE" ] ]
}, }
{ }
"name":"MyNamespace::Table2",
"type":"TABLE",
"privileges":[ "SELECT" ]
}
]
}
}
For detailed information about hdbroles, see Roles (.hdbrole and .hdbroleconfig) in the SAP HANA Platform
documentation.
2. Configure your Cloud Foundry Settings to build to your SAP Datasphere tenant.
3. Edit your mta.yaml file to reference the database id of your SAP Datasphere tenant, your HDI container,
and the name of your user-provided service instance (see Consume Space Objects in Your HDI Container
[page 156]).
4. Build your database module.
Once the build is complete, the HDI container is available to be added to your SAP Datasphere space (see
Add an HDI Container and Access its Objects in Your Space [page 155]).
Context
You must define the roles DWC_CONSUMPTION_ROLE and DWC_CONSUMPTION_ROLE# (with grant option) in the
container to allow you to add it to your space and allow you to exchange data between them (see Prepare Your
HDI Project for Exchanging Data with Your Space [page 153]).
Note
See the blog SAP Datasphere Integrated with SAP SQL Data Warehousing Sharing a Single DB Tenant
(published in July 2021) for a walkthrough of this process.
Procedure
1. In the side navigation area, click (Space Management), locate your space tile, and click Edit to open it.
2. In the HDI Container section, click the + button to open the Add Unassigned HDI Containers dialog.
If the + button is not available, then you need to raise a ticket (see Exchanging Data with SAP SQL Data
Warehousing HDI Containers [page 152]) or no HDI containers have yet been built on the SAP Datasphere
tenant.
3. Select one or more HDI containers and then click Add to add them to your space.
Note
Each HDI container can only be added to one SAP Datasphere space. A database analysis user can
review the assignment of HDI assignments using the following query:
• It may have been created before assignment was enabled or the database_id may need to be
specified (see Set Up an HDI Container in the SAP HANA Service documentation).
4. Deploy your space.
5. Create a view or data flow and select one or more objects from the container as sources (see Import an
Object from a Connection or Other Source).
To consume SAP Datasphere views that are exposed for consumption as sources for your calculation views and
flowgraphs, you must create a database user with appropriate privileges, create a user-provided service in SAP
BTP Cockpit, and configure your HDI project.
Context
Note
See the blog SAP Datasphere Integrated with SAP SQL Data Warehousing Sharing a Single DB Tenant
(published in July 2021) for a walkthrough of this process.
Procedure
1. Add the HDI container to your space (see Add an HDI Container and Access its Objects in Your Space [page
155]).
2. Create a database user with the following privileges:
See Creating User-Provided Service Instances in Cloud Foundry Environment in the SAP Service Manager
documentation.
5. In your SAP Business Application Studio project, do the following:
• Prepare your project for building, including specifying the name of your user-provided service instance
in your mta.yaml file (see Prepare Your HDI Project for Exchanging Data with Your Space [page 153]).
• Create an hdbgrants file to grant read access to the views that you want to consume from your SAP
Datasphere space.
You can, optionally, grant access to the Application Function Library by including the following code in
hdbgrants:
"global_roles" : [
{
"roles" :
[ "AFL__SYS_AFL_AFLPAL_EXECUTE_WITH_GRANT_OPTION"]
}
Note
The SAP HANA Cloud Script Server must be enabled to allow this grant (see Enable the SAP HANA
Cloud Script Server on Your SAP Datasphere Tenant).
• Create hdbsynonym and hdbsynonymconfig files to define synonyms for the consumable views.
6. Build your project.
To allow data flows in your SAP Datasphere space to use tables in your HDI container as targets, you must set
the appropriate roles and add the container to your space.
Procedure
1. Define the roles DWC_WRITE_ROLE and DWC_WRITE_ROLE# (with grant option) in your HDI container (see
Prepare Your HDI Project for Exchanging Data with Your Space [page 153]).
2. Add the container to your space (see Add an HDI Container and Access its Objects in Your Space [page
155]).
3. Create a data flow and select a table in the container as the target for your data flow (see Add or Create a
Target Table in a Data Flow).
Users with an administrator role can connect an SAP Datasphere space with SAP HANA Cloud, data lake to
store and access large amounts of data.
Data lake is a fully managed cloud service for securely storing and analyzing large amounts of data that
is infrequently updated. It leverages inexpensive storage options to lower costs, while maintaining excellent
performance and full SQL access to data.
Data lake includes elastically scalable compute to provide high-performance analysis on-demand, and enabling
cost control when it isn’t in use. It’s fully integrated into SAP HANA Cloud, sharing common security, tenancy,
and tooling.
For more information on data lake, please visit the SAP HANA Cloud, Data Lake product page.
To benefit from data lake, you need to assign a SAP Datasphere space that gets access to the data lake.
Currently, only one space can be assigned to access and use the data lake.
Tables in the data lake are accessed via virtual tables in the open SQL schema. These tables can be consumed
in the Data Builder. SAP Datasphere provides two stored procedures to ease this task. Take a look at Data Lake
API [page 161] for more details and examples on how to use the procedures.
For a detailed step-by-step guide to connect and work with your data lake see Working with Data Lake [page
159].
Assign a SAP Datasphere space to access and work with SAP HANA Cloud, data lake.
Prerequisites
Before you can work with data lake, you need to enable it by selecting a storage size in System
Configuration Tenant Configuration (see Configure the Size of Your SAP Datasphere Tenant.
Context
You can assign a space that connects to data lake. Tables in the data lake can then be accessed via virtual
tables in the open SQL schema. These tables can also be consumed in the Data Builder.
Note
Procedure
1. Go to Space Management and select the space that you want to give access to data lake. Alternatively, you
could create a dedicated space for the data lake.
2. Under Storage Assignment select the check box Use this space to access the data lake. If another space
already has access to the data lake, you won't be able to assign your space.
3. Click Save.
You can now use your preferred SQL tool to create tables in data lake and access these tables via SAP
HANA virtual tables in your open SQL schema. SAP Datasphere offers two stored procedures that you can
use to easily create and access the tables. For more information and examples on the stored procedures
see Data Lake API [page 161].
4. In your SQL tool, use the "DWC_GLOBAL"."DATA_LAKE_EXECUTE" ( IN STMT NCLOB ) stored
procedure in the schema DWC_GLOBAL to create tables in data lake. Please note, that the statements
issued via this procedure are not audited.
5. Then create SAP HANA virtual tables in your open SQL schema that refer to the tables in data lake. The
virtual tables are used to query the tables in data lake.
Use the following procedure to create a virtual table in your open SQL schema:
"DWC_GLOBAL"."DATA_LAKE_CREATE_VIRTUAL_TABLE"
Results
You can then work with your virtual tables in the Data Builder by selecting the open SQL schema used for the
data lake.
Example
CALL "DWC_GLOBAL"."DATA_LAKE_CREATE_VIRTUAL_TABLE"
( VIRTUAL_TABLE_NAME => 'TABLE1_VT',
DATA_LAKE_TABLE_NAME => 'TABLE1'
) ;
CALL "DWC_GLOBAL"."DATA_LAKE_CREATE_VIRTUAL_TABLE"
( VIRTUAL_TABLE_NAME => 'SYSTAB',
DATA_LAKE_TABLE_NAME => 'SYSTAB',
TARGET_SCHEMA_SYS => true
) ;
For more information see System Views in Data Lake Relational Engine (SAP HANA DB-Managed).
This procedure is used to execute statements in data lake. It's a wrapper procedure around the data lake
procedure REMOTE_EXECUTE. This enables you to create or drop tables in data lake, or to load data from
files.
• "DWC_GLOBAL"."DATA_LAKE_CREATE_VIRTUAL_TABLE"
(
IN VIRTUAL_TABLE_NAME NVARCHAR(256),
IN DATA_LAKE_TABLE_NAME NVARCHAR(256),
IN TARGET_SCHEMA_SYS BOOLEAN DEFAULT false
).
• This procedure provides the option to create SAP HANA virtual tables in the open SQL schema that refer to
objects in data lake. These virtual tables are used to query data in the data lake. For more information see
What's the Difference Between Data Lake Relational Engine and SAP HANA Database SQL Statements?.
The procedure can also be used to access system views of data lake in schema SYS. The optional IN
parameter IN_TARGET_SCHEMA_SYS needs to be to true to create a virtual table on a data lake system
view. For more information see System Views in Data Lake Relational Engine (SAP HANA DB-Managed).
Virtual tables can be dropped directly using DROP TABLE <virtual_table_name> syntax.
For more information see Query Data in Data Lake Relational Engine (SAP HANA DB-Managed).
Example
CALL "DWC_GLOBAL"."DATA_LAKE_CREATE_VIRTUAL_TABLE"
( VIRTUAL_TABLE_NAME => 'TABLE1_VT',
DATA_LAKE_TABLE_NAME => 'TABLE1'
) ;
CALL "DWC_GLOBAL"."DATA_LAKE_CREATE_VIRTUAL_TABLE"
( VIRTUAL_TABLE_NAME => 'SYSTAB',
DATA_LAKE_TABLE_NAME => 'SYSTAB',
TARGET_SCHEMA_SYS => true
) ;
Users with a space administrator or integrator role can use the Data Integration Monitor app to schedule, run,
and monitor data replication and persistence tasks for remote tables and views, track queries sent to remote
source systems, and manage other tasks through flows and task chains.
Note
For optimal performance, it is recommended that you consider staggering the scheduled run time of tasks
such as data flows and task chains that may contain these tasks. There is a limit on how many tasks can be
started at the same time. If you come close to this limit, scheduled task runs may be delayed and, if you go
beyond the limit, some scheduled task runs might even be skipped.
The Data Integration Monitor is composed of several monitors that help you track how data is managed within
your space. Note that the available monitors depend on space capabilities.
• Remote Tables (Not available if you are in space of type SAP HANA Cloud, data lake files): Copy data for
remote tables that have been deployed in your space and monitor the replication of the data. Copy or
schedule copying the full set of data from the source, or set up replication of data changes in real-time via
change data capturing (CDC). For more information, see Replicating Data and Monitoring Remote Tables
[page 168].
Note
Remote tables from a source that is configured as data access Remote Only are not displayed in this
monitor. For more information, see Managing and Monitoring Data Integration [page 163]
• Local Tables (Not available if you are in space of type SAP HANA Cloud, data lake files): Monitor all local
tables which have been created in a space and display their metrics. From the detail screen, you can also
delete table records on-demand, using filter conditions or using a schedule.
• Local Tables (Files) (Not available if you are in space of type SAP HANA Cloud, SAP HANA database):
Check how and when your local tables (files) were last updated and if new data has still to be merged. For
more information, see Monitoring Local Tables (File) [page 195]
• Views (Not available if you are in space of type SAP HANA Cloud data lake files): Add, monitor and
schedule persisted views. Turn your views into persisted views to improve performance while accessing
your data. For more information, see Persisting and Monitoring Views [page 196].
• Flows: View and monitor the execution details of the executed data flows, replication flows or
transformation flows. For more information, see Monitoring Flows [page 239].
Note
Data flows are not available for spaces of type SAP HANA Cloud data lake files
• Remote Queries:
• Remote Queries: Check and analyze how the communication is running between the federation layer
of SAP HANA Cloud and the connected remote source systems. For more information, see Monitoring
Remote Queries [page 249].
Depending on your role and the tasks you need to perform in the Data Integration Monitor, you may need
specific privileges. See Privileges and Permissions.
Here are some examples of privileges and permissions you need to perform some tasks:
• To create or update connections, you need the Data Warehouse Connection privilege.
• To perform one-time actions in one of the monitors, such as replicating full set of data, redeploying
persisted views, loading new snapshot, or removing persisted data, you need the privilege Data Warehouse
Data Integration Update.
• To schedule tasks in one of the monitors and perform task execution and log retrieval in general, you need
the privilege Data Warehouse Data Integration Execute.
• To add a new view in the Views monitor, or to set up or change partitioned data loading in the Remote
Tables monitor, you need the privilege Data Warehouse Data Builder Read.
Selecting a specific monitor tab from Data Integration Monitor ( ) displays a tabular list of the scheduled
tasks associated with the selected monitor type. It shows the details of runs executed in last 60 days. You can
personalize the columns you want to display clicking on Select Columns. From the tabular display, you can
click > at the end of a particular task row to display details of that monitored item’s execution history including
the completion date and status of each run. After selecting a specific monitor type, for Remote Tables, Views,
Flows, or Task Chains, each of the monitor displays provides a summary listing of currently scheduled and
previously executed tasks. For example, here is a sample of the monitor display for the Remote Tables monitor
The display provides a summary of all currently scheduled and previously executed remote table replication
tasks, their schedule and refresh frequency, various memory and disk usage statistics, as well as the last task
update or execution. Similar displays are provided for each of the other monitor types.
To view more information on any specific task in the summary list, click > in the right-most column of the row
to access the Run Details page for the task you selected. Depending on the monitor type you selected, the Run
Details page provides a flexible two or three panel display in which you can resize ( and ), expand/collapse
( ), or close ( ) the display of each panel to focus on specific detail information or log messages. In addition,
you can sort ( ) and filter ( ) information displayed in the different panels, as well as search log messages.
The following is an example of the Run Details display for a remote table replication task.
By default, the failed tasks appear on the top. For more information on statuses and substatuses, see
Understanding Statuses and Substatuses [page 165].
Note
From each monitor, you can navigate back to the Data Builder to change your objects by clicking Open in
Data Builder.
You also get information about who runs the task (Run By) and how it has run (Run Type): Direct (the task
started directly while clicking the Run button) or Scheduled (a schedule is created)
When you run an activity (replicate a remote table, persist a view, etc..), the progress of the task is monitored
using statuses and substatuses. Statuses and substatuses are available in the relevant editors as well as in the
System Monitor.
Depending on your activity, there might exist additional statuses. For more information, see Replicating Data
and Monitoring Remote Tables [page 168], Persisting and Monitoring Views [page 196], Monitoring Flows
[page 239], Monitoring Remote Queries [page 249] and Monitoring Task Chains [page 252].
Substatus Description
CANCELLED This task run was canceled, after it had started. In this case,
the data was rolled back and restored to the state that
existed before the task run was initially triggered.
FAIL_CONSENT_NOT_AVAILABLE You have not authorized SAP to run task chains or schedule
data integration tasks on your behalf. Select the option
provided to give your consent (Changing SAP Datasphere
Settings).
FAIL_NOT_TRIGGERED This task job could not be executed due to a system outage
or some part of the database system not being available
at the time of the planned execution. Wait for the next
scheduled job execution time or reschedule the job.
FAIL_OWNER_MISSING This task job could not be executed because it does not
have an assigned system user. Assign an owner user to the
job.
FORBIDDEN The assigned user does not have the privileges necessary to
execute this task.
LOCKED The same task was already running, so this task job cannot
be run in parallel with an existing task’s execution.
PARTIALLY_COMPLETED (PARTIALLY_EXECUTED) This task failed but, for some sub-objects, the execution
ended successfully.
SCHEDULE_CANCELLED This task job could not be executed due to an internal error.
Contact SAP Support and provide them with the correlation
id and timestamp from this task job’s log detail information.
SET_TO_FAILED This task was running but the user set this task’s status to
FAILED.
You can also monitor the statuses and substatuses in the (System Monitor). For more information, see
Monitoring SAP Datasphere
An administrator or a data integrator has configured the source connection as Remote Only to prevent remote
tables from being replicated in SAP Datasphere.
Note
If your connections have been set up in a previous version, they have been updated with data access .
The possibility to replicate the entire source object using remote table replication (snapshot or real-time) is
useful for some use cases such as performance optimization.
However, remote table replication offers limited support for use cases that require more flexibility when dealing
with larger amounts of data. Many of these use cases require ETL (Extract, Transform, Load) capabilities that
can be easily achieved with Replication Flow or Data Flow capabilities.
There are also cases where you don't want the data to be persisted on the Cloud, for example for data
protection or security concerns.
To support use cases where data does not need to be replicated using remote tables, the option data access:
Remote Only can be set to a connection when supported (currently, for SAP HANA Cloud and SAP HANA
on-premise). For more information, see SAP HANA Connections [page 92]. You can use the data access:
Remote Only for connections that both support remote tables and replication flows. This way, you will replicate
data only via replication flows (SAP recommendation) and not via remote tables, and thus avoid replicating the
same data twice.
The table below compares the actions you can do when your remote table is in data access Remote Only
compared to a remote table in data access Remote and Replication or with no particular option selected:
Replication of Data Not possible directly. However, Possible. You can do it via Data
selected data can be replicated via
Replication Start Data Replication
Flows monitor.
or Data Replication Enable
For more information, see Creating a
Real-Time Data Replication when
Data Flow and Creating a Replication
supported. Data can be replicated
Flow
either from the table editor (in the
.
Data Builder) or from the Remote
Tables monitor (in the Data Integration
Monitor)
Scheduling of data load Not possible Possible. For more information, see
Scheduling Data Integration Tasks
[page 256]
Import/Edit/Delete/Refresh/Display Possible. For more information, see Possible. or more information, see
Importing Tables and Views from Importing Tables and Views from
Sources. Sources.
Monitoring of remote tables No specific monitor is available. Possible. You can monitor remote
However, you can monitor the tables via the Remote Tables monitor.
exchange of data between SAP For more information, see Monitoring
Datasphere and the remote source Remote Queries [page 249].
system in the Remote Queries Monitor.
For more information, see Monitoring
Remote Queries [page 249]
Create/Edit/Delete Statistics Yes. For more information, see Creating Not possible
Statistics for Your Remote Tables [page
251]
The monitor shows the remote tables that have already been deployed in your space. Also, it only shows
remote tables for connections that have been created within your space.
Note
If your source data comes from an SAP HANA On-Premise system, select the adaptor following your use
case:
• You want to access the data remotely: SAP HANA smart data access (Data Provisioning Option: Direct)
would be the recommended adaptor to read the data. It allows higher degree of query pushdown to the
remote database, leading to better response times and less resource consumption.
• You want to replicate the data into SAP Datasphere: The preferred option for this is to use Replication
Flows, see Creating a Replication Flow. In case you require replication for remote tables, SAP HANA
smart data integration (Data Provisioning Option: Data Provisioning Agent) is the recommended
adaptor to push the data. It offers more options when loading the data, such as applying filter
conditions or data partitioning.
For more information on these adaptors, see Connecting SAP HANA Cloud, SAP HANA Database to
Remote Data Sources.
Initially, by default, when a remote table is deployed, the data is directly accessed live from the source and no
data is stored in SAP Datasphere.
For connection types that support replication, you can set up data replication to improve performance when
accessing the data.
The Data Access property of a remote table shows you whether data is accessed directly from the source or
whether data is replicated.
Note
Parallel execution is not supported. You can't perform two similar tasks at the same time and therefore
you need to wait until the currently running task is completed before launching a new identical task. For
example, if you want to start a data replication, but another task is already replicating data from the same
table, then you will get a notification and your task will fail.
The Frequency property of a remote table shows you whether or not data for a remote table is replicated and if
yes, at what frequency the replication takes place. Replicated data is stored in a replica table.
As long as the prerequisites are met, you can switch between directly accessing the data from the source
(remote), or accessing to replicated data (data replication in real-time or in direct or schedule mode),
according to your needs.
Data is copied into SAP Datasphere and is read from the replica table:
• For data replication, replicated means that data is read from the replica table but not expected to be
updated in real-time. To get your data updated, you must start a direct data replication or create a schedule
to plan regular replication.
• For real-time replication, replicated means that data is read from the replica table and expected to be
updated in real-time.
Data is not accessed locally in SAP Datasphere. It’s read from a virtual table, which accesses the data directly
from the source via the adaptors. Remote tables are created from the Data Builder.
Replica table
Remote data is read from these replica tables once the data access is replicated. For remote tables connected
via SAP HANA smart data access, only one remote table per source entity of a connection can be in status
"Replicated" (direct or scheduled replication, or real-time replication).
Virtual Table
Virtual tables enable you to access objects in other databases, without having to replicate in SAP HANA.
Replication
Start a data replication to improve performance when accessing the data. For connection types that support
replication, you can run a data replication to improve performance when accessing the data. You can replicate
data via a direct data replication, real-time replication, or you can schedule your replication tasks. You can also
partitions your data loads.
For more information, see Integrating Data via Connections [page 8] and Scheduling Data Integration Tasks
[page 256].
From the Data Builder, you can also restrict data loads to reduce the volume of data loaded in your remote
table.
SAP HANA smart data access and SAP HANA smart data integration allow you to access remote data through
virtual tables without copying the data into SAP HANA.
For more information, see Connecting SAP HANA Cloud, SAP HANA database to Remote Data Sources.
DP Agent
The Data Provisioning Agent is a lightweight component running outside the SAP Datasphere environment. It
hosts data provisioning adapters for connectivity to remote sources, enabling data federation and replication
scenarios. The Data Provisioning Agent acts as a gateway to SAP Datasphere, providing secure connectivity
between the database of your SAP Datasphere tenant and the adapter-based remote sources.
Odata
SAP OData is a standard Web protocol used for querying and updating data present in SAP using ABAP,
applying and building on Web technologies such as HTTP to provide access to information from a variety of
external applications, platforms, and devices. To be able to successfully validate and use a connection to an
OData service for remote tables, you need to configure some properties.
Cloud Connector
The SAP Cloud Connector provides a secure tunnel between SAP Datasphere and on-premise systems. You
need to set up and configure Cloud Connector before creating a connection to an on-premise source that you
want to use for remote tables via SAP HANA smart data integration.
Direct Connection
Some connections, don’t need extra connector (for example Amazon Athena, Google Big Query, etc.).
Remote Sources
You can connect several data sources to SAP Datasphere that support remote table functionality.
In the monitor, for all the remote tables that have been previously deployed in your space, you can see how
the data is accessed, remotely from the source or replicated. For replicated data, the frequency shows how you
are replicating your data: in real-time, using a schedule (scheduled) or on-demand (none) to capture your data
changes. You monitor the size and replication status of the tables and you can see when the latest replication
for a remote table took place.
Connection (Business Name) or Connection (Technical Name Displays the name of the connection the remote table be-
longs to. Technical or Business Name is displayed, depend-
Profile Settings
Technical Name (or Business Name) Displays the name of the remote table. Technical or Business
Name is displayed, depending on how you have configured
Frequency Displays how often data is replicated. The value here shows
the intended frequency that might have not been reached
yet depending on the status of the remote table.
Latest Update Displays when the data was last successfully updated in SAP
Datasphere.
Latest Change (Source) (For remote tables connected with SAP HANA smart data
integration in real-time replication only). Displays the date
and time when the data has been last updated in the source
system. It can take up to 5 minutes to show the information
in SAP Datasphere after the source has been updated.
Note
If the Latest Change (Source) column shows a signifi-
cantly lower value than the Latest Update column, it is
very likely that there is an overload situation in transfer-
ring changes from the source to SAP Datasphere. In
that case, refer to the documentationBest Practices and
Sizing Guide for Smart Data Integration when used in
SAP Data Warehouse Cloud
New Run Displays the next scheduled run (if a schedule is set for the
remote table).
Memory Used for Storage (MiB) Displays the size occupied by the remote table data in mem-
ory.
Disk Used for Storage (MiB) Displays the size occupied by the remote table data on disk.
You can personalize the columns you want to display clicking on Select Columns
Depending on the combination of values in the Data Access, Frequency and Status columns, from the Remote
Tables monitor, you can perform the following actions on remote tables:
Action Description
Start Data Replication Directly start a copy of the full set of data from the source in
the background.
Note
If you start data replication on a remote table whose
data access is Replicated (Real-time), you’ll remove the
real-time replication for this table. Your data will no lon-
ger be updated in real-time and you’ll have to start a new
data replication or create a schedule to get your data
updated.
Remove Replicated Data Stop replication and delete data from replica table.
Note
If you need to remove the replicated data for a remote
table whose data access is Replicated (Real-time), you
must ensure that the following requirements are met to
allow a proper deletion:
Enable Real-Time Data Replication Start replication of data changes in the source in real-time.
Note
When enabling the replication in real-time, the data is
updated in real-time. You don’t need to start a new data
replication or to create a schedule to update your data
Disable Real-Time Data Replication Disable real-time data replication even if your source system
is unreachable and switch the data access from Replicated
(Real-time) to Replicated (snapshot). It preserves the replica
table and its replicated data to allow further actions on the
remote table such as deleting or changing the properties
of the connection. For more information, see Disabling Real-
Time Data Replication [page 184].
Note
The remote table must be connected via SAP HANA
smart data integration, and must not be connected via
SAP HANA smart data access.
Note
If you create a schedule for a remote table whose data
access is Replicated (Real-time), the replication type will
change from real-time replication to batch replication at
the next run of the schedule. The data will no longer be
updated in real-time
Edit Schedule Change how the schedule is specified, or change the owner
of the schedule.
Resume Schedule Resume the scheduled task that was previously paused.
Note
You can select several remote tables to group actions whenever it's applicable:
• Schedule data replication: You can delete, pause or resume schedules for several remote tables that
have schedules defined. You can also assign several schedules to your name and become the owner of
the schedules.
• Remove Replicated Data: You can select several remote tables that have replicated data available and
remove the replicated data for all of them.
• Many actions like Start Data Replication, Remove Replicated Data, Create Schedule, etc. cannot be
grouped. In such cases, they are disabled.
After triggering a group of actions, the status of each action is displayed in the notification area. As these
operations run in an asynchronous mode, the status will not be updated automatically, and you need to
manually Refresh the page to get the updated status.
Other Actions on Remote Tables (connected via SAP HANA Smart Data
Integration/SAP HANA Smart Data Access)
Other actions are available on remote tables, but behavior depends on how your remote table is connected (via
SAP HANA smart data access or SAP HANA smart data integration):
Remote Tables via SAP HANA Remote Tables via SAP HANA
Actions Smart Data Access Smart Data Integration Comments
Filtering data loads Supported but with Supported but with Action to be done from the
restrictions depending on restrictions depending on (Data Builder). For more
data type and adapter used. data type and adapter used. information, see Restrict
Remote Table Data Loads
Refresh Remote Table Supported with some Supported. Action to be done from the
Definition connections. (Data Builder). For more
information, see Process
Source Changes in the Table
Editor
Partitioning data loads Partitioning on key columns Partitioning on key and non- Action to be done from
only and for snapshot key columns with certain Remote Tables monitor
replication only (no real- restrictions depending on details screen. For more
time) data type and adapter used. information, see Partitioning
Remote Table Data Loads
[page 184]
Replicating Remote Tables Connected via SAP HANA Smart Data Access
For remote tables connected via SAP HANA smart data access, only one remote table per remote object can
have data access Replicated.
Example
Example 1: Several remote tables (A, B, and C) are pointing to the same source object but none of the
remote tables is replicated:
• You are allowed to replicate only one remote table in this group. For example, you replicate remote table
B.
• After the replication (or the scheduling) is triggered on this remote table B, the replication and the
scheduling of the other remote tables (A and C) are disabled. Remote tables A and C will have to use
the replicated data of the replicated remote table (B), or you must remove the schedule and replicated
data of the remote table B to allow scheduling and replication of the remote table A or C.
Example
Example 2: Several remote tables (A, B, and C) are pointing to the same source object and one remote
table (A) is replicated (or is scheduled to do so):
• On remote table A, normal replication and scheduling actions are enabled, depending if it is replicated
and/or scheduled.
• The replication and the scheduling of the other remote tables (B and C) are disabled. Remote tables B
and C will have to use the replicated data of the replicated remote table (A), or you must remove the
schedule and replicated data of the remote table A to allow scheduling and replication of the remote
table B or C.
If the replication status shows an error for snapshot or real-time replication, you can repair the load.
1. Select the remote table and choose Data Replication Remove Replicated Data .
2. Select the remote table and choose Data Replication Start Data Replication or Data Replication
Enable Real-Time Data Replication .
Canceling a Run
If you need to cancel a running snapshot replication or an initialization task (for example if the replication or
initialization takes long) you can click the Cancel Run button available in the Remote Tables Details screen:
1. From the Remote Tables , Click of the remote table you need to cancel the run.
2. Click Cancel Run.
Note
A replication task consists of several steps that include exchanges with SAP HANA. Therefore, you
might get different status while clicking Cancel Run, depending where you are in the processing steps.
As a consequence, the replication status of your remote table turns into Failed, and a cancel replication task
starts.
Note
If replica data already exists from a previous run, the remote table continues to use it after the run is
canceled.
Start a data replication to copy the full set of data from your source object (usually a database table or view)
into SAP Datasphere.
Context
Availability of data replication (snapshot) depends on the connection type. For more information on which
connection types support snapshot replication, see Integrating Data via Connections [page 8].
Note
By default, the full set of data is copied and imported in a single background task. For remote tables
with larger data volumes, this can lead to out-of-memory errors or high memory peaks. To avoid such
situations, you can enable partitioned data transfer of large datasets for supported connection types
Procedure
1. Go to Data Integration Monitor and select the space in which the relevant remote table resides.
2. Go to the Remote Tables.
3. Select the remote table that you want to replicate data for.
4. When real-time replication has been used before, select Data Replication Remove Replicated Data .
5. You have two options:
Option Description
Directly start the replication Select Data Replication Start Data Replication .
Schedule the replication Select Schedule Create Schedule . For more infor-
mation, see Schedule a Data Integration Task (Simple
Schedule) [page 256].
Results
When you directly start the replication, it starts in the background and you can see the progress of the
replication in the replication status. Once the replication has finished, you can see it in the Status column. For a
successful load, you will see the status Available and the date and time of the latest update. The Frequency for a
replication you started directly is None. The changed size of the remote table will also be shown in the monitor.
When you scheduled the replication, the columns Frequency and Next Run are updated with the scheduling
information. You can display the scheduling settings by clicking Scheduled in the Frequency column.
You can access the detailed logs for the remote table: Select the relevant remote table and click (Remote
Table Logs).
Once, you start a new data replication (either directly or via a schedule) while replicated data already exists, the
replicated data in SAP Datasphere is deleted and fully replaced by the new one. .
Note
You can still access the current data until the replication ends. Once replication has successfully run, the
data is updated. Please note that this data refresh behavior requires temporarily additional in-memory and
disk-storage to be reserved during the processing of a new data replication.
Prerequisites
Note
Availability of real-time / trigger based replication depends on the connection type. For more information
on which connection types support real-time replication, see Integrating Data via Connections [page 8].
• The object in the source has to be enabled for change data capture (CDC).
• If the database table in the source is CDC enabled but the remote table has been deployed before real-
time replication was possible with SAP Datasphere, you have to re-deploy the table to allow real-time
replication. You can re-deploy the table in the remote table editor in the Data Builder.
• For source views, real-time replication ist not supported.
• For ABAP Dictionary tables, real-time replication is not supported.
• For ABAP ODP sources, real-time replication is supported for in the following cases:
• ODP-BW: InfoProviders with a change log. This can be:
• DataStore objects (advanced) with data activation and change log (object type ADSO)
• Standard DataStore objects (classic) (object type ODSO)
• InfoObjects
Note
To properly display real-time capability for ODP-BW InfoProviders in SAP Datasphere, please
update your SAP BW to:
• SAP BW 7.4: Support Package Stack 23 or higher (see SAPKW74023 )
• SAP BW 7.5: Support Package Stack 17 or higher (see SAPK-75017INSAPBW )
• ODP-CDS: For all ABAP CDS views with primary key and that are delta-enabled
Note
If you apply filters to replicate your remote table, you must apply them on primary key fields
only. Indeed, delete operations on ABAP CDS Views are only recorded with their key fields in the
Operational Delta Queue (ODQ), but no content is recorded for other field types. In case of filtering
on non-key fields, there is strong risk that deletions are not propagated consistently to the target
remote table.
• ODP-SAPI: For all extractors that are delta enabled (delta method ADD* not allowed) and have a
primary key. For more information, see CDS Views Enabled for Data Extraction.
Note
For an overview of all extractors released for ODP replication (ODP SAPI), see SAP note
2232584 .
• For remote tables using an SAP HANA smart data access connection (SAP HANA on-premise via
Cloud Connector or SAP HANA Cloud), you can replicate your data changes in real-time but with some
restrictions as some remote object types are not supported:
• Remote object needs to be a table of type "COLUMN TABLE" (other table types like Row Tables or
Virtual Tables not supported)
• Some data types and SAP HANA table features are not supported. For more information, see
Unsupported Data Types and Other Limitations.
Note
You can’t pause the replication for connections if SAP HANA smart data access is used.
• For remote tables connected via an SAP HANA smart data integration Cloud Data Integrator (CDI) adapter,
we recommend that you use a DP Agent version equal or higher than 2.6.1.
Procedure
1. Go to Data Integration Monitor and select the space in which the relevant remote table resides.
2. Go to the Remote Tables.
3. Select the remote table that you want to replicate data for.
4. When changing from scheduled data replication to real-time replication, you must select Data
Replication Remove Replicated Data to delete the current replicated data from the replica table and
delete the schedule.
Results
Real-time replication gets initialized by copying the data from the source. You will see the status Active in the
Status column once real-time replication has been successfully initialized. Whenever data now changes in the
source, the changes will be copied into SAP Datasphere. You won't be notified on individual replications. The
Latest Update column shows date and time when data has been lastly changed in SAP Datasphere.
Note
No logs are generated when data is replicated in real-time mode. The displayed logs relate to previous
actions.
Whenever you need to perform maintenance operations in one or more underlying source systems, with the
DW Space Administrator role you can pause real-time replication for selected connections that connect to
the sources through a Data Provisioning agent in the Connections application. For more information, see Pause
Real-Time Replication for a Connection [page 126].
Whenever you need to upgrade or patch your Data Provisioning agent, or you need to apply any changes to its
configuration, with the DW Administrator role you can pause real-time replication for all connections that
use a specific Data Provisioning agent in the Configuration Data Integration On-Premise Agents . For
more information, see Pause Real-Time Replication for an Agent.
Note
From remote tables using an SAP HANA Cloud connection, you can’t pause the replication in the
Connections application if SAP HANA smart data access is used.
Context
It can occur that your real-time replication has been paused with errors. To resume your real-time replication,
you need to check the relevant log in the details screen of the relevant remote table to identify what is the
error and how to solve it. Once the error has been corrected, use the Retry button to resume the real-time
replication:
Procedure
1. Go to the Details screen of the remote table using the button Remote Table Logs.
2. Check the log details and correct the error.
3. Click retry to resume the real-time replication.
4. Refresh your screen.
Note
The Retry button must be used once you’ve corrected the errors. It does not serve to correct the error
itself but it serves to resume the real-time replication only. It corresponds to the RETRY statement
Tip
You might need to consider monitoring the delta queues in the ABAP backend in case of real-time
replication exceptions raised in ABAP adapter-based connections (ABAP, SAP BW, SAP ECC, SAP S/
4HANA on premise).
Disable real-time data replication even if your source system is unreachable, and switch the data access from
Replicated (Real-time) to Replicated (Snapshot).
If the source system is not reachable or if you simply don' t need to get your data updated in real-time, you can
disable the real-time data replication.
From the Remote Tables monitor, select one or more remote tables and click Data Replication Disable
Real-Time Data Replication .
Note
The selected remote tables must fulfill the following requirements to support disabling real-time data
replication:
When you disable the real-time data replication, it changes the data access from Replicated (Real-Time) to
Replicated (Snapshot). It preserves the replica table and its replicated data to allow further data consumption.
Switching the data access from Replicated (Real-Time) to Replicated (Snapshot) allows you to:
• Delete the remote table as it is not possible to delete a remote table, which is in data access Replicated
(Real-time).
• Change the properties of the connection. For more information, see Modify or Duplicate Remote Tables.
Note
If your remote table is connected via SAP HANA smart data access, you can delete it. This will stop the data
replication in real-time and you’ll be able to modify the connection details and restart a data replication.
Replication capabilities depend on the data type and size. Larger data volumes can cause replication runs
to terminate because of memory shortage. From the Data Integration Monitor Remote Tables - Details
Replicating remote tables with larger data volumes without having out-of-memory errors or high memory
peaks can sometimes be a challenge for Data modelers. To avoid such situations, you can create partitions
per remote tables and thus enable partitioned data transfer of large datasets for supported connection types
(which are all remote table-enabled connections that use a Data Provisioning agent).
Note
The partition is effective for both snapshot and real-time enabling replication if your remote table is
connected to the source via SAP HANA smart data integration. However, it is not working for real-time
replication if your remote table is connected via SAP HANA smart data access.
You can create partitions from the Data Integration Monitor Remote Tables – Details screen:
Column Select the column from which the It might happen that you can’t see all
partition will be defined. the columns available in your dataset
because:
• Not all data types are supported
for partitioning depending on
your data sources:
• SAP HANA data: Supported
data types include:
• String
• Integer, Integer
64, Decimal,
hana.SMALLINT,
hana.TINYINT
• Date, DateTime, Time,
Timestamp
• Binary
For more information,
see Partitioning Limits in
the SAP HANA Cloud,
SAP HANA Database
documentation.
Restriction
For ODP data sources, you
can create partitions on non-
key columns only if your
remote tables have been
deployed using a DP agent
connection from release
2.5.4.1.
Parallel Processes (SAP HANA smart You can define up to 5 parallel By default, the partitions are loaded
data integration only) processes. one after the other to control
resource consumption.
Note
No parallel processes are
possible for remote tables
connected with an SAP HANA
smart data access adapter.
4. Load a data replication to start the partitioning immediately, or alternatively create a schedule to run the
data replication later (or wait until your next schedule run will occur).
Note
Partitions can be updated or deleted using the corresponding menu. But for remote tables connected
via SAP HANA smart data access, you must first remove the replicated data before you can delete the
partitions.
Whenever it’s necessary, you can update the existing partitions using the menu Partitions Edit .
Example
For example, you want to include a new year in your partitions, or you need to fine granular the existing
partitions.
Note
Changes will be reflected only after a new data replication (snapshot or scheduled), or a real-time
replication.
Restriction
For remote tables connected via SAP HANA smart data access, you must first remove the replicated data
before you can change the partitions.
You can delete an existing partition using the menu Partitions Delete .
Restriction
For remote tables connected via SAP HANA smart data access, you must first remove the replicated data
before you can change the partitions.
Once the partitioning task is completed, messages with information on partitions is created. You have different
places to monitor the partitioned data loads:
• In the task logs that are created: one log is created per partition and contains detailed information.
• In the Data Integration Monitor Remote Queries by displaying the SQL Statement: The SELECT
statements sent by SAP HANA smart data integration to the remote source are listed. The partition
information is part of the WHERE clause of the particular statement. For more information, see Monitoring
Remote Queries [page 249].
Note
If your defined ranges don’t cover all the data, additional partitions are created in the background . For
example, if rows have been inserted or modified since your last data replication and they now don’t match
any of the defined ranges. An additional partition called 'Others' is created to cover this data.
Let's take an example to understand how the partitioning is done. The following partitions for column Calyear, a
4-digit numeric field which can include NULL values, are defined in the UI:
The replica table is partitioned accordingly. An "Others" partition could be added for data, which is not covered
by the defined partitions:
The partition ranges are also used by SAP HANA smart data integration data load runtime as filters in the
SELECT statements sent to the remote source. Usually, the filters are exactly the same as the partition ranges,
with the exception of ABAP ODP Extractors. In this case, the filters are as follows:
SAP HANA smart data integration data load runtime for ABAP ODP Extractors:
Monitor all the local tables created for a space in the Data Builder and check their metrics.
Local tables can be used as source data or target data by SAP Datasphere apps. As a Data Integrator you might
want to monitor all local tables which have been created in the current space to check data size, updates in
these tables, etc.
Go to Data Integration Monitor Local Tables . All local tables that have been created in the Data Builder
are listed.
Note
Tables created in Open SQL schemas are not listed. For more information, see Integrating Data via
Database Users/Open SQL Schemas [page 145].
Technical Name or Business Name Displays the technical name or the business name of the
table, depending on how you have configured your UI
settings in Your Profile Settings .
Growth in 30 Days Displays the growth in % of the number of records since the
last month. It is computed by taking the difference between
the number of records in the table today and the number
of records in the table 30 days ago, and it calculates the
growth in percentage.
Note
Figures are only available for large tables (superior of 1
megabyte). It takes the value N/A if the table is smaller
than 1 megabyte and O if no changes have been made
during the month.
Used Disk (MiB) Displays the disk storage currently used by the local table in
MiB.
Note
Value is set to Not Applicable for local tables that store
data in-memory.
Used In-Memory (MiB) Displays the memory currently used by the local table in
MiB.
Note
Value is set to Not Applicable for local tables that store
data on disk.
Last updated Displays the timestamp when the local table was last
updated.
Restriction
If the local table is updated through a replication flow,
the information is not available if the flow is running via
the load type Initial and Delta.
Last updated by Indicates how the table was last updated. For example,
name of the flow if it was updated by a flow.
Note
If you are authorized to navigate to the relevant object,
a clickable link is provided and you can navigate to the
object details. If you are not authorized to navigate to
the object, you’ll get the value unauthorized.
Restriction
If the local table is updated through a replication flow,
the information is not available if the flow is running via
the load type Initial and Delta.
Number of partitions Displays the number of partitions defined on the local table.
Note
Displays 0 if no partitions have been defined on the
local table. For more information, see Partitioning Local
Tables.
Delta capture Indicates if the local table allows delta capture. For more
information, see Capturing Delta Changes in Your Local
Table.
Delete records for local tables, on-demand, using filter conditions or using a schedule.
Table records can consume a lot of memory and you might need to do a cleanup of older records regularly.
Especially for tables where delta capture is enabled, as when you delete records, they are not physically deleted
but marked as deleted to allow other apps to propagate the changes to the different objects that consume
it in delta mode. For more information, see Capturing Delta Changes in Your Local Table. As these records
continue to consume memory until they are physically deleted from SAP Datasphere, you must clean up your
data regularly.
From the Data Integration Monitor ( ), go to the Local Tables monitor. Navigate to the details screen of the
local table you want to perform the data deletion.
Metric Description
Used Disk Displays the size used by the local table if it is stored on the
disk.
Used In-Memory (MiB) Displays the size used by the table, if it is stored in-memory.
For more information, see Accelerate Table Data Access
with In-Memory Storage
Delta Capture Indicates if the table tracks the changes made on your local
table after you have deployed it. For more information, see
Capturing Delta Changes in Your Local Table
From the Data Deletion tab, you define criteria to delete the records:
Delete All Records You want to delete all records contained in the local table.
Note that in case of tables with delta capture enabled,
all records will be marked as deleted but will not be
deleted physically. For more information, see Capturing
Delta Changes in Your Local Table
Delete Filtered Records You define a filter and based on this filter, the records will be
deleted. Note that for tables with delta capture enabled the
records will be marked as deleted but will not be physically
deleted.
Restriction
SQL based selection criteria are not supported.
Note
Click Show Preview to preview records that match the
filter criteria and will be deleted. You must have the role
DW Integrator or DW Modeler to preview the data.
Delete Records Marked as "Deleted" This option will appear only if your table is a delta capture
table. With this option, you can physically delete records
marked as deleted and that have already been successfully
processed by other apps that consume them. You can
define a period of deletion. For example, records older than
30 days.
Note
Data deletion preview is not supported for records
marked as "deleted", as the data editor does not
In the Preview section, in addition to previewing the data to be deleted when supported, you can see 2
additional metrics:
Metric Description
Number of Deletable records Displays the number of records that will be deleted based
on your previous selection.
Size of Deletable Records Displays the size of the records that will be deleted based
on your previous selection.
Note
Ensure that flows that consume the table are not affected.
From the Data Deletion Schedules tab, you can automate data deletion with a schedule.
1. Define the settings of the schedule: Give a business name and a technical name to your schedule, and
select the type of deletion you would like to apply.
2. Define the Frequency and Time Range for your schedule. For more information, see Scheduling Data
Integration Tasks [page 256].
3. Click Next Step.
4. Review your settings and click Create Schedule.
From this tab, you can also change or delete a schedule, and monitor the existing schedules with the available
metrics:
Metric Description
Profile Settings .
Note
Paused is displayed in the case the schedule has been
paused.
Last Run Start Displays when the last scheduled run has started (date and
time).
Last Run End Displays when the last scheduled run has ended (date and
time).
Records (Last Run) Number of records deleted with the last scheduled run.
From the Data Deletion Logs tab, you can monitor how the deletion task runs:
Metric Description
Start Displays the date and time when the deletion task started.
Duration Displays how long the task ran to delete all records.
Object Type Displays the type of object that contains the records to be
deleted.
Number of records Displays the number of records that have been deleted, or
marked for deletion in case of local table with delta capture
enabled.
Monitor your local tables (file). Check how and when they were last updated and if new data has still to be
merged.
Note
The object store is not enabled by default in SAP Datasphere tenants. To enable it in your tenant, see SAP
note 3525760 .
For additional information on working with data in the object store, see SAP Note 3538038 .
The object store cannot be enabled in SAP Datasphere tenants provisioned prior to version 2021.03. To
request the migration of your tenant, see SAP note 3268282 .
Local tables can be used as source data or target data by SAP Datasphere apps. As a Data Integrator you
might want to monitor all local tables (file) deployed in your space and check how and when data has been last
updated and if some data must still be merged.
Go to Data Integration Monitor Local Tables (Files) . All local tables (file) that have been created in the
Data Builder are listed. For more information on local tables (file), see Creating a Local Table (File).
Property Description
Technical Name or Business Name Displays the technical name or the business name of the
table, depending on how you have configured your UI
settings in Your Profile Settings .
Last updated Displays the timestamp when the local table (file) was last
updated.
Last updated by Indicates how the local table (file) was last updated. For
example, name of the task chain that updated it.
Buffer Last Updated Displays the timestamp when the inbound buffer was last
updated by a replication flow.
Buffer Last Updated by Indicates the name of the replication flow that last updated
the inbound buffer.
Delta capture Indicates if the local table (file) allows delta capture. For
more information, see Capturing Delta Changes in Your
Local Table.
From Data Integration Monitor Views , you can monitor views that have been created in the Data Builder.
You can persist these views (direct run or via a schedule) to make them available locally to improve the
performance when accessing your data. You can monitor the existing persisted views to keep control of your
data sizing and free up memory space.
Note
While persisting complex views, see Persisted Views and Memory Consumption [page 214].
Persisted Views
The definition of the view is stored in SAP Datasphere and data is read from the persistency table to improve
performance when accessing the data.
Persistency Tables
Tables created in SAP HANA Cloud that stored the required data. When a view is persisted, data is read from
this table to improve performance. When you refresh the persisted data of a view, a new persistency table is
created. Data from the technical view are inserted into the new persistency table. If everything runs properly,
the old persisted data is replaced by the new inserted data. In case of issues, previously persisted data are still
available, until you solve the issue and refresh the data with success.
Technical Views
Copy of the virtual views to be persisted into SAP HANA Cloud. The definition of the view is stored (including
associations, hints, etc). It serves to create the persistency table.
Persist Data
Views can be persisted by running a data persistence or scheduling a data persistence task. You can also
partition your data load into smaller parts to improve the performance.
For more information, see Creating Partitions for Your Persisted Views [page 210] and Scheduling Data
Integration Tasks [page 256].
For more information, see Creating a Graphical View and Creating an SQL View.
Remote table
Table created from a connection to allow source data to be accessed in SAP Datasphere. Views can be created
using these tables as source. By default, data access of a remote table is "remote", but it can be "replicated".
For more info, see Replicating Data and Monitoring Remote Tables [page 168].
Local table
Table created from sources that have already been integrated into SAP Datasphere. It serves of source to
create a view.
Flows
Define the flow of your data, starting by reading data from sources and writing data into a target after
performing transformations on the data.
For more information, see Creating a Data Flow and Creating a Replication Flow.
Remote Sources
You can connect several data sources to SAP Datasphere that support remote table functionalities. Views can
then be created from these remote tables.
For more information, see Integrating Data via Connections [page 8].
Local files
Local files can be imported and deployed to your space. They are then used to create a local table.
For more information, see Creating a Local Table from a CSV File.
Action in Views
• All Views: It displays all the deployed views that have been created in the Data Builder for the current space.
• Scheduled (default): It displays the views for which a schedule is defined to persist the data.
• Persisted: It displays the views already persisted.
You can perform the following actions (might depend on your role and privilege, see Privileges and Permissions
and Roles and Privileges by App and Feature:
Data Persistence menu Start Data Persistence: Start a new data persistence to up-
date or create the persisted view.
Note
You can set up a view as persisted even if it has been
created on top of remote data.
Note
You need to authorize SAP to run the recurring sched-
uled tasks on your behalf. You can do so via the message
that is displayed at the top of the monitor, or in your
profile settings under Schedule Consent Settings.
Access the detailed logs to monitor what's happened with Select the relevant view and click: (View Details).
your persisted view.
Note
You can select several views to group actions whenever it's applicable:
• Schedule: You can delete, pause or resume schedules for several views that have schedules defined.
You can also assign several schedules to your name and become the owner of the schedules.
• Remove Persisted Data: You can select several views that have persisted data and remove the persisted
data for all of them.
• Many actions like Start Data Persistence, Run, Create Schedule, etc. cannot be grouped. In such cases,
they are disabled.
After triggering a group of actions, the status of each action is displayed in the notification area. As these
operations run in an asynchronous mode, the status will not be updated automatically, and you need to
manually Refresh the page to get the updated status.
Note
You can still access the current data until the load ends. Once the data persistence has successfully run, the
data is updated. Please note that this data refresh behavior requires temporarily additional in-memory and
disk-storage to be reserved during the processing of a new data persistence.
The list of available persisted views is displayed in a table containing several information:
Column Information
Technical Name (or Business Name) Name of the persisted view.Technical or Business Name is
displayed, depending on how you have configured your UI
Schedule .
• Scheduled: A schedule task is defined for this view. If
you click on Scheduled, you will get detailed informa-
tion on the schedule. You can update the schedule op-
Schedule .
• Paused: Schedule is paused and the data is not re-
freshed.
Last Updated Shows when the persisted view was last updated.
New Run If a schedule is set for the view, see by when the next run is
scheduled.
Memory Used for Storage (MiB) Tracks how much size the view is using in your memory.
Disk Used for Storage (MiB) Tracks how much size the view is taking on your disk.
You can personalize the columns you want to display clicking on Select Columns
Data persistence can be run in case a view contains a parameter only if the following prerequisites are met:
If a view contains several input parameters, or if the input parameter has no default value, the view can’t be
persisted.
Once persisted, the view takes the data access value Partially Persisted because only records that match the
input parameter default value are persisted.
When you deploy a persisted view, you need to consider the following cases:
• If you update or redeploy a persisted view, you need the right permission level (Data Integration - Update).
For more information, see Privileges and Permissions. Note that after a redeployment, the data persistence
might be deleted and the data access could be changed back to virtual for views with structural changes.
This will happen as soon as the underlying database object is recreated as part of the deployment. This can
be seen in the Data Access status in the View Builder or in the Data Integration Monitor Views . The
update of the data persistence has to be triggered again.
• If the view uses other views that were changed but not yet deployed, they are deployed as well. If these
views are persisted, the persistence is deleted as well.
If you need to cancel a running data persistence task, go to the Details screen of the relevant view and click
Cancel Run . This will turn the status of the task into failed and you will be able to restart the task to persist the
view later if needed.
Note
A task to persist a view consists of several steps that includes exchanges with SAP HANA. Therefore, it
might take some time until the task is stopped and the status turns into Failed.
From the Views Details screen, you can change the settings to persist your data to ensure that you choose the
best run mode according to your business case.
When persisting a view, an SQL procedure is called in SAP HANA. In case of high data volume to persist,
processing the SQL procedure can impact the memory consumption and the system resources. In the details
screen of the view under Settings, you can optimize the run mode depending on your needs.
Run Mode
You can change the run mode of your data persistence and choose between performance and memory
consumption:
• Performance-Optimized (Recommended): This is the recommended mode to run the task. SQL queries are
processed directly in-memory. No intermediate result is stored on the disk to ensure a quicker data access
and data consumption. However, this mode consumes memory resources and if your data model is very
complex, it can result in an out of memory and the task failed.
• Memory-Optimized: The mode uses an enhancement of the SAP HANA Execution Engine, which stores
intermediate results of SQL statements on the disk. Memory consumption of SQL statements is then lower
because intermediate results are not stored in-memory but are saved on the disk. It results in slower
runtime, though. For example, you can use it when you get an out of memory error while running the task,
as alternative to manual partitioning.
You must ensure that you have sufficient disk space available. If not, it can result with an error of type
"Disk Full". You can monitor your disk space using the SAP HANA Cockpit. For more information, see
Disk Usage: Monitor Disk Volume
When Data Access Control is defined, it can impact the data persistence. Depending on where the Data Access
Control is set, it might prevent you from persisting a view.
Without Data Access Control (DAC), all users assigned to a space in SAP Datasphere can see and consume all
data available in this space. When DAC is applied to a data layer view or a business layer object, the rows of
data contained in the object are filtered based on the specified criteria and can impact the data persistence.
While defining Data Access Control in your view, you need to consider the impact on the data persistence:
• You can persist a view with Data Access Control set up.
• You can't persist a view consuming a view with Data Access Control set up.
Example
A view A is created as join of view B and C. As Data Access Control is set on view B, all users can see
data from view C, but only authorized users can see data from view A and B. View A can't be persisted.
In such cases, we recommend persisting the data on views B and C. If Data Access Control is set to view
A instead of view B, then also view A can be made persisted. And it may not be necessary to persist
view B and C.
Example
A view A is created on top of view B and view B is created on top of view C. Data Access Control is set on
view C. View B and view A can’t be persisted.
Example
View A is persisted, and it consumes view B. You add a DAC on view B. When redeploying view
B, persistence of view A is removed.
User A has authorizations to see data from "NA" region and User B is authorized to see data from "EMEA"
region.
Taking back the previous example, we now create the "Salary_Bonus" view on top of the "Sales" and "Salary"
views. Data from the "Sales" view is not protected whereas data from the "Salary" view is protected with DAC
on the "Region" column. An inner join is set between the "Product" column issued in the "Sales" view and the
"Product Responsible" column used in the "Salary" view:
Example 3: DAC is defined on one column and left outer join is set.
Again the "Salary_Bonus" view is created on top of the "Sales" and "Salary" views. Data from the "Sales" view is
not protected whereas data from the "Salary" view is protected with DAC on the "Region" column. A left outer
join is set between the "Product" column from the "Sales" view and the "Product Responsible" column from the
"Salary" view:
Due to the left outer join, the "Salary_Bonus" view contains all records of the "Sales" view. Even the
records from "Sales" view, where no authorized join partner exists in the "Salary" view are considered in the
"Salary_Bonus" view. The "Sales" and "Salary" views can be persisted but the Salary_Bonus" view can’t be
persisted due to DAC assigned to the "Salary" view.
The "Salary_Bonus" view is created on top of the "Sales" and "Salary" views. Data from the "Sales" view is
protected with DAC on the "Product" column whereas data from the "Salary" view is protected with DAC on the
"Region" column. A left outer join is set between the "Product" column from the "Sales" view and the "Product
Responsible" column from the "Salary" view:
Due to left outer join, the salary records are selected as long as they are in the authorized region (EMEA),
even when no dedicated "product" authorizations (responsibility) are granted. You can persist the "Sales" and
"Salary" views but you can't persist the "Salary_Bonus" view.
The "Salary_Bonus" view is created on top of "Sales" and "Salary" views. Data from the "Sales" view is
protected with DAC on the "Product" column whereas data from the "Salary" view is protected with DAC
on the "Region" column. An inner join is set between "Product" column from the "Sales" view and "Product
Responsible" column from the "Salary" view:
With a Union, data records are selected from the underlying sources and the respective authorizations are
applied . The result set is the sum of the selected records. You can persist both the "Sales External" and "Sales
Internal" views but you can't persist the "Sales_Bonus" view.
Example 7: Permissions entity for DAC is a view containing run-time dependent logic.
A DAC can use a view or a table as permissions entity. A view used as permissions entity for a DAC can contain
run-time dependent logic. If such a DAC is assigned to a view and this view is persisted, wrong results may
occur when reading data from the persisted view. This behavior occurs if the runtime dependent logic inside
the view used as permissions entity returns different results when loading a snapshot and when reading later
on from the snapshot. In this case, only a subset of the expected entries of the view will be considered. The
subset may even be an empty set.
Example
Let's assume we have the "EMEA_Sales" SQL View in which the "Current day" is fetched allowing users to
see the sales amount per day. A DAC is created using the "EMEA_Sales" view as permissions entity. This
DAC is assigned to a "Global_Sales" view.
Now, let's say that we are on April 1st, 2021 and you decide to persist the "Global_Sales" view. Imagine
that a user would like to consume data of the "Global_Sales" view on April 4, 2021. As a result, no data is
displayed as the view was persisted with entries of data from April 1st and the DAC is configured to always
fetch the entries from the current day (April 4).
Persisting views with larger data volumes without having out-of-memory errors or high memory peaks can
sometimes be a challenge for Data modelers. To avoid such situations, you can create partitions when
persisting a view and thus enable partitioned data transfer of large datasets for supported connection types.
Note
For this action, Data Builder (Read) privilege is required which is not included in the DW Integrator role. To
perform this action, ask your tenant administrator to assign your user to a scoped role that is based either
on the DW Space Administrator role or on a custom role that includes the following required privileges: Data
Warehouse Data Integration (Read, Update, Execute) and Data Builder (Read).
Caution
Some adapters used by the connected sources are not able to push down the filters for the partitioning.
Thus, performance can be impacted by the number of partitions created.
1. Click for the view in virtual access mode you would like to persist and for which you need to create
partitions for.
2. Go to the Partitions tab and click Define Partitions.
Selecting a Column
Select a column from the list, which will then become the partitioning column.
The columns available for selection are a subset of all view output columns. Some columns can be excluded
because:
Note
Partitioning is not recommended on columns that stem from remote tables using ABAP adapter with
ODP, as related filter conditions are not pushed to the source. Instead, use remote table partitioning. The
You can either select on a column that suits your business case, or switch on the option Rate Column
Suitability:
It runs a technical analysis of your view definition to calculate the level of suitability for partitioning for each
column. For example, calculated columns or columns with moving values (such as last changed date) are not
suitable as it might be difficult to do a reliable partitioning on such data.
In a 2nd step, the columns are sorted by their partitioning suitability, and additional information is displayed:
• Rating: It shows the suitability column rating. Possible values can be good, intermediate or bad.
• Details: It shows additional details about important factors that influence the suitability for each column to
help you with your choice. It can be one message or multiple messages.
Caution
Check the Locked option if you don't want to update a partition in the next run.
Caution
If locked partitions contain data filtered by authorizations provided by data access controls, you need to
unlock partitions and load a new snaphot of data, whenever these authorizations are changed.
An OTHERS partition is set by default. It contains data that is not covered by your defined intervals . For
example, if rows have been inserted or modified since your last data persistence and they now don’t match any
of the defined intervals. This OTHERS partitions can't be locked.
Note
• Partitioning requires an in-depth knowledge of the values that are used or are valid for the chosen
partitioning column.
• Negative values are not valid for partition bounds of numeric columns and will be converted to 0 at
runtime.
• The low bound must be lower than the high bound. The order is checked according to the type of the
partition column. Example: for string-like columns, '90' <= DOC_ID < '100' is not valid, whereas '090'
<= DOC_ID < '100' is valid.
• Values entered in the bounds for ABAP ODP remote tables must match the ABAP internal format: for
example, enter "20210714" if the date is "July 14, 2021".
Once the partitions are created and the view is persisted, don't change data type of the column you have
used to create the partition: if the data type is changed, it deletes the partitioning data.
When you think the data will no longer change, or no change is expected for a while, you can check this option
to avoid unnecessary data load.
When I load a new snapshot, only partitions 1, 4 and OTHERS are refreshed:
Note
If you change the partition definition, you can lock unchanged partitions to avoid unnecessary data loads.
In the case of adding partitions, if you lock the unchanged partitions, only the new partitions will be loaded
(Locking new partitions is ineffective, as they need to be updated once after they are introduced. They will
only be locked in the subsequent runs.)
In some cases, you might want to refresh data for a single or for multiple partitions only, independently if they
are locked or not. To do so, select the desired partitions and click Load Selected Partitions.
Example
In the example below, I have selected 2 partitions. Partition 1 is not locked and partition 2 is locked:
Note
If you change the partition definition, then the next load will be a full load. No manual load will be possible,
and the Load Selected Partitions button will be disabled.
You want to persist a complex view and consider how it affects the memory consumption.
When you persist a view, all the data that is provided by the view is stored in SAP Datasphere. When the view
is complex and consume a lot of data, some intermediate results have also to be stored in the memory so
that they can be used as source for further calculations and subsequent views. The amount of memory that is
needed for the runtime of the overall scenario can exceed the memory resources, and thus an out of memory
exception is raised or the memory consumption is very high.
If your view is very complex, using other views or remote tables, you might get an out of memory or increase
the memory consumption in SAP HANA Cloud while persisting it.
You might need to think about the following steps before persisting the view:
• Persist intermediate views: Identify views that combine a lot of data and are perhaps also used in
other scenarios. These views might read data from different sources and combine them. The memory
consumption could already be quite high, and the result of the view may have to be stored as intermediate
result in the memory. If this view is persisted, then the top views that do also heavy calculations can be
optimized: They reach out directly to the same SAP HANA instance and no additional memory for the
pre-calculations is needed. It could also be a good idea to split the large view into two or several smaller
views, and to persist these smaller views first. In most cases, the data persistence on a deeper level in
the view graph is more effective than just persisting the top view, because the data retrieval and the
calculations can be optimized if internal SAP HANA tables can be used. You might need to consider the
View Analyzer to check the memory consumptions and to identify views that can be optimized. For more
information, see Exploring Views with View Analyzer [page 221].
• Partition your data: While partitioning your data, not all the data is retrieved and stored at once in the
persisted views: The data is divided into several batches (partitions). Each partition is retrieved and
stored on its own, and thus consume less memory. For more information, see Creating Partitions for Your
Persisted Views [page 210].
• Handling of remote tables: The memory consumption can increase with usage of remote tables, especially
when filters used in the view definition can't be pushed down to remote tables. Make sure that the statistics
for the remote tables are up to date. If the remote tables are part of joins or aggregations are existing, then
statistics could optimize the data retrieval. The statistics can be run from the Remote Queries Remote
Tables Statistics . For more information, see Creating Statistics for Your Remote Tables [page 251].
You can check the SQL statement that is used to retrieve the data for a remote source in Remote Queries.
For more information, see Monitoring Remote Queries [page 249]. If the needed filter condition is not
part of the SELECT statement, then the filter is applied on all the data. This leads to a higher memory
consumption and a performance decrease. It might be a solution to replicate the data for some remote
tables where the data does not change regularly, for example master data tables. You can also try to persist
views that just do basic joins/unions on some remote tables, and are used in more than one view. If these
remote tables and views are persisted, then the data retrieval for other scenarios is highly improved in
terms of memory consumption and performance, because the top views can use tables in the same SAP
HANA instance.
• You can change the run mode of your data persistence and choose between performance and memory
consumptions. For more information, see Data Persistence and Run Mode [page 204].
Persisting views can consume a lot of memory and it might be useful to keep an eye on memory figures that are
stored with task logs.
You can monitor memory consumption in the task logs if you switch on the option Expensive Statement Tracing
(under (Configuration) → Monitoring).
• M_EXPENSIVE_STATEMENT: This view provides all statements, for which memory consumption, CPU
time or duration exceed the corresponding specified thresholds. For more information, see Configure
Monitoring
• Memory consumed during insert statements can be tracked using the view: M_ACTIVE_PROCEDURES .
For more information on how to use system view, see Working with SAP HANA Monitoring Views.
Access the detailed log of a data persistence task to get statistics information and run details, and navigate to
the remote statements for more information.
When your views are based on remote tables in Remote access, the data is read from a remote source system,
and it can take times to persist the data.
In the details information of a data persistence run, you can view statistics information in the logs: number of
remote statements, volume of data fetched from the remote source system:
When you click on View Details you can have detailed information and you can navigate to the Remote Queries
monitor clicking on the space name.
Note
You must have the relevant privileges to access the Remote Queries monitor (Data Warehouse Data
Integration privilege).
Note
If the data is coming from a space for which you don't have authorization to access, it will be indicated
as "unauthorized spaces".
Clicking on the space name will bring you to the Remote Queries monitor.
Even if you are a member of the space, you must have the relevant privileges to access the Remote
Queries monitor: Data Warehouse Data Integration privilege.
• If your view contains partitions, the log also shows how many partitions and how many remote statements
are involved during data persistence:
You can analyze the remote statements per partitions because each partition has its own statement ID.
Click on View Details on the partition row level of the log for more information or to navigate to the Remote
Queries monitor if you have the relevant permissions.
Check the data persistence metrics to decide on design change to prevent performance issues
Note
Administrators or data integrators want to analyze metrics information for a persisted view. The information
can be used to manually schedule data persistence tasks so that peaks memory can be avoided, or quicky
identify overload memory situations and decide on design change to prevent such situations, for example.
When your data persistence task has completed successfully, in the run details of the data persistence task,
you can see additional tabs: Metrics and Partitions Metrics (if you have defined partitions to persist your view
data).
Number of Persisted Records Displays the number of records persisted for the view.
Number of Records Read From Remote Sources Displays the number of records read from the remote
source.
Remote Sources Processing Time Displays the time (in seconds) spent reading data from the
remote sources.
SAP HANA Peak Memory Displays the maximum amount of memory (in MiB) the task
to persist view data has used during the runtime in SAP
HANA.
Note
You can see this information if the option Enable
Expensive Statement Tracing is enabled, and if the task
exceeds the thresholds specified in (Configuration)
→ Monitoring. Otherwise, no number is displayed.
SAP HANA Total CPU Time Displays the amount of CPU time (in ms) the task to persist
view data has used in SAP HANA.
Note
You can see this information if the option Enable
Expensive Statement Tracing is enabled, and if the task
exceeds the thresholds specified in (Configuration)
→ Monitoring. Otherwise, no number is displayed.
Used Storage Displays how much size (in MiB) the view is using in your
memory.
Number of Partitions Displays how many partitions have been defined for this
view.
Number of Locked Partitions Displays how many partitions have been locked for this
persisted view.
Note
Locked partitions are not updated and metrics related
to these partitions will take the value N/A in the
Partition Metrics tab.
Run Mode Displays the run mode used to persist your view data:
performance-optimized or memory-optimized.
Data Access Controls Displays if view data is protected by data access controls.
Elastic Compute Node Name Displays the technical name of the elastic compute node
used to persist the view. If no elastic compute node is used
the value is empty.
The Partition Metrics tab displays the metrics for each defined partitions of the persisted view:
Partition Range Display the partition range for this partition as defined
during creation.
Note
OTHERS partitions can contain NULL value or values
not included in the defined ranges.
Number of records Displays the number of records persisted with the partition.
SAP HANA Peak Memory Displays the maximum amount of memory (in MiB) the task
to persist view data for this partition range has used during
the runtime in SAP HANA.
Note
You can see this information if the option Enable
Expensive Statement Tracing is enabled, and if the task
exceeds the thresholds specified in (Configuration)
→ Monitoring. Otherwise, no number is displayed.
Data Access Controls Displays if view data is protected by data access controls.
Elastic Compute Node Name Displays the technical name of the elastic compute node
used to persist the view. If no elastic compute node is used
the value is empty.
Use the View Analyzer to explore graphical or SQL views and the entities they consume.
The View Analyzer provides you with statistics and useful information on each entity that composes your view,
no matter the complexity of the view. It analyzes each view definition, the consumed views, the local tables
and the data sources used by remote tables that compose your data model. You can then use this information
to optimize your data model and decide, for example, which view to persist when performance or technical
problems such as out-of-memory errors occur.
Use the View Analyzer to explore graphical or SQL views and the entities they consume.
The View Analyzer provides you with statistics and useful information on each entity that composes your view,
no matter the complexity of the view. It analyzes each view definition, the consumed views, the local tables
and the data sources used by remote tables that compose your data model. You can then use this information
to optimize your data model and decide, for example, which view to persist when performance or technical
problems such as out-of-memory errors occur.
1. Go to Data Integration Monitor Views . Select the view you need to analyze and navigate to the details
screen of this view. You need to select a space if you are assigned to several spaces.
2. Select Data Persistence Start View Analyzer .
Note
For this action, Data Builder (Read) privilege is required which is not included in the DW Integrator role.
To perform this action, ask your tenant administrator to assign your user to a scoped role that is based
either on the DW Space Administrator role or on a custom role that includes the following required
privileges: Data Warehouse Data Integration (Read, Update, Execute) and Data Builder (Read). For more
information see, Create a Scoped Role to Assign Privileges to Users in Spaces.
Note
If your data persistence contains partitions, the number of partitions is displayed in the results, but
the data persistence simulation does not consider it.
2. Generate Explain Plan: In the View Analyzer click the option Generate Explain Plan to get a quick
overview of your model and analyze the performance of a view.
It is a lightweight tool that can be generated in case of data persistence simulation failure or out-of-
memory errors of SQL analyzer plan file fails.
Once the explain plan is generated:
Note
You can start the View Analyzer from the Data Builder, but to view the results, you need to go to the
details page.
Note
You can generate the Explain Plan with the same permission required to run View Analyzer without
memory consumption.
To download the Explain Plan, you must have either the scoped role DW Administrator DW
Administrator or a custom role that includes Data Warehouse Runtime (Read) privilege.
3. Run the View Analyzer with memory consumption: In addition to exploring the entities that compose
your data model, the analyzer will execute the data persistence simulation for non persisted views. It
will report the memory consumption that would be used to get an entity persisted or will provide you
with the actual memory consumption used when you have persisted your data. The simulations are
executed sequentially as they may add some workloads to the system. To reduce these workloads,
you can define a limit to the memory consumption that can be used to persist the view thanks to the
threshold. If one of the analyzed views exceeds the limit, then no further data persistence simulations
are run.
Note
• To start the View Analyzer with memory consumption, you must have the space Read privilege.
If you are missing this privilege, the View Analyzer will run without memory consumption by
default.
• If your data persistence contains partitions, the number of partitions is displayed in the results,
but the data persistence simulation does not consider it.
If your data persistence contains partitions, the number of partitions is displayed in the results, but the
data persistence simulation does not consider it. You can run View Analyzer with memory consumption
for views with one input parameter with default value.
The maximum memory consumption is based on the statement memory limit of the space workload
configuration. For more information on statement limits, see Set Priorities and Statement Limits for
Spaces.
Note
To use the View Analyzer with memory consumption during data persistence simulation, you must
enable Expensive Statement Tracing in (Configuration). For more information, see Configure
Monitoring.
4. Generate SQL Analyzer Plan File: Before using this option, you must consider the following
requirements:
• To download this file, you must have the DW Administrator role or a custom role that includes
the Data Warehouse Runtime (Read) privilege. For more information, see Privileges and
Permissions.
• To open this file, you must install a compatible SQL plan visualization tool, such as SQL
Analyzer Tool for SAP HANA.
• This option requires additional system resources.
With this option, the data persistence simulation is run for the main view only, and the analyzer
creates a file containing detailed information about your data model that you can download for further
analysis.
Note
• If your data persistence contains partitions, the number of partitions is displayed in the results,
but the data persistence simulation does not consider it.
• If you want to analyze one of the underlying views, select the relevant view from the
dependency graph or the entity list, and launch the View Analyzer with SQL analyzer plan
file option.
• You can select only one view for analysis.
The Plan File is stored along with task logs and has a retention period that is the same as the task logs:
Tip
While working with complex views, see Persisted Views and Memory Consumption [page 214].
Cross-space analysis can be analyzed in View Analyzer, but you can only view entities that you are
authorized to view. For more information see: Authorization and Permissions [page 226]
Note
For views with one input parameter with default value, you can execute views with memory
consumption. For views with input parameters with default values, you can generate Explain Plan
and SQL analyzer plan file.
Working in View Analyzer requires different privileges and permissions depending on the tasks you want to
perform. You will not be able to view entities that you are not authorized to view.
Unauthorized Notification
When a user does not have permission to access a space or spaces, a row titled unauthorized will be displayed.
You can only view entities that you are authorized to view. Entities from spaces that you are not allowed to view
will be labeled as unauthorized. Contact your administrator to get authorization to view the hidden entities.
Note
To start the View Analyzer you will need the DW Integrator scoped role and the Data Builder (Read) privilege
which can be found in the DW Modeler scoped role. You will be able to see all entities with either the global
role DW Administrator or a custom role that includes Data Warehouse Runtime (Read) privilege.
Ask your tenant administrator to assign you a scoped or custom role that fits your required task.
The View Analyzer provides you with statistics and useful information on each entity that composes your view.
There are a variety of ways to learn more about your view and make decisions to optimize your data model
and performance. Due to the potential sensitivity of information gathered in View Analyzer, either the DW
Administrator role or a custom role is required to run the view or to read the results of generated files.
Ask your tenant administrator to assign you a scoped or custom role that fits your required task.
Action Permission
Run View Analyzer without mem- To run View Analyzer without memory consumption, you must have a role that is either
ory consumption based on the scoped role DW Administrator or on a custom role that includes the Data
Warehouse Data Integration (Read, Update, Execute) and Data Builder (Read) privileges.
Generate Explain Plan You can generate the Explain Plan with the same permission required to run View
Analyzer without memory consumption.
To generate Explain Plan, you must have either the scoped role DW Administrator or an
additional custom role that includes Data Warehouse Runtime (Read) privilege.
Run View Analyzer with memory To run the View Analyzer with memory consumption you must have a role that is ei-
consumption ther based on the scoped role DW Administrator or a custom role that includes Data
Warehouse Data Integration (Read, Update, Execute) and Data Builder (Read) and the
additional Space (Read) privilege.
Generate SQL Analyzer Plan File You can generate a SQL analyzer plan file with the same permission required to run View
Analyzer without memory consumption.
To download the SQL analyzer plan file, you must have either the scoped role DW
Administrator or an additional custom role that includes Data Warehouse Runtime
(Read) privilege.
Note
Now, let's start the View Analyzer with memory consumption. On the total of 8 GiB that I have allowed in
my statement memory limit, let's decide to limit the memory consumption to 4 GiB. This means that if a
data persistence task exceeds this limit, the next data persistence simulations will be canceled:
Once the View Analyzer has completed the task, 2 tabs are available in the details screen:
Note
The Analyzed Views tab was previously called the Entities tab.
• Task Log: It contains all information related to the View Analyzer task itself. You can see that my 5
views have been analyzed, and an overview of some key indicators is displayed: memory consumption,
number of records, etc.
• Analyzed Views: It contains the details of the View Analyzer findings per analyzed entities. Information
is displayed in a table format:
Note
If partitions have been defined for remote tables,
it's not displayed. Only partitions defined for views
are indicated by the View Analyzer.
Peak Memory (MiB) Indicates the memory used during the simulation, or in
case of the view already persisted, the memory used to
persist the view.
If you click on an entity row or click the >, you'll access the details for the selected entity. For example,
if I click on the table SCARR:
Action Description
(Inspect) Display the whole list of results that you can also export
in CSV format:
Start Analyzer You can select one or more views and start a new view
analyzer.
Note
The analyzer will always start from the main
view. Selecting the views here in the entity list,
or in the lineage graph has an effect only on
the data persistence simulation. In case memory
consumption is chosen, the data persistence
simulation will be executed for the selected views,
regardless of their data persistence status.
Note
It is also possible to select multiple views here and
start the View Analyzer with memory consumption.
The entities results list show the name of the entity and the space that it resides in. The orange bar on the left
side of the table signifies any entities that have warnings. The warnings icon in the task bar opens a popover
that will provide more information about entities with warnings.
Data persistence guidance provides recommendation on what views to persist in View Analyzer. The entire
lineage graph is reviewed and relative metrics are analyzed to provide guidance for the optimal views to persist
to improve performance, run-time, and memory consumption.
The View Analyzer provides statistics and useful information on each entity that composes your view regardless
of the the complexity of the model. A graphical view can look simple, but the underlying lineage graph can be
complex and built with many stacked views. This can lead to technical challenges, performance problems, or
time-out errors.
The best candidate to be persisted can be the main view or a view of the lineage graph.
Note
Data persistence guidance is most useful with models that have complex lineage graphs.
The Data Persistence Candidate Score column gives you recommendations by identifying views that can help
improve your model if they are persisted.
The information tab provides more details of the View Analyzer Results and the metrics of the view.
Persist only one view at a time. Check the performance of the main view after the persistence run
completed. One possibility to check the performance is to simulate the data persistence in the View
Analyzer with the option, 'With memory consumption'. Continue the process in an iterative way until the
optimal performance is reached. The goal is to persist the minimal number of views required to optimize
performance.
Note
To get the best results from data persistence guidance, it is recommended that the user have permission
to see all views in the model from all spaces. If a user is unauthorized to see some views, the view with
the highest rating value may remain hidden (a warning will be displayed). For more information about
authorization and permission in View Analyzer see Authorization and Permissions [page 226].
The Data Persistence Candidate Score provides a recommendation for the best views to persist to increase
performance.
The Data Persistence Candidate Score is not a specific inherent value but a ranked score based on weighted
values from a list of metrics that most affect the performance of a view or a model. The top scores show the
best options to persist with the top five noted. The first is the 'Best Option'. The next are the 'Best Alternative'.
Persist one view at a time, check the performance, and iterate the process if needed.
Note
The values of the score are relative weighted values determined by the primary metrics in the lineage
graph.
Note
Branches of the views that are already persisted are not considered in the metric evaluation of the Data
Persistence Candidate Score. It is best to persist one view at a time and then check the performance.
A model can have a Data Persistence Candidate Score with values ranging from 0 - 10. Any change in the model,
for example, persisting one view, will change the values due to the change in the metrics used to calculate the
data persistence score.
Analysis for data persistence guidance will be run for all entities including shared entities from other spaces.
The best candidate score will be computed regardless of the entities the views belong to. Results of the analysis
will be able to be viewed by those that have the authorization. If the best candidate is from a space the
user does not have the authorization to see the result from, a message will be shown. See Authorization and
Permissions [page 226] for more information.
Further Recommendations
Persisting data is one option to optimize a model or improve performance. Utilizing Transformation Flows is
also a good option to consider instead of persisting data depending on the scenario.
Consider replacing remote tables with local tables by using Replication Flow for replicating the data.
In the Flows monitor, you can find all the deployed flows per space.
Prerequisites
• To use the Flows monitor, you need the DW Integrator role along with the DW Viewer or the DW
Modeler role.
• To run the flows, you need the DWC_DATAINTEGRATION privilege and the Update permission. For more
information, see Managing Roles and Privileges.
• To schedule the flows, you need the DWC_DATAINTEGRATION privilege and the Execute permission. For
more information, see Managing Roles and Privileges.
Context
In the Flows monitor, you can find all the deployed flows (data flows, replication flows, and transformation
flows) per space. Here, you can run the deployed flows, and you can view and monitor the execution details of
the flows.
Note
For optimal performance, it is recommended that you consider staggering the scheduled run time of tasks
such as data flows and task chains that may contain these tasks. There is a limit on how many tasks can be
started at the same time. If you come close to this limit, scheduled task runs may be delayed and, if you go
beyond the limit, some scheduled task runs might even be skipped.
For more information and points to consider when using replication flows, see also SAP Note 3297105 .
Procedure
You can find all the deployed flows per space. You can run the deployed flows under Run, and you can
view and monitor the execution details of the flows. For some flows, you can create, edit, delete, pause
or resume a schedule, or assign a schedule to your name under Schedule, and you can select several
data flows to group the scheduling actions. For more information about scheduling, see Schedule a Data
Integration Task (Simple Schedule) [page 256].
Technical Name or Business Name Name of the flow. Technical or Business Name is dis-
played, depending on how you have configured your UI
settings in Your Profile Settings
Note
If the status of a flow is empty, it means that the
flow has not run yet. The status is updated every 10
minutes for replication flow (except when it is run via
a task chain: It updates every 2 minutes), and every
1 minute for a transformation flow or a data flow. To
get an immediate updated status, navigate to the run
details page.
Note
The initialization time for executing a flow takes an
average of 20 seconds even with smaller data loads,
causing longer runtime for the flow. It is calculated
using the task start time and task end time. For rep-
lication flow, it does not depend on the initial load
duration.
Last Run Start Provides information on when the last run of the flow
started.
Last Run End Provides information on when the last run of the flow
ended.
Next Run If a schedule is set for a flow, you can see here by when the
next run is scheduled.
Schedule Owner Displays the name of the current schedule owners when a
schedule is defined for a flow.
You can personalize the columns you want to display clicking on Select Columns
Note
Parallel execution is not supported: You can't run two identical flows at the same time. You need to wait
until the first run is completed before you can launch another one. Otherwise your second run gets the
status Failed (Run Already in Progress).
2. To see the run details of a flow, click (Details) for the relevant flow.
The run details page opens. It shows the runs of the flow in the left panel and their corresponding
messages in the right panel. Here, you can use the Stop Run (for Replication Flow), Cancel Run (for
Transformation Flow) or Set to failed (for Data Flow) buttons to stop a run that is currently running. The
buttons are available until the run ends.
3. Select a flow run in the left panel to view its details. The run details contain messages for all runs and
metrics for the current run.
• Messages:
For failed runs, click View Details besides the error message to see more information about the error.
• Metrics:
For successful runs, the metrics provide the record count for source and target tables used in the flow.
Note
• It takes a while to load the metrics once the flow run is completed.
• If the flow run has been optimized for better performance, no record count information is
shown.
• For historical runs of a flow, metrics can’t be retrieved.
For data flows, the logs messages and metrics are only displayed in English.
For more information, see Metrics for Transformation Flows [page 247].
4. If your flow failed, you can use the following tools from the details page of the flow run to analyze and solve
your issue:
• [Data Flow only] You can analyze the errors that have occurred during your data flow runs,
downloading the run details as a JSON file you can provide to the SAP Support for troubleshooting:
Click Download Run Details.
• [Data Flow only] You can also perform a complete health analysis and generate detailed logs if you
choose Run Start Run with Checkup . Send this health analysis log to your support team to get
help in case of issues.
• [Transformation Flow only]: You can run your transformation flow with Support Settings. For more
information, see Explore Transformation Flows [page 248].
You can pause a replication flow run and resume it at a later point in time, or you can stop it completely.
Tip
This text explains the options related to replication flow runs (in the context of monitoring). For information
about how to make changes to an existing replication flow in the Data Builder, see Modify a Replication
Flow.
On the left of the monitoring screen, you find the aggregated status information for your replication flow. If the
status for one or more objects in a run differs from the overall status, this is indicated by an additional status
value in brackets.
You can create a schedule for your replication flow and include your replication flow in a task chain if all objects
in it have load type Initial Only. For more information, see Schedule a Data Integration Task (Simple Schedule)
[page 256] and Creating a Task Chain.
You may want to pause a replication flow run, for example, while a system update is running.
Caution
You must always stop or pause a running replication flow before a source system downtime.
When you do so, the flow is stopped in SAP Datasphere, but not in the source. Consequently, the system still
keeps track of changes in the source, but does not copy them to the target.
Note
If you have many data changes in your source, and if you pause the replication flow for a longer time (such
as several hours), the logging tables can become quite big.
To pause a replication flow that has a schedule, you need to pause the schedule. (Pausing the replication flow
run itself is not possible in this case.)
When you resume the flow, the system replicates all source data changes that happened while the replication
flow was paused, which means that the first replication flow run after a pause can take considerably longer
than usual.
When you open the Flows monitor for a replcation flow with premium outbound integration, a value called Used
Premium Outbound Data Volume is shown in the overview panel on the left. This is the total volume for all
replication flow runs with premium outbound integration (replication to a non-SAP target) in this tenant during
the last 360 days.
Note
• The value may change significantly from one day to the other, for example when you run an initial load
for a replication flow, or when an initial load drops out of the statistics because it happened more than
360 days ago.
• The retention period has been increased from 90 days to 360 days starting as of July 11, 2024, so
depending on when you access this feature you may see the data for less than 360 days.
If you do so, the flow run is stopped permanently in SAP Datasphere as well as in the source. You can still run
it again, but it will then start from scratch (rather than from where it left off when you stopped it). If you stop a
replication flow run because you don't need it anymore, you should also delete it so that it does not clutter your
system. For more information, see Delete a Replication Flow.
Space Deletion
If you have replication flows in a space that is about to be deleted, make sure to stop your replication flows
before space deletion starts. This helps to avoid issues during space deletion and makes it possible for you to
start the replication flows again if the space gets restored at a later point in time.
When a space gets permanently deleted, all replication flows in it are deleted as well.
For more information about space deletion, see Delete Your Space [page 325].
This can be done in the Data Builder, see Modify a Replication Flow.
Note
Adding or removing objects without stopping the replication flow first is only possible for replication flows in
status Active.
This can be done in the Data Builder, see Modify a Replication Flow.
Context
You can use the Cancel Run button to stop a transformation flow run that is currently running. This button is
available until the run ends.
Procedure
The Run Details page opens. It shows the runs of the flow in the left panel and their corresponding
messages in the right panel. Under Runs, you can view the status of the transformation flow runs (the Run
activity) and also the status of any attempts to cancel a transformation flow run (the Cancel activity).
3. To stop a run that is currently running, select the relevant run and click the Cancel Run button.
If the run has been canceled successfully, the status of the transformation flow run will have the value
Failed (Canceled).
Canceling a transformation flow involves rolling back database transactions. Therefore, it might take some
time until the transformation flow is canceled and the status changes to Failed (Canceled).
Caution
For a transformation flow in a file space, when you cancel a run, it stops where it is, and does not roll
back what has already been processed. If you restart the run later, it will restart from the beginning, not
where it has stopped.
8.8.3 Watermarks
When you run a transformation flow that loads delta changes to a target table, the system uses a watermark (a
timestamp) to track the data that has been transferred.
On the Delta Capture Settings tab of the Data Integration Monitor, you can view the watermark for a source
table. Note that if a transformation flow run does not load delta data to a target table, no source table
information is displayed.
Local Tables
When you run a transformation flow that loads delta changes to a target table, the system only transfers data
that has a change timestamp greater than or equal to the watermark. Once the transformation flow run has
completed successfully, the system sets the watermark to the start time of the transformation flow run. Note
that if open transactions exist in the source table at the time the transformation flow run started, then the
earliest start timestamp of these transactions is used as the watermark.
When you run a transformation flow that loads delta changes to a target table, the system sets the watermark
to the timestamp of the first request that has a timestamp greater than the value of the watermark. If no
request exists that has a timestamp that is greater than the value of the watermark, the system sets the value
of the watermark to the maximum timestamp (12/31/9999 25:59:59).
If the transformation flow has not yet been run, or if you reset the watermark, then the value of the watermark
is null. If you then run the transformation flow, the system uses the minimum timestamp (0001-01-01
00:00:00) as the watermark.
The system only transfers data that has a change timestamp greater than or equal to the watermark. Once the
transformation flow run has completed successfully, the system sets the watermark to the timestamp of latest
request of the source table that the system could ensure was loaded to the target table.
If you reset a watermark, the system will transfer all data to the target table the next time the transformation
flow runs (using the load type Initial and Delta). This means that you do not need to redeploy the
transformation flow and use the load type Initial Only.
• If a table that can capture delta changes is joined with a second table, and columns in the second table
have been updated, you can reset the watermark to ensure that these changes are reflected in records that
have already been transferred.
• If corrupt data is present in the target table, you can reset the watermark for the transformation flow to
ensure that the latest data from the source table is loaded to the target table.
You can reset the watermark for a transformation flow in the Data Integration Monitor. Open your
transformation flow and navigate to the Delta Capture Settings tab. To reset the watermark, choose the Reset
Watermark button.
Procedure
In the Data Integration Monitor, open your transformation flow. On the Settings tab, you can change the run
mode for a transformation flow and choose between performance and memory consumption:
• Performance-Optimized (Recommended): This is the recommended mode to run the transformation flow.
SQL queries are processed directly in-memory. No intermediate result is stored on the disk to ensure a
quicker data access and data consumption. However, this mode consumes memory resources and if your
data model is very complex, it can result in an out of memory and the task failed.
• Memory-Optimized: The mode uses an enhancement of the SAP HANA Execution Engine, which stores
intermediate results of SQL statements on the disk. Memory consumption of SQL statements is then lower
because intermediate results are not stored in-memory but are saved on the disk. It results in slower
runtime, though. For example, you can use it when you get an out of memory error while running the task,
as alternative to manual partitioning.
Note
You must ensure that you have sufficient disk space available. If not, it can result with an error of type
"Disk Full". You can monitor your disk space using the SAP HANA Cockpit. For more information, see
Disk Usage: Monitor Disk Volume
Metrics provide the record count for source and target tables used in the flow. You can view the following
metrics for a completed transformation flow run:
• LOAD_TYPE
The load type for a transformation flow run. For more information about load types, see Creating a
Transformation Flow.
• TRUNCATE
Indicates whether the Delete All Before Loading option is used for the transformation flow run. For more
information, see Create or Add a Target Table to Your Transformation Flow.
• REMOTE_ACCESS
Shows "true" if a remote access has been used when executing the transformation flow. This is possible
only when the transformation flow consumes a BW Bridge object imported as remote table.
• NUMBER_OF_RECORDS
The number of records written to the target table.
• MEMORY_CONSUMPTION_MIB
The peak memory consumption (in mebibytes) for the SAP HANA database while running the
transformation flow.
Use Run with Settings to explore graphical or SQL views and the entities they consume in a transformation flow.
The Run with Settings option provides you with the ability to simulate a run, but also to download a plan file
containing statistics and useful information on each entity that composes your view, no matter the complexity
of the view. You can analyze each view definition, the consumed views, the local tables and the data sources
used by remote tables that compose your data model. You can then use this information to optimize your
data model and decide, for example, which view to persist when performance or technical problems such as
out-of-memory errors occur.
1. Go to Data Integration Monitor Flows . Select the view you need to analyze and navigate to the details
screen of this view. You need to select a space if you are assigned to several spaces.
2. Select Run Run with Settings . You have two options:
• Simulate Run: Simulating a flow allows you to test a transformation flow and see if you get the desired
outcome. Based on the result, you can decide to resolve errors or to optimize the flow to improve
performances. No changes are saved in the target table.
Note
A simulated run will not take the truncate function into account.
• Generate SQL Analyzer Plan File: Before using this option, you must consider the following
requirements:
Note
This file allows you to analyze your transformation flow to resolve errors and enhance its
performances. With this option, you create a plan file (also called .plv file) containing detailed
information about your data model that you can download for further analysis. This file provides more
details than the Simulate Run option.
Note
The plan file is stored along with task logs and has a retention period that is the same as the task logs.
In the Remote Queries monitor, you track the queries sent to your remote connected source systems for
your space. You can monitor the communication between the federation layer of SAP HANA Cloud and the
connected remote source systems, and analyze them.
SAP Datasphere is connected to remote systems via SAP HANA Cloud using data federation. SAP HANA Cloud
already provides a corresponding monitoring view of actual SAP HANA runtime data, including statistics and
status information related to the execution of Data Manipulation Language (DML) statements. The Remote
Queries monitor retrieves this information, so that you can keep an eye on how the communication between
SAP Datasphere and the connected remote source systems is running and better analyze possibilities for
performance bottlenecks while running queries in SAP Datasphere.
The Remote Queries monitor allows you to monitor the queries sent to your remote connected source systems
for your space. This information is helpful when analyzing executions of story (or on views or previews), since it
provides details of single read requests, like runtimes or result set size.
Note
Data Flows executions are not contained in this monitoring view as Data Flows don't use SAP HANA
Federation.
Example
You build a story on top on a view. In this story, you create charts and set filters on those charts. Let’s say
you remove one of those filters. A query is then triggered to the connected remote source system. A new
entry appears in the Remote Queries monitor and you can see the last statement sent by the story.
The Remote Queries monitor only shows the last 100 statements. It provides you with the following
information:
Remote Statement Duration Displays the duration in milliseconds of the total remote statement (open/
fetch/close).
Rows Displays the number of rows that have been processed by the statement.
Remote Statement Displays the outgoing statement that is sent to the remote connected
source system.
As the statement is most often too long and appears as truncated, you
need to click on More to get the full statement information.
This ID is provided by SAP HANA and can be used to do deep dive analysis
in the SAP HANA Cockpit. For more information, see SAP HANA Cockpit.
Executing SQL Statement Displays the statement that is currently running in SAP HANA Cloud, and
which includes a remote SQL statement.
You might have a look at this statement if something goes wrong to identify
the root cause. As the statement is most often too long and appears as
truncated, you need to click on More to get the full statement information.
Data Provisioning Agent Displays the name of the data provisioning agent. For more information, see
Preparing Data Provisioning Agent Connectivity.
You can personalize the columns you want to display clicking on Select Columns
Create statistics for your remote tables to improve federated query execution.
SAP Datasphere is using SAP HANA smart data integration and SAP HANA smart data access to connect to
remote data sources. Objects from remote data sources are made available with so-called SAP HANA virtual
tables, which are created when deploying remote tables in SAP Datasphere.
To ensure the best possible performance when accessing your remote tables and provide efficient query
execution plans, you can create statistics for your remote tables. SAP Datasphere uses the SAP HANA SQL
Optimizer that relies on accurate data statistics for virtual tables in SAP HANA and help you take decisions. For
more info, see SAP HANA SQL Optimizer.
Restriction
1. From the Data Integration Monitor ( ), select Remote Queries Remote Table Statistics .
Note
With the different tabs, you can navigate between the remote tables in Data Access: Remote or Data
Access: Replicated. However, statistics can be created only on remote tables with Data Access: Remote.
However, you can still navigate in existing statistics from previous actions.
2. Select the remote table for which you need to create statistics for, and choose Create Statistics. Or select
the remote table, click to navigate to the details screen, and click the button Create Statistics.
RECORD COUNT Creates a data statistics object that helps the query
optimizer calculate the number of records (rows) in a table
Statistics are created for each column of the table and can only be created if the data access is "Remote". Only
one data statistics object at a time is allowed per remote table.
Note
Creating statistics can cause a significant workload on the source system, especially if the source object is
a view, like for example a CDS View in ABAP or an SQL View in a Database. RECORD COUNT is the simplest
statistics object, and complexity increases for SIMPLE and HISTOGRAM, which causes heavier workload.
You can update data statistics objects: when you choose the menu Create Statistics on a remote table for which
data statistics are already available, the windows Create Statistics opens, and the existing type of statistics line
is marked as active. You can then choose to create another type of statistics or update the existing ones.
You can delete the data statistics objects for one or several tables with the menu Delete Statistics.
Monitor the status and progress of running and previously run task chains.
To view all task chains that are scheduled to run, are currently running, or have already been run in your space,
go to Data Integration Monitor Task Chains :
Properties Comments
Business Name or Technical Name Name of the task chain. Technical or Business Name is
displayed, depending on how you have configured your UI
settings in Your Profile Settings
Create .
• Scheduled: A schedule is defined for this task chain. If
you click Scheduled, you will get detailed information
on the schedule. You can click Schedule to create, edit,
delete, pause or resume a schedule for a selected task
chain or to assign a schedule to your name.
• Paused: A schedule is defined for this task chain
but it's currently paused. You can resume it if you
click Schedule Resume
Last Run Start Date and Time when the last run started.
Last Run End Date and Time when the last run ended.
You can personalize the columns you want to display clicking on Select Columns
Note
A task chain consists of multiple tasks that run one after another. As it takes some time for the task chain
monitor to collect and update all a task’s status information, the task runtime information displayed here
might not exactly match the task runtime information shown in a respective detail monitor, for example, in
the Views or Flows monitors.
Click Run Start Run to manually start a new run of a selected task chain. If a task chain’s last run failed,
you can also choose the Run Retry Latest Run option to retry the failed task chain’s execution from the
point at which it previously failed.
Note
If you add a remote table whose data access is Replicated (Real-time) in a task chain, the replication type
will change from real-time replication to batch replication at the next run of the task chain. The data will no
longer be updated in real-time.
Note
For optimal performance, it is recommended that you consider staggering the scheduled run time of tasks
such as data flows and task chains that may contain these tasks. There is a limit on how many tasks can be
started at the same time. If you come close to this limit, scheduled task runs may be delayed and, if you go
beyond the limit, some scheduled task runs might even be skipped.
Click in the right-most column to access the Task Chain Log Details page for the selected task chain. The
Details page provides a flexible three-panel display to access all the relevant information about the selected
task chain’s run history. It also provides details about individual child tasks or subtasks in the task chain, and
log messages about a selected task chain’s execution.
Note
The task chain Details page also lets you resize ( and ), expand/collapse ( ), or close ( ) the display of
each panel to focus on specific detail information or log messages. In addition, you can sort ( ) and filter
( ) information displayed in the different panels, as well as search log messages.
To view details about a specific task run, select it in the Runs list in the left-side panel. The Runs list provides a
summary of all the selected task chain’s previous runs as well as those scheduled in the future. It also indicates
the status of each task chain run, whether previous task chains completed successfully, failed, are currently
running, or are scheduled to be run.
Note
For failed task chain runs, you can click on Failed in the Status field to show the likely cause of the failure,
when that information is available.
The center Run Details panel provides a list of task chain objects and child tasks (if any are defined) for a
selected task chain and displays their individual type and execution status.
The Run Details panel also provides a Cancel Run link in the top right-hand corner of the display for a specific
task chain. If you click this link, SAP Datasphere initiates the Cancel operation and immediately remove all
child tasks not already triggered. For any child task that has already started, the effect of the Cancel operation
depends on the individual object. For example, replication flows and Open SQL procedures cannot effectively
In the right-side panel, the task chain monitor displays log messages for a selected task chain. When you click
on an object in the Run Details task list, detailed log messages for the selected task or subtask are displayed in
the right-most panel.
Note
From each monitor, you can navigate back to the Data Builder to change your objects by clicking Open in
Data Builder.
To return to the original Run Details display, click on the task chain name under the Parent Task Chain field in
the upper right corner of the page.
If a task chain’s last run failed, the Run Details panel provides a Retry Run option to retry the failed task
chain’s execution. You can also choose the Run Retry Latest Run option to retry the failed task chain's
execution. Note that the Retry Latest Run option is only active when the last taskchain runs failed.
Any user with authorization to run the task chain can retry a task chain’s execution. Retries of a task chain are
run on behalf of the original user and the retried task chain will have the same log ID as the original task chain
run. The monitor will create a new log ID for retries of failed subtasks and both the original and retried subtasks
appear in the hierarchy of tasks displayed in the Run Details panel.
Retries of a failed task chain will begin at the first failed child or subtask in the task chain, skipping any tasks
already completed successfully. The monitor updates the status of a retried task as the new run is triggered.
You can schedule or unschedule data integration tasks such as remote table replication, persisting views, and
data flow, replication flow, or task chain execution. You may also pause and then later resume execution of
scheduled tasks.
Note
For optimal performance, it is recommended that you consider staggering the scheduled run time of tasks
such as data flows and task chains that may contain these tasks. There is a limit on how many tasks can be
started at the same time. If you come close to this limit, scheduled task runs may be delayed and, if you go
beyond the limit, some scheduled task runs might even be skipped.
Using a dedicated dialog box, you can specify the frequency and time range of the schedule by using a simple
form or by directly entering a cron expression.
Note
When you click Create, the definition of the schedule that is created and saved is the one that is currently
displayed (either in the Simple Schedule area or the Cron Expression area).
Note
If you create a schedule for a remote table whose data access is Replicated (Real-time), the replication type
will change from real-time replication to batch replication at the next run of the schedule. The data will no
longer be updated in real-time.
You can schedule or unschedule data integration tasks such as remote data replication, data persistence, data
flow, replication flow, or task chains runs. You may also pause and then later resume the run of scheduled
tasks.
Prerequisites
• With the DW Integrator role you can create, change, or delete a schedule and let the scheduling
component of SAP Datasphere run tasks on your behalf. For data flows, you need the DW Integrator role
along with the DW Viewer or the DW Modeler role.
• For replication flows, scheduling is possible if all objects in the flow have load type Initial Only.
• To run recurring scheduled tasks on your behalf, you need to authorize the job scheduling component of
SAP Datasphere. In your profile settings under Authorized Consent Settings, you can give and revoke your
consent to SAP Datasphere to run your scheduled tasks in the future. Note that when you don't give your
consent or revoke your consent, tasks that you own won't be executed but will fail.
Your consent is valid for 365 days. If your consent will expire within the next four weeks, when you
attempt to schedule new tasks, SAP Datasphere displays a message warning that your consent is
approaching its expiration date. After the consent has expired, a log message informs you that the
tasks for which you own the schedule won’t be executed anymore. Renew your consent to resume
task execution according to the schedules. Additionally, in Data Integration Monitor, a warning message
appears four weeks before the expiry of consent.
Context
The job scheduling component of SAP Datasphere runs scheduled tasks on behalf of the owner of the
schedule. Per default, the user who creates a schedule is the owner of the schedule. You can takeover the
ownership, however, if required. Scheduled tasks run in the background according to the settings defined in the
schedule.
Procedure
You can also use schedules to automate actions on objects created from the Data Builder
Note
Tasks for which you own a schedule without having authorized SAP Datasphere to run scheduled tasks
on your behalf before won't be executed but will fail.
Frequency • Minutes: Define how often per Let's take a few examples. Imagine
hour that you want to trigger the that we are in April 14, 2021, it's
scheduled run during the desig- 9:00 am and you define your schedule
nated time or date range. You can without a start and end date:
specify scheduled runs every 10, • Minutes: You set frequency of
15, 20, or 30 minutes. scheduled task runs to 15 mi-
• Hourly: Define how often you nutes. Tasks will now be triggered
want to trigger the schedule per to run every 15 minutes within
day (every 1 to 23 hours of the designated time/date range.
day) and the start time. • Hourly: You select every "5" hours
Optionally, you can specify an off- "00" minutes. You need to con-
set from which the tasks will run. sider that a cycle run reset will
For example, so that a task runs occur every day at 0:00. Next
runs will be:
every 2 hours starting from 1:30
AM, specify 2 in Every and 1:30 in • Apr 14, 2021 10:00
• Daily: Define how often you want • Apr 14, 2021 20:00
to trigger the schedule per month • Apr 15, 2021 0:00
(every 1 to 28 days per month) • Apr 15, 2021 5:00
and the start time. If you define that the schedule
By default, the option Reset Every
must run at 0:15 mn, then the
Month is not selected, which
next runs will be
means that you schedule daily
tasks with a fixed frequency (for
• Apr 14, 2021 10:15
not reset on the first day of each • Apr 14, 2021 20:15
month. You can select the option • Apr 15, 2021 0:15
so that the run cycle is reset ev-
• Apr 15, 2021 5:15
ery month.
• Daily: You select every "4" days at
• Weekly: Select which day(s) of "0:00" hours.
the week the schedule must be
Next runs will be:
triggered (multiple selection is al-
lowed) and the start time. • April 17, 2021
Time Zone • UTC: This is the default schedule UTC is the default setting. You can
setting. also choose a regional time zone
• Region/Location: You can choose in the format of Region/Location,
a regional time zone. This will for example, Europe/Berlin, Amer-
automatically adjust for daylight ica/New_York, or Asia/Kolkata.
saving time.
Note
Time Range • Start Date: Specify a date if you Now imagine we are still in April 14,
want to the change the default 2021, but you set as start date "May
start date, which is the current 2, 2020" and as end date "March 31,
date. 2021". If we take the same examples
• End Date: Optionally you can as above the next runs will be:
specify an end date. Otherwise, • Daily:
the schedule runs indefinitely. • May 6, 2021
• May 11, 2021
• May 16, 2021
• May 21, 2021
• May 26, 2021
• Monthly:
• Jun 1, 2021 0:00
• Nov 1, 2021 0:00
• Jan 1, 2022 0:00 (Cycle run
reset)
Preview Show Runs In: Preview your next five runs in ei-
• UTC ther UTC time, local time, or the
Region/Location time selected. The
• Local Time
preview will take into consideration
• Region/Location selected in the changes in daylight saving time.
Time Zone drop down.
Let's look at an example. You schedule
a task to run in the time zone Amer-
ica/Los_Angeles every week on Mon-
day at 20:00. Your start date is March
5, 2025. You want to preview your next
runs.
The column Frequency shows the status Scheduled. By clicking the status you can display the scheduling
settings.
The column Schedule Owners displays the name of the current schedule owner.
The Next Run column shows the start date and time of the next run according to the scheduling settings.
As long as the consent from the owner of the schedule hasn't expired the task will run in the background.
You can access the log by selecting the relevant object and clicking (Remote Table Logs), (Views Logs), or
(Details).
Next Steps
Once a schedule is defined, you can adjust the scheduling settings at any time selecting Schedule Edit
Schedule . The next run of the task will use the adjusted scheduling settings.
The schedule can be removed at any time via the menu Schedule Delete Schedule with the result that no
new run will start anymore for the task. The Next Run column gets cleared.
You may also pause and then later resume execution of scheduled tasks via the menu Schedule Pause
Schedule option. See Pause or Resume a Scheduled Task [page 266] for more information.
Related Information
Schedule data integration tasks such as data replication, data persistence, or data flow runs or task chains
runs, by entering directly a cron expression.
Prerequisites
• With the DW Integrator role you can create, change, or delete a schedule and let the scheduling
component of SAP Datasphere run tasks on your behalf. For data flows, you need the DW Integrator role
along with the DW Viewer or the DW Modeler role.
Note
Your consent is valid for 365 days. After the consent has expired, a log informs you that the tasks
for which you own the schedule won’t be executed anymore. Renew your consent to resume task
execution according to the schedules. Additionally, in To run recurring scheduled tasks on your behalf,
you need to authorize the job scheduling component of Data Integration Monitor, a warning message
appears 10 days before the expiry of consent.
Context
The job scheduling component of SAP Datasphere runs scheduled tasks on behalf of the owner of the
schedule. By default, the user who creates a schedule is the owner of the schedule. You can takeover the
ownership, however, if required. Scheduled tasks run in the background according to the settings defined in the
schedule.
Procedure
You can also use schedules to automate actions on objects created from the Data Builder
Note
Tasks for which you own a schedule without having authorized SAP Datasphere to run scheduled tasks
on your behalf before won't be executed but will fail.
7. In the Frequency area, select Cron Expression from the Enter As drop down list.
8. In the Frequency area, select your preferred Region//Location from the Time Zone drop down list. Cron
expression will reflect the time in the Region/Location you have chosen.
9. Define when the task will be run by using the unix-cron string format (* * * * *) in the five following fields:
Minute, Hour, Day (Month), Month and Day (Week).
Syntax Description
x,y Where x and y are two or more values of a list separated by a comma.
Example: 4,5,6 in the Month field is equivalent to "April, May and June".
x-y Where x and y are the first and last values of a range separated by a dash.
Example: 1-5 in the Day (Week) field is equivalent to "from Monday to Friday".
*/x Where * is all possible values and x is the step value, separated by a slash.
Example: */3 in the Hour field is equivalent to 0,3,6,9,12,15,18,21. The asterisk (*) specifies
"every hour" but the /3 means "every 3 hours".
a-b/x Where a and b are the first and last values of a range and where x is the step value.
Example: 1-15/2 in the Day (Month) field is equivalent to 1,3,5,7,9,11,13,15. 1-15 specifies "from
day 1 to day 15" of the month but /2 means "every 2 days".
Expression Description
0 0 1-3 * * The task runs the first 3 days of each month at 00:00
(UTC).
0 18 1,15,25 * * The task runs every 1st, 15th and 25th of each month at
6:00 PM (UTC).
0 8 */1 * 1-5 The task runs from Monday to Friday at 8:00 AM (UTC).
0 2-8/2 */1 * 1-5 The task runs every 2 hours between 2:00 AM (UTC) and
8:00 AM (UTC), from Monday to Friday.
Note
You can copy to the clipboard the cron expression you've entered by clicking the copy icon button next
to the cron expression.
The Duration area is available only for elastic compute notes. See Run an Elastic Compute Node.
10. In the Start Date area, specify a date if you want to the change the default start date, which is the current
date.
11. Optionally, in the End Date area, specify a date. If you do not specify an end date, the schedule runs
indefinitely.
12. In the right area of the dialog box, you can see a preview of the schedule and when the five next runs are
planned. Schedules are viewed in Coordinated Universal Time (UTC) by default. You can also view the next
runs in local time or the Region/Location time that you have set as your preferrence.
13. Click Create to create the schedule.
Results
The column Frequency shows the status Scheduled. By clicking the status you can display the scheduling
settings.
The column Schedule Owners displays the name of the current schedule owner.
The Next Run colums shows the start date and time of the next run according to the scheduling settings.
As long as the consent from the owner of the schedule hasn't expired the task will run in the background. You
can access the log by selecting the relevant object and clicking .
Next Steps
Once a schedule is defined, you can adjust the scheduling settings at any time selecting Schedule Edit
Schedule . The next run of the task will use the adjusted scheduling settings.
The schedule can be removed at any time via the menu Schedule Delete Schedule with the result that no
new run will start anymore for the task. The Next Run colum gets cleared.
You may also pause and then later resume execution of scheduled tasks via the menu Schedule Pause
Schedule option. See Pause or Resume a Scheduled Task [page 266] for more information.
Per default, the user who creates a task schedule owns the schedule which means that the job scheduling
component runs the task on the owner's behalf according to the defined schedule. You can assign the
ownership of the schedule to yourself.
Prerequisites
• With the DW Integrator role you can change the schedule and let the scheduling component run tasks
on your behalf. For data flows, you need the DW Integrator role along with the DW Viewer or the DW
Modeler role.
• To let the job scheduling component of SAP Datasphere run scheduled tasks on your behalf, you need to
give your consent. For more information, see Changing SAP Datasphere Settings.
Context
Changing the owner of the schedule is required when the user owning the schedule until now left the
organization and the user has been deleted, for example. In such a case the schedule isn't assigned to any
user and can't be executed successfully.
Procedure
Results
The Owner field now shows your user ID as the new owner of the schedule. The scheduling component from
now own will execute the task on your behalf.
Using options available for creating, editing, and removing task scheduling, you can pause or resumed
previously scheduled data integration tasks such as remote table replication, persisting views, data flow or
task chain execution.
After previously scheduling a task, you can select Schedule Pause Schedule to pause the next schedule
run of a task, or select Schedule Resume Schedule to resume a schedule previously paused.
If you pause a schedule, when you return to the task display, the Frequency column displays Paused for the
selected task, and the Next Run column is cleared of any future scheduled runs. The defined schedule for
future scheduled runs is still saved, however, so you can simply go back later and resume the schdule to run
future scheduled task again.
If you pause one or more objects within a replication flow that has a schedule, the schedule continues
unchanged for all other objects in the flow, wheras no replication takes place for the paused objects until
you resume them.
In the Configuration area, you can check how much spaces the task logs are using on your tenant, and decide to
delete the obsolete ones to reduce storage consumption.
Context
Each time an activity is running in SAP Datasphere (for example, replicate a remote table), task logs are
created to allow you to check if the activity is running smoothly or if there is an issue to solve. You access
these detailed task logs by navigating to the Data Integration Monitor - Details screen of the relevant object. For
example, clicking the button of the relevant remote table.
However, task logs can consume a lot of spaces in a tenant. Deleting old task logs that are no longer needed
can be useful to release storage space. This is why SAP Datasphere has a log deletion schedule activated by
default. You can change the schedule defining your own criteria or decide to take immediate deletion actions.
Procedure
• Manually Delete Task Log: SAP Datasphere automatically triggers logs deletion using the following
default criteria:
Users with a space administrator role can create data access controls to allow modelers to apply row-level
security to Data Builder and Business Builder objects. Once a data access control is applied to an object, any
user viewing its data either directly or via an object using it as a source, will see only those records they are
authorized to view, based on the specified criteria.
Permissions Entities
Your criteria are defined in a table or view that lists SAP Datasphere user IDs (in the form required by your
identity provider) and assigns them to one or more criteria. You can define the following types of criteria:
• Single Values - Each user can only see the records that match any of the single values she is authorized for
in the permissions entity. See Create a "Single Values" Data Access Control [page 271].
• Operator and Values - Each user can only see the records that fulfill the operator-value pairs she is
authorized for in the permissions entity, including support for complex AND and OR combinations. See
Create an "Operator and Values" Data Access Control [page 274].
• Hierarchy - Each user can only see the records that match the hierarchy values she is authorized for in
the permissions entity, along with any of their descendents. See Create a "Hierarchy" Data Access Control
[page 277]
You can create one or more data access controls that consume each permissions entity, and select one or
more columns in each data access control to specify the criteria that it will enforce. You can apply a single
data access control to multiple views. Each view protected in this way will filter the results available in its data
preview to only those rows meeting the criteria for the current user.
Note
Permissions entities must not, themselves, be protected by a data access control, or have any protected
view among their sources.
We recommend that you develop clear policies for securing data, and that you:
Note
The row-level security provided by the data access control can be circumvented while the view remains in
its space. It is enforced only when the view is:
• Applying a data access control to a data layer view, see Apply a Data Access Control to a Graphical or SQL
View.
• Using a data access control to create authorization scenarios in the business layer, see Authorization
Scenario.
• Persisting data in a view that has a data access control applied to it, see Persisted Views and Data Access
Control [page 205].
Note
If you experience performance issues with a view protected by a data access control, we recommend
enabling replication for source tables, particularly if any source contains more than 500,000 rows.
This diagram shows a typical environment where a permissions entity is maintained by business users in one
space and shared to a second space, where technical users create data access controls from it and apply them
to views. These views are then shared to other spaces where they can be securely accessed by modelers and
analysts:
In this environment:
• The Permissions space users are a select group of business users who:
• Maintain the Permissions table, assigning users (which, in this case, are identified by their email
address) to the appropriate country, department, and any other relevant criteria:
anne.archer@acme.com US Sales
jennifer.jones@acme.com FR Sales
• Share the Permissions table to the IT space to use as a data source for data access controls.
Tables and views shared to a space cannot be directly used as the permissions entity for a data
access control. Modelers in the space receiving the shared object must encapsulate it in a view,
which can then serve as the permissions entity for one or more data access controls.
Note
The row-level security provided by the Country data access control can still be circumvented while
the view remains in the IT space.
• Share the Sales view to the Sales space where it can be used securely.
• The Sales space users use the protected Sales view to do analytics on sales data:
• Anne is a data modeler in the US sales department. She connects to SAP Datasphere and combines
the Sales view with other data. Whenever she previews data, she can only see US data.
• Jennifer is a business analyst in the French sales department. She connects to SAP Analytics Cloud,
builds a story on the Sales view, and can only see French sales data.
Users with the DW Space Administrator role (or equivalent privileges) can create data access controls in which
criteria are defined as single values. Each user can only see the records that match any of the single values she
is authorized for in the permissions entity.
Context
Before creating your data access control, you must have prepared a permissions entity with the following
columns:
• User ID column - Containing user ids in the format required by your identity provider (email addresses,
logon names, or other identifiers). If you are using SAML authentication, this column must contain values
in the form defined as your User Attribute / IdpUserID (see Enabling a Custom SAML Identity Provider).
This column must be selected as the Identifier Column in your data access control.
If a user has no entries in the permissions entity, then they will not have access to any records in the
protected view.
• One or more columns containing filter criteria (country or region names or ids, department names or ids,
or any other criteria to control how your data is exposed to users). These columns must be selected as
Criteria columns in your data access control.
Procedure
1. In the side navigation area, click (Data Builder), select a space if necessary, and click New Data Access
Control to open the editor.
2. Complete the following properties:
Property Description
Business Name Enter a descriptive name to help users identify the object. This name can be changed at any
time.
Technical Name Displays the name used in scripts and code, synchronized by default with the Business Name.
To override the default technical name, enter a new one in the field. Technical names can
contain only alphanumeric characters and underscores.
Note
Once the object is saved, the technical name can no longer be modified.
Packages are used to group related objects in order to facilitate their transport between
tenants.
Note
Once a package is selected, it cannot be changed here. Only a user with the DW Space
Administrator role (or equivalent privileges) can modify a package assignment in the
Packages editor.
Owner Enter the name of the person responsible for your data access control.
Permissions Entity Select the table or view containing your user ids and criteria.
The permissions entity must match the structure selected in the Structure field. Click the Open
in New Tab button to the right of the field to open the entity in its own editor.
Identifier Column Select a column containing user ids in the format required by your identity provider (email
addresses, logon names, or other identifiers).
These columns will be listed in the Criteria section. When you assign a data access control to a view,
these columns must be mapped to columns in the view to enforce filtering by these criteria.
d. Click OK to confirm the selection.
4. Click Save and then Deploy to deploy your data access control and make it available for use.
For information about attaching a data access control to a view, see Apply a Data Access Control to a
Graphical or SQL View.
Note
You can use the View as User tool in the Data Viewer panel to review the effects of the data access
controls you apply by checking the records that another user will be allowed to see (see Viewing Object
Data).
5. The tools in the editor toolbar help you work with your object throughout its lifecycle:
Tool Description
(Save) Save your changes to the design-time repository. You can use Save As to create a copy of the
object.
(Deploy) Deploy your changes to make them available in the run-time environment.
(Impact and Open the Impact and Lineage Analysis graph for the object.
Lineage Analysis)
See Impact and Lineage Analysis.
Users with the DW Space Administrator role (or equivalent privileges) can create data access controls in
which criteria are defined as operator and value pairs. Each user can only see the records that fulfill the
operator-value pairs she is authorized for in the permissions entity, including support for complex AND and OR
combinations.
Context
Before creating your data access control, you must have prepared a permissions entity with the following
columns:
Note
If a user has no entries in the permissions entity, then they will not have access to any records in the
protected view.
• Restriction Column - Each restriction per user must have a unique name, and one or more criteria can
belong to each restriction. All criteria belonging to a restriction act together as an AND condition. If you
have more than one restriction per user then all restrictions act together as an OR condition.
• Criterion Column - The criteria name must contain only alphanumeric characters and underscores and
appears in the Mappings area when a user applies the data access control to their view. It does not need to
exactly match a column name, but should guide the user to the column to which it should be mapped.
• Operator Column - The following operators are supported:
• ALL (or *) - Provides access to all records. No values required.
• EQ (or =) - Equal to First Value.
• NE (or <> or !=) - Not equal to First Value.
• GT (or >) - Greater than First Value.
• GE (or >=) - Greater than or equal to First Value.
• LT (or <) - Less than First Value.
• LE (or <=) - Less than or equal to First Value.
• CP (or LIKE) - Contains pattern First Value. Optionally Second Value can be set to an escape character.
See LIKE Predicate in the SAP HANA Platform documentation.
• BT (or BETWEEN) - Between First Value and Second Value.
• First Value Column - This value acts as the first argument or lower bound of the operator.
• Second Value Column - This value acts as the second argument or upper bound of the operator.
For example:
1 bob@acme.c 0 Code BT CA CZ
om
2 bob@acme.c 0 Type EQ 1
om
Note
A criterion with an ALL operator can have any name and can be mapped to any column in a view.
Procedure
1. In the side navigation area, click (Data Builder), select a space if necessary, and click New Data Access
Control to open the editor.
2. Complete the properties in the General section:
Property Description
Business Name Enter a descriptive name to help users identify the object. This name can be changed at any
time.
Technical Name Displays the name used in scripts and code, synchronized by default with the Business Name.
To override the default technical name, enter a new one in the field. Technical names can
contain only alphanumeric characters and underscores.
Note
Once the object is saved, the technical name can no longer be modified.
Packages are used to group related objects in order to facilitate their transport between
tenants.
Note
Once a package is selected, it cannot be changed here. Only a user with the DW Space
Administrator role (or equivalent privileges) can modify a package assignment in the
Packages editor.
Permissions Entity Select the table or view containing your user ids and criteria.
The permissions entity must match the structure selected in the Structure field. Click the Open
in New Tab button to the right of the field to open the entity in its own editor.
Identifier Column Select a column containing user ids in the format required by your identity provider (email
addresses, logon names, or other identifiers).
Property Description
Restriction Column Select the permissions entity column that contains the restriction names.
Each restriction per user must have a unique name, and one or more criteria can belong to
each restriction. All criteria belonging to a restriction act together as an AND condition. If you
have more than one restriction per user then all restrictions act together as an OR condition.
Criterion Column Select the permissions entity column that contains the criteria names.
The criteria name must contain only alphanumeric characters and underscores and appears in
the Mappings area when a user applies the data access control to their view. It does not need
to exactly match a column name, but should guide the user to the column to which it should be
mapped.
Operator Column Select the permissions entity column that contains operators.
First Value Column Select the permissions entity column that contains the lower bound values to apply to the
operators.
This value acts as the first argument or lower bound of the operator.
Second Value Column Select the permissions entity column that contains the upper bound values to apply to the
operators.
This value acts as the second argument or upper bound of the operator.
4. Use the Available Criteria table to control which criteria should be applied.
All available criteria are listed in the table and are all selected to be applied by default. To remove a
criterion, deselect it in the list.
5. Click Save and then Deploy to deploy your data access control and make it available for use.
Note
You can use the View as User tool in the Data Viewer panel to review the effects of the data access
controls you apply by checking the records that another user will be allowed to see (see Viewing Object
Data).
6. The tools in the editor toolbar help you work with your object throughout its lifecycle:
Tool Description
(Save) Save your changes to the design-time repository. You can use Save As to create a copy of the
object.
(Deploy) Deploy your changes to make them available in the run-time environment.
(Impact and Open the Impact and Lineage Analysis graph for the object.
Lineage Analysis)
See Impact and Lineage Analysis.
Users with the DW Space Administrator role (or equivalent privileges) can create data access controls in which
criteria are defined as hierarchy values. Each user can only see the records that match the hierarchy values she
is authorized for in the permissions entity, along with any of their descendents.
Context
Before creating your data access control, you must have identified the following entities:
Note
If a user has no entries in the permissions entity, then they will not have access to any records in
the protected view.
• Criterion column - A column containing the criterion data. This criterion data is organized into a
hierarchy via an external entity with a semantic usage of Hierarchy.
• An entity with a semantic usage of Hierarchy containing parent-child relationships for the records in the
permissions entity.
Note
Only external hierarchies with a single pair of parent-child columns are supported. Level-based
hierarchies, dimensions with internal hierarchies, and entities with a semantic usage of Hierarchy with
Directory cannot be used, and the data in your hierarchy must respect the following rules:
• A single root node with a parent value of null
• No nodes with multiple parents
• No circular relationships
For example:
• The Geo Permissions table is used as a permissions entity, and contains the following records:
bob@acme.com Europe
jim@acme.com France
ann@acme.com Paris
• The Geo Hierarchy table has a semantic usage of Hierarchy and contains the following records:
World
Europe World
France Europe
Paris France
• The Geo Hierarchy data access control uses the Geo Permissions table as a permissions entity and,
when it is applied to a view, ensures that the following users will only have access to records with the
appropriate geo location values:
• bob - Europe, France, or Paris.
• jim - France or Paris.
• ann - Paris.
1. In the side navigation area, click (Data Builder), select a space if necessary, and click New Data Access
Control to open the editor.
2. Complete the properties in the General Section:
Property Description
Business Name Enter a descriptive name to help users identify the object. This name can be changed at any
time.
Technical Name Displays the name used in scripts and code, synchronized by default with the Business Name.
To override the default technical name, enter a new one in the field. Technical names can
contain only alphanumeric characters and underscores.
Note
Once the object is saved, the technical name can no longer be modified.
Packages are used to group related objects in order to facilitate their transport between
tenants.
Note
Once a package is selected, it cannot be changed here. Only a user with the DW Space
Administrator role (or equivalent privileges) can modify a package assignment in the
Packages editor.
Permissions Entity Select the table or view containing your user ids and criteria.
The permissions entity must match the structure selected in the Structure field. Click the Open
in New Tab button to the right of the field to open the entity in its own editor.
Identifier Column Select a column containing user ids in the format required by your identity provider (email
addresses, logon names, or other identifiers).
Property Description
Business Name Enter a descriptive name to help users identify the object. This name can be changed at any
time.
Technical Name Displays the name used in scripts and code, synchronized by default with the Business Name.
To override the default technical name, enter a new one in the field. Technical names can
contain only alphanumeric characters and underscores.
Note
Once the object is saved, the technical name can no longer be modified.
Criterion Column Select the column containing the authorized nodes from the permissions entity.
Hierarchy Entity Select the entity with a semantic usage of Hierarchy that defines the parent-child hierarchy
relationships between the records in the permissions entity.
4. Click Save and then Deploy to deploy your data access control and make it available for use.
For information about attaching a data access control to a view, see Apply a Data Access Control to a
Graphical or SQL View.
Note
You can use the View as User tool in the Data Viewer panel to review the effects of the data access
controls you apply by checking the records that another user will be allowed to see (see Viewing Object
Data).
5. The tools in the editor toolbar help you work with your object throughout its lifecycle:
Tool Description
(Save) Save your changes to the design-time repository. You can use Save As to create a copy of the
object.
(Deploy) Deploy your changes to make them available in the run-time environment.
(Impact and Open the Impact and Lineage Analysis graph for the object.
Lineage Analysis)
See Impact and Lineage Analysis.
You can import analysis authorizations defined in SAP BW and SAP BW∕4HANA systems into SAP Datasphere
to provide row-level protection for data imported from these systems.
Introduction
This feature is supported only for SAP BW v7.5 SP16 and higher and SAP BW∕4HANA v2.0 and higher systems
accessed via one of the following connection types:
An SAP BW or SAP BW∕4HANA administrator must have created the report that generates the
RSDWC_RSEC_DAC permissions table on the source system (see SAP Note 3062381 ). This SAP note also
includes instruction on how to implement BAdI RSDWC_DAC_RSEC_USER_UPDATE for situations where your
SAP BW or SAP BW∕4HANA instance does not use email addresses as user IDs.
Note
This SAP note has been updated to support importing permissions via an SAP BW∕4HANA Model Transfer
connection. If you have previously used this feature, you must follow the updated steps to ensure you have
the latest version of the report.
The objects to be protected must have already been imported into your space. For objects imported from:
• SAP ABAP or SAP BW connections - Only analysis authorizations applied to characteristics used on
InfoProviders and CompositeProviders can be imported with this wizard.
Note
• SAP BW∕4HANA Model Transfer connections - Analysis authorizations applied to navigation attributes can
also be imported.
See the following blogs for additional information about this feature:
• For a general introduction, see Introducing Remote Authorizations from SAP BW/4HANA for SAP
Datasphere (updated February 2022).
• For information about leveraging SAP HANA-native analytical privileges in an open SQL schema in case of
slow-running views or out of memory issues, see HANA Analytical Privileges in SAP Datasphere – Part 1
and Part 2 (published January 2023).
The Import Permissions wizard will import the permissions table from the source system, create additional
objects to calculate the permissions, and apply a data access control to each selected object.
1. In the side navigation area, click (Data Builder), select a space if necessary, and click Import Import
Permissions to open the Import Permissions wizard.
2. Select the connection from which you want to import permissions and click Next.
3. Import or accept the permissions table for this connection. If the permissions table is:
• Not imported - Enter a Business Name and Technical Name and click Import and Deploy. Once the
table has been imported and deployed, click Next.
Note
You cannot choose the name of the permissions table for a SAP BW∕4HANA Model Transfer
connection.
Note
For SAP BW∕4HANA Model Transfer connections, if SAP Note 3062381 has been correctly
applied and the wizard cannot find the permissions table, verify that the user specified in
your connection has permission to access the permissions table RSDWC_RSEC_DAC. Your SAP
BW∕4HANA administrator may need to add this permission to the user or provide an alternative
user with the necessary permission.
Note
We recommend that the report generating the permissions table in SAP BW∕4HANA is run at least
once a day and the remote table in SAP Datasphere is kept in remote (federated) access to ensure
that it is always up-to-date. If you decide to replicate the permissions table, you should schedule at
minimum a daily refresh. Real-time replication is not supported for this table.
4. Select the objects for which you want to import permissions, and then click Next.
Only those objects that have previously been imported from this connection, and which can be protected,
are listed.
5. Review the summary of the objects that will be created in SAP Datasphere.
For each object to be protected, new objects are created:
Note
If you have previously protected some or all of the selected objects, they will appear in the wizard with a
Status/Action of Overwrite and will be overwritten.
6. Click Import and Deploy to create, import, and deploy all these objects.
7. Click Finish to exit the wizard and perform all the actions necessary to protect the data.
All the objects are listed in the Repository Explorer and, where appropriate, also in the Data Access Controls
and Data Builder start pages.
If the permissions entity that is consumed by your data access control is modified, then the next time you open
the data access control, you will be asked to process the changes. If the source change has generated warnings
or errors in your data access control, its status will be updated and you will receive a notification inviting you to
review the changes.
Context
Your object's status may be updated when changes to one or more of its sources are saved or deployed:
• If no warnings or errors were generated by the source changes, then the status will not be changed.
• If warnings or errors were generated by the source changes and:
• Source changes were saved but not deployed, your object's status will be set to Design-Time Error.
• Source changes were saved and deployed, your object's status will be set to Run-Time Error.
Procedure
1. If a source change has generated warnings or errors in your view, you will receive a notification inviting you
to review them, and you can click the notification to open it directly.
The Source Updates dialog opens, listing the sources that have been modified.
2. Click OK to dismiss the dialog. The following validation messages may be displayed:
Add Column Adding columns to the permissions entity has no impact on the data access control.
Change Business Name Modifying the business name of the permissions entity or any of its columns has no
impact on the data access control
Change Column Data Type Source column data type changes are not allowed for columns that are used by a data
access control.
Delete Column Deleting columns from the permissions entity generates errors in dependent data ac-
cess controls if the columns are used:
• If the column was selected as the Identifier Column, an error message is displayed.
You must specify a new column in this field.
• If the column was selected as a column in the Criteria list, an error message is
displayed. You must remove this column from the Criteria list.
Note
You cannot add new criteria columns to a data access control that is used by
views. If this change is necessary, we recommend that you create a new data
access control with the additional columns and use it to replace the existing
data access control in the views to which it is attached.
3. Once you have processed the modifications to the Identifier Column or Criteria list, and if your data access
control is applied to any views, warning messages are displayed urging you to review the impacted views to
ensure that the changes have not compromised row level security.
4. Click (Save) to save the changes to your object and dismiss any info and warning messages.
If the source changes were saved but not deployed, your object's status should change from Design-Time
Error to Changes to Deploy.
Note
As the new version of your object depends on the source updates, you should coordinate with the
person who has modified the source about deployment. The new version of your object will work
correctly in the run-time only when both it and its source are deployed.
Your object's status should change from Changes to Deploy (or Run-Time Error) to Deployed.
Users with an administrator role or space administrator role, can use the Semantic Onboarding app to import
business content and sample content from SAP and partners published to the Content Network.
Context
To import business content and sample content from SAP and partners via the Semantic Onboarding app, you
must have either:
A global role that allows you to update any space and to A combination of a global role and a scoped role:
create new spaces if necessary, by granting you the following
• A global role that grants you the following privilege:
privileges:
• Data Warehouse General (-R------) - To access
• Data Warehouse General (-R------) - To access SAP SAP Datasphere.
Datasphere.
• Lifecycle (-----M--) - To use the Content
• Lifecycle (-----M--) - To use the Content Network Network tiles in the Semantic Onboarding app.
tiles in the Semantic Onboarding app.
• A scoped role that grants you access to the space or
• Spaces (-------M) - To update all spaces and space spaces to import into with the following privileges:
properties.
• Data Warehouse General (-R------) - To access
• Spaces (C-------) - To create spaces. SAP Datasphere.
• Spaces Files (-------M) - To create, read, update, and • Spaces (--U-----) - To update your spaces and
delete all objects in all spaces. their properties.
The DW Administrator role, for example, grants these privi- • Spaces Files (CRUD----) - To create, read, up-
leges. date, and delete objects in your spaces.
• Data Warehouse Data Builder (CRUD--S-) - To
create, read, update, delete and share data builder
objects (and any other relevant object privileges
to allow you to create and update other types of
objects contained in the package).
For more information, see Privileges and Permissions and Standard Roles Delivered with SAP Datasphere.
Note
Use of sample content and business content is optional and not part of the business functionality of the
product. Content packages may be discontinued or replaced at any time.
Detailed documentation for all business content is available at SAP Datasphere Business Content.
Procedure
1. In the side navigation area, click (Semantic Onboarding), and then click one of the following tiles:
• Business Content: End-to-end business scenarios created by SAP for various industries and lines of
business.
• 3rd Party Content: End-to-end business scenarios created by SAP partners. Some third-party content
has to be purchased in the SAP Store.
• Samples: Not used in SAP Datasphere.
2. Click a package to open it and see an overview of its content.
3. Review the How to install this package section to understand the requirements for installing the content.
You may be instructed to:
• Create a space with a specific business and technical name before importing the content.
• Create a connection in the new space.
• Visit the SAP Store and, in some cases, make a payment to a partner.
• Contact a partner to complete the installation or obtain data.
4. Review the Import Overview section. For more details, and to modify any of the options, click the Import
Options tab:
Section Description
Overwrite Preferences Controls whether objects that are already present in your space can be overwritten
during an import.
Note
Only object definitions can be transported. Data cannot be transported between
SAP Datasphere tenants via the (Transport) app.
Note
• The Including permissions option is not supported for SAP Datasphere.
• Connections are never overwritten by an import.
• When overwriting objects with a release state, the objects you are import-
ing must have the same release state or another release state that is con-
sistent with their lifecycle (see Releasing Stable Views for Consumption).
Deployment Preference Check the Deploy after import checkbox to automatically deploy objects after import. If
you have selected:
• Don't overwrite objects or data - Only newly imported objects will be deployed.
• Overwrite objects and data - All imported objects in the space will be deployed.
Note
This option does not support packages that contain objects from one or more
spaces where objects are shared from one space to another (see Sharing Entities
and Task Chains to Other Spaces. In this case, you should manually deploy the
objects from the source space and then deploy the objects from the target space.
Content Select the space or spaces to import your content into in the Space Mapping field. You
must choose one space for each space listed in the package:
• You can select any space for which you have the Spaces.Update permission to
import the space's content into.
If you have the Spaces.Manage permission, then you can choose to import into any
space in your tenant.
• If the space does not exist in your tenant and you have the Spaces.Create permis-
sion, then you are invited to create it by default.
• You cannot "merge" content from two or more spaces included in your package. You
must select a separate space for each space listed.
Note
Choosing space mappings is only possible if the source objects for any shared
objects are included in the package. For example if your package contains shared
objects in Space A that have been shared from Space B, then the source objects
from Space B must be included in the package. If the sources of any shared
objects cannot be resolved in the package, then the package must be imported into
spaces with the same technical names as the spaces listed in the package.
You can select and deselect individual objects, or types of objects. The Impacted Object
column shows warnings for objects that will overwrite content on your system.
You will receive a notification when the process is complete, or if an error occurs.
If you selected Don’t overwrite objects or data, a notification will let you know if any of the content wasn’t
imported to avoid overwriting existing content.
To view the log of updates and imports for a package, select the package in the content network and
choose History.
Users with an administrator or space administrator role can use the Transport app to transfer content between
tenants via a private cloud storage area.
The following object types can be exported and imported via the (Transport) app:
Note
Only object definitions can be transported. Data cannot be transported between SAP Datasphere tenants
via the (Transport) app.
Connections The definition of a connection does not have dependencies on any other objects.
Remote Tables The definition of a remote table contains information about its connection.
Local Tables The definition of a local table contains the structure of the table only, and does not have depend-
encies on any other objects.
See Creating a Local
Table.
Flows The definition of a data flow, replication flow, or transformation flow contains the definitions of all
its sources and its target table. When you export a flow, these objects are exported too.
See Creating a Data
Flow, Creating a Rep-
lication Flow, and Cre-
ating a Transformation
Flow.
Views The definition of a view contains the definitions of all its sources and any used data access
controls. When you export a view, these objects are exported too.
See Creating a Graphi-
cal View and Creating
an SQL View.
Intelligent Lookups The definition of an intelligent lookup contains the definitions of its input and lookup entities.
When you export an intelligent lookup, these entities are exported too.
See Creating an Intelli-
gent Lookup.
Analytic Models The definition of an analytic model contains the definitions of its fact and dimension sources.
When you export an analytic model, these entities are exported too.
See Creating an Ana-
lytic Model.
E/R Models The definition of an E/R model does not include the objects that it visualizes as dependencies.
These objects must be selected manually.
See Creating an Entity-
Relationship Model.
Data Access Controls The definition of a data access control contains the definition of its permissions entity. When you
export a data access control, the permissions entity is exported too.
See Securing Data with
Data Access Controls
[page 268].
Task Chains The definition of a task chain does not include the objects that it automates. These objects must
be selected manually.
See Creating a Task
Chain.
Business Entities / The definition of a business entity contains all its versions (which are also listed separately), along
Business Entity Ver- with the definition of its source data entity and any authorization scenarios. When you export a
sions business entity (or one of its versions), these objects are exported too.
Fact Models The definition of a fact model contains all its versions (which are also listed separately), along with
the definition of all its source fact models and business entities. When you export a fact model (or
See Creating a Fact
one of its versions), these objects are exported too.
Model.
Consumption Models The definition of a consumption model contains all its perspectives (which are also listed sepa-
rately), along with the definition of all its source fact models and business entities. When you
See Creating a Con-
export a consumption model (or one of its perspectives), these objects are exported too.
sumption Model.
Authorization Scenar- The definition of an authorization scenario contains the definition of its data access control. When
ios you export an authorization scenario, the data access control is exported too.
Note
Only local and remote tables, views, data flows, and E/R models can be selected for exporting in SAP
Datasphere tenants provisioned prior to version 2021.03. While objects created in SAP Datasphere tenants
provisioned prior to version 2021.03 can be transported to tenants provisioned after that date, transport of
objects in the other direction is not possible.
Note
You can also export content from and import content to your space via:
• Export to CSN/JSON File buttons in selected Data Builder editors (see Importing and Exporting Objects
in CSN/JSON Files).
• The datasphere command line interface objects commands (see Manage Modeling Objects via the
Command Line).
Users with the DW Space Administrator role can create packages to model groups of related objects for
transport between tenants. Modelers can add objects to packages via the Package field, which appears in
editors when a package is created in their space. Once a package is complete and validated, the space
administrator can export it to the Content Network. The structure of your package is preserved and, as the
objects it contains evolve, you can easily export updated versions of it.
Introduction to Packages
Create a Package
1. In the side navigation area, click (Transport) (Packages) to open the list of packages.
Use the Space filter field to show packages from all spaces or only one space.
2. Click Add and specify a space, if necessary, to create a new package and open it in the editor.
3. Enter the following properties on the General tab:
Business Name Enter a descriptive name to help users identify the object. This name can be changed at
any time.
Technical Name Displays the name used in scripts and code, synchronized by default with the Business
Name.
To override the default technical name, enter a new one in the field. Technical names can
contain only alphanumeric characters and underscores.
Note
Once the object is saved, the technical name can no longer be modified.
Business Purpose Provide a description, purpose, contacts, and tags to help other users understand the
package.
Version Enter the version number of the package in the format 1.0.0 (major version, minor
version, patch number).
You can re-export your package with the same version number to overwrite the current
exported version or change the version to export a package update. The version number
must always go up and never down.
Current Exported Version Displays the last exported version of the package.
4. [optional] Add any packages that contain objects that the objects you add to your package depend on as
part of their lineage.
1. Click the Required Packages tab.
2. Click Add to open the Select Packages dialog, which lists all packages you have permission to see.
The list includes packages created in your space and packages shared to your space.
3. Select one or more packages and click Select to add them as required packages.
5. Add objects from your space to the package.
1. Click the Objects tab.
2. Click Add to open the Add Objects dialog, which lists all objects in your space.
Note
Though modelers can add objects to a package in the object editors, they cannot remove objects from
a package.
Export a Package
Update a Package
Once your package is exported, you can continue to evolve it by adding, removing, or updating objects.
Modelers in your space may also deploy changes to the objects it contains and add new objects to it.
When any updates are made to a package that has been exported, its status changes to Changes to Export. You
can choose at any time to export these changes. If you:
Delete a Package
You cannot delete a package if it is listed among the Required Packages for any other package.
You can use the Transport app to export content from one or more spaces for sharing with other tenants.
Prerequisites
To export content from a space via the Export app, you must have either:
A global role that allows you to access and export content A combination of a global role and a scoped role:
from any space, by granting you the following privileges:
• A global role that grants you the following privilege:
• Data Warehouse General (-R------) - To access SAP • Data Warehouse General (-R------) - To access
Datasphere. SAP Datasphere.
• Lifecycle (-----M--) - To use the Transport apps. • Lifecycle (-----M--) - To use the Transport apps.
• Spaces (-------M) - To update all spaces and space • A scoped role that grants you access to the space or
properties. spaces to export from with the following privileges:
• Spaces Files (-------M) - To create, read, update, and • Data Warehouse General (-R------) - To access
delete all objects in all spaces. SAP Datasphere.
The DW Administrator role, for example, grants these privi- • Spaces Files (-R------) - To view specific spaces
leges. and their contents.
• Data Warehouse Data Builder (-R------) - To
view Data Builder objects (and any other relevant
object privileges to allow you to create and update
other types of objects contained in the package).
For more information, see Privileges and Permissions and Standard Roles Delivered with SAP Datasphere.
You can share content across tenants from one or more landscapes and set access rights for each sharing
destination, for example, to allow other tenants to update a package and export it again.
Each tenant can store up to 2 GB of exported content for free in the My Content area.
The following object types can be exported and imported via the (Transport) app:
Note
Only object definitions can be transported. Data cannot be transported between SAP Datasphere tenants
via the (Transport) app.
Connections The definition of a connection does not have dependencies on any other objects.
Remote Tables The definition of a remote table contains information about its connection.
Local Tables The definition of a local table contains the structure of the table only, and does not have depend-
encies on any other objects.
See Creating a Local
Table.
Flows The definition of a data flow, replication flow, or transformation flow contains the definitions of all
its sources and its target table. When you export a flow, these objects are exported too.
See Creating a Data
Flow, Creating a Rep-
lication Flow, and Cre-
ating a Transformation
Flow.
Views The definition of a view contains the definitions of all its sources and any used data access
controls. When you export a view, these objects are exported too.
See Creating a Graphi-
cal View and Creating
an SQL View.
Intelligent Lookups The definition of an intelligent lookup contains the definitions of its input and lookup entities.
When you export an intelligent lookup, these entities are exported too.
See Creating an Intelli-
gent Lookup.
Analytic Models The definition of an analytic model contains the definitions of its fact and dimension sources.
When you export an analytic model, these entities are exported too.
See Creating an Ana-
lytic Model.
E/R Models The definition of an E/R model does not include the objects that it visualizes as dependencies.
These objects must be selected manually.
See Creating an Entity-
Relationship Model.
Data Access Controls The definition of a data access control contains the definition of its permissions entity. When you
export a data access control, the permissions entity is exported too.
See Securing Data with
Data Access Controls
[page 268].
Task Chains The definition of a task chain does not include the objects that it automates. These objects must
be selected manually.
See Creating a Task
Chain.
Business Entities / The definition of a business entity contains all its versions (which are also listed separately), along
Business Entity Ver- with the definition of its source data entity and any authorization scenarios. When you export a
sions business entity (or one of its versions), these objects are exported too.
Fact Models The definition of a fact model contains all its versions (which are also listed separately), along with
the definition of all its source fact models and business entities. When you export a fact model (or
See Creating a Fact
one of its versions), these objects are exported too.
Model.
Consumption Models The definition of a consumption model contains all its perspectives (which are also listed sepa-
rately), along with the definition of all its source fact models and business entities. When you
See Creating a Con-
export a consumption model (or one of its perspectives), these objects are exported too.
sumption Model.
Authorization Scenar- The definition of an authorization scenario contains the definition of its data access control. When
ios you export an authorization scenario, the data access control is exported too.
Note
Only local and remote tables, views, data flows, and E/R models can be selected for exporting in SAP
Datasphere tenants provisioned prior to version 2021.03. While objects created in SAP Datasphere tenants
provisioned prior to version 2021.03 can be transported to tenants provisioned after that date, transport of
objects in the other direction is not possible.
Procedure
1. In the side navigation area, click (Transport) (Export) and then click New Export.
You can, alternatively, use thePackages app to prepare and export packages (see Creating Packages to
Export [page 292]).
2. On the Choose content page, expand the SPACE folder to show your spaces, select the objects you want to
export, and then click Next.
You can select a space to include all its objects, or expand it to select individual objects within it.
You can select objects from any space of which you are a member, and select objects from more than one
space.
If an object depends on other objects, then they will also be selected. For example, when you select a view,
all its sources and any used data access controls will be selected.
Note
If an object selected for export depends on one or more objects that have been shared from other
spaces (see Sharing Entities and Task Chains to Other Spaces), then you must be a member of each of
those spaces in order to include all the necessary objects in your package.
If you are not a member of one or more sharing spaces, then you can still export your package (no error
will be displayed), but it will fail to import correctly unless all the missing objects are already present in
the target tenant.
You can review all the dependencies of your object by using the Impact and Lineage Analysis tool (see
Impact and Lineage Analysis).
3. On the Set package properties page, complete the properties, and then click Next.
Property Description
Details Enter more detailed information about the contents of the package, including terms and
conditions, if appropriate.
Location By default, the package is exported to the My Content folder. To export to a subfolder,
click Browse and select or create a new folder.
Sharing Select the sharing destinations for the export and their permissions on the content (see
Adding Sharing Destinations [page 300]).
Terms and Conditions Enable this switch to require users to accept terms and conditions before importing the
content.
4. On the Review and export page, review the overview of your package.
If there are any problems, select the warning icon to see how to fix them.
5. Click Export to export your package.
You can check the progress of your export in the Notifications list. You'll also receive a message when it is
complete, and you can check the Export Summary to see the status of all the objects in the package.
You can edit the content, share it with other tenants, or delete it from the Transport / Export / My Content
page (see Managing Exported Content [page 301]).
You can add sharing destinations to your export and specify their level of permissions for the content in the
Package Properties page of the Export app.
Procedure
1. Click the (Add destination) button above the Destinations list to open the Destinations dialog.
2. Complete the dialog as follows:
Property Description
Destination Type Select the category of the destination that you want to add:
• System: Add an individual tenant by URL. Select a Landscape, and copy the URLs to
the Systems list.
Note
SAP Datasphere is only available on the Cloud Foundry platform. You can share
to tenants that use the same type of Cloud Foundry data center as yours. These
host systems use two digits, such as eu10 or us30.
• Customer: Add all tenants that belong to one customer based on their ERP ID. Type
the ID in the Customer field.
Access Select one or more permissions that you want to grant to the sharing destination:
• View: Your package will appear in the My Content page for the destination tenants,
and their users will be able to view the name and summary.
• Import: The tenants' users can import the package from the My Content list.
• Edit: Your package will appear in the Manage Packages page for the destination
tenants, where users can make changes to the package and export it again.
• Delete: From the Manage Packages page, the tenants' users can delete the package.
You can add multiple destinations and edit access or remove destinations from the list.
4. Select OK to finish adding destinations and return to the wizard.
If you've exported packages for sharing with other tenants, or have edit or delete permissions for another
tenant's exports, you can work on them in the Transport / Export / My Content page.
In the side navigation area, click (Transport) (Export) and then click the My Content folder to
display the list of packages that you have exported and those from other systems that you have permission to
edit or delete.
For packages:
For folders:
• To create a new folder, select Create New Folder and type a name and description.
• To share a folder and all its contents, select the checkbox next to a folder and select Share. See the
access steps in Exporting Content for Sharing with Other Tenants [page 296] for more information.
• To delete a folder and all its contents, select the checkbox next to a folder and select Delete. The folder
and its contents will be removed for any tenants they were shared with.
You can use the Transport app to import content that has been shared from another tenant.
Context
To import content into a space via the Import app, you must have either:
For more information, see Privileges and Permissions and Standard Roles Delivered with SAP Datasphere.
Procedure
1. In the side navigation area, click (Transport) (Import) to open the list of content available for
import.
If you don’t see expected packages to import, click the Settings button and review the import settings. You
can:
• Enable the Show content shared by all systems option and optionally exclude content from specific
tenants by adding them to the Add systems to block content sharing list.
• Disable the Show content shared by all systems option and choose specific tenants to import from by
adding them to the Add systems to allow content sharing list.
If, after checking these settings, an expected package is still not listed, contact an administrator for help.
2. Click a package to open it and see an overview of its content.
3. Review the Import Overview section. For more details, and to modify any of the options, click the Import
Options tab:
Overwrite Preferences Controls whether objects that are already present in your space can be overwritten
during an import.
Note
Only object definitions can be transported. Data cannot be transported between
SAP Datasphere tenants via the (Transport) app.
Note
• The Including permissions option is not supported for SAP Datasphere.
• Connections are never overwritten by an import.
• When overwriting objects with a release state, the objects you are import-
ing must have the same release state or another release state that is con-
sistent with their lifecycle (see Releasing Stable Views for Consumption).
Deployment Preference Check the Deploy after import checkbox to automatically deploy objects after import. If
you have selected:
• Don't overwrite objects or data - Only newly imported objects will be deployed.
• Overwrite objects and data - All imported objects in the space will be deployed.
Note
This option does not support packages that contain objects from one or more
spaces where objects are shared from one space to another (see Sharing Entities
and Task Chains to Other Spaces. In this case, you should manually deploy the
objects from the source space and then deploy the objects from the target space.
Content Select the space or spaces to import your content into in the Space Mapping field. You
must choose one space for each space listed in the package:
• You can select any space for which you have the Spaces.Update permission to
import the space's content into.
If you have the Spaces.Manage permission, then you can choose to import into any
space in your tenant.
• If the space does not exist in your tenant and you have the Spaces.Create permis-
sion, then you are invited to create it by default.
• You cannot "merge" content from two or more spaces included in your package. You
must select a separate space for each space listed.
Note
Choosing space mappings is only possible if the source objects for any shared
objects are included in the package. For example if your package contains shared
objects in Space A that have been shared from Space B, then the source objects
from Space B must be included in the package. If the sources of any shared
objects cannot be resolved in the package, then the package must be imported into
spaces with the same technical names as the spaces listed in the package.
You can select and deselect individual objects, or types of objects. The Impacted Object
column shows warnings for objects that will overwrite content on your system.
You will receive a notification when the process is complete, or if an error occurs.
If you selected Don’t overwrite objects or data, a notification will let you know if any of the content wasn’t
imported to avoid overwriting existing content.
To view the log of updates and imports for a package, select the package in the content network and
choose History.
5. If your content included one or more connections that didn't exist in your space before the import, then
you should:
a. In the Connections app, open each connection, complete the configuration and enter appropriate
credentials to connect to your system (see Edit a Connection [page 123]).
b. In the Repository Explorer or in the object editors, select the imported objects that depend on an
imported connection (and which could not, therefore, be automatically deployed) and click (Deploy)
to deploy them.
Note
If you checked the Deploy after import checkbox in the Import Options, you can find a notification
about the failed deployment of objects in the Notifications panel. Clicking the notification directly
leads you to the list of imported objects in the Repository Explorer.
Integrate SAP Datasphere with SAP Cloud Transport Management service to transport content packages
across different landscapes.
SAP Cloud Transport Management service lets you manage software deliverables between accounts of
different environments (such as Cloud Foundry, ABAP, and Neo), by transporting them across various
runtimes.
To transport your SAP Datasphere content through SAP Cloud Transport Management, you first must
configure your SAP Cloud Transport Management system.
Once you’ve configured the system, you must connect it to your SAP Datasphere tenant.
Then you can use the SAP Datasphere Transport app to transfer your content to SAP Cloud Transport
Management.
Note
Users with the DW Administrator global role (or users with both a scoped DW Space Administrator role and
a global role providing the Lifecycle privilege), can use the Transport app to export content.
Finally, you can return to SAP Cloud Transport Management to ensure the transferred content is available and
ready to be imported to another SAP Datasphere tenant.
Tip
• Go to System→Administration→App Integration to find the token URL and add the extension
<Token URL>?grant_type=client_credentials.
• You must create the OAuth Client (client ID and secret) with the purpose API Access and
the access Analytics Content Network Interaction. See Create OAuth2.0 Clients to Authenticate
Against SAP Datasphere for more information.
1. Create a destination in the SAP Cloud Transport Management subaccount for each target environment
of SAP Datasphere. See Create Transport Destinations .
2. Create nodes and routes. See Create Transport Nodes and Create Transport Routes . Make sure to
select the checkbox Allow Upload to Node when creating a source node. When defining a target node,
the content type must be Application Content to be able to select the destination name you've created
in the previous step.
3. Go to your SAP Business Technology Platform subaccount. Choose the SAP Cloud Transport Management
instance of your choice. In the instance details page, choose … → Download to download the SAP Cloud
Transport Management instance key.
Using the SAP Datasphere Transport App to Transfer Content to SAP Cloud
Transport Management
1. In the side navigation area, click Transport Export and select the package you want to export.
2. Click Upload to SAP Cloud Transport Management Node from the menu bar: in the dialog box, you can
see the list of selected packages.
1. Log into your SAP Cloud Transport Management tenant: in the Transport Nodes panel, select the target
node where you have uploaded the SAP Datasphere package.
2. In the Target Node Details page, you can see your packages that are ready for import: select the package
and click Import Selected.
Once the import is completed, you see the content of the package available in its corresponding destination in
your SAP Datasphere account.
Note
The user who uploaded the package must be a member of the target tenant.
The authentication method (defined by the tenant owner) must be the same on the source and on the
target tenant. If the selected method is single sign-on (SSO), make sure to define the same user attribute
for the source and the target tenant. For more information on authentication methods, see Configuring
Identity Provider Settings.
Troubleshooting
In case of import failure, you can access the logs to investigate the error by adding this endpoint to the target
tenant URL in your browser:
{TenantURL}/contentmanager/v4/jobs/{ContentReferenceID}?brief=false
Note
The Content Reference ID is retrieved from the SAP Cloud Transport Management import logs.
Users with a space administrator role are responsible for controlling user access to their space, and monitoring
and otherwise managing the space.
All data acquisition, preparation, and modeling in SAP Datasphere happens inside spaces. A space is a secure
area - space data cannot be accessed outside the space unless it is shared to another space or exposed for
consumption.
An administrator must create one or more spaces. They allocate disk and memory storage to the space, set
its priority, and can limit how much memory and how many threads its statements can consume. See Creating
Spaces and Allocating Storage.
If an administrator has assigned you the role of space administrator for a certain space via a scoped role,
you can assign other users, create connections to source systems, secure data with data access controls, and
manage other aspects of the space.
To view information on the properties of a space that have been set by an administrator, see:
• Create a Space
• Allocate Storage to a Space
• Set Priorities and Statement Limits for Spaces
Prerequisites
To assign users to your space and manage them from the Users area in your space details page, you must have
a scoped role that grants you access to your space with the following privileges:
The DW Space Administrator role template, for example, grants these privileges. For more information, see
Privileges and Permissions and Standard Roles Delivered with SAP Datasphere.
If a user with an administrator role has assigned you to a space with space administration privileges via
a scoped role, you can assign and manage users in your space. For more information about how a DW
Administrator assigns users to roles in particular spaces via scoped roles, see Create a Scoped Role to Assign
Privileges to Users in Spaces.
In this example, based on scoped roles that the DW Administrator has created, users are assigned to spaces
with scoped roles as follows:
Joan has the role of space administrator for the spaces Sales US and Sales Asia. She can change the user
assignment. She can assign any user to her 2 spaces with the scoped roles to which her spaces are assigned:
Sales Modeler and Sales Spaces Admin.
• As Bob no longer needs to work in the space Sales US, Joan unassigns Bob from the space.
• As Jim needs to work in space Sales US with the modeling privileges, Joan assigns Jim to the space with
the Sales Modeler scoped role.
• As Bob then needs to manage the space Sales Asia, Joan assigns Bob to the space with the Sales Space
Admin scoped role.
With Joan's changes, users are now assigned to spaces with scoped roles as follows:
In the side navigation area, click (Space Management), locate your space tile, and click Edit to open it.
In the Users area of the space page, you can see and edit which users are assigned to your space with which
scoped roles. You can assign users to your space among all the users in the tenant, with one or more scoped
roles to which your space is assigned.
Note
• Space Administrator column - A green tick appears for each user who has space administration
privileges for the space, which includes having the privilege (permission) Scoped Role User Assignment
(Manage).
• Scoped Role column - The one or more scoped roles that are assigned with each user to the space are
displayed. Only a DW Administrator can click a scoped role, which opens the scoped role page.
All user assignment changes you do in the Users area of your space page are reflected in the scoped roles to
which your space is assigned.
You can add a user who will have certain privileges in your space by selecting the user and one or more scoped
roles.
1. Click Add. All the users created for the tenant are displayed in the dialog that opens.
2. Select one or more users and click Next. All scoped roles to which your space is assigned are displayed.
3. Select one or more scoped roles and click Create.
You can change the privileges that the user has in your space by assigning it to another scoped role. A user can
be assigned to one or more scoped roles in a space.
1. Select a user and click Edit. All scoped roles to which your space is assigned are displayed.
2. Select one or more scoped roles and click Select.
You can unassign a user from your space. The user will be unassigned from the space in the one or more
scoped roles which gave access to the space.
You can also use the SAP Datasphere command line interface to assign users to a space. For more information,
see Manage Spaces via the Command Line.
Create a time table and dimension views in your space to provide standardized time data for your analyses.
The time table contains a record for each day in the specified period (by default from 1900 to 2050), and the
dimension views allow you to work with this date data at a granularity of day, week, month, quarter, and year,
and to drill down and up in hierarchies.
Prerequisites
To work with the time table and dimensions in your space, you must have a scoped role that grants you access
to your space with the following privileges:
The DW Space Administrator role template, for example, grants these privileges. For more information, see
Privileges and Permissions and Standard Roles Delivered with SAP Datasphere.
Procedure
1. In the side navigation area, click (Space Management), locate your space tile, and click Edit to open it.
2. In the Time Data section, click Create Time Tables and Dimensions to open the dialog.
The dialog lists all the objects that will be created:
• Time Table - A table containing one record for each day from, by default, the year 1900 to the year
2050.
• Four views with a semantic usage of Dimension:
• Time Dimension - Day - Suitable for an association from a column containing dates in the form Dec
31, 2020, and providing hierarchies with levels:
Note
• Time Dimension - Month - Suitable for an association from a column containing months in the form
202012 (for December 2020), and providing hierarchies with levels:
• Year, Quarter, Month
• Year, Month
• Time Dimension - Quarter - Suitable for an association from a column containing quarters in the
form 20204 (for the fourth quarter of 2020), and providing one hierarchy with levels:
• Year, Quarter
• Time Dimension - Year - Suitable for an association from a column containing years in the form
2020.
• Three translation tables - which contain the names of days, months, and quarters in English, French,
German, and Spanish.
3. [optional] Modify the values in any of the following fields:
• Enter different years in the From Year and To Year fields to expand or reduce the scope of the time table
data.
Note
The To Year cannot be more than 100 years later than the current year.
• Edit any of the proposed business names for the tables and views to be created. The technical names
cannot be modified.
4. Click Create.
When creation is complete, the content of the section changes to show that the objects are now present
in the space. You can use the tools to edit the Time Table values, refresh the time data, or delete these
entities.
Note
From v2023.08, we use the ISOWEEK() function to generate ISO-standard year and week numbers
(see ISOWEEK Function (Datetime) in the SAP HANA Platform documentation. To regenerate time data
generated before this version, click Edit, modify your From Year or To Year, and click Save.
The tables and views are listed with your other space objects and can be opened (in read-only form) from
the Data Builder and Repository Explorer.
Users assigned to your space can now use the views as dimensions by creating associations to them from
their entities containing time data.
You can enable audit logs for your space so that read and change actions (policies) are recorded.
Administrators can then analyze who performed which action at which point in time.
Prerequisites
To enable audit logs for your space, you must have a scoped role that grants you access to your space with the
following privileges:
The DW Space Administrator role template, for example, grants these privileges. For more information, see
Privileges and Permissions and Standard Roles Delivered with SAP Datasphere.
• Enable audit logs of SAP Datasphere objects for read and change operations on the space level in Space
Management <Your Space> Edit Space Auditing . You can set the retention time in days for each
audit log policy (read or change). The default and minimum retention time is 7 days and the maximum
retention time is 10 000 days.
If audit logs have been enabled, entries of all SAP Datasphere related objects are saved in an SAP HANA
audit log. These logs don't include the objects of the database access schemas, like open SQL schemas,
for example.
Note
If you choose to enable audit logs, be aware that they can consume a large quantity of GB of disk
in your SAP Datasphere tenant database, especially when combined with long retention periods.
Administrators can delete audit logs when needed, which will free up disk space. For more information,
see Delete Audit Logs.
• For individual database schemas (open SQL schemas), you can enable auditing for read and
change operations in Space Management <Your Space> Database Access Database Users Edit
Privileges . The retention time can be defined per schema.
Data Lake
Please note, that the statements issued via the execute-procedure for data lake are currently not
audited in SAP Datasphere.
As a result, SAP HANA policies are created for the schemas of the space.
• DWC_DPP_<space name>_READ
• DWC_DPP_<space name>_CHANGE
When you disable an audit policy for a space or for a database user (open SQL schema), all related audit log
entries are deleted.
Note
You can keep the audit log entries generated for your space by exporting them before the space is deleted.
If you want to keep the audit log entries generated for your space, you can export them before they are
automatically deleted by these actions: deleting a space, deleting a database user (open SQL schema),
disabling an audit policy for a space, disabling an audit policy for a database user (open SQL schema) and
unassigning an HDI container from a space.
To export audit log entries using the SAP HANA Database Explorer, follow these steps:
1. Expose for consumption the view containing the audit log entries that you want to export. See Exposing
Data For Consumption.
2. Create a database user in your space. See Create a Database User [page 147]
3. Connect to your Open SQL schema with SAP HANA database explorer. See Connect to Your Open SQL
Schema [page 148]
4. Export the data. See Export Schemas, Tables, and Other Catalog Objects in the SAP HANA Database
Explorer guide.
Related Information
Users with a scoped role containing the Translation privilege can translate metadata such as business names
and column names for dimensions and analytic models, and hierarchy dimension labels for SAP Analytics
Cloud stories.
For example, you may have stories that have column names in a default language, but you'd like to have those
names displayed in other languages so that your SAP Analytics Cloud story can be viewed in the language of
your choice.
To translate objects' metadata, you can have it translated into multiple target languages using third-party
translation tools or you can have those descriptions translated in-line using the Translation tool. You can use
the tool to review translations, and edit or add translations from scratch for some or all languages.
To manage translation, you must be assigned to a scoped role that inherits a template (such as DW Space
Administrator), which grants the Translation privilege (see Create a Scoped Role to Assign Privileges to Users in
Spaces).
You can refer to the list below for the supported languages in SAP Datasphere.
Prerequisites
To enable translation for your space, and to manage translation in the Translation tool, you must have a scoped
role that grants you access to your space with the following privileges:
The DW Space Administrator role template, for example, grants these privileges. For more information, see
Privileges and Permissions and Standard Roles Delivered with SAP Datasphere.
Enable Translation
DW Space Administrators can turn on the toggle to enable content translation for a space:
Note
Only Analytic Models can be selected in the Add Objects dialog, but the dimensions and hierarchies
attached to added Analytic Models are automatically selected to the Translation tool.
3. Click Add Objects. The selected Analytic Models and their dependencies are added to the tool.
You can translate large amounts of metadata with the help of XLIFF files. The (Download) button in the
Translation tool allows you to download the strings to be translated to a XLIFF file, which can be used to
translate the content. Then, the translated strings must be uploaded back to the Translation tool in the XLIFF
file format with the (Upload) button.
1. In the Translation tool, select at least one entity that you want to translate.
2. Click (Download) to download the XLIFF files. If one object is selected to be downloaded, the
corresponding XLIFF file can be downloaded as a .zip file.
3. You can choose to select All Strings or Outstanding Strings while exporting the XLIFF file(s). The
Outstanding Strings option is enabled only for partially translated objects.
The XLIFF file is downloaded to the default location in your system.
Note
The downloaded XLIFF file contains the source locale and the list of text that needs translation. You
can use this XLIFF file in any external translation tool to generate the translated content as a separate
XLIFF file per locale.
4. Once the translation is done, click (Upload) to open the Import Translations dialog.
5. Click Upload Files to upload the translated XLIFF files back into the Translation tool.
6. Click Import to upload the XLIFF files.
Note
You can import multiple XLIFF files. Each XLIFF file contains the translated content in one language, which
is automatically fetched from the language code embedded in the name of the uploaded XLIFF files. These
You can view the source and the translated text in the Translation tool, and add or edit the translated text
in-line.
1. Select the object you want to translate manually and click (Edit). The list of source text appears.
2. Choose the Target Language and start adding the translated text in the empty field.
3. Select Save. The message Translation updated successfully appears, and the user who had marked the
model or dimension for translation is notified.
Delete Translations
To delete translations, select one or more objects in the Translation tool and click (Delete).
Note
Deleting translations removes the object and attached translated strings from the Translation tool. You'll be
able to add the object back to the tool, but no translation will be attached to it any longer.
Manage Translation
The Translation tool shows the spaces you are assigned to with enabled translation. Each space contains
entities that are selected for translation. The tool allows you to perform and manage these translations.
From the side navigation, choose (Translation) and select a space. The spaces available here are the ones
enabled for translation. For more information, see Enable Translation [page 315].
Note
• If the model has changed, new strings will automatically appear in the translation tool as Not
Translated. Expired strings will still be available in the Translation tool.
• You can upload the XLIFF files even if there is a mismatch in the HTML tags between the source and
target language strings.
(Add) Adds objects to the Translation tool. Only Analytic Models can be selected in the Add
Objects dialog, but the dimensions and hierarchies attached to selected Analytic Models
are automatically added to the Translation tool. See Select Objects to Translate [page 316].
(Edit) Open the selected object for manual in-line translation. Users with the Create permission
can access this functionality. See Translate Metadata Manually [page 317].
(Delete) Deletes translations for the selected objects. See Delete Translations [page 317].
(Download) (Upload) Downloads translatable (source) XLIFF files and uploads the translated XLIFF files. See
Translate Metadata via XLIFF Files [page 316].
(Search) Searches objects based on Name, Description, Translated by, Translated Language, Re-
quested on, Changed on, Status, and Owner.
Tool Fields
Title Description
• remote table
• local table
• view
• analytic model
• hierarchy
Creator Name of the person who seletected the object for translation
Changed On When the strings or translations for strings were last up-
dated
When you save a space, it is stored in the SAP Datasphere repository, which contains the design-time
definitions of all your spaces. When you deploy a space, you are creating a run-time version for use in the
SAP Datasphere database.
Prerequisites
To save and deploy the changes you've made in your space, you must have a scoped role that grants you
access to your space with the following privileges:
The DW Space Administrator role template, for example, grants these privileges. For more information, see
Privileges and Permissions and Standard Roles Delivered with SAP Datasphere.
Space Deployment
The current status of a space is shown in its Deployment Status property, which can have the following values:
• Not Deployed - The space has never been deployed and exists only as a design-time version.
• Deployed - The space is deployed to the run-time database and its design-time and run-time versions are
identical.
Note
For spaces that have been created before the deploy feature exists, the Status area will display “No
Information”. Spaces are deployed anyway. If you click Deploy, the status will display Deployed.
You can also view the date of the last deployment in the Deployed On area.
Note
• When you assign a user to the space, the user becomes active when you save the space. You can view
this in the Active column of the Users area in your space.
• Any change that you make in the Time Data area of your space is immediately deployed even if you
haven’t saved nor deployed the space. A message is displayed in this area, which indicates that the
changes are immediately applied
• The option Expose for Consumption by Default does not need a redeployment to be active.
Note
On rare occasions when the run-time database is not available, an information message is displayed and
certain features that depend on the run-time are disabled:
• The Deploy, Delete and Monitor buttons or commands are greyed out.
• In the Space Management start page, dashes (-) appear instead of numbers in the bars (for example,
the Used Disk bar at the top of the page and in the space tiles), in the space status area at the top of the
page and space status colors are not displayed in the individual space tiles.
• In a space page, the Time Data area is disabled.
You can see the status of all spaces at the top of the Space Management page.
You also can see the status of an individual space at the top-left of the space.
Each status shows the used storage and may indicate whether a space needs extra attention. The status also
shows whether a space is active or not (locked).
This space is close to exceeding it's storage quota. It might make sense to delete unnecessary
data in your space or to extend the assigned storage.
If your space exceeds it's storage quota, it might change to a locked state.
For more information on how to unlock a space, see Unlock a Locked Space [page 321].
When a space is locked, its capabilities are limited. Users can continue to create and modify
objects and save their changes in the repository, but they cannot deploy their changes to the
run-time database.
Restrictions include:
When a space exceeds its assigned storage or when the audit logs enabled in the space consume too much
disk storage, the space is automatically locked after 60 minutes if you do not free up space.
When a space is locked, users assigned to the space can continue to create and modify objects and save their
changes in the repository, but they cannot deploy their changes to the run-time database.
Note
To manually lock or unlock your space, you must have a scoped role that grants you access to your space with
the following privileges:
The DW Space Administrator role template, for example, grants these privileges. For more information, see
Privileges and Permissions and Standard Roles Delivered with SAP Datasphere.
If a space exceeds its allocations of memory or disk storage, it will be locked until a space user deletes the
excess data or an administrator assigns additional storage.
• Users assigned to a space can delete data to bring the space back under the limit of its assigned storage.
• A space administrator can use the Unlock button on the Space Management page or on the space page to
unlock the space for a 24-hour grace period, in case urgent changes must be deployed.
• An administrator can assign more disk and/or memory storage to the space (see Allocate Storage to a
Space).
If the total size of all audit logs across all spaces has reached 40% of the tenant disk storage, the system
automatically locks any spaces - and deactivate any database analysis users - whose audit logs consume more
than 30% of the total audit log size.
• A space administrator can use the Unlock button on the Space Management page or on the space page to
unlock the space for a 24-hour grace period, in case urgent changes must be deployed.
• An administrator can delete audit logs and free up disk space (see Delete Audit Logs).
• A space administrator can disable audit logs or decrease the number of days audit logs are kept for (see
Logging Read and Change Actions for Audit [page 313]).
Lock a Space
If you want to put a space on hold and you are a space administrator, you can manually lock it.
Note
You can unlock one or more spaces that have been manually locked by using the Unlock button.
Prerequisites
To monitor the storage consumption of your space, you must have a scoped role that grants you access to your
space with the following privileges:
The DW Space Administrator role template, for example, grants these privileges. For more information, see
Privileges and Permissions and Standard Roles Delivered with SAP Datasphere.
Procedure
To open the monitoring page, select your space and click Monitor from the Space Management page or
alternatively open your space and click Monitor on the upper-right side of your space.
Note
Action Example
You can use the SAP Datasphere command line interface, datasphere, to read, update, and delete your
space. You can set space properties, assign (or remove) users, create database users, create or update objects
(tables, views, and data access controls), and associate HDI containers to your space.
To use datasphere to manage your space you must have an SAP Datasphere user with the DW Space
Administrator role or equivalent permissions (see Roles and Privileges by App and Feature).
For more information, see Manage Spaces via the Command Line.
You can export the data contained in your space at any time. For example, you may want to export data before
deleting your space.
Prerequisites
To export the data contained in your space, you need to connect to an Open SQL schema or to allow the space
to access the open SQL schema. To do so, you must have a database user and a password (see Create a
Database User [page 147]).
Procedure
To export data using the SAP HANA Database Explorer, follow these steps:
1. Expose for consumption the views containing the data you want to export (see Exposing Data For
Consumption).
2. Create a database user in your space (see Create a Database User [page 147]).
3. Connect to your Open SQL schema with SAP HANA database explorer (see Connect to Your Open SQL
Schema [page 148]).
4. Export the data (see Export Schemas, Tables, and Other Catalog Objects in the SAP HANA Database
Explorer guide).
Delete a space if you are sure that you no longer need any of its content or data.
When you delete a space, the space is moved to the Recycle Bin area. From the Recycle Bin, a user with an
administrator role can either restore the space or permanently delete the space from the database to recover
the disk storage used by the data in the space (see Restore Spaces from, or Empty the Recycle Bin).
Note
For spaces that have been deleted before version 2023.05, all related audit logs have been kept. A user with
an administrator role can decide to delete them (see Delete Audit Logs ).
Note
• Export the data contained in your space (see Export Your Space Data [page 325]).
Only users with an administrator role can create spaces. Users with an administrator or a space administrator
role can delete spaces.
Prerequisites
To move your space to the Recycle Bin area, you must have a scoped role that grants you access to your space
with the following privileges:
The DW Space Administrator role template, for example, grants these privileges. For more information, see
Privileges and Permissions and Standard Roles Delivered with SAP Datasphere.
Procedure
Note
If your space contains replication flows with objects of load type “Initial and Delta”, you should make sure
that these replication flows are stopped before you delete the space. If you restore the space at a later point
in time, the replication flows can then be started again. For more information about replication flows, see
Working With Existing Replication Flow Runs [page 242].
Note
Hyperlinks
Some links are classified by an icon and/or a mouseover text. These links provide additional information.
About the icons:
• Links with the icon : You are entering a Web site that is not hosted by SAP. By using such links, you agree (unless expressly stated otherwise in your
agreements with SAP) to this:
• The content of the linked-to site is not SAP documentation. You may not infer any product claims against SAP based on this information.
• SAP does not agree or disagree with the content on the linked-to site, nor does SAP warrant the availability and correctness. SAP shall not be liable for any
damages caused by the use of such content unless damages have been caused by SAP's gross negligence or willful misconduct.
• Links with the icon : You are leaving the documentation for that particular SAP product or service and are entering an SAP-hosted Web site. By using
such links, you agree that (unless expressly stated otherwise in your agreements with SAP) you may not infer any product claims against SAP based on this
information.
Example Code
Any software coding and/or code snippets are examples. They are not for productive use. The example code is only intended to better explain and visualize the syntax
and phrasing rules. SAP does not warrant the correctness and completeness of the example code. SAP shall not be liable for errors or damages caused by the use of
example code unless damages have been caused by SAP's gross negligence or willful misconduct.
Bias-Free Language
SAP supports a culture of diversity and inclusion. Whenever possible, we use unbiased language in our documentation to refer to people of all cultures, ethnicities,
genders, and abilities.
SAP and other SAP products and services mentioned herein as well as
their respective logos are trademarks or registered trademarks of SAP
SE (or an SAP affiliate company) in Germany and other countries. All
other product and service names mentioned are the trademarks of their
respective companies.