Documenti di Didattica
Documenti di Professioni
Documenti di Cultura
General Questions:
1. What is DATA Provisioning?
DATA Provisioning is a process of creating, preparing, and enabling a network to
provide data to its user. Data needs to be loaded to SAP HANA before data reaches to
the user via a front-end tool.
All these processes are referred as ETL (Extract, Transform, and Load), and detail is
as below-
Extract – This is first and sometimes most difficult part of ETL, in which data are
extracted from the different source system.
Transform – In the Transformation Part, series of rules or functions is defined
for the data extracted from the source system, for loading data into the target
system.
Load – The Load phase loads the data in the target system.
Data provisioning is all about ways to get the data from SAP /NON-SAP systems
into SAP –HANA.
Load:
Starts an initial load of replication data from the source system. The procedure is a
one-time event. After it is completed, further changes to the source system database
will not be replicated.
For the initial load procedure, neither database triggers nor logging tables are
created in the source system. Default settings use reading type 3 (DB_SETGET) with
up to 3 background jobs in parallel to load tables in parallel or subsequently into the
HANA system.
Replicate:
Combines an initial load procedure and the subsequent replication procedure (real
time or scheduled).
Before the initial load procedure will start, database trigger and related logging
table are created for each table in the source system as well as in SLT replication
server.
Stop:
Stops any current load or replication process of a table.
The stop function will remove the database trigger and related logging tables
completely. Only use this function if you want to continue a selected table otherwise
you must initially load the table again to ensure data consistency.
Suspend:
Pauses a table from a running replication. The database trigger will not be deleted
from the source system. The recording of changes will continue and related
information is stored in the related logging tables in the source system.
If you suspend tables for a long time the size of logging tables may increase and
adjust the table space if required.
Resume:
Restarts the application for a suspended table. The previous suspended replication
will be resumed (no new initial load required).
A rule specified in the Advanced Replication settings transaction for source tables
such that data is transformed during the replication process. Example you can
specify rule to
Convert fields
Fill empty fields
Skip records
INTERVIEW QUESTIONS FOR SAP HANA-“DATA PROVISIONING”
8. Is it possible to load and replicate data from one source system to multiple target
database schemas of HANA system?
Yes. It is possible for up to 4.
10. Is there any pre-requisite before creating the configuration and replication?
For the SAP source systems DMIS add-on is installed in SLT replication server. User
for RFC connection has the role IUUC_REPL_REMOTE assigned but not DDIC.
For non-SAP source systems DMIS add-on is not required and grant a database user
sufficient authorization for data replication.
ERP stands for Enterprise Resource Planning Software and is a integrated computer
used to manage a company's resources effectively. It ensures smooth information
amongst various departments in an enterprise or a company and manages
workflows.
1. SAP
2. Baan
3. JD Edwards (now acquired by Oracle)
4. Siebel
5. People soft (now acquired by Oracle)
6. Microsoft Dynamics
SLT Questions:
1. What is SLT?
The SAP Landscape Transformation (LT) Replication Server is the SAP technology that
allows us to load and replicate data in real-time from SAP source systems and non-SAP
source systems to an SAP HANA environment.
The SAP LT Replication Server uses a trigger-based replication approach to pass
data from the source system to the target system.
A table in the source system that records any changes to a table that is being replicated.
This ensures that SLT replication server can replicate these changes to the target
system.
INTERVIEW QUESTIONS FOR SAP HANA-“DATA PROVISIONING”
Advantages:
SAP LT uses trigger based approach. Trigger-based approach has no
measureable performance impact in source system.
It provides transformation and filtering capability.
It allows real-time (and scheduled) data replication, replicating only
relevant data into HANA from SAP and non-SAP source systems.
It is fully integrated with HANA Studio.
Replication from multiple source systems to one HANA system is allowed,
also from one source system to multiple HANA systems.
4. Is it possible to use a replication for multiple sources and target SAP HANA systems?
Yes, the SAP LT Replication Server supports both 1:N replication and and N:1 replication.
Multiple source system can be connected to one SAP HANA system.
One source system can be connected to multiple SAP HANA systems. Limited to 1:4 only.
Status Yellow: It may occur due to triggers which are not yet created successfully.
Status Red: It may occur if master job is aborted (manually in transaction SM37).
When a new configuration in the SAP LT Replication Server is created, the database
connection is automatically created along with a schema GUID and Mass transfer id
(MT_ID).
A schema GUID ensures that configurations with the same schema name can be created.
The Mass transfer ID is used in the naming of SLT jobs and the system can uniquely
identify a schema.
The mapping of mass transfer IDs and related schema are stored in the SLT control table
DMC_MT_HEADER in the SAP LT Replication Server.
8. Does SLT for SAP HANA support data compression like SAP HANA database?
Yes, this is automatically covered by the RFC connection used for data replication from
the SAP source system.
The information required to create the connection between the source system, SLT
system, and the SAP HANA system is specified within the SLT system as a Configuration.
A new configuration in Configuration & Monitoring Dashboard (transaction LTR) can
hence be defined.
10. What ETL server is used for ECC system?
Performance optimization for operational reporting
Processing of large volumes of data
Provision of real-time analytics
12. What method is used to extract tables from ecc server in real time?
BODS Questions:
1. What is BODS?
It is an ETL tool currently acquired by SAP used for integration with all types of
disparate systems, extracting data from them, transforming them into meaningful
information and loading them into all types of systems.
Component of Data Services:
SAP DATA services have below component –
You can reuse the ERP application logic by reading extractors or utilizing SAP
function modules. In addition, the ETL-based method offers options for the
integration of third-party data providers.
Business Objects Data Services provides a graphical interface that allows you to easily
create jobs that extract data from heterogeneous sources, transform that data to meet
the business requirements of your organization, and load the data into a single location.
4. What is Repository?
The HANA Repository is storage system for development objects and is built into SAP
HANA. The repository supports Version control, Transport, and Sharing of objects
between multiple developers. We can add objects to the repository, update the objects,
publish the objects, and compile these objects into runtime objects.
In order to load data from external sources to SAP HANA we need to establish a
connection with the server. To connect, we need to provide details of Business Objects
Data Services repository and ODBC drivers. Once the connection is established, we can
import the tables definition and then load data into table definitions.
ETL based replication(data services)Which all Business Objects tools can be directly
connected to HANA Information views
SAP Lumira
Analysis edition for OLAP
SAP Crystal Reports
Design Studio
INTERVIEW QUESTIONS FOR SAP HANA-“DATA PROVISIONING”
Data Service Workbench was introduced, we used to perform the following steps for data
loading in SAP HANA via Data Services: Import Metadata into Data Services
Workbench (source system HANA relevant) | create DS jobs to push data in SAP HANA |
schedule/run the DS job | data is then available in the relevant schema in SAP HANA.
Job is the smallest unit of work that you can schedule independently for execution.
Work flow defines the decision-making process for executing data flows.
Data flows extract, transform, and load data. Everything having to do with data,
including reading sources, transforming data, and loading targets, occurs inside a
data flow.
9. What is a transform?
During the extraction step the data is extracted from the source systems and is
made available for the next steps. The main goal is to obtain all the data from the
sources so that the impact on the performance of the source systems is
minimized. Generally, the data from the source systems is stored 1-on-1 in a
‘Staging Area’, which will serve as the input for the processing in the
transformation phase.
DXC Questions:
SAP HANA DXC is batch driven data replication technique i.e. it can execute after a time
interval.
SAP HANA DXC uses existing ETL (Extract, Transform, and Load) method of
SAP Business Suite Application via a HTTPS connection.
In SAP Business suite application content Data Source Extractors have been
available for data modeling and data acquisition for SAP Business
Warehouse.
SAP DXC uses these Data Source Extractor to deliver data directly to SAP
HANA.
This method requires no additional server or application in the SAP HANA system
landscape.
DXC method reduces complexity of data modeling in SAP HANA as data sends to
HANA after applying all business extractor logics in Source System.
It speeds up the time lines for SAP HANA implementation project
It provides semantically rich data from SAP Business Suite to SAP HANA
It reuses existing proprietary extraction, transformation, and load mechanism built
into SAP business Suite systems over a simple HTTP(S) connection to SAP HANA.
Data Source must have a predefined mechanism for extraction, transformation and
load and if not we need to define one.
INTERVIEW QUESTIONS FOR SAP HANA-“DATA PROVISIONING”
It requires a Business Suite System based on Net Weaver 7.0 or higher with at least
below SP: Release 700 SAPKW70021 (SP stack 19, from Nov 2008).
4. What is an Extractor?
In the SAP source system, extractors are a data retrieval mechanism. It can fill the extract
structure of a data source with the data from the SAP source system datasets.
SDA Questions:
Prerequisites
To use SAP HANA Smart Data Access as the source type for Open ODS views or for query
access to near-line storage data, the following prerequisites apply:
You have installed the ODBC drivers for the databases that you want to connect
with SAP HANA Smart Data Access.
You have connected the databases SAP HANA Smart Data Access as remote
sources.Pay particulate attention here to which databases are supported in which
versions for SAP HANA Smart Data Access.
Note that the SAP HANA system authorization CREATE REMOTE SOURCE is
required to create a remote source. If the remote source is not created with the
SAP<SID> user but with a different database user instead, then this database user
must assign the corresponding object authorizations to the SAP<SID> user (CREATE
VIRTUAL TABLE and DROP).
The remote data is accessed from the BW system with the database user that is used
for the connection of the BW system to the SAP HANA database. When creating a
remote source in SAP HANA, you specified a user for the connection to the source
database. SAP HANA passes the SQL statements on to this user. Make sure that this
user has sufficient authorizations in the relevant schemas and tables in the source
database.
SAP HANA support uploading data from a file without ETL tools (SLT, BODS, and DXC). It is
a new feature of HANA Studio Revision 28 (SPS04).
SAP HANA Support following type of files which will be available on client system as below