Technology Blogs by Members
Explore a vibrant mix of technical expertise, industry insights, and tech buzz in member blogs covering SAP products, technology, and events. Get in the mix!
cancel
Showing results for 
Search instead for 
Did you mean: 
luv_sharma
Newcomer

In this blog, we will explore one of the latest innovations from SAP Datasphere. Effectively, from version 2021.03 onwards, SAP Datasphere has introduced the concept of ‘Replication Flows’.  

Think of replication flows as your trusty sidekick in the world of data management. This feature is all about efficiency and ease. It lets you effortlessly copy multiple tables from a source to a target, without breaking a sweat. 

The beauty in this story is that it is cloud-based all the way! Say goodbye to the hassle of dealing with on-premises components like installing and maintaining data provisioning agents. Thanks to SAP Datasphere's cloud-based replication tool, those headaches are a thing of the past. Replication flows support the following source and target systems: 

Supported Source Systems: 

  • SAP Datapshere (Local tables) 
  • SAP S/4 HANA: Cloud and on premise 
  • SAP BW and SAP BW/4 HANA 
  • SAP HANA: Cloud and on premise 
  • SAP Business Suite 
  • MS Azure SQL 

Supported Target Systems: 

  • SAP Datasphere 
  • SAP HANA: Cloud and on premise 
  • SAP HANA Data Lake 
  • Google BigQuery 
  • Amazon S3 
  • Google Cloud Storage 
  • MS Azure Data Lake Storage 
  • Apache Kafka 

 

Technical architecture diagram: 

luv_sharma_0-1711958893569.png

 

 So now, let us dive into how we can replicate S/4 HANA data to Google BigQuery using these newly released replication flows in SAP Datasphere! 

Pre-requisites: 

  1. Establish a source connection (e.g., S4 Hana Cloud) within Datasphere. 
  1. Establish a connection between SAP Datasphere and Google BigQuery. 

Replication process: 

Follow these steps to configure the replication flows: 

STEP1: Ensure that a BigQuery folder is ready to receive replicated tables.  

luv_sharma_1-1711958893578.png

 

 

STEP2: Create the CDS view to be replicated from S/4 HANA on-premise system. 

 

 

luv_sharma_2-1711958893583.png

 

 

STEP3: Head to SAP Datasphere's Data Builder and select "New Replication Flow." 

luv_sharma_3-1711958893587.jpeg

 

 

STEP4: Choose desired source views from the SAP system and a relevant target in BigQuery dataset. 

 

luv_sharma_4-1711958893592.png

 

 

luv_sharma_5-1711958893596.jpeg

 

 

luv_sharma_6-1711958893598.jpeg

 

 

luv_sharma_7-1711958893600.jpeg

 

 

STEP 5: Set load type, configure, save, and deploy. 

 

You can configure the settings for both Source and target systems. The Replication Thread Limit refers to the no. of parallel jobs that the replication flow can run. As per best practises, the thread limit on both the source and target system setting should be equal.   

luv_sharma_8-1711958893603.jpeg

 

 

luv_sharma_9-1711958893605.png

 

 

luv_sharma_10-1711958893606.png

 

 

STEP 6: You can monitor the execution process from Tools section. 

 

luv_sharma_11-1711958893608.png

 

 

luv_sharma_12-1711958893611.png

 

 

Step 7: Head over to the Google BigQuery folder which was selected in step 3 to verify. 

 

luv_sharma_13-1711958893614.jpeg

 

 

 

RESULT: Congratulations on setting up replication from SAP S4 HANA Cloud to Google BigQuery! 

 

What happens behind the scenes is that the data from source system is written in a component called Data Buffer in the form of packages. Datasphere reads these data packages and transfers it to the target system. Once all the packages have been transferred to the target system, a commit message is sent to the source system indicating that all the packages have been transferred and the data buffer can now be dropped. 

luv_sharma_14-1711958893617.png

Additional considerations: 

  • Premium outbound integration is necessary to replicate data to a non-SAP target. A Datasphere admin must assign capacity units to Premium Outbound Integration. 
  • Google BigQuery does not support the data types DECFLOAT16, DECFLOAT34 and UINT64. The system automatically converts it to DECIMAL (20,0).  
  • In a scenario where a target structure already exists in BigQuery, changes such as renaming a column or changing a data type must take place directly in BigQuery. 
  • CDS View not having the "dataExtraction" annotation are not supported for data transfer. Please refer to SAP note 3343889 for more information. 

 

3 Comments
Seanosky
Explorer
0 Kudos

Very insightful post @luv_sharma ! Thanks for sharing this step-by-step procedure!

Reza
Participant
0 Kudos

Thank you for sharing the procedure it really helps!

jj
Active Contributor
0 Kudos

Good one 👍

Labels in this area