Technology Blogs by Members
Explore a vibrant mix of technical expertise, industry insights, and tech buzz in member blogs covering SAP products, technology, and events. Get in the mix!
cancel
Showing results for 
Search instead for 
Did you mean: 
RonaldKonijnenb
Contributor
In October 2018, SAP announced Data Hub 2.3. In the below video,I will take you through the new features that this release brings and will also demonstrate the most important functionality.

The video will demonstrate the following:

  • Connecting with BW, HANA and an Azure Data lake

  • Data profiling and indexing

  • Connecting and loading data to HANA. In my example I will use a dataset which you can find in this excellent blog

  • Create an ETL process, using the Azure Data Lake and load the results back to the Data Lake

  • Report the data in SAP Analytics Cloud


In the demo I will use a very extensive data set which is made available by "Berkeley Earth":

The Berkeley Earth's Surface Temperature Study combines 1.6 billion temperature reports from 16 pre-existing archives. Date: starts in 1750 for average land temperature and 1850 for max and minus land and global ocean and land temperatures

I will enrich this data with GEO information and then visualize the new data set in SAP Analytics Cloud to see if we can indeed proof that temperatures are rising.



Enjoy!
4 Comments
schneidertho
Advisor
Advisor
Hi Ronald,

nice blog/video. Cheers

Thorsten
anil_vetsa
Advisor
Advisor
Excellent Demo.

Just a quick one from my side- what is the precedure for connecting SAC & DATAHUB .

Are there any out of the box connectors are available ?

Thank you

 
korayyersel_q
Participant
0 Kudos
Good question. From the description my guess is that SAC consumes the data Azure Data lake. At the end this is where the final data is persisted
Create an ETL process, using the Azure Data Lake and load the results back to the Data Lake

 
RonaldKonijnenb
Contributor
0 Kudos
In the future (at least that was the story before) you can send your data into a SAC data model.

 

For now, you could send it to ADL but need a 3rd party connector to connect to it (APOS)
Labels in this area