Hello Relay
To bring data from SAP Ariba and SAP HANA to Azure Data Lake Storage Gen2 (ADLS Gen2), you can use a combination of Azure Data Factory (ADF), SAP connectors because SAP HANA data can be copied to ADLS Gen2 using Azure Data Factory’s SAP HANA connector.
- In ADF, go to Manage > Linked Services > New.
- Select SAP HANA.
- Configure: Server: SAP HANA server address, Port: Default is 3XX15 (replace XX with instance number). Authentication: Use Basic (username/password) or Windows authentication. Integration Runtime: Choose Azure IR or self-hosted IR (for on-premises).
- Test the connection and save
-
Post which in ADF, create a dataset for SAP HANA.
- Select the SAP HANA linked service.
- Specify the schema and table or write a custom SQL query to extract data (e.g., from Analytic/Calculation views or Row/Column tables).
- Go to Linked Services > New > Azure Data Lake Storage Gen2.
- Configure: Storage Account Name: Select your ADLS Gen2 account. Authentication: Use Account Key, Service Principal, or Managed Identity. Test the connection and save.
-
Create a dataset for ADLS Gen2.
- Specify the file format (e.g., Parquet, CSV) and folder path (e.g., data/saphana/).
- Create a pipeline in ADF and add a Copy Activity.
- Set the SAP HANA dataset as the Source.
- Set the ADLS Gen2 dataset as the Sink.
SAP Ariba data is typically accessed via APIs (REST or SOAP) or exported reports. Since ADF does not have a native SAP Ariba connector, you can use the HTTP or REST connector or third-party tools like CData Sync.
- In ADF, go to Linked Services > New > REST.
- Configure: Base URL: SAP Ariba API endpoint (e.g., https://api.ariba.com/v2/), Authentication: OAuth 2.0 with Client Credentials flow. Enter Client ID, Client Secret, and token endpoint.
- Create a dataset for the REST service.
- Specify the relative URL and parameters for the API call (e.g., /reports for reporting data).
Post which Add a Copy Activity in an ADF pipeline.
- Set the REST dataset as the Source.
- Set the ADLS Gen2 dataset as the Sink.
- Map the API response fields to the ADLS Gen2 file structure.
- Use pagination rules if the API returns paginated data.
Use API filters (e.g., modifiedDate) to fetch only new or updated records. Store the last processed timestamp in a control table or ADLS Gen2 file to track increments.
References: https://learn.microsoft.com/en-us/azure/data-factory/connector-sap-hana?tabs=data-factory
Hope the above answer helps! Please let us know do you have any further queries.
Please do not forget to "Accept the answer” and “up-vote” wherever the information provided helps you, this can be beneficial to other community members.