Greetings & Welcome to Microsoft Q&A forum! Thanks for posting your query!
As I understand that you want to modify your existing Azure Data Factory pipeline to copy data from your on-premises SQL Server to Delta Lake format in your Blob Storage Gen2 (Hot storage). While ADF doesn't directly handle this conversion, we can achieve it using Azure Synapse Notebooks.
Here is the step-by-step procedure to achieve your requirement.
Maintain Existing Copy Activity configuration (Parquet Format): Create a pipeline in Azure Synapse Analytics that copies data from the on-premises SQL Server to Parquet format in your Blob Storage.
Add Synapse Notebook Activity for Conversion: After the copy activity, add a new Synapse Notebook activity. This notebook will handle the Parquet to Delta Lake conversion.
Create Conversion Code in Synapse Notebook: The notebook will use Spark SQL to read the Parquet files and write them as Delta Lake tables. Here is the conversion code.
Note: Please replace your storage details and file paths.
I have tried to repro the scenario, please refer to the gif file for more details.
I hope this information helps. Please do let us know if you have any further queries.
If this answers your query, do click Accept Answer
and Yes
for was this answer helpful. And, if you have any further query do let us know.