I would like to setup ADF pipeline in such a way that I need to load all the Parquet files hosted for 2+ years on ADLS Gen2 with a hierarchy of Year -> Month -> Day -> Hour - > Min. Over the period, we did have some file structure changes with a variance of 2-3 columns. I would like to pull all the common columns and load entire data in a sql table. Can someone please point me to the resources which could help with my requirement.
Thank you!
In the Azure data factory pipeline,
Get Metadata
activity to get the list of parquet files.ForEach
activity to loop each current item.You can refer to this document to copy data to the SQL table.