ADF Project - 1
ADF Project - 1
From the On-premise we will move the raw data to cloud using ADF pipeline
Here we have 3 layers Raw, Processed Data, Predefined data.( in data lake storage gen2)
Dataflows will read the data from raw layer, there it will check the duplicate and everything. If data is
good then it will move that to processed data.
From the Processed data will join some data and move it to refined data.
From refined data some of the data will move to azure synapse analytics.
Create tables
Create a connection strings (2 Linked services) for SQL and azure blob. Here we need to use the
keyvaults option.
Create pipeline – lookup activity – write a query select * from table where status<>’succeeded’.
Pass this output to foreach activity and inside this use copy activity – new data set and create
parameter(source schema and source table)
Then sink create dataset referring to Data Lake Storage Gen2.
Then pass this values to stored procedure success and failure. Along with failure stored procedure need
to add web activity
Create logic app( logic app designer – http – gmail – send email.
Here in this scenario it will through error for each time it fails, so to over come this we can do one this is
we need to place the web activity after the for each activity.
One more thing can do is we can add stored procedure and web activity after the for each.
Explanation
Open a Data flow – select source and sink – windows – turn on the dataflow debug once it is done -
import
In data flow – windows ( used to check duplicate) select each column – add filter then add condition
select rownumber is equal to 1 – select then remove row number column.
After doing it for all the tables create pipeline processed to refined – create new data flow add two
tables and join them.
Then create pipeline for loading refined data into a synapse analytics
Then create new pipeline as master pipeline and add all the pipelines then add trigger