Azure Data Factory is a managed cloud data integration service. With this service, we can create automated pipelines to transform, analyze data, and much more.

Create a Datafactory Project

In the following steps, we will create a Data Factory project with one pipeline with a copy data activity and two datasets, a dynamics entity and an azuresqltable.

Step 1. Search for the Azure Datafactory service

Search for ‘Data factory’ service and click Enter.

Step 2. Create the service

Select Create to begin the creation of the service.

In the “New data factory” form, we must fill up all the mandatory fields.

Setting Value
Name Type a valid name for the Data Factory service
Subscription Select a valid Subscription
Resource Group Select an existing or create a new Resource Group
Version Select the Data Factory version, V2
Location Select a Location for the Data Factory service
Git Hub (If Enabled Then Needs Configuration)
GIT URL Type the URL of an existing Git Hub repository. For ex. https://github.com/myusername or https://myAccount.visualstudio.com/myProject
Repo name Type the name of an existing repository.
Branch name This is the name of an existing Git branch to use for collaboration, usually is master
Root folder Folder in the collaboration branch where Factory’s entities would be stored. For ex: ‘/factorydata’. ‘/’ would indicate the root folder.

Create a Pipeline

When the Data Factory deployment is completed we can start to deploy the pipeline. From the left main blade select Overview – Auditor & Monitor.

Step 1. Create pipeline

A new page will open and begin the deployment of the pipeline we have to select Create pipeline.

Step 2. Add Copy Data activity

From the Activities Search box, type “copy”, and drag & drop the Copy Data activity.

Step 3. Configure the Source parameters

Then, select Source and from the Source dataset click +New, to add a Dynamics Data Store

In the search box type “Dynamics”, select Dynamics 365 and click Continue.

Click +New to add Linked service

Type a Name for the New Linked Service, e.g Dynamics1 (default), type the Service UriUsernamePassword and then select Finish.

At this point we can Test connection and if works then select an Entity name, e.g Account.

Select Schema and click Import schema, at this point we can keep or remove the unnecessary fields that we want to get data from Dynamics.

When we complete the changes, we wait for a few seconds and select the button Publish All.

Step 4. Configure the Sink parameters

Type a Name for the New Linked Service, e.g AzureSqlDatabase1 (default), select a valid Azure Subscription, a Server name, a Database name, type UsernamePassword and then select Finish.

In the search box type “Azure SQL Database”, select Azure SQL Database and click Continue.

Next, we have to create the Dataset, type a Name, e.g AzureSqlTable1 (default), select Linked serviceTable, click Import schema {From connection/store} and select Finish.

Select Schema, and click Import schema.

After a few seconds select Publish All.

Test the pipeline

To test the pipeline, click Add trigger – Trigger Now.

After the Trigger executes the pipeline we must make sure that the execution was succeeded.

To make sure that the pipeline worked, open an SSMS and run a Select statement from the Azure SQL Database table (Accounts).

Schedule the Trigger

To schedule the Trigger run periodically, click Add trigger – New/Edit.

In the Add Triggers page, choose +New

In the New Trigger page, type a Name, e.g trigger1 (default), leave the Type as Schedule (default), specify a Start Date (UTC) the Recurrence (Minutes, Hours, Days, Weeks, Months ). To execute the trigger on a specific time, from the Advanced recurrence options, set Hours and Minutes. Before click Next, we have to be sure that the trigger is Activated.

Select Finish, to save the scheduled task, and Publish it.

Conclusion

By using two simple connectors for the datasets we create a copy of dynamics entity to an Azure SQL table. This can be useful if we want to transform the data for a custom report e.g Power BI, Azure SQL, etc. Of course, their many other solutions that we can deploy using this service.

Share This