Bring data to Microsoft Fabric Onelake from D365F&O (no Azure SQL needed)

 

Hi freinds, in one of recent blogs, I chalked out an integration stratgey pattern of setting up of D365F&O with Microsoft Fabric, by routing it through a DMF batch job to Azure SQL. This currrent blog talks about even a cool hack of direct integration between D365F&O and Microsoft Fabric, just harnessing Synapse Link itself -- no additional/intermeidary steps are required.

Here are the steps:

Step 1: Create a Datalake Storage account in Azure Portal, with the following settings:

 Don't forget to mark the following setting, from the advanced tab:

Click Review + create >> Create to complete the wizard to create the Storage account.

Step 2: Come to PowerApps page (https://make.powerapps.com/environments), and check if you are able to view the  Azure Synapse Link:
If not, then you have to click on More >> Discover all >> You should be able to see it as under the Data management:

Ensure that you are on the correct environment:


Step 3:
Once this is done click on Tables >> and choose Azure Fabric from the highlighted Link:

Allow it some time to spin up from the capacity planning and the following screen will load up:


Click Save connection.
Make sure that you choose the same region for your Azure Fabric as you have selected for Capacity planning for the Microsoft fabric.
 
Step 4: 
In the next step, you need to select the Workspace which you have created for the Microsoft Fabric:

Step-5: 
In the next step, it will create all the needful of setting up connects with Onelake and connects with Microsoft Fabric:

Step 5:
It will appear as a link as under Azure Synapse Link:

Just click on this link: it will also show up under the D365 Financed & Operations tab:
For our demo purpose, I am just selecting Item Group. Click Save and allow it some time for queuing and initial Synch.

Once completed, the data will automaticlally show under Onelake Datahub:


Whew -- that's it. Your data is now in Microsoft Fabric, without doing anything as an extra step 👲👲
What's next? Power up your BI reports, prepare Spark Jobs, route it Azure Data modelling pieplines, to analyse it further.

With that, I am taking your leave, and would come back with more ideas on Data engineering, Azure and AI/ML studios. Much love and namaste 💓💓💓

Comments

Popular posts from this blog

Make your menu items visible on main menu, conditionally,, using this cool feature of D365FO

Are you still using macros? Be sure you read this.

Modifying a DMF data entity query dynamically, from another query, for filteration in D365FO