Activity
Mon
Wed
Fri
Sun
Mar
Apr
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
Jan
Feb
What is this?
Less
More

Owned by Will

Fabric Dojo ็ป‡็‰ฉ

349 members โ€ข $67/month

Build robust data engineering solutions with Microsoft Fabric & get hired as a Fabric Data Engineer.

Fabric AI Workflows

54 members โ€ข $24/month

Ultimate environment to learn, share & engage with the Microsoft Fabric AI Workflows that are changing the enterprise analytics game - forever.

Memberships

Power BI Park

261 members โ€ข $69/m

Azure Innovation Station

809 members โ€ข Free

1496 contributions to Learn Microsoft Fabric
Power BI to MS Fabric Migration
Hello i am new Power BI and have to migrate Power BI to MS Fabric. What user permissions needed for the migration is workspace admin enough or we need even tenant admin as well. once the tenant of the MS fabric created just reassigning the workspace from Power BI MS fabric will do ? or do we have any step by step documents which can help me to achieve this. Also i do have some of the workspace where i am the ownder and for am not the owner. How do i identify who is owner when i do see the contact details for that workspace. what all are the pre checks i need to take care
0 likes โ€ข 1h
Hi Shobha, this depends on what you mean by migrating Power BI to MS Fabric. Power BI is part of Fabric - it can be as simple as changing the assigned capacity of the workspace from a Power BI license to a Fabric Capacity. For this you will need to be Workspace Admin, and Capacity Contributor (in Azure), as a minimum.
Direct Lake model stuck with SQL endpoint connection
Hi, I have a model that should be direct lake but when looking at the Gateway and Cloud connections in settings is it using a SQL endpoint. Normally this would show a string containing AzureDataLakeStorage as my other direct lake models do. I'm using git integration so I could look at changing .tmdl configurations. Any suggestions appreciated.
0 likes โ€ข 8h
Hi there @Renรฉ Tachon you can use Power BI Desktop with TMDL live edit mode. The steps are shown in this article - scroll down to "Migrating an existing Direct Lake on SQL semantic model to Direct Lake on OneLake"
Table Column Encryption
Is there any way to encrypt the warehouse table column in Fabric?
1 like โ€ข 15h
Hi there Hiren, an individual column? I don't think that's possible. You can check out this in-depth overview of all encryption options in the Warehouse here
Dataverse n Fabric integration
Hi All, anyone dabbled with dataverse to fabric integration? Any gotchas that I would need to be wary of ?
1 like โ€ข 2d
This post / article might be relevant for you Mark: Gotchaโ€™s of bringing Dataverse data into Microsoft Fabric Also lots of other previous conversations via the search: https://www.skool.com/microsoft-fabric/-/search?q=dataverse It's not something I've needed to do though personally yet, so can't help first hand on that one!
1 like โ€ข 1d
@Carl Thompson thanks for sharing Carl!
API Pagination in Copy Data pipeline activity
I am working with an API that does pagination via bookmarks. In the response JSON file there are 2 string value pairs that are used for identifying more records and the parameter to pass in the next request URL. The JSON is structured like this. { "Items": [ { ... }, { ... } ], "Bookmark": "<B><P><p>job</p><p>suffix</p><p>site_ref</p></P><D><f>false</f><f>false</f><f>false</f></D><F><v> 1</v><v>0</v><v>BLGS</v></F><L><v> 783</v><v>0</v><v>BLGS</v></L></B>", "MoreRowsExist": true, "Success": true, "Message": null } When setting up the Pagination rules in the Copy Data activity, how can I pass the value in the Bookmark back to the Relative URL? My current Relative URL looks like this, @concat('IDORequestService/ido/load/',item().source_table,'?filter=',item().FILTER,'&loadtype=NEXT&Properties=',item().cols,'&recordcap=',pipeline().parameters.recordcount,'&bookmark=') I need to add the bookmark parameter at the very end. In Azure Data Factory it is done via a parameter set in the Linked Service but since that doesn't exist in Fabric I'm stuck at getting this to work. I can handle the pagination in a notebook easily, but I wanted to explore using a Data Pipeline to try a more Metadata driven approach.
API Pagination in Copy Data pipeline activity
0 likes โ€ข Jul '24
@Bob Samuelson Nice, well done for getting to a solution with that one Bob!
0 likes โ€ข 1d
@Damien O'Connor great input, thanks for sharing Damien!
1-10 of 1,496
Will Needham
8
23,257points to level up
@will-needham-6328
I simplify Microsoft Fabric for data professionals. Passionate about teaching Data & AI.

Active 14m ago
Joined Jan 15, 2024
Powered by