<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>Forem: Reiji Otake</title>
    <description>The latest articles on Forem by Reiji Otake (@_d2a1ea24c442526a9777).</description>
    <link>https://forem.com/_d2a1ea24c442526a9777</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F2869291%2Ff59d12a5-f3e2-4e3c-8ebf-09e9758be4dc.jpg</url>
      <title>Forem: Reiji Otake</title>
      <link>https://forem.com/_d2a1ea24c442526a9777</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://forem.com/feed/_d2a1ea24c442526a9777"/>
    <language>en</language>
    <item>
      <title>[FabCon Atlanta 2026 Report] My Take on Fabric IQ Ontology</title>
      <dc:creator>Reiji Otake</dc:creator>
      <pubDate>Wed, 06 May 2026 23:03:45 +0000</pubDate>
      <link>https://forem.com/_d2a1ea24c442526a9777/fabcon-atlanta-2026-report-my-take-on-fabric-iq-ontology-4hp3</link>
      <guid>https://forem.com/_d2a1ea24c442526a9777/fabcon-atlanta-2026-report-my-take-on-fabric-iq-ontology-4hp3</guid>
      <description>&lt;p&gt;I attended &lt;strong&gt;FabCon Atlanta 2026&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;I also created a few short videos that show the atmosphere of the venue, so feel free to check them out first.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://youtube.com/shorts/EZ_yla-xrx8?si=3KB9uRHlrqR33dPp" rel="noopener noreferrer"&gt;FabCon Atlanta 2026 Day1-Day3 morning workshops 、KeyNone&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://www.youtube.com/shorts/mkqowCw-KF0" rel="noopener noreferrer"&gt;FabCon Atlanta 2026 Day3 noon-Day5CoreNote session power hour&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;In this article, based on what I saw and heard at FabCon, I would like to focus especially on &lt;strong&gt;Ontology&lt;/strong&gt; within Fabric IQ and share how I think we should understand it at this point in time.&lt;/p&gt;

&lt;p&gt;Fabric IQ is described as a workload that organizes data in OneLake using business language, enabling analytics and AI agents to use that data with consistent meaning.&lt;/p&gt;

&lt;p&gt;The Fabric IQ workload includes semantic models and Data Agents, and Ontology is one part of it.&lt;/p&gt;

&lt;p&gt;I think many people may currently understand “Fabric IQ” as almost the same thing as “Ontology.”&lt;/p&gt;

&lt;p&gt;That is not completely wrong. However, &lt;strong&gt;Fabric IQ is a broader term&lt;/strong&gt;, so in this article I will mainly use the word “Ontology” to avoid confusion.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fsd6np72xkctozxz90nnw.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fsd6np72xkctozxz90nnw.png" alt=" " width="701" height="286"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://learn.microsoft.com/en-us/fabric/iq/overview" rel="noopener noreferrer"&gt;What is Fabric IQ (preview)?&lt;/a&gt;&lt;/p&gt;

&lt;h1&gt;
  
  
  The Atmosphere Around Fabric IQ at FabCon
&lt;/h1&gt;

&lt;p&gt;At FabCon, I felt that everyone also highly interested in Fabric IQ.&lt;/p&gt;

&lt;p&gt;At the same time, some of the questions were very basic, such as “What is IQ?”&lt;/p&gt;

&lt;p&gt;In other words, my honest impression was that Fabric IQ is attracting a lot of attention, but even in the United States, understanding of it has not yet become widespread.&lt;/p&gt;

&lt;p&gt;I also attended several IQ-related sessions. Based on the sessions I joined, I cannot say that I clearly saw exactly which real-world projects should use it and how.&lt;/p&gt;

&lt;p&gt;Of course, there were Ontology demos, and there were discussions about how AI will be able to understand business meaning more easily and how the semantic layer will become more important. Officially, Ontology is also described as a way to represent a business in a machine-readable form through entities, properties, relationships, and rules.&lt;/p&gt;

&lt;p&gt;However, to be honest, my current impression is that &lt;strong&gt;the concept itself is very attractive, but common implementation patterns are not yet widely understood&lt;/strong&gt;.&lt;/p&gt;

&lt;h1&gt;
  
  
  My Conclusion First: I Would Still Take a Wait-and-See Approach for Production Use
&lt;/h1&gt;

&lt;p&gt;My conclusion is that, at this point, I would still take a &lt;strong&gt;wait-and-see approach before placing Ontology at the center of a production environment&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;The reason is simple.&lt;/p&gt;

&lt;p&gt;First, it is still officially in &lt;strong&gt;preview&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;Second, when it comes to improving the accuracy of Data Agents by giving them business context, I feel that many use cases can already be covered quite well by using &lt;strong&gt;semantic models&lt;/strong&gt;.&lt;/p&gt;




&lt;h2&gt;
  
  
  A Common Misunderstanding
&lt;/h2&gt;

&lt;p&gt;Ontology allows you to create entities as business objects and define relationships using natural language to represent business meaning.&lt;/p&gt;

&lt;p&gt;On the other hand, based on the current specification, you cannot simply write natural-language descriptions for tables and columns inside Ontology in the same way you can with semantic model properties.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fs1gjrqe2mim89cdt22v2.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fs1gjrqe2mim89cdt22v2.png" alt="image.png" width="800" height="449"&gt;&lt;/a&gt;&lt;/p&gt;




&lt;p&gt;Of course, I am not saying that Ontology is unnecessary.&lt;/p&gt;

&lt;p&gt;Rather, I believe Microsoft will continue to invest heavily in this area, and I personally have high expectations for it.&lt;/p&gt;

&lt;p&gt;However, at least for now, I think the right stage is:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Development teams should try it in a test environment&lt;/li&gt;
&lt;li&gt;Organizations should watch it as a future architecture option&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;On the other hand, I think it is still a little early to talk about adopting it broadly in production right away.&lt;/p&gt;

&lt;h1&gt;
  
  
  Semantic Models Will Continue to Be Important for AI
&lt;/h1&gt;

&lt;p&gt;So, does that mean the semantic layer is still something for the future?&lt;/p&gt;

&lt;p&gt;I do not think so.&lt;/p&gt;

&lt;p&gt;Rather, even right now, building a well-designed semantic model is very effective. I also believe that &lt;strong&gt;even after Ontology becomes generally available in the future, the importance of semantic models will not disappear&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;Officially, Ontology can be generated from semantic models. In other words, it feels more natural to see Ontology not as something that replaces semantic models, but as something that extends business meaning and relationships on top of semantic models as one of its foundations.&lt;/p&gt;

&lt;h2&gt;
  
  
  What Semantic Models Can Already Do Today
&lt;/h2&gt;

&lt;p&gt;With the arrival of Data Agent, semantic models are no longer just models for BI.&lt;/p&gt;

&lt;p&gt;You can specify a semantic model as a data source for a Data Agent, and through Data Agent customization, you can provide business metadata to AI.&lt;/p&gt;

&lt;p&gt;For example:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;p&gt;Semantic model&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Use the “Prep for AI” feature&lt;/li&gt;
&lt;li&gt;Write the business meaning of tables and columns in properties such as table names, column names, table descriptions, and column descriptions&lt;/li&gt;
&lt;li&gt;Predefine calculations and business logic with DAX&lt;/li&gt;
&lt;/ul&gt;


&lt;/li&gt;

&lt;li&gt;

&lt;p&gt;Data Agent&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Clarify the role of the agent through instructions&lt;/li&gt;
&lt;li&gt;Add descriptions for data sources so that the agent can choose the right source depending on the question&lt;/li&gt;
&lt;li&gt;Use example query sets for expected questions

&lt;ul&gt;
&lt;li&gt;Note: this is not available for semantic models&lt;/li&gt;
&lt;/ul&gt;


&lt;/li&gt;

&lt;/ul&gt;

&lt;/li&gt;

&lt;/ul&gt;

&lt;p&gt;For more details, I recommend starting with the following documentation.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://learn.microsoft.com/en-us/fabric/data-science/semantic-model-best-practices" rel="noopener noreferrer"&gt;Semantic model best practices for data agent&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://learn.microsoft.com/en-us/fabric/data-science/data-agent-configuration-best-practices" rel="noopener noreferrer"&gt;Best practices for configuring your data agent&lt;/a&gt;&lt;/p&gt;




&lt;p&gt;Also, a Data Agent does not necessarily need to have only one data source.&lt;/p&gt;

&lt;p&gt;When the data volume is large, or when you want to use example query sets, combining a semantic model with a lakehouse or warehouse can be a very realistic design.&lt;/p&gt;

&lt;p&gt;For example:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Store large volumes of data in a lakehouse or warehouse&lt;/li&gt;
&lt;li&gt;Organize the metrics and definitions you want AI to use in a semantic model&lt;/li&gt;
&lt;/ul&gt;




&lt;p&gt;If you want to add business metadata to each table or column, my personal recommendation at this point is to write it in the semantic model properties.&lt;/p&gt;

&lt;p&gt;Data Agent can refer to semantic model properties.&lt;/p&gt;

&lt;p&gt;Related article:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://qiita.com/ReijiOtake/items/a06dcc9967121d652282" rel="noopener noreferrer"&gt;Editing Semantic Model Metadata Properties from a Notebook with Semantic Link in Fabric&lt;/a&gt;&lt;/p&gt;

&lt;h1&gt;
  
  
  When Would Ontology Become Necessary?
&lt;/h1&gt;

&lt;p&gt;At this point, you might think, “Then isn’t a semantic model enough?”&lt;/p&gt;

&lt;p&gt;In fact, I think semantic models can cover a large part of many use cases.&lt;/p&gt;

&lt;p&gt;That said, based on my current understanding, I feel that Ontology becomes especially useful in the following two scenarios.&lt;/p&gt;

&lt;p&gt;In other words, if your use case does not fall into these two patterns, a semantic model may be enough for now.&lt;/p&gt;

&lt;h2&gt;
  
  
  1. When You Want to Query Across Multi-Layered Relationships Like a Graph
&lt;/h2&gt;

&lt;p&gt;The first case is when you want to ask questions that go across multiple layers of relationships.&lt;/p&gt;

&lt;p&gt;Semantic models can also express relationships. However, as the relationships become more complex, the thinking tends to become more JOIN-oriented.&lt;/p&gt;

&lt;p&gt;Ontology, on the other hand, uses a graph-based approach, so it seems better suited to graph-like operations such as path exploration.&lt;/p&gt;

&lt;p&gt;For example, imagine you have the following tables:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Customers&lt;/li&gt;
&lt;li&gt;Orders&lt;/li&gt;
&lt;li&gt;Products&lt;/li&gt;
&lt;li&gt;Contracts&lt;/li&gt;
&lt;li&gt;Support history&lt;/li&gt;
&lt;li&gt;Responsible organizations&lt;/li&gt;
&lt;li&gt;Related events&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;If you want to ask, “What is related to this customer?” across multiple business domains, Ontology seems like a more natural way to express that.&lt;/p&gt;

&lt;p&gt;In other words, Ontology becomes meaningful when &lt;strong&gt;the relationships themselves are valuable&lt;/strong&gt;, rather than when you only need simple aggregations or KPI questions.&lt;/p&gt;

&lt;h2&gt;
  
  
  2. When You Want to Treat Historical Data and Real-Time Data as One Business Entity
&lt;/h2&gt;

&lt;p&gt;The second case is when you want to treat historical data and real-time data not as separate systems, but as the same business object.&lt;/p&gt;

&lt;p&gt;Officially, Fabric IQ is described as a way to unify data in OneLake using business language and give consistent meaning to analytics and AI agents.&lt;/p&gt;

&lt;p&gt;For example:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Recent order events stored in Eventhouse&lt;/li&gt;
&lt;li&gt;Historical order data accumulated in Lakehouse&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;If you want to handle these together in the context of a single business entity such as “Order,” the idea of Ontology seems to be a very good fit.&lt;/p&gt;

&lt;p&gt;This feels less like a simple BI model, or physical model, and more like a foundation that helps AI understand the meaning structure of the business, in other words, a logical model.&lt;/p&gt;

&lt;h1&gt;
  
  
  We Do Not Need to Rush Ontology. For Now, This Is a Preparation Phase
&lt;/h1&gt;

&lt;p&gt;As I have written so far, I believe Ontology has great potential.&lt;/p&gt;

&lt;p&gt;However, I personally do not think it is something that must be introduced as the highest priority right now.&lt;/p&gt;

&lt;p&gt;Ontology can be seen as a mechanism for &lt;strong&gt;strengthening the business meaning layer afterward&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;Therefore, rather than seeing it as a foundation that must be introduced from the beginning, it feels more natural to think of it as something that organizations can add after their data platform and semantic organization have reached a certain level of maturity.&lt;/p&gt;

&lt;p&gt;In fact, even if you want to use Ontology, there will likely be many cases where the organization’s data itself is not yet ready.&lt;/p&gt;

&lt;p&gt;For example:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Required tables do not exist&lt;/li&gt;
&lt;li&gt;Key definitions and meanings differ across systems&lt;/li&gt;
&lt;li&gt;Tables that should be related cannot be connected cleanly through relationships&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;In such a state, the problem exists before Ontology can even be built.&lt;/p&gt;

&lt;p&gt;That is why I believe the most important thing right now is to &lt;strong&gt;prepare and organize the organization’s data&lt;/strong&gt; so that it can take advantage of Ontology in the future.&lt;/p&gt;

&lt;p&gt;Microsoft will likely continue to invest heavily in this area, and the concept of Ontology itself will become increasingly important.&lt;/p&gt;

&lt;p&gt;In that sense, I think we should see the current phase not as “the time to rush Ontology into production,” but as &lt;strong&gt;a preparation period for creating the conditions where Ontology can be used effectively&lt;/strong&gt;.&lt;/p&gt;




&lt;p&gt;In addition, I also feel that &lt;strong&gt;building Ontology requires a surprisingly high level of skill&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;It is not enough to have only data modeling knowledge.&lt;/p&gt;

&lt;p&gt;You need both:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;An understanding of the &lt;strong&gt;business meaning&lt;/strong&gt; behind the organization’s operations and data&lt;/li&gt;
&lt;li&gt;The &lt;strong&gt;data modeling knowledge&lt;/strong&gt; required to turn that meaning into a structure&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;In other words, Ontology cannot be built only by the IT department.&lt;/p&gt;

&lt;p&gt;At the same time, it also cannot be fully defined only by the business department.&lt;/p&gt;

&lt;p&gt;Collaboration between IT and business will be important, and people who understand both sides to some extent will become increasingly valuable.&lt;/p&gt;

&lt;h1&gt;
  
  
  Bonus 1: Foundry IQ Already Feels More Practical
&lt;/h1&gt;

&lt;p&gt;As a side note, based on my experience, Foundry IQ felt more practical at this point.&lt;/p&gt;

&lt;p&gt;For example, use cases such as the following are relatively easy to imagine even now:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Using OneLake as a knowledge source&lt;/li&gt;
&lt;li&gt;Using SharePoint as a knowledge source&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Fabric Ontology still looks like something that may become very interesting in the future.&lt;/p&gt;

&lt;p&gt;On the other hand, Foundry IQ already feels easier to connect to concrete use cases.&lt;/p&gt;

&lt;p&gt;Of course, these two are not competitors. I believe they will become more connected over time.&lt;/p&gt;

&lt;h1&gt;
  
  
  Bonus 2: Data Agent Development Works Well with CI/CD and Should Use Git Integration
&lt;/h1&gt;

&lt;p&gt;This is slightly separate from Ontology, but through FabCon, I was reminded again that &lt;strong&gt;Data Agent works very well with CI/CD&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;Are you using Git integration in Fabric?&lt;/p&gt;

&lt;p&gt;As mentioned earlier, when developing a Data Agent, you define items such as instructions, data source descriptions, and example query sets.&lt;/p&gt;

&lt;p&gt;Among these, data source descriptions may not change very frequently.&lt;/p&gt;

&lt;p&gt;However, I feel that instructions and example query sets are things that will continue to evolve once the agent starts being used.&lt;/p&gt;

&lt;p&gt;For example, in actual operation, the following situations are likely to happen:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;A user asks an unexpected question, and you want to add a query set for that pattern&lt;/li&gt;
&lt;li&gt;You adjust the instruction prompt, but the accuracy becomes worse&lt;/li&gt;
&lt;li&gt;You want to roll back to a previous version and check the behavior&lt;/li&gt;
&lt;li&gt;You want to compare the previous version and the latest version while testing&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;In other words, a Data Agent is not something you configure once and forget.&lt;/p&gt;

&lt;p&gt;It is something that should be &lt;strong&gt;continuously improved during operation&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;That is why it works very well with Git integration, where you can manage change history, track differences, and roll back when necessary.&lt;/p&gt;

&lt;p&gt;If you want to use Data Agent seriously in Fabric, I believe it is important not only to create the agent, but also to grow it with Git integration in mind.&lt;/p&gt;

&lt;p&gt;Related articles:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://qiita.com/ReijiOtake/items/5d37c75f3c9e753a131d" rel="noopener noreferrer"&gt;Microsoft Fabric Git Integration × Azure DevOps: How to Release Fabric Items Across Different Tenants&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://qiita.com/ReijiOtake/items/3a7d4acf36065023bd2d" rel="noopener noreferrer"&gt;How to Reflect Changes to Another Repository with Azure DevOps Pipeline: A Minimal Memo for Repo A → Repo B&lt;/a&gt;&lt;/p&gt;

&lt;h1&gt;
  
  
  Summary
&lt;/h1&gt;

&lt;p&gt;Finally, here is my current understanding.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Expectations for Fabric IQ / Ontology are high&lt;/li&gt;
&lt;li&gt;However, it is still in preview, so I would be cautious about using it in production at this stage&lt;/li&gt;
&lt;li&gt;In many cases, the combination of semantic models and Data Agent is already quite effective&lt;/li&gt;
&lt;li&gt;Ontology will become especially useful in scenarios such as:

&lt;ul&gt;
&lt;li&gt;Queries across multi-layered relationships&lt;/li&gt;
&lt;li&gt;Use cases where accumulated data and real-time data need to be handled in one business context&lt;/li&gt;
&lt;/ul&gt;


&lt;/li&gt;

&lt;/ul&gt;

&lt;p&gt;I believe this is definitely an area where Microsoft will continue to invest.&lt;/p&gt;

&lt;p&gt;Therefore, now is a good time to catch up on Ontology and prepare your organization’s data platform so that you can adopt it quickly when the right timing comes.&lt;/p&gt;

&lt;p&gt;Thank you for reading this long article!&lt;/p&gt;

&lt;h1&gt;
  
  
  I Also Have a YouTube Channel!
&lt;/h1&gt;

&lt;p&gt;&lt;a href="https://www.youtube.com/@msfabricreijiotake" rel="noopener noreferrer"&gt;https://www.youtube.com/@msfabricreijiotake&lt;/a&gt;&lt;/p&gt;

</description>
      <category>ai</category>
      <category>analytics</category>
      <category>data</category>
      <category>microsoft</category>
    </item>
    <item>
      <title>Fabric &amp; Databricks Interoperability (4): Using Databricks Tables in Fabric for Viewing, Analysis, and Editing</title>
      <dc:creator>Reiji Otake</dc:creator>
      <pubDate>Sun, 16 Feb 2025 10:52:27 +0000</pubDate>
      <link>https://forem.com/_d2a1ea24c442526a9777/fabric-databricks-interoperability-4-using-databricks-tables-in-fabric-for-viewing-analysis-2n2m</link>
      <guid>https://forem.com/_d2a1ea24c442526a9777/fabric-databricks-interoperability-4-using-databricks-tables-in-fabric-for-viewing-analysis-2n2m</guid>
      <description>&lt;h1&gt;
  
  
  Introduction
&lt;/h1&gt;

&lt;p&gt;Is it possible to seamlessly reference and edit tables created in Fabric within Databricks?&lt;br&gt;&lt;br&gt;
Many people may have this question.&lt;/p&gt;

&lt;p&gt;In this article, we will specifically explore the use case of:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Utilizing tables created in Databricks within Fabric.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;For prerequisite settings and configurations, please refer to previous articles.&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;This article is part of a four-part series:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;&lt;a href="https://qiita.com/ReijiOtake/items/48c7b1e54796f4f569f3" rel="noopener noreferrer"&gt;Overview and Purpose of Interoperability&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://qiita.com/ReijiOtake/items/7ce8e1a743d05efe925c" rel="noopener noreferrer"&gt;Detailed Configuration of Hub Storage&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://qiita.com/ReijiOtake/items/86b44b2c30986c65db08" rel="noopener noreferrer"&gt;Using Tables Created in Fabric in Databricks&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Using Tables Created in Databricks in Fabric (this article)&lt;/li&gt;
&lt;/ol&gt;
&lt;/blockquote&gt;
&lt;h1&gt;
  
  
  Linking Tables Created in Databricks to Fabric
&lt;/h1&gt;
&lt;h3&gt;
  
  
  Creating a New Table in Databricks
&lt;/h3&gt;

&lt;p&gt;Create a new empty external table from Databricks.&lt;br&gt;&lt;br&gt;
Specify the folder path of the hub storage as the &lt;code&gt;Location&lt;/code&gt;.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F9gw6t5m71lji56wl67td.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F9gw6t5m71lji56wl67td.png" alt="image.png" width="800" height="183"&gt;&lt;/a&gt;&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight sql"&gt;&lt;code&gt;&lt;span class="k"&gt;CREATE&lt;/span&gt; &lt;span class="k"&gt;TABLE&lt;/span&gt; &lt;span class="n"&gt;create_from_Databricks_sales&lt;/span&gt;
&lt;span class="k"&gt;USING&lt;/span&gt; &lt;span class="n"&gt;DELTA&lt;/span&gt;
&lt;span class="k"&gt;LOCATION&lt;/span&gt; &lt;span class="s1"&gt;'abfss://&amp;lt;container_name&amp;gt;@&amp;lt;ADLS2_name&amp;gt;.dfs.core.windows.net/folder_name/create_from_Databricks_sales'&lt;/span&gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;

&lt;h1&gt;
  
  
  Checking the Created Table
&lt;/h1&gt;

&lt;p&gt;You can verify that the external table created from the Catalog Explorer contains data.&lt;br&gt;
&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F12p0agyv80gwm8ruy3oe.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F12p0agyv80gwm8ruy3oe.png" alt="image.png" width="800" height="362"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;A folder named &lt;code&gt;create_from_Databricks_sales&lt;/code&gt; is created in the &lt;code&gt;ext&lt;/code&gt; folder of the hub storage.&lt;br&gt;
(This means that the newly created external table physically exists in the hub storage.)&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fi5mqmy10qlypxwbotd56.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fi5mqmy10qlypxwbotd56.png" alt="image.png" width="800" height="200"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;It can also be confirmed that the table is in Delta format.&lt;br&gt;
&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fjkcb8byqyzrt10yh5m99.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fjkcb8byqyzrt10yh5m99.png" alt="image.png" width="800" height="165"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;At this point, the &lt;code&gt;create_from_Databricks_sales&lt;/code&gt; table also becomes visible from Fabric's Lakehouse.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F5vbjfzqstq9km8y11su5.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F5vbjfzqstq9km8y11su5.png" alt="image.png" width="800" height="298"&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h1&gt;
  
  
  Viewing and Analyzing Tables Created in Databricks in Fabric (Creating BI)
&lt;/h1&gt;

&lt;p&gt;From the Semantic Model, select the &lt;code&gt;create_from_Databricks_sales&lt;/code&gt; table (created in Databricks) and click &lt;strong&gt;[Confirm]&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fsi9a6ftlv5iw6fz8m0zv.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fsi9a6ftlv5iw6fz8m0zv.png" alt="image.png" width="800" height="477"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Now, the table created in Databricks can be analyzed in Fabric.&lt;/strong&gt;&lt;br&gt;
&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F6byru9bnamh8n4hepg4f.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F6byru9bnamh8n4hepg4f.png" alt="image.png" width="800" height="374"&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h1&gt;
  
  
  Editing (DML) Tables Created in Databricks from Fabric
&lt;/h1&gt;

&lt;p&gt;Execute an &lt;code&gt;UPDATE&lt;/code&gt; statement (DML statement) from Fabric's Notebook.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fwml9wx811uvq2v5jb7a4.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fwml9wx811uvq2v5jb7a4.png" alt="image.png" width="800" height="320"&gt;&lt;/a&gt;&lt;br&gt;
&lt;/p&gt;
&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight sql"&gt;&lt;code&gt;&lt;span class="k"&gt;UPDATE&lt;/span&gt; &lt;span class="n"&gt;Fabric_Lakehouse&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;ext&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;create_from_Databricks_sales&lt;/span&gt;
&lt;span class="k"&gt;SET&lt;/span&gt; &lt;span class="n"&gt;Item&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="s1"&gt;'No.1 Quantity Water Bottle - 30 oz.'&lt;/span&gt;
&lt;span class="k"&gt;WHERE&lt;/span&gt; &lt;span class="n"&gt;Item&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="s1"&gt;'Water Bottle - 30 oz.'&lt;/span&gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;


&lt;p&gt;Of course, it was confirmed that changes were reflected from Fabric.&lt;br&gt;&lt;br&gt;
&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F3dmp5cfgs18eifpmgat5.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F3dmp5cfgs18eifpmgat5.png" alt="image.png" width="800" height="405"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Editing was performed from Fabric, and the changes were also reflected on the Databricks side.&lt;br&gt;&lt;br&gt;
&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fmlpsnk1519v4v2xko504.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fmlpsnk1519v4v2xko504.png" alt="image.png" width="800" height="348"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Therefore, it is possible to edit (DML statements) in Fabric for tables created in Databricks.&lt;/strong&gt;&lt;/p&gt;
&lt;h1&gt;
  
  
  Issues and Specific Operational Methods
&lt;/h1&gt;

&lt;p&gt;The method introduced here has the advantage that both Fabric and Databricks can edit data. However, this can also be a weakness, as it makes table updates too easy.&lt;/p&gt;

&lt;p&gt;Additionally, in this case, an external table in Databricks was used.&lt;br&gt;&lt;br&gt;
However, predictive optimization is currently only available for managed tables, making it ideal to use managed tables rather than external ones.&lt;/p&gt;


&lt;div class="crayons-card c-embed text-styles text-styles--secondary"&gt;
    &lt;div class="c-embed__content"&gt;
        &lt;div class="c-embed__cover"&gt;
          &lt;a href="https://docs.databricks.com/aws/en/optimizations/predictive-optimization" class="c-link align-middle" rel="noopener noreferrer"&gt;
            &lt;img alt="" src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdocs.databricks.com%2Faws%2Fen%2Fimg%2Fog-image.png" height="420" class="m-0" width="800"&gt;
          &lt;/a&gt;
        &lt;/div&gt;
      &lt;div class="c-embed__body"&gt;
        &lt;h2 class="fs-xl lh-tight"&gt;
          &lt;a href="https://docs.databricks.com/aws/en/optimizations/predictive-optimization" rel="noopener noreferrer" class="c-link"&gt;
            Predictive optimization for Unity Catalog managed tables | Databricks on AWS
          &lt;/a&gt;
        &lt;/h2&gt;
          &lt;p class="truncate-at-3"&gt;
            Learn how predictive optimization improves data layout and query performance for Unity Catalog managed tables on Databricks.
          &lt;/p&gt;
        &lt;div class="color-secondary fs-s flex items-center"&gt;
            &lt;img alt="favicon" class="c-embed__favicon m-0 mr-2 radius-0" src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdocs.databricks.com%2Faws%2Fen%2Ffavicon.ico" width="64" height="64"&gt;
          docs.databricks.com
        &lt;/div&gt;
      &lt;/div&gt;
    &lt;/div&gt;
&lt;/div&gt;



&lt;p&gt;We will continue to examine these challenges and share specific operational methods in the future.&lt;/p&gt;

&lt;p&gt;I also think that table cloning in Databricks might provide some useful hints.&lt;/p&gt;


&lt;div class="crayons-card c-embed text-styles text-styles--secondary"&gt;
    &lt;div class="c-embed__content"&gt;
        &lt;div class="c-embed__cover"&gt;
          &lt;a href="https://qiita.com/taka_yayoi/items/50c5a75caff8a6fb721d" class="c-link align-middle" rel="noopener noreferrer"&gt;
            &lt;img alt="" src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fqiita-user-contents.imgix.net%2Fhttps%253A%252F%252Fqiita-user-contents.imgix.net%252Fhttps%25253A%25252F%25252Fcdn.qiita.com%25252Fassets%25252Fpublic%25252Farticle-ogp-background-afbab5eb44e0b055cce1258705637a91.png%253Fixlib%253Drb-4.0.0%2526w%253D1200%2526blend64%253DaHR0cHM6Ly9xaWl0YS11c2VyLXByb2ZpbGUtaW1hZ2VzLmltZ2l4Lm5ldC9odHRwcyUzQSUyRiUyRmxoMy5nb29nbGV1c2VyY29udGVudC5jb20lMkZhLSUyRkFPaDE0R2lIRHVtdWNzM282Zm16ZnJFc1NjdjJ4UkNIbE9sVXBuNXpOcTV1JTNEczUwP2l4bGliPXJiLTQuMC4wJmFyPTElM0ExJmZpdD1jcm9wJm1hc2s9ZWxsaXBzZSZiZz1GRkZGRkYmZm09cG5nMzImcz1mNmM1MzJjNTRmNWU1NjFiZWNjMDRhNWZlMTBhMjRlMA%2526blend-x%253D120%2526blend-y%253D462%2526blend-w%253D90%2526blend-h%253D90%2526blend-mode%253Dnormal%2526mark64%253DaHR0cHM6Ly9xaWl0YS1vcmdhbml6YXRpb24taW1hZ2VzLmltZ2l4Lm5ldC9odHRwcyUzQSUyRiUyRnMzLWFwLW5vcnRoZWFzdC0xLmFtYXpvbmF3cy5jb20lMkZxaWl0YS1vcmdhbml6YXRpb24taW1hZ2UlMkZiODdjZTQ0N2RjZWRiZGFhM2UzOTFmOTFlYjIzZjZiMzE4ZjM4ZjAxJTJGb3JpZ2luYWwuanBnJTNGMTYzNDA5MzAxNT9peGxpYj1yYi00LjAuMCZ3PTQ0Jmg9NDQmZml0PWNyb3AmbWFzaz1jb3JuZXJzJmNvcm5lci1yYWRpdXM9OCZiZz1GRkZGRkYmYm9yZGVyPTIlMkNGRkZGRkYmZm09cG5nMzImcz1iNWM2ZWFmMGJmNmExYjhkNjMwNDdiNGUzMDI5ZWU1Yw%2526mark-x%253D186%2526mark-y%253D515%2526mark-w%253D40%2526mark-h%253D40%2526s%253Da53ec2df2c9a6d25bf50e8a03dcd9e77%3Fixlib%3Drb-4.0.0%26w%3D1200%26fm%3Djpg%26mark64%3DaHR0cHM6Ly9xaWl0YS11c2VyLWNvbnRlbnRzLmltZ2l4Lm5ldC9-dGV4dD9peGxpYj1yYi00LjAuMCZ3PTk2MCZoPTMyNCZ0eHQ9RGF0YWJyaWNrcyVFMyU4MSVBQiVFMyU4MSU4QSVFMyU4MSU5MSVFMyU4MiU4QiVFMyU4MyU4NiVFMyU4MyVCQyVFMyU4MyU5NiVFMyU4MyVBQiVFMyU4MSVBRSVFMyU4MiVBRiVFMyU4MyVBRCVFMyU4MyVCQyVFMyU4MyVCMyZ0eHQtYWxpZ249bGVmdCUyQ3RvcCZ0eHQtY29sb3I9JTIzMUUyMTIxJnR4dC1mb250PUhpcmFnaW5vJTIwU2FucyUyMFc2JnR4dC1zaXplPTU2JnR4dC1wYWQ9MCZzPWQwYWU5MzEyNTc2MzVkYzBhNzYwYjljN2EwOTAyMjYz%26mark-x%3D120%26mark-y%3D112%26blend64%3DaHR0cHM6Ly9xaWl0YS11c2VyLWNvbnRlbnRzLmltZ2l4Lm5ldC9-dGV4dD9peGxpYj1yYi00LjAuMCZ3PTgzOCZoPTU4JnR4dD0lNDB0YWthX3lheW9pJnR4dC1jb2xvcj0lMjMxRTIxMjEmdHh0LWZvbnQ9SGlyYWdpbm8lMjBTYW5zJTIwVzYmdHh0LXNpemU9MzYmdHh0LXBhZD0wJnM9N2IxNWFlODUxN2VkZTFjYmVjZDkwMWE4NzA0NDk3MjA%26blend-x%3D242%26blend-y%3D454%26blend-w%3D838%26blend-h%3D46%26blend-fit%3Dcrop%26blend-crop%3Dleft%252Cbottom%26blend-mode%3Dnormal%26txt64%3D44OH44O844K_44OW44Oq44OD44Kv44K544O744K444Oj44OR44Oz5qCq5byP5Lya56S-%26txt-x%3D242%26txt-y%3D539%26txt-width%3D838%26txt-clip%3Dend%252Cellipsis%26txt-color%3D%25231E2121%26txt-font%3DHiragino%2520Sans%2520W6%26txt-size%3D28%26s%3Dd1fae4a778d2864e1b7e73fa11d568a7" height="630" class="m-0" width="1200"&gt;
          &lt;/a&gt;
        &lt;/div&gt;
      &lt;div class="c-embed__body"&gt;
        &lt;h2 class="fs-xl lh-tight"&gt;
          &lt;a href="https://qiita.com/taka_yayoi/items/50c5a75caff8a6fb721d" rel="noopener noreferrer" class="c-link"&gt;
            Databricksにおけるテーブルのクローン #deltalake - Qiita
          &lt;/a&gt;
        &lt;/h2&gt;
          &lt;p class="truncate-at-3"&gt;
            Clone a table on Databricks | Databricks on AWS [2022/10/28時点]の翻訳です。 本書は抄訳であり内容の正確性を保証するものではありません。正確な内容に関しては原文を参照ください。 cloneコマンドを用いてD...
          &lt;/p&gt;
        &lt;div class="color-secondary fs-s flex items-center"&gt;
            &lt;img alt="favicon" class="c-embed__favicon m-0 mr-2 radius-0" src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fcdn.qiita.com%2Fassets%2Ffavicons%2Fpublic%2Fproduction-c620d3e403342b1022967ba5e3db1aaa.ico" width="120" height="120"&gt;
          qiita.com
        &lt;/div&gt;
      &lt;/div&gt;
    &lt;/div&gt;
&lt;/div&gt;


&lt;h1&gt;
  
  
  Conclusion
&lt;/h1&gt;

&lt;p&gt;Based on the above,&lt;br&gt;&lt;br&gt;
it was confirmed that &lt;strong&gt;"tables created in Databricks can be used in Fabric."&lt;/strong&gt;&lt;br&gt;&lt;br&gt;
Once the hub storage is set up, achieving interoperability between Fabric and Databricks is relatively simple.&lt;/p&gt;

&lt;p&gt;▽ Previous article&lt;br&gt;&lt;br&gt;
&lt;/p&gt;
&lt;div class="crayons-card c-embed text-styles text-styles--secondary"&gt;
    &lt;div class="c-embed__content"&gt;
        &lt;div class="c-embed__cover"&gt;
          &lt;a href="https://qiita.com/ReijiOtake/items/86b44b2c30986c65db08" class="c-link align-middle" rel="noopener noreferrer"&gt;
            &lt;img alt="" src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fqiita-user-contents.imgix.net%2Fhttps%253A%252F%252Fqiita-user-contents.imgix.net%252Fhttps%25253A%25252F%25252Fcdn.qiita.com%25252Fassets%25252Fpublic%25252Farticle-ogp-background-afbab5eb44e0b055cce1258705637a91.png%253Fixlib%253Drb-4.0.0%2526w%253D1200%2526blend64%253DaHR0cHM6Ly9xaWl0YS11c2VyLXByb2ZpbGUtaW1hZ2VzLmltZ2l4Lm5ldC9odHRwcyUzQSUyRiUyRnMzLWFwLW5vcnRoZWFzdC0xLmFtYXpvbmF3cy5jb20lMkZxaWl0YS1pbWFnZS1zdG9yZSUyRjAlMkYzOTIxOTk5JTJGOGY0NTg5ODc0MzczNDE5ZTJiZjhkMTJmNmQ0OGUyODcwZGU2OTJlMSUyRnhfbGFyZ2UucG5nJTNGMTc0MzgyMTEwMT9peGxpYj1yYi00LjAuMCZhcj0xJTNBMSZmaXQ9Y3JvcCZtYXNrPWVsbGlwc2UmYmc9RkZGRkZGJmZtPXBuZzMyJnM9OTRiZmVjNTRhNDllMDdlY2JiNjg5NzZlYWUwZGY5OTI%2526blend-x%253D120%2526blend-y%253D467%2526blend-w%253D82%2526blend-h%253D82%2526blend-mode%253Dnormal%2526s%253D728e1be30549a85eb776dc99c3b55165%3Fixlib%3Drb-4.0.0%26w%3D1200%26fm%3Djpg%26mark64%3DaHR0cHM6Ly9xaWl0YS11c2VyLWNvbnRlbnRzLmltZ2l4Lm5ldC9-dGV4dD9peGxpYj1yYi00LjAuMCZ3PTk2MCZoPTMyNCZ0eHQ9RmFicmljJUUzJTgxJUE4RGF0YWJyaWNrcyVFMyU4MSVBRSVFNyU5QiVCOCVFNCVCQSU5MiVFOSU4MSU4QiVFNyU5NCVBOCVFNiU4MCVBNyVFMiU5MSVBMiVFRiVCQyU5QUZhYnJpYyUyMCVFMyU4MSVBNyVFNCVCRCU5QyVFNiU4OCU5MCVFMyU4MSU5NyVFMyU4MSU5RiVFMyU4MyU4NiVFMyU4MyVCQyVFMyU4MyU5NiVFMyU4MyVBQiVFMyU4MiU5MkRhdGFicmlja3MlRTMlODElQTclRTUlODglQTklRTclOTQlQTglRTMlODElOTklRTMlODIlOEIlRUYlQkMlODhEYXRhYnJpYyVFMiU4MCVBNiZ0eHQtYWxpZ249bGVmdCUyQ3RvcCZ0eHQtY29sb3I9JTIzMUUyMTIxJnR4dC1mb250PUhpcmFnaW5vJTIwU2FucyUyMFc2JnR4dC1zaXplPTU2JnR4dC1wYWQ9MCZzPWQwMjM0NmMzNTEzZTY4MmRiMGI0ODkzMTRkZDVkZGE3%26mark-x%3D120%26mark-y%3D112%26blend64%3DaHR0cHM6Ly9xaWl0YS11c2VyLWNvbnRlbnRzLmltZ2l4Lm5ldC9-dGV4dD9peGxpYj1yYi00LjAuMCZ3PTgzOCZoPTU4JnR4dD0lNDBSZWlqaU90YWtlJnR4dC1jb2xvcj0lMjMxRTIxMjEmdHh0LWZvbnQ9SGlyYWdpbm8lMjBTYW5zJTIwVzYmdHh0LXNpemU9MzYmdHh0LXBhZD0wJnM9YzZmYWVkZTgyMDU5YzQwNDk3NWU4MjAwNDZmMjA4NGM%26blend-x%3D242%26blend-y%3D480%26blend-w%3D838%26blend-h%3D46%26blend-fit%3Dcrop%26blend-crop%3Dleft%252Cbottom%26blend-mode%3Dnormal%26s%3D64e9c79c4982522989a41bf18e74c7c0" height="630" class="m-0" width="1200"&gt;
          &lt;/a&gt;
        &lt;/div&gt;
      &lt;div class="c-embed__body"&gt;
        &lt;h2 class="fs-xl lh-tight"&gt;
          &lt;a href="https://qiita.com/ReijiOtake/items/86b44b2c30986c65db08" rel="noopener noreferrer" class="c-link"&gt;
            FabricとDatabricksの相互運用性③：Fabric で作成したテーブルをDatabricksで利用する（Databrickで閲覧・分析・編集可能） #BI - Qiita
          &lt;/a&gt;
        &lt;/h2&gt;
          &lt;p class="truncate-at-3"&gt;
            この記事は以下の記事の補足記事となります。 【総集編】Microsoft Fabric と Databricksをつなぐデータ総合運用術 -hubストレージにDelta Lake形式で保管する- YouTubeのアーカイブでも、16分57秒~あたりからこの記事に...
          &lt;/p&gt;
        &lt;div class="color-secondary fs-s flex items-center"&gt;
            &lt;img alt="favicon" class="c-embed__favicon m-0 mr-2 radius-0" src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fcdn.qiita.com%2Fassets%2Ffavicons%2Fpublic%2Fproduction-c620d3e403342b1022967ba5e3db1aaa.ico" width="120" height="120"&gt;
          qiita.com
        &lt;/div&gt;
      &lt;/div&gt;
    &lt;/div&gt;
&lt;/div&gt;


</description>
      <category>databricks</category>
      <category>azure</category>
      <category>deltalake</category>
      <category>sql</category>
    </item>
    <item>
      <title>Fabric &amp; Databricks Interoperability (3): Using Fabric Tables in Databricks for Viewing, Analyzing, and Editing</title>
      <dc:creator>Reiji Otake</dc:creator>
      <pubDate>Sun, 16 Feb 2025 10:34:06 +0000</pubDate>
      <link>https://forem.com/_d2a1ea24c442526a9777/fabric-databricks-interoperability-3-using-fabric-tables-in-databricks-for-viewing-analyzing-39o6</link>
      <guid>https://forem.com/_d2a1ea24c442526a9777/fabric-databricks-interoperability-3-using-fabric-tables-in-databricks-for-viewing-analyzing-39o6</guid>
      <description>&lt;h1&gt;
  
  
  Introduction
&lt;/h1&gt;

&lt;p&gt;Can tables created in Fabric be seamlessly referenced and edited in Databricks?&lt;br&gt;&lt;br&gt;
Many people may have this question.&lt;/p&gt;

&lt;p&gt;In this article, we will specifically introduce the use case of:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Using tables created in Fabric within Databricks.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;For details on the prerequisite settings, please refer to the previous article.&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;This article is part of a four-part series:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;&lt;a href="https://qiita.com/ReijiOtake/items/48c7b1e54796f4f569f3" rel="noopener noreferrer"&gt;Overview and Purpose of Interoperability&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://qiita.com/ReijiOtake/items/7ce8e1a743d05efe925c" rel="noopener noreferrer"&gt;Detailed Configuration of Hub Storage&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Using tables created in Fabric within Databricks (this article)&lt;/li&gt;
&lt;li&gt;&lt;a href="https://qiita.com/ReijiOtake/items/088abbd5f5ce06035501" rel="noopener noreferrer"&gt;Using tables created in Databricks within Fabric&lt;/a&gt;&lt;/li&gt;
&lt;/ol&gt;
&lt;/blockquote&gt;
&lt;h1&gt;
  
  
  Linking Tables Created in Fabric to Databricks
&lt;/h1&gt;
&lt;h3&gt;
  
  
  Creating a New Table in Fabric
&lt;/h3&gt;

&lt;p&gt;Upload a CSV file to the Fabric Lakehouse.&lt;br&gt;&lt;br&gt;
&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fcxzzllf6mte1ccquj4n6.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fcxzzllf6mte1ccquj4n6.png" alt="image.png" width="800" height="390"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;:::note info&lt;br&gt;
The CSV file used in this article is &lt;code&gt;sales.csv&lt;/code&gt; from the following Microsoft documentation:&lt;br&gt;&lt;br&gt;
&lt;a href="https://microsoftlearning.github.io/mslearn-fabric.ja-jp/Instructions/Labs/01-lakehouse.html" rel="noopener noreferrer"&gt;Create a Microsoft Fabric Lakehouse&lt;/a&gt;&lt;br&gt;
:::&lt;/p&gt;

&lt;p&gt;From the CSV file, select &lt;strong&gt;[Load to Table] &amp;gt; [New Table]&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F0x58re7crc7th2sg6f04.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F0x58re7crc7th2sg6f04.png" alt="image.png" width="800" height="195"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Specify &lt;code&gt;ext&lt;/code&gt;, which is a shortcut created in the hub storage, as the schema.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F2whh73kt23pxpx003rcd.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F2whh73kt23pxpx003rcd.png" alt="image.png" width="511" height="373"&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h3&gt;
  
  
  Verifying the Created Table
&lt;/h3&gt;

&lt;p&gt;You can confirm that a new table has been created in the Lakehouse.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F1c8e7xm0dpavkuoow7wd.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F1c8e7xm0dpavkuoow7wd.png" alt="image.png" width="800" height="298"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;A &lt;code&gt;create_from_fabric_sales&lt;/code&gt; folder is created in the &lt;code&gt;ext&lt;/code&gt; folder of the hub storage.&lt;br&gt;&lt;br&gt;
(This means that the newly created table physically exists in the hub storage.)&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7154fjipv5h6btc4amsh.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7154fjipv5h6btc4amsh.png" alt="image.png" width="800" height="354"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;You can also confirm that the table is in Delta format.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fwd27bit08i4eewi7a6qn.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fwd27bit08i4eewi7a6qn.png" alt="image.png" width="800" height="187"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;At this point, as expected, the table created in Fabric is not yet visible in Databricks.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7uv6nmxm67bowc9fuhk9.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7uv6nmxm67bowc9fuhk9.png" alt="image.png" width="800" height="266"&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h3&gt;
  
  
  Enabling Databricks to Access Fabric Tables
&lt;/h3&gt;

&lt;p&gt;Use the Databricks SQL Editor to create an external table.&lt;br&gt;&lt;br&gt;
Specify the &lt;strong&gt;hub storage folder path&lt;/strong&gt; (the folder of the table created in Fabric) in the &lt;strong&gt;Location&lt;/strong&gt; field.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F95p8lkdin7q8bwmfxr1b.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F95p8lkdin7q8bwmfxr1b.png" alt="image.png" width="800" height="205"&gt;&lt;/a&gt;&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight sql"&gt;&lt;code&gt;&lt;span class="k"&gt;CREATE&lt;/span&gt; &lt;span class="k"&gt;TABLE&lt;/span&gt; &lt;span class="o"&gt;&amp;lt;&lt;/span&gt;&lt;span class="k"&gt;table_name&lt;/span&gt;&lt;span class="o"&gt;&amp;gt;&lt;/span&gt;
&lt;span class="k"&gt;USING&lt;/span&gt; &lt;span class="n"&gt;DELTA&lt;/span&gt;
&lt;span class="k"&gt;LOCATION&lt;/span&gt; &lt;span class="s1"&gt;'abfss://&amp;lt;container_name&amp;gt;@&amp;lt;ADLS2_name&amp;gt;.dfs.core.windows.net/folder_name/&amp;lt;table_folder_name&amp;gt;'&lt;/span&gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;


&lt;p&gt;Then, you can view tables created in Fabric from the [Catalog].&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fm82dbs21vcwarkrama3z.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fm82dbs21vcwarkrama3z.png" alt="image.png" width="800" height="376"&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h1&gt;
  
  
  Viewing and Analyzing Tables Created in Fabric with Databricks (BI Creation)
&lt;/h1&gt;

&lt;p&gt;From the [Dashboard] in Databricks, you can create a new dashboard and select an external table (i.e., a table created in Fabric) from [Data] &amp;gt; [Select Table].&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ftisgx3r0sku69kvqdvg0.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ftisgx3r0sku69kvqdvg0.png" alt="image.png" width="800" height="508"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Thus, it is possible to analyze tables created in Fabric using Databricks.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F2udlz9uzlmgc1cr1bgq6.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F2udlz9uzlmgc1cr1bgq6.png" alt="image.png" width="800" height="366"&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h1&gt;
  
  
  Editing Tables Created in Fabric with Databricks (DML)
&lt;/h1&gt;

&lt;p&gt;Try executing an &lt;code&gt;UPDATE&lt;/code&gt; statement (DML statement) from the SQL Editor in Databricks.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fs5a34tc4u42anmnh4q7y.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fs5a34tc4u42anmnh4q7y.png" alt="image.png" width="800" height="180"&gt;&lt;/a&gt;&lt;br&gt;
&lt;/p&gt;
&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight sql"&gt;&lt;code&gt;&lt;span class="k"&gt;UPDATE&lt;/span&gt; &lt;span class="n"&gt;create_from_fabric_sales&lt;/span&gt;
&lt;span class="k"&gt;SET&lt;/span&gt; &lt;span class="n"&gt;Item&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="s1"&gt;'No.1 Item'&lt;/span&gt;
&lt;span class="k"&gt;WHERE&lt;/span&gt; &lt;span class="n"&gt;Item&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="s1"&gt;'Road-150 Red, 48'&lt;/span&gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;


&lt;p&gt;Of course, you can confirm that the changes have been reflected on the Databricks side.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Flasox5tqd3m4fntjbtdb.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Flasox5tqd3m4fntjbtdb.png" alt="image.png" width="800" height="511"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Although the edit was made from Databricks, the changes were successfully reflected on the Fabric side as well.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F65r3o65g0pli04xkiqqz.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F65r3o65g0pli04xkiqqz.png" alt="image.png" width="800" height="373"&gt;&lt;/a&gt;&lt;br&gt;
&lt;/p&gt;
&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight sql"&gt;&lt;code&gt;&lt;span class="k"&gt;SELECT&lt;/span&gt; &lt;span class="n"&gt;Item&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="k"&gt;SUM&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="n"&gt;Quantity&lt;/span&gt; &lt;span class="o"&gt;*&lt;/span&gt; &lt;span class="n"&gt;UnitPrice&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt; &lt;span class="k"&gt;AS&lt;/span&gt; &lt;span class="n"&gt;Revenue&lt;/span&gt;
&lt;span class="k"&gt;FROM&lt;/span&gt; &lt;span class="n"&gt;Fabric_Lakehouse&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;ext&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;create_from_fabric_sales&lt;/span&gt;
&lt;span class="k"&gt;GROUP&lt;/span&gt; &lt;span class="k"&gt;BY&lt;/span&gt; &lt;span class="n"&gt;Item&lt;/span&gt;
&lt;span class="k"&gt;ORDER&lt;/span&gt; &lt;span class="k"&gt;BY&lt;/span&gt; &lt;span class="n"&gt;Revenue&lt;/span&gt; &lt;span class="k"&gt;DESC&lt;/span&gt;&lt;span class="p"&gt;;&lt;/span&gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;


&lt;p&gt;Therefore, it is possible to edit tables created in Fabric using Databricks (DML statements).&lt;/p&gt;
&lt;h1&gt;
  
  
  Conclusion
&lt;/h1&gt;

&lt;p&gt;From the above, we have confirmed that&lt;br&gt;&lt;br&gt;
&lt;strong&gt;"Tables created in Fabric can be used in Databricks."&lt;/strong&gt;&lt;br&gt;&lt;br&gt;
Once the hub storage is set up, it is relatively easy to achieve interoperability between Fabric and Databricks.&lt;/p&gt;

&lt;p&gt;In the next article, we will introduce the reverse case:&lt;br&gt;&lt;br&gt;
&lt;strong&gt;"Using tables created in Databricks in Fabric."&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;▽ Next article&lt;br&gt;&lt;br&gt;
&lt;/p&gt;
&lt;div class="crayons-card c-embed text-styles text-styles--secondary"&gt;
    &lt;div class="c-embed__content"&gt;
        &lt;div class="c-embed__cover"&gt;
          &lt;a href="https://qiita.com/ReijiOtake/items/088abbd5f5ce06035501" class="c-link align-middle" rel="noopener noreferrer"&gt;
            &lt;img alt="" src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fqiita-user-contents.imgix.net%2Fhttps%253A%252F%252Fqiita-user-contents.imgix.net%252Fhttps%25253A%25252F%25252Fcdn.qiita.com%25252Fassets%25252Fpublic%25252Farticle-ogp-background-afbab5eb44e0b055cce1258705637a91.png%253Fixlib%253Drb-4.0.0%2526w%253D1200%2526blend64%253DaHR0cHM6Ly9xaWl0YS11c2VyLXByb2ZpbGUtaW1hZ2VzLmltZ2l4Lm5ldC9odHRwcyUzQSUyRiUyRnMzLWFwLW5vcnRoZWFzdC0xLmFtYXpvbmF3cy5jb20lMkZxaWl0YS1pbWFnZS1zdG9yZSUyRjAlMkYzOTIxOTk5JTJGOGY0NTg5ODc0MzczNDE5ZTJiZjhkMTJmNmQ0OGUyODcwZGU2OTJlMSUyRnhfbGFyZ2UucG5nJTNGMTc0MzgyMTEwMT9peGxpYj1yYi00LjAuMCZhcj0xJTNBMSZmaXQ9Y3JvcCZtYXNrPWVsbGlwc2UmYmc9RkZGRkZGJmZtPXBuZzMyJnM9OTRiZmVjNTRhNDllMDdlY2JiNjg5NzZlYWUwZGY5OTI%2526blend-x%253D120%2526blend-y%253D467%2526blend-w%253D82%2526blend-h%253D82%2526blend-mode%253Dnormal%2526s%253D728e1be30549a85eb776dc99c3b55165%3Fixlib%3Drb-4.0.0%26w%3D1200%26fm%3Djpg%26mark64%3DaHR0cHM6Ly9xaWl0YS11c2VyLWNvbnRlbnRzLmltZ2l4Lm5ldC9-dGV4dD9peGxpYj1yYi00LjAuMCZ3PTk2MCZoPTMyNCZ0eHQ9RmFicmljJUUzJTgxJUE4RGF0YWJyaWNrcyVFMyU4MSVBRSVFNyU5QiVCOCVFNCVCQSU5MiVFOSU4MSU4QiVFNyU5NCVBOCVFNiU4MCVBNyVFMiU5MSVBMyVFRiVCQyU5QURhdGFicmlja3MlMjAlRTMlODElQTclRTQlQkQlOUMlRTYlODglOTAlRTMlODElOTclRTMlODElOUYlRTMlODMlODYlRTMlODMlQkMlRTMlODMlOTYlRTMlODMlQUIlRTMlODIlOTJGYWJyaWMlRTMlODElQTclRTUlODglQTklRTclOTQlQTglRTMlODElOTklRTMlODIlOEIlRUYlQkMlODhGYWJyaWMlRTMlODElQTclRTIlODAlQTYmdHh0LWFsaWduPWxlZnQlMkN0b3AmdHh0LWNvbG9yPSUyMzFFMjEyMSZ0eHQtZm9udD1IaXJhZ2lubyUyMFNhbnMlMjBXNiZ0eHQtc2l6ZT01NiZ0eHQtcGFkPTAmcz0xMjBjZmFlYWVhZjNkM2I1MGE4NzZmNWFiNDg1ZjNkMA%26mark-x%3D120%26mark-y%3D112%26blend64%3DaHR0cHM6Ly9xaWl0YS11c2VyLWNvbnRlbnRzLmltZ2l4Lm5ldC9-dGV4dD9peGxpYj1yYi00LjAuMCZ3PTgzOCZoPTU4JnR4dD0lNDBSZWlqaU90YWtlJnR4dC1jb2xvcj0lMjMxRTIxMjEmdHh0LWZvbnQ9SGlyYWdpbm8lMjBTYW5zJTIwVzYmdHh0LXNpemU9MzYmdHh0LXBhZD0wJnM9YzZmYWVkZTgyMDU5YzQwNDk3NWU4MjAwNDZmMjA4NGM%26blend-x%3D242%26blend-y%3D480%26blend-w%3D838%26blend-h%3D46%26blend-fit%3Dcrop%26blend-crop%3Dleft%252Cbottom%26blend-mode%3Dnormal%26s%3D69721fab8d2f11a8e46198ddc1fe2902" height="630" class="m-0" width="1200"&gt;
          &lt;/a&gt;
        &lt;/div&gt;
      &lt;div class="c-embed__body"&gt;
        &lt;h2 class="fs-xl lh-tight"&gt;
          &lt;a href="https://qiita.com/ReijiOtake/items/088abbd5f5ce06035501" rel="noopener noreferrer" class="c-link"&gt;
            FabricとDatabricksの相互運用性④：Databricks で作成したテーブルをFabricで利用する（Fabricで閲覧・分析・編集可能） #SQL - Qiita
          &lt;/a&gt;
        &lt;/h2&gt;
          &lt;p class="truncate-at-3"&gt;
            この記事は以下の記事の補足記事となります。 【総集編】Microsoft Fabric と Databricksをつなぐデータ総合運用術 -hubストレージにDelta Lake形式で保管する- YouTubeのアーカイブでも、13分22秒~あたりからこの記事に...
          &lt;/p&gt;
        &lt;div class="color-secondary fs-s flex items-center"&gt;
            &lt;img alt="favicon" class="c-embed__favicon m-0 mr-2 radius-0" src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fcdn.qiita.com%2Fassets%2Ffavicons%2Fpublic%2Fproduction-c620d3e403342b1022967ba5e3db1aaa.ico" width="120" height="120"&gt;
          qiita.com
        &lt;/div&gt;
      &lt;/div&gt;
    &lt;/div&gt;
&lt;/div&gt;



&lt;p&gt;▽ Previous article&lt;br&gt;&lt;br&gt;
&lt;/p&gt;
&lt;div class="crayons-card c-embed text-styles text-styles--secondary"&gt;
    &lt;div class="c-embed__content"&gt;
        &lt;div class="c-embed__cover"&gt;
          &lt;a href="https://qiita.com/ReijiOtake/items/7ce8e1a743d05efe925c" class="c-link align-middle" rel="noopener noreferrer"&gt;
            &lt;img alt="" src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fqiita-user-contents.imgix.net%2Fhttps%253A%252F%252Fqiita-user-contents.imgix.net%252Fhttps%25253A%25252F%25252Fcdn.qiita.com%25252Fassets%25252Fpublic%25252Farticle-ogp-background-afbab5eb44e0b055cce1258705637a91.png%253Fixlib%253Drb-4.0.0%2526w%253D1200%2526blend64%253DaHR0cHM6Ly9xaWl0YS11c2VyLXByb2ZpbGUtaW1hZ2VzLmltZ2l4Lm5ldC9odHRwcyUzQSUyRiUyRnMzLWFwLW5vcnRoZWFzdC0xLmFtYXpvbmF3cy5jb20lMkZxaWl0YS1pbWFnZS1zdG9yZSUyRjAlMkYzOTIxOTk5JTJGOGY0NTg5ODc0MzczNDE5ZTJiZjhkMTJmNmQ0OGUyODcwZGU2OTJlMSUyRnhfbGFyZ2UucG5nJTNGMTc0MzgyMTEwMT9peGxpYj1yYi00LjAuMCZhcj0xJTNBMSZmaXQ9Y3JvcCZtYXNrPWVsbGlwc2UmYmc9RkZGRkZGJmZtPXBuZzMyJnM9OTRiZmVjNTRhNDllMDdlY2JiNjg5NzZlYWUwZGY5OTI%2526blend-x%253D120%2526blend-y%253D467%2526blend-w%253D82%2526blend-h%253D82%2526blend-mode%253Dnormal%2526s%253D728e1be30549a85eb776dc99c3b55165%3Fixlib%3Drb-4.0.0%26w%3D1200%26fm%3Djpg%26mark64%3DaHR0cHM6Ly9xaWl0YS11c2VyLWNvbnRlbnRzLmltZ2l4Lm5ldC9-dGV4dD9peGxpYj1yYi00LjAuMCZ3PTk2MCZoPTMyNCZ0eHQ9RmFicmljJUUzJTgxJUE4RGF0YWJyaWNrcyVFMyU4MSVBRSVFNyU5QiVCOCVFNCVCQSU5MiVFOSU4MSU4QiVFNyU5NCVBOCVFNiU4MCVBNyUyMCVFMiU5MSVBMSVFRiVCQyU5QUZhYnJpYyVFMyU4MSVBOERhdGFicmlja3MlRTMlODIlOTIlRTMlODMlOEYlRTMlODMlOTYlRTMlODIlQjklRTMlODMlODglRTMlODMlQUMlRTMlODMlQkMlRTMlODIlQjglRUYlQkMlODhEZWx0YSUyMExha2UlRUYlQkMlODklRTMlODElQTclRTYlOEUlQTUlRTclQjYlOUElRTMlODElOTklRTIlODAlQTYmdHh0LWFsaWduPWxlZnQlMkN0b3AmdHh0LWNvbG9yPSUyMzFFMjEyMSZ0eHQtZm9udD1IaXJhZ2lubyUyMFNhbnMlMjBXNiZ0eHQtc2l6ZT01NiZ0eHQtcGFkPTAmcz01Y2MzNmQ4NWMxMjc4NTJhM2Q5YmFlOTE0YTI3YTE2ZQ%26mark-x%3D120%26mark-y%3D112%26blend64%3DaHR0cHM6Ly9xaWl0YS11c2VyLWNvbnRlbnRzLmltZ2l4Lm5ldC9-dGV4dD9peGxpYj1yYi00LjAuMCZ3PTgzOCZoPTU4JnR4dD0lNDBSZWlqaU90YWtlJnR4dC1jb2xvcj0lMjMxRTIxMjEmdHh0LWZvbnQ9SGlyYWdpbm8lMjBTYW5zJTIwVzYmdHh0LXNpemU9MzYmdHh0LXBhZD0wJnM9YzZmYWVkZTgyMDU5YzQwNDk3NWU4MjAwNDZmMjA4NGM%26blend-x%3D242%26blend-y%3D480%26blend-w%3D838%26blend-h%3D46%26blend-fit%3Dcrop%26blend-crop%3Dleft%252Cbottom%26blend-mode%3Dnormal%26s%3D6c4da4c178ad4cf1c7d03b5770bf24fc" height="630" class="m-0" width="1200"&gt;
          &lt;/a&gt;
        &lt;/div&gt;
      &lt;div class="c-embed__body"&gt;
        &lt;h2 class="fs-xl lh-tight"&gt;
          &lt;a href="https://qiita.com/ReijiOtake/items/7ce8e1a743d05efe925c" rel="noopener noreferrer" class="c-link"&gt;
            FabricとDatabricksの相互運用性 ②：FabricとDatabricksをハブストレージ（Delta Lake）で接続する設定手順 #Azure - Qiita
          &lt;/a&gt;
        &lt;/h2&gt;
          &lt;p class="truncate-at-3"&gt;
            この記事は以下の記事の補足記事となります。 【総集編】Microsoft Fabric と Databricksをつなぐデータ総合運用術 -hubストレージにDelta Lake形式で保管する- YouTubeのアーカイブでも、9分45秒~あたりからこの記事につ...
          &lt;/p&gt;
        &lt;div class="color-secondary fs-s flex items-center"&gt;
            &lt;img alt="favicon" class="c-embed__favicon m-0 mr-2 radius-0" src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fcdn.qiita.com%2Fassets%2Ffavicons%2Fpublic%2Fproduction-c620d3e403342b1022967ba5e3db1aaa.ico" width="120" height="120"&gt;
          qiita.com
        &lt;/div&gt;
      &lt;/div&gt;
    &lt;/div&gt;
&lt;/div&gt;


</description>
      <category>databricks</category>
      <category>azure</category>
      <category>bi</category>
      <category>venderfree</category>
    </item>
    <item>
      <title>Fabric &amp; Databricks Interoperability (2): Configuring Hub Storage</title>
      <dc:creator>Reiji Otake</dc:creator>
      <pubDate>Sun, 16 Feb 2025 09:48:01 +0000</pubDate>
      <link>https://forem.com/_d2a1ea24c442526a9777/fabric-databricks-interoperability-2-configuring-hub-storage-4l85</link>
      <guid>https://forem.com/_d2a1ea24c442526a9777/fabric-databricks-interoperability-2-configuring-hub-storage-4l85</guid>
      <description>&lt;h1&gt;
  
  
  Introduction
&lt;/h1&gt;

&lt;p&gt;This article provides a detailed guide on &lt;strong&gt;how to configure settings&lt;/strong&gt; for the following use cases:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;strong&gt;Using tables created in Databricks within Fabric&lt;/strong&gt;&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Using tables created in Fabric within Databricks&lt;/strong&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;:::note info&lt;br&gt;
This article is part of a four-part series:&lt;br&gt;
1. &lt;a href="https://qiita.com/ReijiOtake/items/48c7b1e54796f4f569f3" rel="noopener noreferrer"&gt;Overview &amp;amp; Purpose of Interoperability&lt;/a&gt;&lt;br&gt;&lt;br&gt;
2. &lt;strong&gt;Detailed Configuration of Hub Storage (This Article)&lt;/strong&gt;&lt;br&gt;&lt;br&gt;
3. &lt;a href="https://qiita.com/ReijiOtake/items/86b44b2c30986c65db08" rel="noopener noreferrer"&gt;Using Tables Created in Fabric within Databricks&lt;/a&gt;&lt;br&gt;&lt;br&gt;
4. &lt;a href="https://qiita.com/ReijiOtake/items/088abbd5f5ce06035501" rel="noopener noreferrer"&gt;Using Tables Created in Databricks within Fabric&lt;/a&gt;&lt;br&gt;&lt;br&gt;
:::&lt;/p&gt;
&lt;h1&gt;
  
  
  Preparing Azure Data Lake Gen2 (ADLS2) as the Hub
&lt;/h1&gt;
&lt;h3&gt;
  
  
  ① Deploy a storage account in Azure Portal as the hub
&lt;/h3&gt;

&lt;p&gt;:::note warn&lt;br&gt;
Enable hierarchical namespace.&lt;br&gt;
:::&lt;/p&gt;
&lt;h3&gt;
  
  
  ② Create a container named 'hub' and a directory named 'ext'
&lt;/h3&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F59atbvryl38i2jroh241.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F59atbvryl38i2jroh241.png" alt="image.png" width="800" height="151"&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h1&gt;
  
  
  Connecting Fabric to Hub Storage
&lt;/h1&gt;
&lt;h3&gt;
  
  
  ① Create a Lakehouse
&lt;/h3&gt;

&lt;p&gt;:::note warn&lt;br&gt;
Enable Lakehouse schema (public preview).&lt;br&gt;
:::&lt;br&gt;
&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F17nlumakj3nx01q7mgp9.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F17nlumakj3nx01q7mgp9.png" alt="image.png" width="800" height="256"&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h3&gt;
  
  
  ② Specify the hub storage in the new schema shortcut of the Lakehouse
&lt;/h3&gt;

&lt;p&gt;From [Tables] in the Lakehouse, click the three-dot menu and select [New Schema Shortcut].&lt;br&gt;
&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fiw1kf6ja3nijf42bnx65.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fiw1kf6ja3nijf42bnx65.png" alt="image.png" width="800" height="626"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Select [Azure Data Lake Gen2].&lt;br&gt;
&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fp2cqeyvynuj0m9v23x9b.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fp2cqeyvynuj0m9v23x9b.png" alt="image.png" width="800" height="448"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Enter the details for creating a new connection.&lt;br&gt;
&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fob8ff4v75goe4962p13a.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fob8ff4v75goe4962p13a.png" alt="image.png" width="800" height="445"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;:::note info&lt;/p&gt;
&lt;h3&gt;
  
  
  How to check ADLS2 access URL:
&lt;/h3&gt;

&lt;p&gt;You can confirm it from the storage account's [Endpoints] section under 'Data Lake Storage'.&lt;br&gt;
&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fncngv3z9ggqxgml10ewq.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fncngv3z9ggqxgml10ewq.png" alt="image.png" width="600" height="892"&gt;&lt;/a&gt;&lt;br&gt;
:::&lt;/p&gt;

&lt;p&gt;Enable the 'ext' directory and click [Next].&lt;br&gt;
&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fdn2r7tb9ypeaj28dibld.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fdn2r7tb9ypeaj28dibld.png" alt="image.png" width="800" height="445"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Click [Create].&lt;br&gt;
&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fmoyldilqcyv59e17xeuv.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fmoyldilqcyv59e17xeuv.png" alt="image.png" width="800" height="449"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;The 'ext' directory is created as an external schema shortcut.&lt;br&gt;
&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F34nrtxwf95t0wsw4t1pn.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F34nrtxwf95t0wsw4t1pn.png" alt="image.png" width="800" height="719"&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h1&gt;
  
  
  Connecting Databricks to Hub Storage
&lt;/h1&gt;
&lt;h3&gt;
  
  
  ① Create an access connector for Azure Databricks in the Azure Portal
&lt;/h3&gt;

&lt;p&gt;Follow the steps under &lt;a href="https://learn.microsoft.com/ja-jp/azure/databricks/data-governance/unity-catalog/azure-managed-identities#config-managed-id" rel="noopener noreferrer"&gt;"Step 1: Create an Access Connector for Azure Databricks" in the "Use Azure Managed Identity to Access Storage in Unity Catalog"&lt;/a&gt; guide, using a system-assigned managed identity.&lt;/p&gt;
&lt;h3&gt;
  
  
  ② Grant the connector access to the hub storage from the Azure Portal
&lt;/h3&gt;

&lt;p&gt;Follow &lt;a href="https://learn.microsoft.com/ja-jp/azure/databricks/data-governance/unity-catalog/azure-managed-identities#config-managed-id" rel="noopener noreferrer"&gt;"Step 2: Grant Managed Identity Access to the Storage Account" in the same guide&lt;/a&gt;.&lt;/p&gt;
&lt;h3&gt;
  
  
  ③ Create storage credentials in Databricks
&lt;/h3&gt;

&lt;p&gt;Log in to Databricks and navigate to [Catalog] &amp;gt; [+] &amp;gt; [Add Storage Credentials].&lt;br&gt;
&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fo848dqgr4hdtkxajfmzs.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fo848dqgr4hdtkxajfmzs.png" alt="image.png" width="800" height="307"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Add new storage credentials.&lt;br&gt;
|  | Input Value |&lt;br&gt;
|:-:|:-:|&lt;br&gt;
|Storage Credentials or Service Credentials| Storage Credentials |&lt;br&gt;
|Credential Name| Any name |&lt;br&gt;
|Access Connector ID| Resource ID of the connector created in step ① (can be confirmed in Azure Portal) |&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fvyzntl21l0ds0eiafyy6.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fvyzntl21l0ds0eiafyy6.png" alt="image.png" width="800" height="740"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;After creation, click the newly created credential name.&lt;br&gt;
&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Flc56yb266v7ymkz9dmn1.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Flc56yb266v7ymkz9dmn1.png" alt="image.png" width="800" height="177"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Click [Permissions] &amp;gt; [Grant].&lt;br&gt;
&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ft5921d1irf5cotp74d5i.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ft5921d1irf5cotp74d5i.png" alt="image.png" width="800" height="194"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Grant [ALL PRIVILEGES] to necessary users.&lt;br&gt;
&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fvge8jyfa5jcol2q8xvfp.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fvge8jyfa5jcol2q8xvfp.png" alt="image.png" width="800" height="456"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;:::note info&lt;br&gt;
Reference for steps and required permissions:&lt;br&gt;&lt;br&gt;
&lt;a href="https://learn.microsoft.com/ja-jp/azure/databricks/connect/unity-catalog/cloud-storage/storage-credentials#next-steps" rel="noopener noreferrer"&gt;Create Storage Credentials for Connecting to Azure Data Lake Storage Gen2&lt;/a&gt;&lt;br&gt;
:::&lt;/p&gt;
&lt;h3&gt;
  
  
  ④ Add an external location in Databricks
&lt;/h3&gt;

&lt;p&gt;Log in to Databricks and navigate to [Catalog] &amp;gt; [+] &amp;gt; [Add External Location].&lt;br&gt;
&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyz2ua8koc5ywms36u7oc.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyz2ua8koc5ywms36u7oc.png" alt="image.png" width="800" height="222"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Create a new external location.&lt;br&gt;
|  | Input Value |&lt;br&gt;
|:-:|:-:|&lt;br&gt;
|External Location Name| Any name |&lt;br&gt;
|URL| abfss://directory-name (hub) @ storage-account-name.dfs.windows.net |&lt;br&gt;
|Storage Credentials| Select the credentials created in step ③ |&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F9uggh8hwfbamejjke773.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F9uggh8hwfbamejjke773.png" alt="image.png" width="800" height="570"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;:::note info&lt;/p&gt;
&lt;h3&gt;
  
  
  How to determine the URL:
&lt;/h3&gt;

&lt;p&gt;Refer to the storage account's [Endpoints] section used in step ② of "Connecting Fabric to Hub Storage".&lt;br&gt;
:::&lt;/p&gt;

&lt;p&gt;:::note info&lt;br&gt;
Reference for steps and required permissions:&lt;br&gt;&lt;br&gt;
&lt;a href="https://learn.microsoft.com/ja-jp/azure/databricks/connect/unity-catalog/cloud-storage/external-locations" rel="noopener noreferrer"&gt;Create an External Location to Connect Cloud Storage to Azure Databricks&lt;/a&gt;&lt;br&gt;
:::&lt;/p&gt;
&lt;h1&gt;
  
  
  Conclusion
&lt;/h1&gt;

&lt;p&gt;Now everything is set up!&lt;br&gt;&lt;br&gt;
Next, let's proceed with the actual interoperability of tables.&lt;/p&gt;

&lt;p&gt;▽ Next Article&lt;br&gt;&lt;br&gt;
&lt;/p&gt;
&lt;div class="crayons-card c-embed text-styles text-styles--secondary"&gt;
    &lt;div class="c-embed__content"&gt;
        &lt;div class="c-embed__cover"&gt;
          &lt;a href="https://qiita.com/ReijiOtake/items/86b44b2c30986c65db08" class="c-link align-middle" rel="noopener noreferrer"&gt;
            &lt;img alt="" src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fqiita-user-contents.imgix.net%2Fhttps%253A%252F%252Fqiita-user-contents.imgix.net%252Fhttps%25253A%25252F%25252Fcdn.qiita.com%25252Fassets%25252Fpublic%25252Farticle-ogp-background-afbab5eb44e0b055cce1258705637a91.png%253Fixlib%253Drb-4.0.0%2526w%253D1200%2526blend64%253DaHR0cHM6Ly9xaWl0YS11c2VyLXByb2ZpbGUtaW1hZ2VzLmltZ2l4Lm5ldC9odHRwcyUzQSUyRiUyRnMzLWFwLW5vcnRoZWFzdC0xLmFtYXpvbmF3cy5jb20lMkZxaWl0YS1pbWFnZS1zdG9yZSUyRjAlMkYzOTIxOTk5JTJGOGY0NTg5ODc0MzczNDE5ZTJiZjhkMTJmNmQ0OGUyODcwZGU2OTJlMSUyRnhfbGFyZ2UucG5nJTNGMTc0MzgyMTEwMT9peGxpYj1yYi00LjAuMCZhcj0xJTNBMSZmaXQ9Y3JvcCZtYXNrPWVsbGlwc2UmYmc9RkZGRkZGJmZtPXBuZzMyJnM9OTRiZmVjNTRhNDllMDdlY2JiNjg5NzZlYWUwZGY5OTI%2526blend-x%253D120%2526blend-y%253D467%2526blend-w%253D82%2526blend-h%253D82%2526blend-mode%253Dnormal%2526s%253D728e1be30549a85eb776dc99c3b55165%3Fixlib%3Drb-4.0.0%26w%3D1200%26fm%3Djpg%26mark64%3DaHR0cHM6Ly9xaWl0YS11c2VyLWNvbnRlbnRzLmltZ2l4Lm5ldC9-dGV4dD9peGxpYj1yYi00LjAuMCZ3PTk2MCZoPTMyNCZ0eHQ9RmFicmljJUUzJTgxJUE4RGF0YWJyaWNrcyVFMyU4MSVBRSVFNyU5QiVCOCVFNCVCQSU5MiVFOSU4MSU4QiVFNyU5NCVBOCVFNiU4MCVBNyVFMiU5MSVBMiVFRiVCQyU5QUZhYnJpYyUyMCVFMyU4MSVBNyVFNCVCRCU5QyVFNiU4OCU5MCVFMyU4MSU5NyVFMyU4MSU5RiVFMyU4MyU4NiVFMyU4MyVCQyVFMyU4MyU5NiVFMyU4MyVBQiVFMyU4MiU5MkRhdGFicmlja3MlRTMlODElQTclRTUlODglQTklRTclOTQlQTglRTMlODElOTklRTMlODIlOEIlRUYlQkMlODhEYXRhYnJpYyVFMiU4MCVBNiZ0eHQtYWxpZ249bGVmdCUyQ3RvcCZ0eHQtY29sb3I9JTIzMUUyMTIxJnR4dC1mb250PUhpcmFnaW5vJTIwU2FucyUyMFc2JnR4dC1zaXplPTU2JnR4dC1wYWQ9MCZzPWQwMjM0NmMzNTEzZTY4MmRiMGI0ODkzMTRkZDVkZGE3%26mark-x%3D120%26mark-y%3D112%26blend64%3DaHR0cHM6Ly9xaWl0YS11c2VyLWNvbnRlbnRzLmltZ2l4Lm5ldC9-dGV4dD9peGxpYj1yYi00LjAuMCZ3PTgzOCZoPTU4JnR4dD0lNDBSZWlqaU90YWtlJnR4dC1jb2xvcj0lMjMxRTIxMjEmdHh0LWZvbnQ9SGlyYWdpbm8lMjBTYW5zJTIwVzYmdHh0LXNpemU9MzYmdHh0LXBhZD0wJnM9YzZmYWVkZTgyMDU5YzQwNDk3NWU4MjAwNDZmMjA4NGM%26blend-x%3D242%26blend-y%3D480%26blend-w%3D838%26blend-h%3D46%26blend-fit%3Dcrop%26blend-crop%3Dleft%252Cbottom%26blend-mode%3Dnormal%26s%3D64e9c79c4982522989a41bf18e74c7c0" height="" class="m-0" width=""&gt;
          &lt;/a&gt;
        &lt;/div&gt;
      &lt;div class="c-embed__body"&gt;
        &lt;h2 class="fs-xl lh-tight"&gt;
          &lt;a href="https://qiita.com/ReijiOtake/items/86b44b2c30986c65db08" rel="noopener noreferrer" class="c-link"&gt;
            FabricとDatabricksの相互運用性③：Fabric で作成したテーブルをDatabricksで利用する（Databrickで閲覧・分析・編集可能） #BI - Qiita
          &lt;/a&gt;
        &lt;/h2&gt;
          &lt;p class="truncate-at-3"&gt;
            この記事は以下の記事の補足記事となります。 【総集編】Microsoft Fabric と Databricksをつなぐデータ総合運用術 -hubストレージにDelta Lake形式で保管する- YouTubeのアーカイブでも、16分57秒~あたりからこの記事に...
          &lt;/p&gt;
        &lt;div class="color-secondary fs-s flex items-center"&gt;
            &lt;img alt="favicon" class="c-embed__favicon m-0 mr-2 radius-0" src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fcdn.qiita.com%2Fassets%2Ffavicons%2Fpublic%2Fproduction-c620d3e403342b1022967ba5e3db1aaa.ico" width="120" height="120"&gt;
          qiita.com
        &lt;/div&gt;
      &lt;/div&gt;
    &lt;/div&gt;
&lt;/div&gt;
  

&lt;p&gt;▽ Previous Article&lt;br&gt;&lt;br&gt;
&lt;/p&gt;
&lt;div class="crayons-card c-embed text-styles text-styles--secondary"&gt;
    &lt;div class="c-embed__content"&gt;
        &lt;div class="c-embed__cover"&gt;
          &lt;a href="https://qiita.com/ReijiOtake/items/48c7b1e54796f4f569f3" class="c-link align-middle" rel="noopener noreferrer"&gt;
            &lt;img alt="" src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fqiita-user-contents.imgix.net%2Fhttps%253A%252F%252Fqiita-user-contents.imgix.net%252Fhttps%25253A%25252F%25252Fcdn.qiita.com%25252Fassets%25252Fpublic%25252Farticle-ogp-background-afbab5eb44e0b055cce1258705637a91.png%253Fixlib%253Drb-4.0.0%2526w%253D1200%2526blend64%253DaHR0cHM6Ly9xaWl0YS11c2VyLXByb2ZpbGUtaW1hZ2VzLmltZ2l4Lm5ldC9odHRwcyUzQSUyRiUyRnMzLWFwLW5vcnRoZWFzdC0xLmFtYXpvbmF3cy5jb20lMkZxaWl0YS1pbWFnZS1zdG9yZSUyRjAlMkYzOTIxOTk5JTJGOGY0NTg5ODc0MzczNDE5ZTJiZjhkMTJmNmQ0OGUyODcwZGU2OTJlMSUyRnhfbGFyZ2UucG5nJTNGMTc0MzgyMTEwMT9peGxpYj1yYi00LjAuMCZhcj0xJTNBMSZmaXQ9Y3JvcCZtYXNrPWVsbGlwc2UmYmc9RkZGRkZGJmZtPXBuZzMyJnM9OTRiZmVjNTRhNDllMDdlY2JiNjg5NzZlYWUwZGY5OTI%2526blend-x%253D120%2526blend-y%253D467%2526blend-w%253D82%2526blend-h%253D82%2526blend-mode%253Dnormal%2526s%253D728e1be30549a85eb776dc99c3b55165%3Fixlib%3Drb-4.0.0%26w%3D1200%26fm%3Djpg%26mark64%3DaHR0cHM6Ly9xaWl0YS11c2VyLWNvbnRlbnRzLmltZ2l4Lm5ldC9-dGV4dD9peGxpYj1yYi00LjAuMCZ3PTk2MCZoPTMyNCZ0eHQ9RmFicmljJUUzJTgxJUE4RGF0YWJyaWNrcyVFMyU4MSVBRSVFNyU5QiVCOCVFNCVCQSU5MiVFOSU4MSU4QiVFNyU5NCVBOCVFNiU4MCVBNyVFMiU5MSVBMCVFRiVCQyU5QWh1YiVFMyU4MiVCOSVFMyU4MyU4OCVFMyU4MyVBQyVFMyU4MyVCQyVFMyU4MiVCOCVFMyU4MSVBRSVFNyU5QiVBRSVFNyU5QSU4NCUyMC1EYXRhYnJpY2tzJTIwJUUzJTgxJUE3JUU0JUJEJTlDJUU2JTg4JTkwJUUzJTgxJTk3JUUzJTgxJTlGJUUzJTgzJTg2JUUzJTgzJUJDJUUzJTgzJTk2JUUzJTgzJUFCJUUzJTgyJTkyRmFicmljJUUyJTgwJUE2JnR4dC1hbGlnbj1sZWZ0JTJDdG9wJnR4dC1jb2xvcj0lMjMxRTIxMjEmdHh0LWZvbnQ9SGlyYWdpbm8lMjBTYW5zJTIwVzYmdHh0LXNpemU9NTYmdHh0LXBhZD0wJnM9NTUxOTY4M2NiNmQ1MDg1ZTBkM2Y0MTEzYmFiNDc5MWY%26mark-x%3D120%26mark-y%3D112%26blend64%3DaHR0cHM6Ly9xaWl0YS11c2VyLWNvbnRlbnRzLmltZ2l4Lm5ldC9-dGV4dD9peGxpYj1yYi00LjAuMCZ3PTgzOCZoPTU4JnR4dD0lNDBSZWlqaU90YWtlJnR4dC1jb2xvcj0lMjMxRTIxMjEmdHh0LWZvbnQ9SGlyYWdpbm8lMjBTYW5zJTIwVzYmdHh0LXNpemU9MzYmdHh0LXBhZD0wJnM9YzZmYWVkZTgyMDU5YzQwNDk3NWU4MjAwNDZmMjA4NGM%26blend-x%3D242%26blend-y%3D480%26blend-w%3D838%26blend-h%3D46%26blend-fit%3Dcrop%26blend-crop%3Dleft%252Cbottom%26blend-mode%3Dnormal%26s%3De3b7be4c153cdeefdfac49984abcf016" height="630" class="m-0" width="1200"&gt;
          &lt;/a&gt;
        &lt;/div&gt;
      &lt;div class="c-embed__body"&gt;
        &lt;h2 class="fs-xl lh-tight"&gt;
          &lt;a href="https://qiita.com/ReijiOtake/items/48c7b1e54796f4f569f3" rel="noopener noreferrer" class="c-link"&gt;
            FabricとDatabricksの相互運用性①：hubストレージの目的 -Databricks で作成したテーブルをFabricで利用する、Fabricで作成したテーブルをDatabricksで利用する- #Azure - Qiita
          &lt;/a&gt;
        &lt;/h2&gt;
          &lt;p class="truncate-at-3"&gt;
            最新版の記事にアップデートしているのでそちらをご覧ください。 【総集編】Microsoft Fabric と Databricksをつなぐデータ総合運用術 -hubストレージにDelta Lake形式で保管する- YouTubeのアーカイブでも、この記事について...
          &lt;/p&gt;
        &lt;div class="color-secondary fs-s flex items-center"&gt;
            &lt;img alt="favicon" class="c-embed__favicon m-0 mr-2 radius-0" src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fcdn.qiita.com%2Fassets%2Ffavicons%2Fpublic%2Fproduction-c620d3e403342b1022967ba5e3db1aaa.ico" width="120" height="120"&gt;
          qiita.com
        &lt;/div&gt;
      &lt;/div&gt;
    &lt;/div&gt;
&lt;/div&gt;


</description>
      <category>databricks</category>
      <category>azure</category>
      <category>microsoftfabric</category>
      <category>interoperability</category>
    </item>
    <item>
      <title>Fabric &amp; Databricks Interoperability (1): Purpose of Hub Storage for Table Sharing</title>
      <dc:creator>Reiji Otake</dc:creator>
      <pubDate>Sun, 16 Feb 2025 09:38:46 +0000</pubDate>
      <link>https://forem.com/_d2a1ea24c442526a9777/fabric-databricks-interoperability-1-purpose-of-hub-storage-for-table-sharing-30pl</link>
      <guid>https://forem.com/_d2a1ea24c442526a9777/fabric-databricks-interoperability-1-purpose-of-hub-storage-for-table-sharing-30pl</guid>
      <description>&lt;h1&gt;
  
  
  Introduction
&lt;/h1&gt;

&lt;p&gt;Although they each have their own characteristics, Microsoft Fabric and Databricks are broadly similar in what they can do.&lt;/p&gt;

&lt;p&gt;Through &lt;a href="https://learn.microsoft.com/ja-jp/fabric/database/mirrored-database/azure-databricks" rel="noopener noreferrer"&gt;Azure Databricks Unity Catalog mirroring&lt;/a&gt;, we are now able to reference Databricks-managed data in Fabric, but editing the data is still not possible.&lt;/p&gt;

&lt;p&gt;This brings up the following concerns:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Is it possible for our department to use tables managed by other departments in Databricks via our Fabric?&lt;/li&gt;
&lt;li&gt;I want to make the tables I create open for modification and reference, regardless of the tool used!&lt;/li&gt;
&lt;li&gt;We are currently using Databricks, but we might migrate to Fabric in the future... we want to maintain a vendor-free stance.&lt;/li&gt;
&lt;li&gt;Business-side employees use Fabric, but engineers use Databricks; there are times when we need to reference the same table.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;In this article, I will introduce use cases for seamlessly utilizing Fabric and Databricks:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;strong&gt;Using tables created in Databricks in Fabric&lt;/strong&gt;&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Using tables created in Fabric in Databricks&lt;/strong&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The goal of this article:&lt;br&gt;
&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F6lp4nkc7c21gx1jxdmux.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F6lp4nkc7c21gx1jxdmux.png" alt="image.png" width="800" height="260"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;:::note info&lt;br&gt;
This article consists of four parts:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Overview and purpose of interoperability (this article)&lt;/li&gt;
&lt;li&gt;&lt;a href="https://qiita.com/ReijiOtake/items/7ce8e1a743d05efe925c" rel="noopener noreferrer"&gt;Detailed setup of hub storage&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://qiita.com/ReijiOtake/items/86b44b2c30986c65db08" rel="noopener noreferrer"&gt;Using tables created in Fabric in Databricks&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;
&lt;a href="https://qiita.com/ReijiOtake/items/088abbd5f5ce06035501" rel="noopener noreferrer"&gt;Using tables created in Databricks in Fabric&lt;/a&gt;
:::&lt;/li&gt;
&lt;/ol&gt;
&lt;h1&gt;
  
  
  Prerequisite: Fabric and Databricks have similar functions... which one should we actually use?
&lt;/h1&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyimt27gcmj2amjs7y2pk.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyimt27gcmj2amjs7y2pk.png" alt="image.png" width="800" height="427"&gt;&lt;/a&gt;&lt;br&gt;
Fabric and Databricks are both attracting attention as Lakehouse platforms that handle data end-to-end.&lt;/p&gt;

&lt;p&gt;As someone new to the industry, my first impression of using both was that &lt;strong&gt;"They can probably do about the same things?"&lt;/strong&gt;.&lt;br&gt;
They both support ETL processes and AI model creation.&lt;/p&gt;

&lt;p&gt;Fabric is appealing because of its beginner-friendly GUI, designed for intuitive operations.&lt;br&gt;
On the other hand, Databricks is more code-based, so it seems to require a slightly higher skill level.&lt;br&gt;
Additionally, Databricks offers more customization options for computer resources, and if you stop the cluster frequently, it can be more cost-effective than Fabric.&lt;/p&gt;

&lt;p&gt;I believe the choice between these platforms depends on whether you prioritize ease of use or flexibility.&lt;/p&gt;

&lt;p&gt;▽Reference&lt;/p&gt;


&lt;div class="crayons-card c-embed text-styles text-styles--secondary"&gt;
    &lt;div class="c-embed__content"&gt;
        &lt;div class="c-embed__cover"&gt;
          &lt;a href="https://qiita.com/akihiro_suto/items/afaadd078c5a87772417" class="c-link align-middle" rel="noopener noreferrer"&gt;
            &lt;img alt="" src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fqiita-user-contents.imgix.net%2Fhttps%253A%252F%252Fqiita-user-contents.imgix.net%252Fhttps%25253A%25252F%25252Fcdn.qiita.com%25252Fassets%25252Fpublic%25252Fadvent-calendar-ogp-background-7940cd1c8db80a7ec40711d90f43539e.jpg%253Fixlib%253Drb-4.0.0%2526w%253D1200%2526blend64%253DaHR0cHM6Ly9xaWl0YS11c2VyLXByb2ZpbGUtaW1hZ2VzLmltZ2l4Lm5ldC9odHRwcyUzQSUyRiUyRnFpaXRhLWltYWdlLXN0b3JlLnMzLmFwLW5vcnRoZWFzdC0xLmFtYXpvbmF3cy5jb20lMkYwJTJGODA0NjclMkZwcm9maWxlLWltYWdlcyUyRjE3NDg2OTA1MDg_aXhsaWI9cmItNC4wLjAmYXI9MSUzQTEmZml0PWNyb3AmbWFzaz1lbGxpcHNlJmJnPUZGRkZGRiZmbT1wbmczMiZzPTRjODI0Y2NjYzY5ZDUyZDQzMGUwYjI3NzBlMmFkMGFi%2526blend-x%253D120%2526blend-y%253D467%2526blend-w%253D82%2526blend-h%253D82%2526blend-mode%253Dnormal%2526s%253D59c3587294dc9f02a5a32dbfb47a02e7%3Fixlib%3Drb-4.0.0%26w%3D1200%26fm%3Djpg%26mark64%3DaHR0cHM6Ly9xaWl0YS11c2VyLWNvbnRlbnRzLmltZ2l4Lm5ldC9-dGV4dD9peGxpYj1yYi00LjAuMCZ3PTk2MCZoPTMyNCZ0eHQ9QXp1cmUlMjBEYXRhYnJpY2tzJTIwJUUzJTgxJUE4JTIwTWljcm9zb2Z0JTIwRmFicmljJTIwJUUzJTgxJUFFJUU5JTk2JUEyJUU0JUJGJTgyJUUzJTgyJTkyJUU4JTgwJTgzJUUzJTgxJTg4JUUzJTgyJThCJUYwJTlGJUE3JTkwJnR4dC1hbGlnbj1sZWZ0JTJDdG9wJnR4dC1jb2xvcj0lMjMzQTNDM0MmdHh0LWZvbnQ9SGlyYWdpbm8lMjBTYW5zJTIwVzYmdHh0LXNpemU9NTYmdHh0LXBhZD0wJnM9MTliMTFiZDdiOTRjMWM5NDlhZjA2MzQyNDYzZmZkMzE%26mark-x%3D120%26mark-y%3D112%26blend64%3DaHR0cHM6Ly9xaWl0YS11c2VyLWNvbnRlbnRzLmltZ2l4Lm5ldC9-dGV4dD9peGxpYj1yYi00LjAuMCZ3PTgzOCZoPTU4JnR4dD0lNDBha2loaXJvX3N1dG8mdHh0LWNvbG9yPSUyMzNBM0MzQyZ0eHQtZm9udD1IaXJhZ2lubyUyMFNhbnMlMjBXNiZ0eHQtc2l6ZT0zNiZ0eHQtcGFkPTAmcz1lODU5ZDAzMTNjM2Y0ZmIwMjBjYjJlZmIwMWNhNmUyNg%26blend-x%3D242%26blend-y%3D480%26blend-w%3D838%26blend-h%3D46%26blend-fit%3Dcrop%26blend-crop%3Dleft%252Cbottom%26blend-mode%3Dnormal%26s%3D5d88657265f2b4a559ae148167e4e853" height="630" class="m-0" width="1200"&gt;
          &lt;/a&gt;
        &lt;/div&gt;
      &lt;div class="c-embed__body"&gt;
        &lt;h2 class="fs-xl lh-tight"&gt;
          &lt;a href="https://qiita.com/akihiro_suto/items/afaadd078c5a87772417" rel="noopener noreferrer" class="c-link"&gt;
            Azure Databricks と Microsoft Fabric の関係を考える🧐 #PowerBI - Qiita
          &lt;/a&gt;
        &lt;/h2&gt;
          &lt;p class="truncate-at-3"&gt;
            はじめに 本記事はDatabricks アドベントカレンダー2024 7日目の記事です。 本記事ではAzure Databricksを扱っています。 投稿日時点ではDatabricks on AWSなどでは利用できない機能もでてきます。 Azure Dat...
          &lt;/p&gt;
        &lt;div class="color-secondary fs-s flex items-center"&gt;
            &lt;img alt="favicon" class="c-embed__favicon m-0 mr-2 radius-0" src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fcdn.qiita.com%2Fassets%2Ffavicons%2Fpublic%2Fproduction-c620d3e403342b1022967ba5e3db1aaa.ico" width="120" height="120"&gt;
          qiita.com
        &lt;/div&gt;
      &lt;/div&gt;
    &lt;/div&gt;
&lt;/div&gt;


&lt;h1&gt;
  
  
  Considering interoperability methods
&lt;/h1&gt;

&lt;p&gt;Here, I will explore methods for achieving interoperability between Fabric and Databricks.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F71a20yxueztqddbfhaep.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F71a20yxueztqddbfhaep.png" alt="image.png" width="800" height="394"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  ① Unity Catalog Mirroring (Currently DML from Fabric to Databricks is not supported)
&lt;/h3&gt;

&lt;p&gt;Using &lt;a href="https://learn.microsoft.com/ja-jp/fabric/database/mirrored-database/azure-databricks" rel="noopener noreferrer"&gt;Azure Databricks Unity Catalog mirroring (preview)&lt;/a&gt;, it is possible to &lt;strong&gt;reference (SELECT statements) Databricks tables from Fabric&lt;/strong&gt;, but &lt;strong&gt;editing (DML statements)&lt;/strong&gt; is currently not supported.&lt;/p&gt;

&lt;p&gt;Thus, this method is not suitable for interoperability.&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Please let me know if my understanding is incorrect 🙇&lt;/em&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  ② Specifying OneLake as an external location from Databricks (Currently not supported)
&lt;/h3&gt;

&lt;p&gt;I tried configuring an external location for OneLake via &lt;a href="https://learn.microsoft.com/ja-jp/azure/databricks/connect/unity-catalog/cloud-storage/external-locations" rel="noopener noreferrer"&gt;cloud storage connection to Azure Databricks&lt;/a&gt;, but it is not currently supported.&lt;/p&gt;

&lt;p&gt;I was hoping this method would work, but unfortunately, it doesn't...&lt;br&gt;
So this method is also not suitable for interoperability.&lt;/p&gt;

&lt;p&gt;▽Reference&lt;/p&gt;


&lt;div class="crayons-card c-embed text-styles text-styles--secondary"&gt;
    &lt;div class="c-embed__content"&gt;
        &lt;div class="c-embed__cover"&gt;
          &lt;a href="https://qiita.com/ryoma-nagata/items/39fd52ab81015e3c9527#3-%E5%A4%96%E9%83%A8%E3%83%AD%E3%82%B1%E3%83%BC%E3%82%B7%E3%83%A7%E3%83%B3" class="c-link align-middle" rel="noopener noreferrer"&gt;
            &lt;img alt="" src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fqiita-user-contents.imgix.net%2Fhttps%253A%252F%252Fqiita-user-contents.imgix.net%252Fhttps%25253A%25252F%25252Fcdn.qiita.com%25252Fassets%25252Fpublic%25252Fadvent-calendar-ogp-background-7940cd1c8db80a7ec40711d90f43539e.jpg%253Fixlib%253Drb-4.0.0%2526w%253D1200%2526blend64%253DaHR0cHM6Ly9xaWl0YS11c2VyLXByb2ZpbGUtaW1hZ2VzLmltZ2l4Lm5ldC9odHRwcyUzQSUyRiUyRnFpaXRhLWltYWdlLXN0b3JlLnMzLmFwLW5vcnRoZWFzdC0xLmFtYXpvbmF3cy5jb20lMkYwJTJGMjgxODE5JTJGcHJvZmlsZS1pbWFnZXMlMkYxNjQ0MzAyODYyP2l4bGliPXJiLTQuMC4wJmFyPTElM0ExJmZpdD1jcm9wJm1hc2s9ZWxsaXBzZSZiZz1GRkZGRkYmZm09cG5nMzImcz1lZTU2ODE3Y2ZhNzY4YTU5NTI5ZTcwMTZhN2YyOWI1OQ%2526blend-x%253D120%2526blend-y%253D467%2526blend-w%253D82%2526blend-h%253D82%2526blend-mode%253Dnormal%2526s%253D3896d3e542a809393a85c3414dad7a97%3Fixlib%3Drb-4.0.0%26w%3D1200%26fm%3Djpg%26mark64%3DaHR0cHM6Ly9xaWl0YS11c2VyLWNvbnRlbnRzLmltZ2l4Lm5ldC9-dGV4dD9peGxpYj1yYi00LjAuMCZ3PTk2MCZoPTMyNCZ0eHQ9QXp1cmUlMjBEYXRhYnJpY2tzJTIwJUUzJTgxJThCJUUzJTgyJTg5JTIwT25lTGFrZSUyMCVFNCVCOCU4QSVFMyU4MSVBRSVFMyU4MyU4NyVFMyU4MyVCQyVFMyU4MiVCRiVFMyU4MSVBQiVFMyU4MiVBMiVFMyU4MiVBRiVFMyU4MiVCQiVFMyU4MiVCOSVFMyU4MSU5OSVFMyU4MiU4QiVFNiU5NiVCOSVFNiVCMyU5NSUyMDIwMjQlMkYxMiUyMCVFNyU4OSU4OCZ0eHQtYWxpZ249bGVmdCUyQ3RvcCZ0eHQtY29sb3I9JTIzM0EzQzNDJnR4dC1mb250PUhpcmFnaW5vJTIwU2FucyUyMFc2JnR4dC1zaXplPTU2JnR4dC1wYWQ9MCZzPTM2NTdmMjI4NjdiZmQ1ZGUwMTU0ODcwY2E4YWE1NTU2%26mark-x%3D120%26mark-y%3D112%26blend64%3DaHR0cHM6Ly9xaWl0YS11c2VyLWNvbnRlbnRzLmltZ2l4Lm5ldC9-dGV4dD9peGxpYj1yYi00LjAuMCZ3PTgzOCZoPTU4JnR4dD0lNDByeW9tYS1uYWdhdGEmdHh0LWNvbG9yPSUyMzNBM0MzQyZ0eHQtZm9udD1IaXJhZ2lubyUyMFNhbnMlMjBXNiZ0eHQtc2l6ZT0zNiZ0eHQtcGFkPTAmcz0xNTRkMWM1OWE4OGFmOTAwYzU0MTRiZDgzODY0Y2FlNg%26blend-x%3D242%26blend-y%3D480%26blend-w%3D838%26blend-h%3D46%26blend-fit%3Dcrop%26blend-crop%3Dleft%252Cbottom%26blend-mode%3Dnormal%26s%3D87c8d3d8e029cfe69fcf17784d4bb8b5" height="630" class="m-0" width="1200"&gt;
          &lt;/a&gt;
        &lt;/div&gt;
      &lt;div class="c-embed__body"&gt;
        &lt;h2 class="fs-xl lh-tight"&gt;
          &lt;a href="https://qiita.com/ryoma-nagata/items/39fd52ab81015e3c9527#3-%E5%A4%96%E9%83%A8%E3%83%AD%E3%82%B1%E3%83%BC%E3%82%B7%E3%83%A7%E3%83%B3" rel="noopener noreferrer" class="c-link"&gt;
            Azure Databricks から OneLake 上のデータにアクセスする方法 2024/12 版 #Microsoft - Qiita
          &lt;/a&gt;
        &lt;/h2&gt;
          &lt;p class="truncate-at-3"&gt;
            はじめに Azure Databricks の Unity Catalog ミラーリング を通して、Databricks の管理するデータについて Fabric で利用できるようになりましたが、Fabric レイクハウスなどの OneLake に格納されたデータに Azu...
          &lt;/p&gt;
        &lt;div class="color-secondary fs-s flex items-center"&gt;
            &lt;img alt="favicon" class="c-embed__favicon m-0 mr-2 radius-0" src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fcdn.qiita.com%2Fassets%2Ffavicons%2Fpublic%2Fproduction-c620d3e403342b1022967ba5e3db1aaa.ico" width="120" height="120"&gt;
          qiita.com
        &lt;/div&gt;
      &lt;/div&gt;
    &lt;/div&gt;
&lt;/div&gt;


&lt;h3&gt;
  
  
  ③ Hub storage as a solution
&lt;/h3&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F0go23z4rkvvzwxfxuemj.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F0go23z4rkvvzwxfxuemj.png" alt="image.png" width="800" height="447"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Since mirroring and external locations didn't work, I decided to store the actual tables in Azure Data Lake Gen 2 (ADLS2). &lt;br&gt;
By using a schema shortcut to ADLS2 from Fabric, and specifying ADLS2 as the storage location in Databricks' catalog, both Fabric and Databricks can perform SELECT and DML operations.&lt;/p&gt;

&lt;p&gt;This means that interoperability between Fabric and Databricks is now possible!&lt;/p&gt;

&lt;p&gt;From this point forward, I will refer to this ADLS2 as &lt;strong&gt;hub storage&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;▽For the specific setup method, refer to the following article:&lt;/p&gt;


&lt;div class="crayons-card c-embed text-styles text-styles--secondary"&gt;
    &lt;div class="c-embed__content"&gt;
        &lt;div class="c-embed__cover"&gt;
          &lt;a href="https://qiita.com/ReijiOtake/items/7ce8e1a743d05efe925c" class="c-link align-middle" rel="noopener noreferrer"&gt;
            &lt;img alt="" src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fqiita-user-contents.imgix.net%2Fhttps%253A%252F%252Fqiita-user-contents.imgix.net%252Fhttps%25253A%25252F%25252Fcdn.qiita.com%25252Fassets%25252Fpublic%25252Farticle-ogp-background-afbab5eb44e0b055cce1258705637a91.png%253Fixlib%253Drb-4.0.0%2526w%253D1200%2526blend64%253DaHR0cHM6Ly9xaWl0YS11c2VyLXByb2ZpbGUtaW1hZ2VzLmltZ2l4Lm5ldC9odHRwcyUzQSUyRiUyRnMzLWFwLW5vcnRoZWFzdC0xLmFtYXpvbmF3cy5jb20lMkZxaWl0YS1pbWFnZS1zdG9yZSUyRjAlMkYzOTIxOTk5JTJGOGY0NTg5ODc0MzczNDE5ZTJiZjhkMTJmNmQ0OGUyODcwZGU2OTJlMSUyRnhfbGFyZ2UucG5nJTNGMTc0MzgyMTEwMT9peGxpYj1yYi00LjAuMCZhcj0xJTNBMSZmaXQ9Y3JvcCZtYXNrPWVsbGlwc2UmYmc9RkZGRkZGJmZtPXBuZzMyJnM9OTRiZmVjNTRhNDllMDdlY2JiNjg5NzZlYWUwZGY5OTI%2526blend-x%253D120%2526blend-y%253D467%2526blend-w%253D82%2526blend-h%253D82%2526blend-mode%253Dnormal%2526s%253D728e1be30549a85eb776dc99c3b55165%3Fixlib%3Drb-4.0.0%26w%3D1200%26fm%3Djpg%26mark64%3DaHR0cHM6Ly9xaWl0YS11c2VyLWNvbnRlbnRzLmltZ2l4Lm5ldC9-dGV4dD9peGxpYj1yYi00LjAuMCZ3PTk2MCZoPTMyNCZ0eHQ9RmFicmljJUUzJTgxJUE4RGF0YWJyaWNrcyVFMyU4MSVBRSVFNyU5QiVCOCVFNCVCQSU5MiVFOSU4MSU4QiVFNyU5NCVBOCVFNiU4MCVBNyUyMCVFMiU5MSVBMSVFRiVCQyU5QUZhYnJpYyVFMyU4MSVBOERhdGFicmlja3MlRTMlODIlOTIlRTMlODMlOEYlRTMlODMlOTYlRTMlODIlQjklRTMlODMlODglRTMlODMlQUMlRTMlODMlQkMlRTMlODIlQjglRUYlQkMlODhEZWx0YSUyMExha2UlRUYlQkMlODklRTMlODElQTclRTYlOEUlQTUlRTclQjYlOUElRTMlODElOTklRTIlODAlQTYmdHh0LWFsaWduPWxlZnQlMkN0b3AmdHh0LWNvbG9yPSUyMzFFMjEyMSZ0eHQtZm9udD1IaXJhZ2lubyUyMFNhbnMlMjBXNiZ0eHQtc2l6ZT01NiZ0eHQtcGFkPTAmcz01Y2MzNmQ4NWMxMjc4NTJhM2Q5YmFlOTE0YTI3YTE2ZQ%26mark-x%3D120%26mark-y%3D112%26blend64%3DaHR0cHM6Ly9xaWl0YS11c2VyLWNvbnRlbnRzLmltZ2l4Lm5ldC9-dGV4dD9peGxpYj1yYi00LjAuMCZ3PTgzOCZoPTU4JnR4dD0lNDBSZWlqaU90YWtlJnR4dC1jb2xvcj0lMjMxRTIxMjEmdHh0LWZvbnQ9SGlyYWdpbm8lMjBTYW5zJTIwVzYmdHh0LXNpemU9MzYmdHh0LXBhZD0wJnM9YzZmYWVkZTgyMDU5YzQwNDk3NWU4MjAwNDZmMjA4NGM%26blend-x%3D242%26blend-y%3D480%26blend-w%3D838%26blend-h%3D46%26blend-fit%3Dcrop%26blend-crop%3Dleft%252Cbottom%26blend-mode%3Dnormal%26s%3D6c4da4c178ad4cf1c7d03b5770bf24fc" height="630" class="m-0" width="1200"&gt;
          &lt;/a&gt;
        &lt;/div&gt;
      &lt;div class="c-embed__body"&gt;
        &lt;h2 class="fs-xl lh-tight"&gt;
          &lt;a href="https://qiita.com/ReijiOtake/items/7ce8e1a743d05efe925c" rel="noopener noreferrer" class="c-link"&gt;
            FabricとDatabricksの相互運用性 ②：FabricとDatabricksをハブストレージ（Delta Lake）で接続する設定手順 #Azure - Qiita
          &lt;/a&gt;
        &lt;/h2&gt;
          &lt;p class="truncate-at-3"&gt;
            この記事は以下の記事の補足記事となります。 【総集編】Microsoft Fabric と Databricksをつなぐデータ総合運用術 -hubストレージにDelta Lake形式で保管する- YouTubeのアーカイブでも、9分45秒~あたりからこの記事につ...
          &lt;/p&gt;
        &lt;div class="color-secondary fs-s flex items-center"&gt;
            &lt;img alt="favicon" class="c-embed__favicon m-0 mr-2 radius-0" src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fcdn.qiita.com%2Fassets%2Ffavicons%2Fpublic%2Fproduction-c620d3e403342b1022967ba5e3db1aaa.ico" width="120" height="120"&gt;
          qiita.com
        &lt;/div&gt;
      &lt;/div&gt;
    &lt;/div&gt;
&lt;/div&gt;


&lt;h1&gt;
  
  
  Hub storage works thanks to Delta Lake
&lt;/h1&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fp1x17twwkdfvc8fgxlcg.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fp1x17twwkdfvc8fgxlcg.png" alt="image.png" width="370" height="302"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;As explained above, hub storage allows interoperability between Fabric and Databricks.&lt;/p&gt;

&lt;p&gt;But why does this interoperability work?&lt;/p&gt;

&lt;p&gt;The key lies in &lt;strong&gt;Delta Lake&lt;/strong&gt;, the mechanism behind it.&lt;/p&gt;

&lt;h3&gt;
  
  
  The mechanism of Delta Lake
&lt;/h3&gt;

&lt;p&gt;Delta Lake is an open-source storage layer that provides transaction and schema management on top of a data lake. It uses a combination of Parquet and JSON as its underlying data formats. Parquet is a columnar compression format that enables fast queries and data compression, while JSON is used as a transaction log to record data change history and versioning.&lt;/p&gt;

&lt;p&gt;By leveraging the Delta Lake mechanism, hub storage enables advanced data sharing and operations. When using Fabric or Databricks, it’s crucial to understand the underlying infrastructure to fully take advantage of the features provided by Delta Lake.&lt;/p&gt;

&lt;p&gt;▽For more details on Delta Lake, refer to the official documentation:&lt;/p&gt;


&lt;div class="crayons-card c-embed text-styles text-styles--secondary"&gt;
    &lt;div class="c-embed__content"&gt;
        &lt;div class="c-embed__cover"&gt;
          &lt;a href="https://learn.microsoft.com/ja-jp/azure/databricks/delta/" class="c-link align-middle" rel="noopener noreferrer"&gt;
            &lt;img alt="" src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Flearn.microsoft.com%2Fen-us%2Fmedia%2Fopen-graph-image.png" height="420" class="m-0" width="800"&gt;
          &lt;/a&gt;
        &lt;/div&gt;
      &lt;div class="c-embed__body"&gt;
        &lt;h2 class="fs-xl lh-tight"&gt;
          &lt;a href="https://learn.microsoft.com/ja-jp/azure/databricks/delta/" rel="noopener noreferrer" class="c-link"&gt;
            Azure Databricks の Delta Lake とは - Azure Databricks | Microsoft Learn
          &lt;/a&gt;
        &lt;/h2&gt;
          &lt;p class="truncate-at-3"&gt;
            Databricks レイクハウスに電源を供給するために使用される Delta Lake ストレージ プロトコルについて説明します。
          &lt;/p&gt;
        &lt;div class="color-secondary fs-s flex items-center"&gt;
          learn.microsoft.com
        &lt;/div&gt;
      &lt;/div&gt;
    &lt;/div&gt;
&lt;/div&gt;


&lt;h1&gt;
  
  
  Conclusion
&lt;/h1&gt;

&lt;p&gt;In this article, I summarized the purpose and methods of interoperability between Fabric and Databricks.&lt;/p&gt;

&lt;p&gt;The next article will provide a detailed guide for setting up hub storage.&lt;/p&gt;

&lt;p&gt;▽Next article:&lt;/p&gt;


&lt;div class="crayons-card c-embed text-styles text-styles--secondary"&gt;
    &lt;div class="c-embed__content"&gt;
        &lt;div class="c-embed__cover"&gt;
          &lt;a href="https://qiita.com/ReijiOtake/items/7ce8e1a743d05efe925c" class="c-link align-middle" rel="noopener noreferrer"&gt;
            &lt;img alt="" src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fqiita-user-contents.imgix.net%2Fhttps%253A%252F%252Fqiita-user-contents.imgix.net%252Fhttps%25253A%25252F%25252Fcdn.qiita.com%25252Fassets%25252Fpublic%25252Farticle-ogp-background-afbab5eb44e0b055cce1258705637a91.png%253Fixlib%253Drb-4.0.0%2526w%253D1200%2526blend64%253DaHR0cHM6Ly9xaWl0YS11c2VyLXByb2ZpbGUtaW1hZ2VzLmltZ2l4Lm5ldC9odHRwcyUzQSUyRiUyRnMzLWFwLW5vcnRoZWFzdC0xLmFtYXpvbmF3cy5jb20lMkZxaWl0YS1pbWFnZS1zdG9yZSUyRjAlMkYzOTIxOTk5JTJGOGY0NTg5ODc0MzczNDE5ZTJiZjhkMTJmNmQ0OGUyODcwZGU2OTJlMSUyRnhfbGFyZ2UucG5nJTNGMTc0MzgyMTEwMT9peGxpYj1yYi00LjAuMCZhcj0xJTNBMSZmaXQ9Y3JvcCZtYXNrPWVsbGlwc2UmYmc9RkZGRkZGJmZtPXBuZzMyJnM9OTRiZmVjNTRhNDllMDdlY2JiNjg5NzZlYWUwZGY5OTI%2526blend-x%253D120%2526blend-y%253D467%2526blend-w%253D82%2526blend-h%253D82%2526blend-mode%253Dnormal%2526s%253D728e1be30549a85eb776dc99c3b55165%3Fixlib%3Drb-4.0.0%26w%3D1200%26fm%3Djpg%26mark64%3DaHR0cHM6Ly9xaWl0YS11c2VyLWNvbnRlbnRzLmltZ2l4Lm5ldC9-dGV4dD9peGxpYj1yYi00LjAuMCZ3PTk2MCZoPTMyNCZ0eHQ9RmFicmljJUUzJTgxJUE4RGF0YWJyaWNrcyVFMyU4MSVBRSVFNyU5QiVCOCVFNCVCQSU5MiVFOSU4MSU4QiVFNyU5NCVBOCVFNiU4MCVBNyUyMCVFMiU5MSVBMSVFRiVCQyU5QUZhYnJpYyVFMyU4MSVBOERhdGFicmlja3MlRTMlODIlOTIlRTMlODMlOEYlRTMlODMlOTYlRTMlODIlQjklRTMlODMlODglRTMlODMlQUMlRTMlODMlQkMlRTMlODIlQjglRUYlQkMlODhEZWx0YSUyMExha2UlRUYlQkMlODklRTMlODElQTclRTYlOEUlQTUlRTclQjYlOUElRTMlODElOTklRTIlODAlQTYmdHh0LWFsaWduPWxlZnQlMkN0b3AmdHh0LWNvbG9yPSUyMzFFMjEyMSZ0eHQtZm9udD1IaXJhZ2lubyUyMFNhbnMlMjBXNiZ0eHQtc2l6ZT01NiZ0eHQtcGFkPTAmcz01Y2MzNmQ4NWMxMjc4NTJhM2Q5YmFlOTE0YTI3YTE2ZQ%26mark-x%3D120%26mark-y%3D112%26blend64%3DaHR0cHM6Ly9xaWl0YS11c2VyLWNvbnRlbnRzLmltZ2l4Lm5ldC9-dGV4dD9peGxpYj1yYi00LjAuMCZ3PTgzOCZoPTU4JnR4dD0lNDBSZWlqaU90YWtlJnR4dC1jb2xvcj0lMjMxRTIxMjEmdHh0LWZvbnQ9SGlyYWdpbm8lMjBTYW5zJTIwVzYmdHh0LXNpemU9MzYmdHh0LXBhZD0wJnM9YzZmYWVkZTgyMDU5YzQwNDk3NWU4MjAwNDZmMjA4NGM%26blend-x%3D242%26blend-y%3D480%26blend-w%3D838%26blend-h%3D46%26blend-fit%3Dcrop%26blend-crop%3Dleft%252Cbottom%26blend-mode%3Dnormal%26s%3D6c4da4c178ad4cf1c7d03b5770bf24fc" height="630" class="m-0" width="1200"&gt;
          &lt;/a&gt;
        &lt;/div&gt;
      &lt;div class="c-embed__body"&gt;
        &lt;h2 class="fs-xl lh-tight"&gt;
          &lt;a href="https://qiita.com/ReijiOtake/items/7ce8e1a743d05efe925c" rel="noopener noreferrer" class="c-link"&gt;
            FabricとDatabricksの相互運用性 ②：FabricとDatabricksをハブストレージ（Delta Lake）で接続する設定手順 #Azure - Qiita
          &lt;/a&gt;
        &lt;/h2&gt;
          &lt;p class="truncate-at-3"&gt;
            この記事は以下の記事の補足記事となります。 【総集編】Microsoft Fabric と Databricksをつなぐデータ総合運用術 -hubストレージにDelta Lake形式で保管する- YouTubeのアーカイブでも、9分45秒~あたりからこの記事につ...
          &lt;/p&gt;
        &lt;div class="color-secondary fs-s flex items-center"&gt;
            &lt;img alt="favicon" class="c-embed__favicon m-0 mr-2 radius-0" src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fcdn.qiita.com%2Fassets%2Ffavicons%2Fpublic%2Fproduction-c620d3e403342b1022967ba5e3db1aaa.ico" width="120" height="120"&gt;
          qiita.com
        &lt;/div&gt;
      &lt;/div&gt;
    &lt;/div&gt;
&lt;/div&gt;


</description>
      <category>interoperability</category>
      <category>azure</category>
      <category>databricks</category>
      <category>microsoftfabric</category>
    </item>
    <item>
      <title>How a New Graduate with No Industry Experience Passed the Databricks Certified Data Engineer Associate Exam in 3 Weeks</title>
      <dc:creator>Reiji Otake</dc:creator>
      <pubDate>Sun, 16 Feb 2025 07:56:12 +0000</pubDate>
      <link>https://forem.com/_d2a1ea24c442526a9777/how-a-new-graduate-with-no-industry-experience-passed-the-databricks-certified-data-engineer-j8</link>
      <guid>https://forem.com/_d2a1ea24c442526a9777/how-a-new-graduate-with-no-industry-experience-passed-the-databricks-certified-data-engineer-j8</guid>
      <description>&lt;h1&gt;
  
  
  Introduction
&lt;/h1&gt;

&lt;p&gt;As the title suggests, I am posting this as a memorandum to keep a record of my learning.&lt;br&gt;&lt;br&gt;
It's just a simple summary, but I hope it can be helpful for your studies as well.  &lt;/p&gt;

&lt;p&gt;If you have any additional questions, please feel free to leave a comment.  &lt;/p&gt;

&lt;h1&gt;
  
  
  Personal Impressions &amp;amp; Key Points
&lt;/h1&gt;

&lt;h3&gt;
  
  
  Grasping the Overall Picture
&lt;/h3&gt;

&lt;p&gt;By reading the &lt;a href="https://www.amazon.co.jp/%E3%83%87%E3%83%BC%E3%82%BF%E3%83%96%E3%83%AA%E3%83%83%E3%82%AF%E3%82%B9-%E3%82%AF%E3%82%A4%E3%83%83%E3%82%AF%E3%82%B9%E3%82%BF%E3%83%BC%E3%83%88%E3%82%AC%E3%82%A4%E3%83%89-%E3%83%87%E3%83%BC%E3%82%BF%E3%83%96%E3%83%AA%E3%83%83%E3%82%AF%E3%82%B9%E3%83%BB%E3%82%B8%E3%83%A3%E3%83%91%E3%83%B3-ebook/dp/B09V1YXFVQ/ref=sr_1_3?__mk_ja_JP=%E3%82%AB%E3%82%BF%E3%82%AB%E3%83%8A&amp;amp;crid=1B8GH67RI7HP1&amp;amp;dib=eyJ2IjoiMSJ9.QjgTf6G7XxomSI9f9hCE1K2qYT1U7IZIh47ExDCXNszKVaBTr_Z4GGJHOz4CG5IwyUn5ieAazLo8vLyGko-HKFbvsy69Wv-5RtjgXMhJ60h_C4-kOkMUPFbeuY7YBT6y0BJEw4UoKmML9hCZntFsVOsfsey_Pvw2CXddGPhE_rqzdqQwHkdR_I4c9vNxsOdEj1INDE93secmQ3SOoA9KEHxTGPWeWe1phgKmfwjolec6OBZq1QpqcyYztj6M0oK9eIlt3QVlNcp4QBaIZWtvMj_sy_DhYwd5FGPITqR9cyP-beIYhV1_NZ8j6RiNxzK9Y-xUVdY8M_-CePTy5bRqKQxtj-IN1fEItWjvniROMfwARoWjOqiGINi_pYxtMhsDXVxxU4Mu2LOTXJ8BRTnu2nMmzxVAFivSK8z-kRlJxj5dxpE_xV5aWc5uclkGeL8a.PXtPo4zK3rMLq5npaos2WoNM2L192fU3bhVfq5eskkE&amp;amp;dib_tag=se&amp;amp;keywords=Databricks&amp;amp;qid=1735199303&amp;amp;sprefix=databricks%2Caps%2C170&amp;amp;sr=8-3" rel="noopener noreferrer"&gt;Databricks Quick Start Guide&lt;/a&gt;, you can get an overview of the exam's flow and key points.  &lt;/p&gt;

&lt;h3&gt;
  
  
  Dealing with SQL Questions
&lt;/h3&gt;

&lt;p&gt;For those who have obtained Oracle Silver, the SQL section is relatively easy.&lt;br&gt;&lt;br&gt;
There were more than five SQL questions on the exam.  &lt;/p&gt;

&lt;h3&gt;
  
  
  Utilizing the Databricks Environment
&lt;/h3&gt;

&lt;p&gt;If you have access to a Databricks environment, actually using it and working through tutorials will deepen your understanding.&lt;br&gt;&lt;br&gt;
Fortunately, I had access to a study environment where I could freely experiment.  &lt;/p&gt;

&lt;h3&gt;
  
  
  Leveraging Official Documentation
&lt;/h3&gt;

&lt;p&gt;Going back to the official documentation is ultimately the fastest way to find answers.&lt;br&gt;&lt;br&gt;
In my case, I had a shallow understanding of Unity Catalog, DLT, structured streaming, and Git, so I studied by reading the official documentation.  &lt;/p&gt;

&lt;h1&gt;
  
  
  Study Timeline
&lt;/h1&gt;

&lt;h3&gt;
  
  
  24/12/23 Asked Senior Employees Who Passed for Recommended Study Materials
&lt;/h3&gt;

&lt;p&gt;&lt;a href="https://qiita.com/nttd-saitouyun/items/e7d1ca77e23b8e635518#%E3%83%87%E3%83%BC%E3%82%BF%E3%82%A8%E3%83%B3%E3%82%B8%E3%83%8B%E3%82%A2-%E3%82%A2%E3%82%BD%E3%82%B7%E3%82%A8%E3%82%A4%E3%83%88--databricks-certified-data-engineer-associate" rel="noopener noreferrer"&gt;Databricks Certifications: I Took Them All and Summarized Systematically&lt;/a&gt;  &lt;/p&gt;

&lt;h3&gt;
  
  
  2024/12/27 Databricks Quick Start Guide
&lt;/h3&gt;

&lt;p&gt;&lt;a href="https://www.amazon.co.jp/%E3%83%87%E3%83%BC%E3%82%BF%E3%83%96%E3%83%AA%E3%83%83%E3%82%AF%E3%82%B9-%E3%82%AF%E3%82%A4%E3%83%83%E3%82%AF%E3%82%B9%E3%82%BF%E3%83%BC%E3%83%88%E3%82%AC%E3%82%A4%E3%83%89-%E3%83%87%E3%83%BC%E3%82%BF%E3%83%96%E3%83%AA%E3%83%83%E3%82%AF%E3%82%B9%E3%83%BB%E3%82%B8%E3%83%A3%E3%83%91%E3%83%B3-ebook/dp/B09V1YXFVQ/ref=sr_1_3?__mk_ja_JP=%E3%82%AB%E3%82%BF%E3%82%AB%E3%83%8A&amp;amp;crid=1B8GH67RI7HP1&amp;amp;dib=eyJ2IjoiMSJ9.QjgTf6G7XxomSI9f9hCE1K2qYT1U7IZIh47ExDCXNszKVaBTr_Z4GGJHOz4CG5IwyUn5ieAazLo8vLyGko-HKFbvsy69Wv-5RtjgXMhJ60h_C4-kOkMUPFbeuY7YBT6y0BJEw4UoKmML9hCZntFsVOsfsey_Pvw2CXddGPhE_rqzdqQwHkdR_I4c9vNxsOdEj1INDE93secmQ3SOoA9KEHxTGPWeWe1phgKmfwjolec6OBZq1QpqcyYztj6M0oK9eIlt3QVlNcp4QBaIZWtvMj_sy_DhYwd5FGPITqR9cyP-beIYhV1_NZ8j6RiNxzK9Y-xUVdY8M_-CePTy5bRqKQxtj-IN1fEItWjvniROMfwARoWjOqiGINi_pYxtMhsDXVxxU4Mu2LOTXJ8BRTnu2nMmzxVAFivSK8z-kRlJxj5dxpE_xV5aWc5uclkGeL8a.PXtPo4zK3rMLq5npaos2WoNM2L192fU3bhVfq5eskkE&amp;amp;dib_tag=se&amp;amp;keywords=Databricks&amp;amp;qid=1735199303&amp;amp;sprefix=databricks%2Caps%2C170&amp;amp;sr=8-3" rel="noopener noreferrer"&gt;Databricks Quick Start Guide&lt;/a&gt;  &lt;/p&gt;

&lt;p&gt;Recommended for getting a broad overview.  &lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;Recommended Points&lt;/strong&gt;  &lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Affordable Price&lt;/strong&gt;

&lt;ul&gt;
&lt;li&gt;The Kindle version is available for 99 yen (this is the regular price, not a sale).
&lt;/li&gt;
&lt;li&gt;The same content is available for free on Qiita, but the Kindle version is compiled into a single volume, making it significantly easier to read. (Well worth the 99 yen.)
&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Reliable Information Source&lt;/strong&gt;

&lt;ul&gt;
&lt;li&gt;Published by Databricks Japan, ensuring high reliability.
&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Easy-to-Understand Explanations&lt;/strong&gt;

&lt;ul&gt;
&lt;li&gt;Clearly explains the background and history of Lakehouse.
&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Practical Content&lt;/strong&gt;

&lt;ul&gt;
&lt;li&gt;Includes ETL hands-on exercises, making it easy to apply in real-world scenarios.
&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;/blockquote&gt;

&lt;p&gt;　&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;Caution&lt;/strong&gt;&lt;br&gt;&lt;br&gt;
Since it was published in 2022, some information may be slightly outdated.  &lt;/p&gt;
&lt;/blockquote&gt;

&lt;h3&gt;
  
  
  2025/1/7~8 Official Practice Questions
&lt;/h3&gt;

&lt;p&gt;&lt;a href="https://qiita.com/kohei-arai/items/5b54a89cbaec801f1972" rel="noopener noreferrer"&gt;Translated &amp;amp; Explained Databricks Certified Data Engineer Associate Practice Questions&lt;/a&gt;&lt;br&gt;&lt;br&gt;
Questions 1-30 have been translated into Japanese and explained.  &lt;/p&gt;

&lt;p&gt;&lt;a href="https://qiita.com/nakazax/items/8f35cecb8f658b35e314" rel="noopener noreferrer"&gt;Unofficial Explanation of Databricks Certified Data Engineer Associate Practice Exam Answers (As of January 2024)&lt;/a&gt;&lt;br&gt;&lt;br&gt;
Questions 31-45 are in English, but explanations are provided in Japanese.  &lt;/p&gt;

&lt;h3&gt;
  
  
  2025/1/8~6 Udemy Course Recommended by Senior Employees
&lt;/h3&gt;

&lt;p&gt;&lt;a href="https://www.udemy.com/course/practice-exams-databricks-certified-data-engineer-associate/?couponCode=ST16MT28125CROW" rel="noopener noreferrer"&gt;Practice Exams: Databricks Certified Data Engineer Associate&lt;/a&gt;  &lt;/p&gt;

&lt;p&gt;Accuracy Rates:&lt;br&gt;&lt;br&gt;
Practice 1: 51% (23/45)&lt;br&gt;&lt;br&gt;
Practice 2: 62% (28/45)  &lt;/p&gt;

&lt;h3&gt;
  
  
  2025/1/16~19 Struggled with Structured Streaming, Delta Live Tables, and Git as a New Graduate
&lt;/h3&gt;

&lt;p&gt;I found them too difficult to understand, so I searched through documentation and tutorials.  &lt;/p&gt;

&lt;h4&gt;
  
  
  Relevant Resources:
&lt;/h4&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;a href="https://docs.databricks.com/ja/delta-live-tables/tutorial-pipelines.html#language-sql" rel="noopener noreferrer"&gt;Delta Live Tables SQL Tutorial&lt;/a&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;a href="https://docs.databricks.com/ja/structured-streaming/incremental.html" rel="noopener noreferrer"&gt;Incremental Processing with Structured Streaming&lt;/a&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;a href="https://docs.databricks.com/ja/structured-streaming/tutorial.html" rel="noopener noreferrer"&gt;Structured Streaming Tutorial&lt;/a&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;a href="https://docs.databricks.com/ja/jobs/jobs-quickstart.html" rel="noopener noreferrer"&gt;Quickstart Guide for Databricks Jobs&lt;/a&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;a href="https://docs.databricks.com/ja/jobs/index.html" rel="noopener noreferrer"&gt;Overview of Databricks Jobs&lt;/a&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;a href="https://learn.microsoft.com/ja-jp/azure/databricks/delta-live-tables/updates#development-and-production-modes" rel="noopener noreferrer"&gt;Azure Databricks Delta Live Tables Update Modes&lt;/a&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;a href="https://www.databricks.com/jp/spark/about" rel="noopener noreferrer"&gt;Introduction to Apache Spark&lt;/a&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;a href="https://docs.databricks.com/ja/delta-live-tables/index.html" rel="noopener noreferrer"&gt;Databricks Delta Live Tables Documentation&lt;/a&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;a href="https://www.databricks.com/jp/product/delta-live-tables" rel="noopener noreferrer"&gt;Databricks Delta Live Tables Overview&lt;/a&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;a href="https://qiita.com/taka_yayoi/items/e881769270f0ec0b7d06#databricks%E3%82%B8%E3%83%A7%E3%83%96%E3%81%A8%E3%81%AF" rel="noopener noreferrer"&gt;What is Databricks Jobs?&lt;/a&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;a href="https://qiita.com/ryoma-nagata/items/74e1bd9ebaf0413c9fd6" rel="noopener noreferrer"&gt;Understanding Databricks Workflows&lt;/a&gt;
&lt;/li&gt;
&lt;/ul&gt;

&lt;h4&gt;
  
  
  Git/GitHub Resources:
&lt;/h4&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;a href="https://www.youtube.com/watch?v=Dqgyc_S3L0s" rel="noopener noreferrer"&gt;Git/GitHub Introduction: Learn the Basics in 30 Minutes&lt;/a&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;a href="https://www.youtube.com/watch?v=1l8oBEown8c" rel="noopener noreferrer"&gt;What is Git/GitHub? 10-Minute Beginner Guide&lt;/a&gt;
&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  2025/1/20 Retook the Official Practice Exam
&lt;/h3&gt;

&lt;p&gt;Accuracy Rate: &lt;strong&gt;86%&lt;/strong&gt;  &lt;/p&gt;

&lt;h3&gt;
  
  
  2025/1/20 Registered for the Exam
&lt;/h3&gt;

&lt;p&gt;&lt;a href="https://www.databricks.com/jp/learn/certification/data-engineer-associate" rel="noopener noreferrer"&gt;Databricks Certified Data Engineer Associate Exam Registration&lt;/a&gt;  &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Warning:&lt;/strong&gt;&lt;br&gt;&lt;br&gt;
I do not recommend using a JCB card for payment.&lt;br&gt;&lt;br&gt;
(My confirmation email never arrived, but the payment was deducted. The follow-up with support was very time-consuming.)  &lt;/p&gt;

&lt;h3&gt;
  
  
  1/22 Passed the Exam!! 🎉
&lt;/h3&gt;

&lt;h1&gt;
  
  
  Conclusion
&lt;/h1&gt;

&lt;p&gt;Even as an industry newcomer with only 9 months of experience, I was able to pass on the first attempt!&lt;br&gt;&lt;br&gt;
I encourage you to give it a try!  &lt;/p&gt;

&lt;p&gt;Since then, I have continued using Databricks in my work and have found that the knowledge gained from certification studies has been directly applicable.  &lt;/p&gt;

&lt;p&gt;This was just a brief summary, but I hope it helps with your studies!  &lt;/p&gt;

</description>
      <category>databricks</category>
      <category>beginners</category>
      <category>certification</category>
    </item>
  </channel>
</rss>
