LakeFusion Deployment Guide via Azure Marketplace

LakeFusion Deployment Guide via Azure Marketplace

Prerequisite: Register Required Azure Resource Providers

Before deploying LakeFusion from the Azure Marketplace, ensure that the necessary resource providers are registered for your Azure subscription. Follow the steps below:

  1. Navigate to Subscriptions
    Go to the Azure portal and search for "Subscriptions" in the search bar.

  1. Select Your Subscription
    Choose the subscription you intend to use for the deployment (e.g., the one with ID 61c9c4c5-c556-4edf-88d6-0707e258fa78 in the screenshot).

  1. Access Resource Providers
    In the selected subscription view, click on "Resource providers" from the menu on the left pane.

  1. Register Required Providers
    Use the search bar at the top of the page to find and register the following providers. For each one:

    • Enter the provider name (e.g., Microsoft.Kubernetes)

  • Click on the provider name from the list

  • Click the "Register" button at the top of the page

  1. Repeat for Each of the Following Providers:

    • Microsoft.Kubernetes

    • Microsoft.KubernetesConfiguration

    • Microsoft.KubernetesRuntime

    • Microsoft.AzureTerraform

    • Microsoft.DBforMySQL

   All required providers must be registered to ensure a successful deployment of LakeFusion.         


Deployment Guide

Here is the step-by-step process to deploy LakeFusion, our Databricks-native Master Data Management (MDM) platform, using the Microsoft Azure Marketplace.

Step 1: Access the Azure Marketplace Listing

  • Visit the Azure Marketplace and search for “LakeFusion”, or use the direct link: LakeFusion on Azure Marketplace

  • Click the "Create" or “Get it Now” button to begin the deployment process.

Step 2: Complete the Deployment Form

You will be prompted to fill in necessary configuration details. The key fields include your Databricks authentication and workspace details.

2.1 Obtain Databricks Credentials

To connect LakeFusion with your Databricks workspace:

  1. Navigate to Azure Databricks Account Console.

  2. Log in using your Azure Databricks credentials.

  3. Go to Settings and select App Connections.

  1. Create a new app registration to obtain:

    • Client ID

    • Client Secret

2.2 Retrieve Workspace URL and Token

  • In your Azure Databricks workspace, locate the workspace you want to connect with LakeFusion.

  • Retrieve the Workspace URL.
    Example format:
    https://<databricks-instance>.azuredatabricks.net

  • Generate a Personal Access Token (PAT) for authentication:

    1. Go to Developer Settings > Access Tokens

    2. Click on Generate New Token

    3. Save the token securely

Step 3: Finalize and Create

  • Fill in the form with the retrieved Databricks details:

    • Workspace URL

    • Client ID

    • Client Secret

    • Token

  • Click "Review + create" to begin the deployment.

Step 4: Wait for Deployment Completion

  • Once the deployment process completes in Azure, allow approximately 5 minutes for the underlying resources to initialize properly before proceeding to the next steps.

Step 5: Retrieve the Public DNS for Access

  • Go to your Resource Group in the Azure Portal.

  • Locate the resource named aks-public-ip.

  • Click on the resource and copy the DNS name.


Step 6: Update DNS in LakeFusion App Connection

  • Paste the copied DNS into the Databricks Connection Settings in your LakeFusion instance.

  • This step is essential for proper API routing and platform functionality.

Deployment Complete!

LakeFusion is now deployed and ready to use within your Azure and Databricks environment. For help with configuration or accessing your deployed instance reach out to support.

    • Related Articles

    • Troubleshooting Guide

      This section outlines known issues encountered during setup or usage of LakeFusion, along with steps to resolve them. 1. Error Accessing the Match Maven Screen Issue: Users encounter an error when attempting to access the Match Maven screen within ...
    • LakeFusion Environment

      The LakeFusion Environment is the core of the platform and consists of two main layers: Micro Frontends: The user interface components. Microservices: The backend logic and processing components. a. Micro Frontends The LakeFusion Micro Frontends are ...
    • Data Flow in LakeFusion

      This section provides a structured overview of the LakeFusion Data Flow, outlining the key stages and enabling technologies that support seamless data ingestion, preprocessing, and Master Data Management (MDM). Each stage ensures data is unified, ...
    • Who is LakeFusion MDM for?

      LakeFusion is ideal for data-driven enterprises seeking to solve challenges related to fragmented data, poor data quality, and unreliable analytics. It serves a wide range of users across business and technical teams. Business Users Chief Data ...
    • Customer Environment

      The Customer Environment represents the user's access point to the LakeFusion platform. The key steps include: User Access: Customers access LakeFusion via their web browser. Authentication: Authentication is handled using Databricks OpenID Connect ...