MENU
    Microsoft Fabric Integration
    • 09 May 2024
    • 1 Minute to read
    • Contributors

    Microsoft Fabric Integration


    Article summary

    Streamline sending data from Tulip to Microsoft Fabric for broader analytics opportunities

    Purpose

    This guide walks through step by step how to send Tulip tables data to Fabric via a paginated REST API query. This is useful for analyzing Tulip tables data in an Azure Data Lakehouse, Synapse Data Warehouse or other Azure Data storage location

    The broader architecture for Fabric -- Microsoft Cloud for Manufacturing is shown below

    image

    Setup

    Detailed setup video is embedded below:



    Overall Requirements:

    • Usage of Tulip Tables API (Get API Key and Secret in Account Settings)
    • Tulip Table (Get the Table Unique ID

    Process:

    1. On Fabric homepage, go to Data Factory
    2. Create a new Data Pipeline on Data Factory
    3. Start with the "Copy Data Assistant" to streamline creation process
    4. Copy Data Assistant Details:
      1. Data Source: REST
      2. Base URL: https://[instance].tulip.co/api/v3
      3. Authentication type: Basic
      4. Username: API Key from Tulip
      5. Password: API Secret from Tulip
      6. Relative URL: tables/[TABLE_UNIQUE_ID]/records?limit=100&offset={offset}
      7. Request: GET
      8. Pagination Option Name: QueryParameters.{offset}
      9. Pagination Option Value: RANGE:0:10000:100

    Note: Limit can be lower than 100 if needed, but the increment in the pagination needs to match
    Note: the Pagination Value for the range needs to be greater than the number of records in the table

    image.png

    1. Then you can send this data to a variety of data destinations such as Azure Lakehouse
    2. Connect to the data destination, review the fields / rename as needed, and you are all set!

    Example Architecture

    image.png

    Use Cases and Next Steps

    Once you have finalized the connection, you can easily analyze the data with a spark notebook, PowerBI, or a variety of other tools.

    1. Defect prediction

    • Identify production defects before they happen and increase right first time.
    • Identify core production drivers of quality in order to implement improvements

    2. Cost of quality optimization

    • Identify opportunities to optimize product design without impact customer satisfaction

    3. Production energy optimization

    • Identify production levers to optimal energy consumption

    4. Delivery and planning prediction and optimization

    • Optimize production schedule based on customer demand and real time order schedule

    5. Global Machine / Line Benchmarking

    • Benchmark similar machines or equipment with normalization

    6. Global / regional digital performance management
    Consolidated data to create real time dashboards


    Was this article helpful?