Monday, April 20, 2026

Deep Dives, Best Practices, and What's Coming in Microsoft Fabric - FabCon 2026 - Part 2

 FabCon 2026 — Recap Part 2

Deep Dives, Best Practices, and What's Coming in Microsoft Fabric

My notes from Fabcon sessions covering Delta Lake internals, SQL in Fabric, Agentic AI with Fabric IQ, and the new OneLake Spark Catalog.


Published after FabCon 2026  ·  Sessions: Delta Lake • SQL Fun in Fabric • Fabric IQ & Foundry IQ • OneLake Spark Catalog

FabCon packed in a lot of technical depth in its sessions. These are my session notes, organized by topic. There's a lot of practical guidance here that I plan to take back and apply directly to our Fabric environment — especially around Delta Lake optimization and the OneLake Spark Catalog migration.


Session 1

Delta Lake Internals & Best Practices

This session covered a lot of ground on how Delta Lake actually works under the hood, and more importantly, what you should be doing (and not doing) to get the best performance.

What Delta Lake gives you

Delta Lake adds transactional features on top of Parquet — merge, update, delete, and insert — tracked via a JSON-based transaction log. The transaction log is the heart of Delta, providing:

  • Full ACID compliance
  • Decoupled compute and storage
  • Time travel and snapshot generation
  • Data skipping with min/max metadata on the first 32 columns (like an index)
  • Change tracking and logical deletes with recovery

V-order: write-time optimization

V-order is a write-time organization technique that makes reads up to 10x faster. It's particularly valuable for Power BI Direct Lake, and for silver and gold layers in a medallion architecture.

File size: it's complicated

One of the most nuanced parts of the session. The right file size strategy depends on your query pattern:

  • Power BI prefers fewer, larger files — one 20 GB file with 160 row groups of 128 MB each outperforms 160 individual 128 MB files.
  • Selective lookups benefit from more smaller files, because each file has its own min/max values for skipping irrelevant data.
  • The worst scenario: 5,000 small files with 4 MB row groups perform about 50x worse than 160 files with 128 MB row groups. The fix is to run Optimize.
Key recommendation
Always use Adaptive File Target Size: set spark.microsoft.delta.targetFileSize.adaptive.enabled = True. This automatically evaluates your table size and applies the best file size settings for the data.

Optimize, Fast Optimize, and Auto Compaction

  • Fast Optimize reduces write amplification and only compacts bins when estimated as optimal — it's about 5x faster than standard Optimize. Use Fast Optimize.
  • Auto Compaction (spark.databricks.delta.autoCompact.enabled=True) evaluates automatically when compaction is needed. Set minNumFiles = 50 as a starting point. Always use Auto Compaction.
  • Optimize Write is a pre-write compaction that's faster for streaming, small batch, and partitioned tables — but don't use it when adaptive file size is enabled.

Deletion vectors

Parquet files are immutable. Historically, a delete meant rewriting an entire file. Deletion vectors solve this by writing a small sidecar file that says "skip this record when reading" — the original Parquet file is untouched. Enable them with:

spark.databricks.delta.properties.defaults.enableDeletionVectors = True

Liquid clustering

Liquid clustering improves file skipping by physically ordering data on disk by a chosen key. A few important rules:

  • The clustering unit (zcube) is 100 GB — so data will be rewritten.
  • Don't use it for tables under 100 GB. It's most valuable at 1 TB and above.
  • Similarly, don't partition tables under 1 TB.

Delta Lake 4.1 highlights

  • Optimized liquid clustering: now incremental, with auto-reclustering — 2.6x cheaper and faster, and it works with Fast Optimize.
  • Type widening: automatically promotes column types (e.g., int to bigint) — enabled by default.
  • Multi-table transactions: coming soon.
Golden configuration
Enable deletion vectors + adaptive target file size + auto compaction. Use Spark resource profiles to ensure data is saved according to profile requirements.

Session 2

SQL Fun in Fabric

This session (presented by Shabnam Watson — check out shabnamwatson.com and the O'Reilly book Analytics Engineering with MS Fabric and Power BI) covered the landscape of SQL options in Fabric and the practical differences between them.

Read-only vs. read-write

  • Lakehouse: read-only from a SQL analytics endpoint perspective.
  • Data Warehouse and Fabric SQL: read/write.

Fabric SQL database

Fabric SQL is the relational database option in Fabric. Key characteristics:

  • Performance is controlled by capacity size.
  • Uses MDF & LDF files, with a SQL analytics endpoint that mirrors those files as Parquet in OneLake in near real-time.
  • Restores are only possible within Fabric.
  • Has SQL Audit Logs.
  • Enforces foreign key constraints (unlike the SQL Warehouse).
  • Has a long GUID in its connection string name; the analytics endpoint does not.

SQL Warehouse

The Warehouse is the distributed, OLAP-oriented engine — equivalent to Synapse Analytics. Key notes:

  • Two execution engine pools: one for SELECT, one for DML.
  • Parquet column storage means no clustered indexes — only nonclustered indexes are available.
  • Foreign key constraints are not enforced.
  • Built-in backups with restore points that are pointers to Delta history. Manual creation is available, but restore is in-place only.
  • Use COPY INTO <Table> for bulk inserts — Bulk Insert is deprecated and slower.
  • Use OPENROWSET to query across workspaces.
  • Use bigint for ID columns.

Session 3

Fabric IQ & Foundry IQ: Making AI Agents Productive

This session introduced Microsoft's vision for AI Agents in the Fabric ecosystem — the goal being to make agents "as trusted and productive as the best employees."

Three layers of IQ

  • Fabric IQ: Context on how your business operates — business entities, systems of record, and actions that bind data to Fabric IQ through an Ontology (i.e., a knowledge graph that models your environment and how your business works).
  • Foundry IQ: The tools for building and deploying enterprise AI apps securely — point-and-click data transformation that automatically chunks data, with network isolation including custom VNet and Private Links.
  • Work IQ: Context on people — collaboration and workflows.
Ontology = Knowledge Graph
The "Ontology" in Fabric IQ is essentially a knowledge graph. You search the ontology to create a knowledge source for your AI agents. One token ≈ 4 characters of text.

Agentic AI with Fabric IQ

The broader vision here is a shift from human operators to human governors — moving from 100% human-driven operations toward a model where roughly 45–55% of the work is AI-powered. Fabric IQ's ontology provides the business meaning that makes this possible.

Status Preview
Fabric IQ & Foundry integration is currently in preview.
Further reading
ai.azure.com →

Session 4

OneLake Spark Catalog & Migrating to Schema-Enabled Lakehouses

This was probably the most operationally impactful session of the day for anyone running Fabric at scale. It covered how Spark discovers data and the significant improvements coming with the new OneLake Spark Catalog.

The old vs. new Spark catalog

The legacy Spark catalog maps table names to file locations and works within a single workspace. Its problems: chatty, slow, no row-level or column-level security, no VNet support.

The new OneLake Spark Catalog is a fundamentally different architecture:

  • Metadata lives with the data in the Delta log — no separate metastore.
  • Immediate table detection after ingestion.
  • Multi-workspace queries.
  • Compatible with schema-enabled lakehouses.
  • Can resolve anything the tenant can access — including Databricks catalogs and external tables via shortcuts.
  • Supports OneLake row-level security and column-level security.

Naming conventions matter

  • Local contextSELECT * FROM product — only for the linked lakehouse.
  • 2-part namemarket.product — ambiguous (schema or another lakehouse?).
  • 3-part name: cross data-source.
  • 4-part name: fabric-wide cross-workspace query.

For non-schema lakehouses, use dbo as the schema name.

Should you switch to schema-enabled lakehouses?

Switch now if...
All Fabric resources are governed by a single group, shortcut usage is well-known, and notebooks and other code are well-documented.
Delay the switch if...
Governance is fragmented, or most tables are interlinked between multiple storage items.

What breaks when you switch

The switch is instant and irreversible. Tables move under the dbo schema — no data files are moved, but the following will break:

  • Notebook code referencing tables by old names.
  • RLS/CLS policies (they need schema references updated).
  • Semantic models.
  • Shortcuts pointing to the lakehouse.

Migration checklist

  • Copy catalog elements to OneLake (Microsoft has a script for this).
  • Enable the OneLake catalog.
  • Test notebook runs.
  • Refactor notebooks to reference tables with the dbo schema.
  • Run PySpark scripts to schema-enable the lakehouses.
  • Update OneLake security policies, semantic models, and Direct Lake references using the provided scripts.
  • Update all shortcuts pointing to updated lakehouses. Note: there is currently no lineage information to help find shortcuts automatically.
  • The SQL analytics endpoint won't change.

What's coming

Microsoft is working on a migration assistant, a Copilot agent with Fabric skills, and a fully automated switch process.

Scripts & instructions
aka.ms/LakehouseSwitch →Notebook scripts and property switch API instructions for enabling schema-enabled lakehouses.

These are personal session notes from FabCon 2026 — Part 2. Accuracy reflects my understanding at the time; always verify details against official Microsoft documentation before making architectural decisions.

Friday, April 17, 2026

Data, Agents, and the Fabric of Everything - FabCon 2026 - Part 1

FabCon 2026 · Recap - Part 1

Data, Agents, and the Fabric of Everything

My notes and takeaways from the Microsoft Fabric Conference 2026 — from the keynote to sessions on AI agents, Purview, and responsible AI.


Keynote

Microsoft IQ: A Unified Intelligence Layer

The keynote set the tone for the entire conference: AI agents should be empowered with the same knowledge and context as your employees — understanding how the business works, how employees work, and drawing on curated, governed data to do so.



Microsoft framed this vision around three pillars of intelligence:

  • Work IQ — AI agents embedded in Office 365 applications
  • Foundry IQ — curated enterprise knowledge, with visibility into SharePoint and other knowledge sources
  • Fabric IQ — the state of your core business data, surfaced through OneLake and Power BI semantic models

One memorable aside from the keynote: Fabric releases happen every single week. If Fabric feels different every time you log in, that's by design.

CI/CD and DevOps in Fabric

End-to-end CI/CD capabilities got a meaningful upgrade. Highlights include selective branching, the ability to compare workspace changes before committing, visualization of git branch structure, parameterized deployments, and a diff view before check-in. A variable library feature now lets you select environment values from a dropdown rather than typing IDs manually — a small but welcome quality-of-life improvement.



OneLake Catalog

The OneLake Catalog gains AI-powered data descriptions and governance tabs for sensitivity labels. Security is now define-once-enforce-everywhere — a significant step for large organizations managing sprawling data estates.

Capacity Management

New capacity tooling includes real-time alerting and usage tracking via the Real-Time Hub, workspace surge protection to keep mission-critical workloads alive, and automatic overage billing controls to prevent surprise throttling.


SQL & Azure SQL

SQL Gets an AI Makeover

SQL Server 2025 arrives with AI capabilities built in from the ground up, along with JSON support baked directly into the engine. GitHub Copilot is now available inside SSMS 2022, enabling chat-with-your-data experiences, schema-aware grouping, improved export options, and SSDT project support directly in the IDE.

Azure SQL highlights

  • SQL MCP (Model Context Protocol) support
  • Larger core options for demanding workloads
  • Vector index improvements for AI retrieval scenarios

SQL database in Fabric

The SQL database experience inside Fabric now reaches engine parity, is positioned as enterprise-ready, and includes a migration assistant to help teams move existing workloads in.

Database Hub

One of the more exciting announcements: a Database Hub providing a unified view of all databases in your environment — including on-premises databases. From here, database agents can monitor and manage an entire fleet, track capacity utilization, handle compliance and security, and proactively alert on and remediate issues before they become outages.


Data Platform

Unifying the Data Estate with OneLake

A recurring theme of the day was the idea of OneLake as the "OneDrive for data" — a single logical store where data can either be physically stored or referenced via shortcuts. The session walked through several major capabilities:

Shortcuts and mirroring

  • Cross-cloud shortcuts let you point to data wherever it lives
  • SAP and Oracle mirroring are generally available
  • SharePoint mirroring is coming soon
  • Shortcut transformations allow in-place views: CSV to table, Excel sheets to individual tables, Parquet and JSON transformations

Mirroring improvements

Delta change data feed support, fine-grained incremental change capture, and — notable for Databricks shops — bidirectional OneLake sharing, meaning you can now read and write both ways between Fabric and Databricks.

Data Factory and Data Warehouse

The Data Factory Migration Assistant now offers full parity with the Fabric Data Factory, removing a major blocker for teams looking to migrate from ADF. New capabilities in the data warehouse include materialized lake views (enabling cross-lakehouse queries), multimodal AI functions, custom SQL pools, AI functions, and real-time alerts. Graph capabilities are also coming to Fabric.

Fabric for planning and analysis

Fabric is expanding into enterprise planning territory — budgets, forecasts, goals, plans vs. actuals. Two agent types got their own spotlight: Data Agents act as virtual analysts that answer questions about your data, while Operation Agents provide 24/7 monitoring and can act on data autonomously. Semantic models remain the key to making data visible and understandable to AI.

Live Pools — pre-warmed compute pools that can be scheduled to spin up at a specific time — are a welcome addition for teams running time-sensitive workloads.


AI Agents & Azure AI Foundry

The Agent Revolution: From Single Models to Orchestrated Fleets

"Possible to have 1.3 billion AI agents by 2028."

This session reframed where we are in the agent evolution. Early AI agents were single-model, manually coded. Next-generation agents are multi-model with automated workflows. The direction is toward dynamic, AI-driven orchestration with end-to-end security — eventually reaching a state where AI creates its own CI/CD pipelines.

Microsoft Azure AI Foundry

Foundry is Microsoft's platform for building, orchestrating, and governing AI agents and applications. Its pillars include:

  • AI App and Agent Orchestration
  • Model selection and hosting
  • Knowledge tools and retrieval
  • Observability and agent controls
  • Fine-tuning, customization, and edge/local deployment

Foundry Agent Service

Open source and designed for interoperability, Foundry Agent Service supports "hosted agents" so other systems can hook in. Microsoft introduced the Microsoft Agent Framework — an open-source SDK with open standards for building and orchestrating intelligent agents.

Foundry IQ

Built on Azure AI Search, Foundry IQ is specialized for enterprise AI retrieval. It can ingest an AI-optimized copy of your data, index remote sources, and serve as the knowledge backbone for your agents. The OneLake Catalog is also surfaced within Foundry, connecting data governance and AI retrieval in a single plane.

The new Foundry URL is https://ai.azure.com.


Purview & Governance

Governing Data at Scale: Purview and AI-Driven MDM

Fabric enables teams to work with data closer to the source, but that freedom requires a strong governance layer. Purview is that centralized control plane, spanning:

  • Data security
  • Data governance
  • Risk and compliance
  • Data loss prevention

Key capabilities

  • Information Protection — sensitivity labels and policies (Public / Internal / Confidential)
  • Data Loss Prevention — monitor, detect, and act on sensitive data movement
  • Insider Risk Management — protection against both malicious and inadvertent insider risk
  • Data Security Posture Management — surface, discover, and assess risks proactively

AI-Driven Master Data Management (MDM) on Fabric

Perhaps the most impactful announcement in this session: AI-driven MDM built directly on Fabric, designed to create the "Golden Record" — a single, trusted source of truth for key business entities. Without MDM, organizations suffer from duplicates across systems, inconsistent formats and attributes, and limited cross-system visibility. AI-driven MDM addresses this by:

  • Cleaning data using AI-powered rule suggestions
  • Enriching records by filling in missing data
  • Standardizing values (e.g., "US" vs. "USA")
  • Consolidating from multiple source systems

AI & SQL Deep Dive (Bob Ward)

Practical AI: Models, MCPs, and Responsible Deployment

This session — featuring Bob Ward, whose open-source presentations are available at https://aka.ms/bobwardms — was a grounding counterbalance to the big-vision keynote energy. Key takeaways:

The architecture of AI

  • AI models are algorithms — your AI apps control everything
  • Smaller models are cheaper and often sufficient; bigger is not always better
  • MCP (Model Context Protocol) is the standard method to run and discover tools — described as "COM objects for AI" or "USB for AI"
  • Agents use tools and/or MCP servers to act on your behalf

Tools mentioned

  • LM Studio — for experimenting with local models
  • Azure AI Gateway (AIPM) — can protect against jailbreaking by scanning prompts and responses
  • GitHub Copilot Instructions — use as a system prompt to guide AI behavior in your workflows

Responsible AI principles from the session

  • Use vetted and trusted AI models; audit and govern AI model usage
  • Tune the temperature for responses appropriate to your use case
  • Better prompts = better responses; evaluate the quality of your RAG pipeline
  • Always consider what data permissions are required
  • Ask models questions — don't tell them what to do — and do not turn "auto approve" on
  • Create a responsible AI policy for your organization
Bob Ward signing his new book. The photographer didn't wait for me to finish turning before she snapped the picture, so it's blurry. 



That wraps Part 1 of my report. It was a lot to absorb — Fabric's surface area keeps expanding, and the AI agent story is moving fast. More session recaps to follow. 


Mark your calendar: the next FabCon is in Atlanta, March 8–12, 2027.


Full agenda: https://aka.ms/FabCon-Agenda 



Deep Dives, Best Practices, and What's Coming in Microsoft Fabric - FabCon 2026 - Part 2

  FabCon 2026 — Recap Part 2 Deep Dives, Best Practices, and What's Coming in Microsoft Fabric My notes from Fabcon sessions covering De...