Edit

Share via


Transform structured files into Delta tables

Use shortcut transformations to convert structured files into queryable Delta tables. If your source data is already in a tabular format like CSV, Parquet, or JSON, file transformations automatically copy and convert that data into Delta Lake format so you can query it by using SQL, Spark, or Power BI without building ETL pipelines.

For unstructured text files that need AI processing like summarization, translation, or sentiment analysis, see Shortcut transformations (AI-powered).

Shortcut transformations stay always in sync with the source data. Fabric Spark compute executes the transformation and copies the data referenced by a OneLake shortcut into a managed Delta table. With automatic schema handling, deep flattening capabilities, and support for multiple compression formats, shortcut transformations eliminate the complexity of building and maintaining ETL pipelines.

Note

Shortcut transformations are currently in public preview and are subject to change.

Why use shortcut transformations?

  • Automatic conversion – Fabric copies and converts source files to Delta format without manual pipeline orchestration.
  • Frequent sync – Fabric polls the shortcut every two minutes and synchronizes changes.
  • Delta Lake output – The resulting table is compatible with any Apache Spark engine.
  • Inherited governance – The shortcut inherits OneLake lineage, permissions, and Microsoft Purview policies.

Prerequisites

Requirement Details
Microsoft Fabric SKU Capacity or trial that supports Lakehouse workloads.
Source data A folder that contains homogeneous CSV, Parquet, or JSON files.
Workspace role Contributor or higher.

Supported sources, formats, and destinations

All data sources supported in OneLake are supported.

Source file format Destination Supported Extensions Supported Compression types Notes
CSV (UTF-8, UTF-16) Delta Lake table in the Lakehouse / Tables folder .csv, .txt (delimiter), .tsv (tab-separated), .psv (pipe-separated) .csv.gz, .csv.bz2 .csv.zip and .csv.snappy aren't supported.
Parquet Delta Lake table in the Lakehouse / Tables folder .parquet .parquet.snappy, .parquet.gzip, .parquet.lz4, .parquet.brotli, .parquet.zstd
JSON Delta Lake table in the Lakehouse / Tables folder .json, .jsonl, .ndjson .json.gz, .json.bz2, .jsonl.gz, .ndjson.gz, .jsonl.bz2, .ndjson.bz2 .json.zip and .json.snappy aren't supported.

Set up a shortcut transformation

  1. In your lakehouse, select New Table Shortcut in the Tables section, which is Shortcut transformation (preview). Choose your source (for example, Azure Data Lake, Azure Blob Storage, Dataverse, Amazon S3, GCP, SharePoint, OneDrive, and more).

    Screenshot that shows creating 'table shortcut'.

  2. Choose file, Configure transformation, and create shortcut – Browse to an existing OneLake shortcut that points to the folder with your CSV files, configure parameters, and initiate creation.

    • Delimiter in CSV files – Select the character used to separate columns (comma, semicolon, pipe, tab, ampersand, space).
    • First row as headers – Indicate whether the first row contains column names.
    • Table Shortcut name – Provide a friendly name; Fabric creates it under /Tables.
  3. Track refreshes and view logs for transparency in Manage Shortcut monitoring hub.

Fabric Spark compute copies the data into a Delta table and shows progress in the Manage shortcut pane. Shortcut transformations are available in Lakehouse items. They create Delta Lake tables in the Lakehouse / Tables folder.

How synchronization works

After the initial load, Fabric Spark compute:

  • Polls the shortcut target every two minutes.
  • Detects new or modified files and appends or overwrites rows accordingly.
  • Detects deleted files and removes corresponding rows.

Monitor and troubleshoot

Shortcut transformations include monitoring and error handling to help you track ingestion status and diagnose problems.

  1. Open the lakehouse and right-click the shortcut that feeds your transformation.

  2. Select Manage shortcut.

  3. In the details pane, you can view:

    • Status – Last scan result and current sync state.
    • Refresh history – Chronological list of sync operations with row counts and any error details.

    Screenshot that shows 'monitoring hub' for viewing transformation status.

  4. View more details in logs to troubleshoot

    Screenshot that shows how to access 'log file' to troubleshoot.

Limitations

Current limitations of shortcut transformations:

  • Source format: Only CSV, JSON, and Parquet files are supported.
    • Unsupported datatypes for CSV: Mixed data type columns, Timestamp_Nanos, Complex logical types - MAP/LIST/STRUCT, Raw binary
    • Unsupported datatypes for Parquet: Timestamp_nanos, Decimal with INT32/INT64, INT96, Unassigned integer types - UINT_8/UINT_16/UINT_64, Complex logical types - MAP/LIST/STRUCT
    • Unsupported datatypes for JSON: Mixed data types in an array, Raw binary blobs inside JSON, Timestamp_Nanos
  • File schema consistency: Files must share an identical schema.
  • Workspace availability: Available only in Lakehouse items (not Data Warehouses or KQL databases).
  • Write operations: Transformations are read-optimized; direct MERGE INTO or DELETE statements on the transformation target table aren't supported.
  • Flattening of Array data type in JSON: Array data type is retained in delta table and data accessible with Spark SQL and Pyspark. For further transformations, Fabric Materialized Lake Views could be used for silver layer.
  • Flattening depth in JSON: Nested structures are flattened up to five levels deep. Deeper nesting requires preprocessing.

Use the Fabric Roadmap and Fabric Updates Blog to learn about new features and releases.

Clean up

To stop synchronization, delete the shortcut transformation from Lakehouse Explorer.

Deleting the transformation doesn't remove the underlying files.