-2

I'm attempting to copy a table from an on-premises SQL Server to Azure Storage using a data pipeline in Fabric. The table consists of 5 columns and 1000 rows, with all columns being of type varchar. The primary key is also a string (a combination of varchar and uniqueString). When I tried to enable dynamic range partitioning, I encountered the error "SqlParallelNotSupportedDataType". Below is the structure of my table:

Column A Column B Column C Column D
unqStr1 value1 value2 value2
unqStr2 value3 value5 value1
unqStr3 value3 value5 value1
unqStr4 value3 value5 value1
unqStr5 value3 value5 value1

Is there a way to copy this type of data to Azure Storage using multiple threads to improve data read performance? Specifically, I'm looking for a solution that leverages parallel threads for better performance. Any guidance or suggestions would be greatly appreciated.

There's a sink setting called "Max rows per file," which I believe is intended for writing. However, my primary goal is to achieve better performance when reading from the database.

1
  • Have you tried using other data types (like int, datetime, or uniqueidentifier) for your primary key or other columns to see if the parallel copy works without the SqlParallelNotSupportedDataType error? Commented Mar 12 at 5:51

1 Answer 1

0

The error SqlParallelNotSupportedDataType occurred because dynamic range partitioning in Azure Data Factory (ADF) or Fabric only supports numeric or date-based columns for partitioning

Since all the columns in your table are of type VARCHAR, dynamic partitioning cannot be applied directly.

I've tried the same process by adding a DATE column in the table and it got succeeded without any data-type error:

When Degree of Copy Parallelism is set to Auto:

When it is set to 5 or higher:

Techniques to enhance data read performance:

  1. Increase Degree of Copy Parallelism:
  • Set Degree of Copy Parallelism as 5 or higher, this will allow multiple threads to execute concurrently.

  1. Optimize Sink Performance:
  • Set Block Size as 4 MB, this controls the size of each data block written to storage.

You can also check the attached Microsoft documentation for more information regarding performance optimization features:

https://learn.microsoft.com/en-us/azure/data-factory/copy-activity-performance-features

Sign up to request clarification or add additional context in comments.

Comments

Your Answer

By clicking “Post Your Answer”, you agree to our terms of service and acknowledge you have read our privacy policy.

Start asking to get answers

Find the answer to your question by asking.

Ask question

Explore related questions

See similar questions with these tags.