naceur ben said
naceur ben said

Reputation: 11

How to write partitioned parquet files to blob storage

I want to load data from On Premise SQL SERVER to blob storage with copy activity in ADF, the target file is parquet, the size of this one is 5 Gb.

The pipeline work well and he wrote one parquet file, now i need to split this file in multiple parquet file to optimise loading data with Poly base and for another uses.

With Spark we can partition file in multiple file by this syntaxe :

df.repartition(5).write.parquet("path")

Upvotes: 0

Views: 7175

Answers (1)

Related Questions