Reputation: 171
Is there a way to calculate the size in bytes of an Apache spark Data Frame using pyspark?
Upvotes: 12
Views: 7318
Reputation: 2767
why don't you just cache the df and then look in the spark UI under storage and convert the units to bytes
df.cache()
Upvotes: 3