Mihai Tache
Mihai Tache

Reputation: 171

how can you calculate the size of an apache spark data frame using pyspark?

Is there a way to calculate the size in bytes of an Apache spark Data Frame using pyspark?

Upvotes: 12

Views: 7318

Answers (1)

thePurplePython
thePurplePython

Reputation: 2767

why don't you just cache the df and then look in the spark UI under storage and convert the units to bytes

df.cache()

Upvotes: 3

Related Questions