Reputation: 1
I’m facing an issue when trying to submit a PySpark job to Google Cloud Dataproc. The goal is to run a script on the Dataproc cluster that uses the minio module. However, I keep encountering the following error:
This is my code that i submit in dataproc:
My Dataproc cluster consists of 1 master node and 2 worker nodes.
How can I correctly install and use the minio module in a PySpark job on Google Cloud Dataproc?
Upvotes: 0
Views: 113
Reputation: 56
Please share more information about your cluster and your submit command (is it serverless or more standard cluster ext.?). The potential reasons that come to my mind from what you've shared so far are:
I hope your problem will be solved!
Upvotes: 0