Reputation: 11
I got an error when trying to write data to Redshift using PySpark on an EMR cluster.
df.write.format("jdbc") \
.option("url", "jdbc:redshift://clustername.yyyyy.us-east-1.redshift.amazonaws.com:5439/db") \
.option("driver", "com.amazon.redshift.jdbc42.Driver") \
.option("dbtable", "public.table") \
.option("user", user_redshift) \
.option("password", password_redshift) \
.mode("overwrite") \
.save()
The error I have got is:
py4j.protocol.Py4JJavaError: An error occurred while calling o143.save.
: org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 0.0 failed 4 times, most recent failure: Lost task 0.3 in stage 0.0 (TID 6, , executor 1):
java.sql.SQLException: [Amazon](500310) Invalid operation: The session is read-only;
at com.amazon.redshift.client.messages.inbound.ErrorResponse.toErrorException(Unknown Source)
at com.amazon.redshift.client.PGMessagingContext.handleErrorResponse(Unknown Source)
at com.amazon.redshift.client.PGMessagingContext.handleMessage(Unknown Source)
at com.amazon.jdbc.communications.InboundMessagesPipeline.getNextMessageOfClass(Unknown Source)
at com.amazon.redshift.client.PGMessagingContext.doMoveToNextClass(Unknown Source)
at com.amazon.redshift.client.PGMessagingContext.getParameterDescription(Unknown Source)
at com.amazon.redshift.client.PGClient.prepareStatement(Unknown Source)
at com.amazon.redshift.dataengine.PGQueryExecutor.<init>(Unknown Source)
at com.amazon.redshift.dataengine.PGDataEngine.prepare(Unknown Source)
at com.amazon.jdbc.common.SPreparedStatement.<init>(Unknown Source)
...
I appreciate any help. Thanks!
Upvotes: 1
Views: 949
Reputation: 41
We also faced the same issue for our EMR pySpark cluster. EMR with "ReleaseLabel": "emr-5.33.0" and Spark version 2.4.7
We resolved it with the following changes
jdbc:redshift://clustername.yyyyy.us-east-1.redshift.amazonaws.com:5439/db?user=username&password=password;ReadOnly=false
You can then try to run your spark-submit with the following:
spark-submit --jars s3://jars/redshift-jdbc42-2.0.0.7.jar s3://scripts/scriptname.py
where scriptname.py has
df.write\
.format('jdbc')\
.option("driver", "com.amazon.redshift.jdbc42.Driver")\
.option("url", jdbcUrl)\
.option("dbtable", "schema.table")\
.option("aws_iam_role", "XXXX") \
.option("tempdir", f"s3://XXXXXX") \
.mode('append')\
.save()
Upvotes: 0