Reputation: 8634
I'm trying to deploy a Spark standalone cluster using a custom built JVM. Running the Spark master and Spark worker processes is fine, but once the driver starts and sends work to the master, the worker crashes because it tries to fork an Executor by looking in JAVA_HOME/bin/java. Is there a way to customize how the executor gets started so that it uses a custom JVM?
Upvotes: 1
Views: 153
Reputation: 8634
I figured out the easiest way of doing this was to inherit org.apache.spark.deploy.worker.Worker. I did that in Java, then created a PartialFunction that handled only the LaunchExecutor message; when I received that message I created another custom instance of a class I inherited from ExecutorRunner.
Inside my CustomExecutorRunner I built up my own special Java command and executed that rather than letting ExecutorRunner do it.
It was quite messy and required a fair bit of reflection to access private fields that I needed to get everything running smoothly, but it worked.
Upvotes: 0