How to manually deploy 3rd party utility jar for Apache Spark cluster? -


i have apache spark cluster (multi-nodes) , manually deploy utility jars each spark node. should put these jars to? example: spark-streaming-twitter_2.10-1.6.0.jar

i know can use maven build fat jar including these jars, deploy these utilities manually. in way, programmers not have deploy these utilities jars.

any suggestion?

1, copy 3rd party jars reserved hdfs directory;
example hdfs://xxx-ns/user/xxx/3rd-jars/

2, in spark-submit, specify these jars using hdfs path;
hdfs: - executors pull down files , jars hdfs directory

--jars hdfs://xxx-ns/user/xxx/3rd-jars/xxx.jar   

3, spark-submit not repleatly upload these jars

client: source , destination file systems same. not copying hdfs://xxx-ns/user/xxx/3rd-jars/xxx.jar 

Comments

Popular posts from this blog

authentication - Mongodb revoke acccess to connect test database -

r - Update two sets of radiobuttons reactively - shiny -

ios - Realm over CoreData should I use NSFetchedResultController or a Dictionary? -