先将spark根目录下的jars文件夹下的所有的jar包上传到hdfs上
hdfs dfs -mkdir -p /spark-yarn/jars/ hdfs dfs -put /server/spark/jars/* /spark-yarn/jars/
修改spark-defaults.conf (注意:要修改所有节点的),增加此配置:spark.yarn.jars hdfs://node22:9000/spark-yarn/jars/*.jar
2.配置spark.yarn.archive将spark根目录下jars里的所有jar包上传到HDFS
打包要注意所有的jar都在zip包的根目录中(如果提示没有zip命令,则需要先安装:yum install zip)
zip -q -r spark_jars.zip * hdfs dfs -mkdir /spark-yarn/zip hdfs dfs -put spark_jars.zip /spark-yarn/zip/
修改spark-defaults.conf (注意:要修改所有节点的),增加此配置:spark.yarn.archive hdfs://node22:9000/spark-yarn/zip/spark_jars.zip
然后在提交任务时,执行如下命令:
spark-submit --class access.Data2Hbase --master yarn --deploy-mode cluster --name xxx hdfs://node22:9000/intelligover/intelligover-1.0-SNAPSHOT-jar-with-dependencies.jar xxxx
(为了修改方便,不建议将自己的程序与spark自身的jar包放置在一起,避免在更新的时候误删)



