spark-submit官方文档 :http://spark.apache.org/docs/latest/submitting-applications.html

spark-properties官方文档:http://spark.apache.org/docs/latest/configuration.html

Launching Applications with spark-submit

./bin/spark-submit \--class <main-class> \--master <master-url> \--deploy-mode <deploy-mode> \--conf <key>=<value> \ ... # other options <application-jar> \ [application-arguments]
Spark shell和spark-submit工具支持两种动态加载配置的方法。 第一个是命令行选项,例如--master,如上所示。 spark-submit可以使用--conf标志接受任何Spark属性,但是对于在启动Spark应用程序中起作用的属性使用特殊标志。 运行./bin/spark-submit --help将显示这些选项的完整列表
Usage: spark-submit [options] <app jar | python file | R file> [app arguments]Usage: spark-submit --kill [submission ID] --master [spark://...]Usage: spark-submit --status [submission ID] --master [spark://...]Usage: spark-submit run-example [options] example-class [example args]
Some of the commonly used options are:
|Options:|  --master MASTER_URL         spark://host:port, mesos://host:port, yarn,|                              k8s://https://host:port, or local (Default: local[*]).|  --deploy-mode DEPLOY_MODE   Whether to launch the driver program locally ("client") or|                              on one of the worker machines inside the cluster ("cluster")|                              (Default: client).|  --class CLASS_NAME          Your application's main class (for Java / Scala apps).|  --name NAME                 A name of your application.|  --jars JARS                 Comma-separated list of jars to include on the driver|                              and executor classpaths.|  --packages                  Comma-separated list of maven coordinates of jars to include|                              on the driver and executor classpaths. Will search the local|                              maven repo, then maven central and any additional remote|                              repositories given by --repositories. The format for the|                              coordinates should be groupId:artifactId:version.|  --exclude-packages          Comma-separated list of groupId:artifactId, to exclude while|                              resolving the dependencies provided in --packages to avoid|                              dependency conflicts.|  --repositories              Comma-separated list of additional remote repositories to|                              search for the maven coordinates given with --packages.|  --py-files PY_FILES         Comma-separated list of .zip, .egg, or .py files to place|                              on the PYTHONPATH for Python apps.|  --files FILES               Comma-separated list of files to be placed in the working|                              directory of each executor. File paths of these files|                              in executors can be accessed via SparkFiles.get(fileName).||  --conf PROP=VALUE           Arbitrary Spark configuration property.|  --properties-file FILE      Path to a file from which to load extra properties. If not|                              specified, this will look for conf/spark-defaults.conf.||  --driver-memory MEM         Memory for driver (e.g. 1000M, 2G) (Default: ${mem_mb}M).|  --driver-java-options       Extra Java options to pass to the driver.|  --driver-library-path       Extra library path entries to pass to the driver.|  --driver-class-path         Extra class path entries to pass to the driver. Note that|                              jars added with --jars are automatically included in the|                              classpath.||  --executor-memory MEM       Memory per executor (e.g. 1000M, 2G) (Default: 1G).||  --proxy-user NAME           User to impersonate when submitting the application.|                              This argument does not work with --principal / --keytab.||  --help, -h                  Show this help message and exit.|  --verbose, -v               Print additional debug output.|  --version,                  Print the version of current Spark.|| Cluster deploy mode only:|  --driver-cores NUM          Number of cores used by the driver, only in cluster mode|                              (Default: 1).|| Spark standalone or Mesos with cluster deploy mode only:|  --supervise                 If given, restarts the driver on failure.|  --kill SUBMISSION_ID        If given, kills the driver specified.|  --status SUBMISSION_ID      If given, requests the status of the driver specified.|| Spark standalone and Mesos only:|  --total-executor-cores NUM  Total cores for all executors.|| Spark standalone and YARN only:|  --executor-cores NUM        Number of cores per executor. (Default: 1 in YARN mode,|                              or all available cores on the worker in standalone mode)|| YARN-only:|  --queue QUEUE_NAME          The YARN queue to submit to (Default: "default").|  --num-executors NUM         Number of executors to launch (Default: 2).|                              If dynamic allocation is enabled, the initial number of|                              executors will be at least NUM.|  --archives ARCHIVES         Comma separated list of archives to be extracted into the|                              working directory of each executor.|  --principal PRINCIPAL       Principal to be used to login to KDC, while running on|                              secure HDFS.|  --keytab KEYTAB             The full path to the file that contains the keytab for the|                              principal specified above. This keytab will be copied to|                              the node running the Application Master via the Secure|                              Distributed Cache, for renewing the login tickets and the|                              delegation tokens periodically.

转载于:https://www.cnblogs.com/jqbai/p/10783805.html

spark-submit(spark版本2.3.2)相关推荐

  1. 如何修改 Linkis 依赖的 Spark、Hive 版本?

    如何修改Linkis的依赖的Hadoop.Hive.Spark版本 请注意:Hadoop 作为大数据基础服务,Linkis 必须依赖 Hadoop 进行编译: 而 Spark.Hive等计算存储引擎则 ...

  2. zeppelin报错master is set as yarn, but spark.submit.deployMode is not specified

    具体报错如下: java.lang.RuntimeException: master is set as yarn, but spark.submit.deployMode is not specif ...

  3. spark submit参数及调优

    spark submit参数及调优 原文地址 spark submit参数介绍 你可以通过spark-submit --help或者spark-shell --help来查看这些参数. 使用格式:  ...

  4. 【原创】大数据基础之Spark(1)Spark Submit即Spark任务提交过程

    Spark2.1.1 一 Spark Submit本地解析 1.1 现象 提交命令: spark-submit --master local[10] --driver-memory 30g --cla ...

  5. Spark On Yarn --jars/spark.yarn.jars 踩坑 与 提升spark submit速度

    一.引言 Spark submit 可以选择 --jars 传入本地的jar也可以 --jars 传入HDFS的jar包,经过半下午的实验,终于搞清了两者的关系以及 spark.yarn.jars 和 ...

  6. Spark Submit任务提交流程

    1,简介 在上一篇博客中,我们详细介绍了Spark Standalone模式下集群的启动流程.在Spark 集群启动后,我们要想在集群上运行我们自己编写的程序,该如何做呢?本篇博客就主要介绍Spark ...

  7. spark submit参数及调优(转载)

    spark submit参数介绍 你可以通过spark-submit --help或者spark-shell --help来查看这些参数. 使用格式:  ./bin/spark-submit \--c ...

  8. spark submit参数及调试

    原文:http://www.cnblogs.com/haoyy/p/6893943.html spark submit参数介绍 你可以通过spark-submit --help或者spark-shel ...

  9. Spark Submit提交及参数调优

    https://www.cnblogs.com/LHWorldBlog/p/8414342.html https://www.cnblogs.com/haoyy/p/6893943.html spar ...

  10. Spark:spark submit参数及调优 perfect

    先看下spark运行原理流程: 我们使用spark-submit提交一个Spark作业之后,这个作业就会启动一个对应的Driver进程. 根据你使用的部署模式(deploy-mode)不同,Drive ...

最新文章

  1. 将课程中的所有动手动脑的问题以及课后实验性的问题,整理成一篇文档,以博客形式发表在博客园...
  2. Java VisualVM插件地址
  3. ​ICML 2021 Long Oral | 顺序不可知的交叉熵函数
  4. Membership学习记录
  5. 目标检测矩形框与polygon数据增加--裁剪,拓展,旋转
  6. RetinaNet和Focal Loss论文笔记
  7. Python案例:获取全国城市列表(区号与名称)
  8. tar.xz文件怎样解压
  9. 计算机培训实践反思模板,基于问题解决式课堂教学模式的反思
  10. BMS锂电池管理系统如何增加蓝牙模块芯片
  11. 为什么计算机三分技术七分管理,如何理解“七分管理,三分技术,运作贯穿始终”?...
  12. Simulated Binary Crossover(SBX)的学习
  13. Apache APISIX 社区成员助力 openEuler 发布第一个社区创新版
  14. vue3基础 —— 子传父
  15. Laravel OAuth2 (一) ---简单获取用户信息
  16. 重磅!道翰天琼破解认知智能核心秘密三大核心技术,机器人大脑API平台。
  17. Golang 判断IPv4和IPv6是否合法
  18. 超150亿元市场启动!智能汽车中间件出圈背后仍存挑战
  19. 嵌入式中绕不开的CAN总线
  20. 空间域名多少钱?选择域名空间要注意什么?

热门文章

  1. session放入缓存(redis)、DB
  2. 如何在Spring MVC工程中进行单元测试
  3. 如何正确复制CSDN文章到自己的博客
  4. .NET Framework 4.5 ZipArchive类压缩解压
  5. MySQL 大数据量分页性能优化
  6. 排序算法性能和使用场景总结
  7. php能从事什么岗位,ps能从事什么工作岗位
  8. pandas提取某两列的值_Pandas进阶修炼120题第五期
  9. EsayExcel简单的读和写
  10. oracle esb 灾备,两地三中心双活系统灾备切换场景和数据补录问题?