spark-submit(spark版本2.3.2)
spark-submit官方文档 :http://spark.apache.org/docs/latest/submitting-applications.html
spark-properties官方文档:http://spark.apache.org/docs/latest/configuration.html
Launching Applications with spark-submit
./bin/spark-submit \--class <main-class> \--master <master-url> \--deploy-mode <deploy-mode> \--conf <key>=<value> \ ... # other options <application-jar> \ [application-arguments]
Spark shell和spark-submit工具支持两种动态加载配置的方法。 第一个是命令行选项,例如--master,如上所示。 spark-submit可以使用--conf标志接受任何Spark属性,但是对于在启动Spark应用程序中起作用的属性使用特殊标志。 运行./bin/spark-submit --help将显示这些选项的完整列表
Usage: spark-submit [options] <app jar | python file | R file> [app arguments]Usage: spark-submit --kill [submission ID] --master [spark://...]Usage: spark-submit --status [submission ID] --master [spark://...]Usage: spark-submit run-example [options] example-class [example args]
Some of the commonly used options are:
|Options:| --master MASTER_URL spark://host:port, mesos://host:port, yarn,| k8s://https://host:port, or local (Default: local[*]).| --deploy-mode DEPLOY_MODE Whether to launch the driver program locally ("client") or| on one of the worker machines inside the cluster ("cluster")| (Default: client).| --class CLASS_NAME Your application's main class (for Java / Scala apps).| --name NAME A name of your application.| --jars JARS Comma-separated list of jars to include on the driver| and executor classpaths.| --packages Comma-separated list of maven coordinates of jars to include| on the driver and executor classpaths. Will search the local| maven repo, then maven central and any additional remote| repositories given by --repositories. The format for the| coordinates should be groupId:artifactId:version.| --exclude-packages Comma-separated list of groupId:artifactId, to exclude while| resolving the dependencies provided in --packages to avoid| dependency conflicts.| --repositories Comma-separated list of additional remote repositories to| search for the maven coordinates given with --packages.| --py-files PY_FILES Comma-separated list of .zip, .egg, or .py files to place| on the PYTHONPATH for Python apps.| --files FILES Comma-separated list of files to be placed in the working| directory of each executor. File paths of these files| in executors can be accessed via SparkFiles.get(fileName).|| --conf PROP=VALUE Arbitrary Spark configuration property.| --properties-file FILE Path to a file from which to load extra properties. If not| specified, this will look for conf/spark-defaults.conf.|| --driver-memory MEM Memory for driver (e.g. 1000M, 2G) (Default: ${mem_mb}M).| --driver-java-options Extra Java options to pass to the driver.| --driver-library-path Extra library path entries to pass to the driver.| --driver-class-path Extra class path entries to pass to the driver. Note that| jars added with --jars are automatically included in the| classpath.|| --executor-memory MEM Memory per executor (e.g. 1000M, 2G) (Default: 1G).|| --proxy-user NAME User to impersonate when submitting the application.| This argument does not work with --principal / --keytab.|| --help, -h Show this help message and exit.| --verbose, -v Print additional debug output.| --version, Print the version of current Spark.|| Cluster deploy mode only:| --driver-cores NUM Number of cores used by the driver, only in cluster mode| (Default: 1).|| Spark standalone or Mesos with cluster deploy mode only:| --supervise If given, restarts the driver on failure.| --kill SUBMISSION_ID If given, kills the driver specified.| --status SUBMISSION_ID If given, requests the status of the driver specified.|| Spark standalone and Mesos only:| --total-executor-cores NUM Total cores for all executors.|| Spark standalone and YARN only:| --executor-cores NUM Number of cores per executor. (Default: 1 in YARN mode,| or all available cores on the worker in standalone mode)|| YARN-only:| --queue QUEUE_NAME The YARN queue to submit to (Default: "default").| --num-executors NUM Number of executors to launch (Default: 2).| If dynamic allocation is enabled, the initial number of| executors will be at least NUM.| --archives ARCHIVES Comma separated list of archives to be extracted into the| working directory of each executor.| --principal PRINCIPAL Principal to be used to login to KDC, while running on| secure HDFS.| --keytab KEYTAB The full path to the file that contains the keytab for the| principal specified above. This keytab will be copied to| the node running the Application Master via the Secure| Distributed Cache, for renewing the login tickets and the| delegation tokens periodically.
转载于:https://www.cnblogs.com/jqbai/p/10783805.html
spark-submit(spark版本2.3.2)相关推荐
- 如何修改 Linkis 依赖的 Spark、Hive 版本?
如何修改Linkis的依赖的Hadoop.Hive.Spark版本 请注意:Hadoop 作为大数据基础服务,Linkis 必须依赖 Hadoop 进行编译: 而 Spark.Hive等计算存储引擎则 ...
- zeppelin报错master is set as yarn, but spark.submit.deployMode is not specified
具体报错如下: java.lang.RuntimeException: master is set as yarn, but spark.submit.deployMode is not specif ...
- spark submit参数及调优
spark submit参数及调优 原文地址 spark submit参数介绍 你可以通过spark-submit --help或者spark-shell --help来查看这些参数. 使用格式: ...
- 【原创】大数据基础之Spark(1)Spark Submit即Spark任务提交过程
Spark2.1.1 一 Spark Submit本地解析 1.1 现象 提交命令: spark-submit --master local[10] --driver-memory 30g --cla ...
- Spark On Yarn --jars/spark.yarn.jars 踩坑 与 提升spark submit速度
一.引言 Spark submit 可以选择 --jars 传入本地的jar也可以 --jars 传入HDFS的jar包,经过半下午的实验,终于搞清了两者的关系以及 spark.yarn.jars 和 ...
- Spark Submit任务提交流程
1,简介 在上一篇博客中,我们详细介绍了Spark Standalone模式下集群的启动流程.在Spark 集群启动后,我们要想在集群上运行我们自己编写的程序,该如何做呢?本篇博客就主要介绍Spark ...
- spark submit参数及调优(转载)
spark submit参数介绍 你可以通过spark-submit --help或者spark-shell --help来查看这些参数. 使用格式: ./bin/spark-submit \--c ...
- spark submit参数及调试
原文:http://www.cnblogs.com/haoyy/p/6893943.html spark submit参数介绍 你可以通过spark-submit --help或者spark-shel ...
- Spark Submit提交及参数调优
https://www.cnblogs.com/LHWorldBlog/p/8414342.html https://www.cnblogs.com/haoyy/p/6893943.html spar ...
- Spark:spark submit参数及调优 perfect
先看下spark运行原理流程: 我们使用spark-submit提交一个Spark作业之后,这个作业就会启动一个对应的Driver进程. 根据你使用的部署模式(deploy-mode)不同,Drive ...
最新文章
- 将课程中的所有动手动脑的问题以及课后实验性的问题,整理成一篇文档,以博客形式发表在博客园...
- Java VisualVM插件地址
- ​ICML 2021 Long Oral | 顺序不可知的交叉熵函数
- Membership学习记录
- 目标检测矩形框与polygon数据增加--裁剪,拓展,旋转
- RetinaNet和Focal Loss论文笔记
- Python案例:获取全国城市列表(区号与名称)
- tar.xz文件怎样解压
- 计算机培训实践反思模板,基于问题解决式课堂教学模式的反思
- BMS锂电池管理系统如何增加蓝牙模块芯片
- 为什么计算机三分技术七分管理,如何理解“七分管理,三分技术,运作贯穿始终”?...
- Simulated Binary Crossover(SBX)的学习
- Apache APISIX 社区成员助力 openEuler 发布第一个社区创新版
- vue3基础 —— 子传父
- Laravel OAuth2 (一) ---简单获取用户信息
- 重磅!道翰天琼破解认知智能核心秘密三大核心技术,机器人大脑API平台。
- Golang 判断IPv4和IPv6是否合法
- 超150亿元市场启动!智能汽车中间件出圈背后仍存挑战
- 嵌入式中绕不开的CAN总线
- 空间域名多少钱?选择域名空间要注意什么?