You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
|`type`| string | Type of the job configuration |`"pythonJobConfiguration"`|
181
-
|`appPath`| string | Project path to notebook (e.g `Resources/foo.ipynb`) |`null`|
182
-
|`environmentName`| string | Name of the python environment |`"pandas-training-pipeline"`|
183
-
|`resourceConfig.cores`| number (float) | Number of CPU cores to be allocated |`1.0`|
184
-
|`resourceConfig.memory`| number (int) | Number of MBs to be allocated |`2048`|
185
-
|`resourceConfig.gpus`| number (int) | Number of GPUs to be allocated |`0`|
186
-
|`logRedirection`| boolean | Whether logs are redirected |`true`|
187
-
|`jobType`| string | Type of job |`"PYTHON"`|
188
-
|`files`| string | HDFS path(s) to files to be provided to the Notebook Job. Multiple files can be included in a single string, separated by commas. <br>Example: `"hdfs:///Project/<project_name>/Resources/file1.py,hdfs:///Project/<project_name>/Resources/file2.txt"`|`null`|
181
+
The following table describes the job configuration parameters for a PYTHON job.
| <nobr>`conf['defaultArgs']`</nobr> | string | Arguments to pass to the notebook.<br>Will be overridden if arguments are passed explicitly via `Job.run(args="...")`.<br>Must conform to Papermill format `-p arg1 val1`|`null`|
190
+
| <nobr>`conf['environmentName']`</nobr> | string | Name of the project Python environment to use |`"pandas-training-pipeline"`|
191
+
| <nobr>`conf['resourceConfig']['cores']`</nobr> | float | Number of CPU cores to be allocated |`1.0`|
192
+
| <nobr>`conf['resourceConfig']['memory']`</nobr> | int | Number of MBs to be allocated |`2048`|
193
+
| <nobr>`conf['resourceConfig']['gpus']`</nobr> | int | Number of GPUs to be allocated |`0`|
194
+
| <nobr>`conf['logRedirection']`</nobr> | boolean | Whether logs are redirected |`true`|
195
+
| <nobr>`conf['jobType']`</nobr> | string | Type of job |`"PYTHON"`|
196
+
| <nobr>`conf['files']`</nobr> | string | Comma-separated string of HDFS path(s) to files to be made available to the application. Example: `hdfs:///Project/<project>/Resources/file1.py,...`|`null`|
|`type`| string | Type of the job configuration |`"sparkJobConfiguration"`|
224
-
|`appPath`| string | Project path to script (e.g `Resources/foo.py`) |`null`|
225
-
|`environmentName`| string | Name of the project spark environment |`"spark-feature-pipeline"`|
226
-
|`spark.driver.cores`| number (float) | Number of CPU cores allocated for the driver |`1.0`|
227
-
|`spark.driver.memory`| number (int) | Memory allocated for the driver (in MB) |`2048`|
228
-
|`spark.executor.instances`| number (int) | Number of executor instances |`1`|
229
-
|`spark.executor.cores`| number (float) | Number of CPU cores per executor |`1.0`|
230
-
|`spark.executor.memory`| number (int) | Memory allocated per executor (in MB) |`4096`|
231
-
|`spark.dynamicAllocation.enabled`| boolean | Enable dynamic allocation of executors |`true`|
232
-
|`spark.dynamicAllocation.minExecutors`| number (int) | Minimum number of executors with dynamic allocation |`1`|
233
-
|`spark.dynamicAllocation.maxExecutors`| number (int) | Maximum number of executors with dynamic allocation |`2`|
234
-
|`spark.dynamicAllocation.initialExecutors`| number (int) | Initial number of executors with dynamic allocation |`1`|
235
-
|`spark.blacklist.enabled`| boolean | Whether executor/node blacklisting is enabled |`false`|
236
-
|`files`| string | HDFS path(s) to files to be provided to the Spark application. Multiple files can be included in a single string, separated by commas. <br>Example: `"hdfs:///Project/<project_name>/Resources/file1.py,hdfs:///Project/<project_name>/Resources/file2.txt"`|`null`|
237
-
|`pyFiles`| string | HDFS path(s) to Python files to be provided to the Spark application. These will be added to the `PYTHONPATH` so they can be imported as modules. Multiple files can be included in a single string, separated by commas. <br>Example: `"hdfs:///Project/<project_name>/Resources/module1.py,hdfs:///Project/<project_name>/Resources/module2.py"`|`null`|
238
-
|`jars`| string | HDFS path(s) to JAR files to be provided to the Spark application. These will be added to the classpath. Multiple files can be included in a single string, separated by commas. <br>Example: `"hdfs:///Project/<project_name>/Resources/lib1.jar,hdfs:///Project/<project_name>/Resources/lib2.jar"`|`null`|
239
-
|`archives`| string | HDFS path(s) to archive files to be provided to the Spark application. Multiple files can be included in a single string, separated by commas. <br>Example: `"hdfs:///Project/<project_name>/Resources/archive1.zip,hdfs:///Project/<project_name>/Resources/archive2.tar.gz"`|`null`|
222
+
The following table describes the job configuration parameters for a PYSPARK job.
| <nobr>`conf['type']`</nobr> | string | Type of the job configuration |`"sparkJobConfiguration"`|
229
+
| <nobr>`conf['appPath']`</nobr> | string | Project path to spark program (e.g `Resources/foo.py`) |`null`|
230
+
| <nobr>`conf['defaultArgs']`</nobr> | string | Arguments to pass to the program. Will be overridden if arguments are passed explicitly via `Job.run(args="...")`|`null`|
231
+
| <nobr>`conf['environmentName']`</nobr> | string | Name of the project spark environment to use |`"spark-feature-pipeline"`|
232
+
| <nobr>`conf['spark.driver.cores']`</nobr> | float | Number of CPU cores allocated for the driver |`1.0`|
233
+
| <nobr>`conf['spark.driver.memory']`</nobr> | int | Memory allocated for the driver (in MB) |`2048`|
234
+
| <nobr>`conf['spark.executor.instances']`</nobr> | int | Number of executor instances |`1`|
235
+
| <nobr>`conf['spark.executor.cores']`</nobr> | float | Number of CPU cores per executor |`1.0`|
236
+
| <nobr>`conf['spark.executor.memory']`</nobr> | int | Memory allocated per executor (in MB) |`4096`|
| <nobr>`conf['files']`</nobr> | string | Comma-separated string of HDFS path(s) to files to be made available to the application. Example: `hdfs:///Project/<project_name>/Resources/file1.py,...`|`null`|
243
+
| <nobr>`conf['pyFiles']`</nobr> | string | Comma-separated string of HDFS path(s) to python modules to be made available to the application. Example: `hdfs:///Project/<project_name>/Resources/file1.py,...`|`null`|
244
+
| <nobr>`conf['jars']`</nobr> | string | Comma-separated string of HDFS path(s) to jars to be included in CLASSPATH. Example: `hdfs:///Project/<project_name>/Resources/app.jar,...`|`null`|
245
+
| <nobr>`conf['archives']`</nobr> | string | Comma-separated string of HDFS path(s) to archives to be made available to the application. Example: `hdfs:///Project/<project_name>/Resources/archive.zip,...`|`null`|
0 commit comments