You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
[HWORKS-2190][APPEND] Updating job configuration to include file, pyfiles, archives and jars (#478)
* updating docs for jobs configs to include files, pyFiles, jars and archives
* updating based on review comments
* updating documentation for notebooks and python Jobs
Copy file name to clipboardExpand all lines: docs/user_guides/projects/jobs/notebook_job.md
+1Lines changed: 1 addition & 0 deletions
Original file line number
Diff line number
Diff line change
@@ -179,6 +179,7 @@ The following table describes the JSON payload returned by `jobs_api.get_configu
179
179
|`resourceConfig.gpus`| number (int) | Number of GPUs to be allocated |`0`|
180
180
|`logRedirection`| boolean | Whether logs are redirected |`true`|
181
181
|`jobType`| string | Type of job |`"PYTHON"`|
182
+
|`files`| string | HDFS path(s) to files to be provided to the Notebook Job. Multiple files can be included in a single string, separated by commas. <br>Example: `"hdfs:///Project/<project_name>/Resources/file1.py,hdfs:///Project/<project_name>/Resources/file2.txt"`|`null`|
|`type`| string | Type of the job configuration |`"sparkJobConfiguration"`|
220
-
|`appPath`| string | Project path to script (e.g `Resources/foo.py`) |`null`|
220
+
|`appPath`| string | Project path to script (e.g `Resources/foo.py`) |`null`|
221
221
|`environmentName`| string | Name of the project spark environment |`"spark-feature-pipeline"`|
222
222
|`spark.driver.cores`| number (float) | Number of CPU cores allocated for the driver |`1.0`|
223
223
|`spark.driver.memory`| number (int) | Memory allocated for the driver (in MB) |`2048`|
@@ -229,6 +229,10 @@ The following table describes the JSON payload returned by `jobs_api.get_configu
229
229
|`spark.dynamicAllocation.maxExecutors`| number (int) | Maximum number of executors with dynamic allocation |`2`|
230
230
|`spark.dynamicAllocation.initialExecutors`| number (int) | Initial number of executors with dynamic allocation |`1`|
231
231
|`spark.blacklist.enabled`| boolean | Whether executor/node blacklisting is enabled |`false`|
232
+
|`files`| string | HDFS path(s) to files to be provided to the Spark application. Multiple files can be included in a single string, separated by commas. <br>Example: `"hdfs:///Project/<project_name>/Resources/file1.py,hdfs:///Project/<project_name>/Resources/file2.txt"`|`null`|
233
+
|`pyFiles`| string | HDFS path(s) to Python files to be provided to the Spark application. These will be added to the `PYTHONPATH` so they can be imported as modules. Multiple files can be included in a single string, separated by commas. <br>Example: `"hdfs:///Project/<project_name>/Resources/module1.py,hdfs:///Project/<project_name>/Resources/module2.py"`|`null`|
234
+
|`jars`| string | HDFS path(s) to JAR files to be provided to the Spark application. These will be added to the classpath. Multiple files can be included in a single string, separated by commas. <br>Example: `"hdfs:///Project/<project_name>/Resources/lib1.jar,hdfs:///Project/<project_name>/Resources/lib2.jar"`|`null`|
235
+
|`archives`| string | HDFS path(s) to archive files to be provided to the Spark application. Multiple files can be included in a single string, separated by commas. <br>Example: `"hdfs:///Project/<project_name>/Resources/archive1.zip,hdfs:///Project/<project_name>/Resources/archive2.tar.gz"`|`null`|
Copy file name to clipboardExpand all lines: docs/user_guides/projects/jobs/python_job.md
+1Lines changed: 1 addition & 0 deletions
Original file line number
Diff line number
Diff line change
@@ -177,6 +177,7 @@ The following table describes the JSON payload returned by `jobs_api.get_configu
177
177
|`resourceConfig.gpus`| number (int) | Number of GPUs to be allocated |`0`|
178
178
|`logRedirection`| boolean | Whether logs are redirected |`true`|
179
179
|`jobType`| string | Type of job |`"PYTHON"`|
180
+
|`files`| string | HDFS path(s) to files to be provided to the Python Job. Multiple files can be included in a single string, separated by commas. <br>Example: `"hdfs:///Project/<project_name>/Resources/file1.py,hdfs:///Project/<project_name>/Resources/file2.txt"`|`null`|
|`files`| string | HDFS path(s) to files to be provided to the Spark application. Multiple files can be included in a single string, separated by commas. <br>Example: `"hdfs:///Project/<project_name>/Resources/file1.py,hdfs:///Project/<project_name>/Resources/file2.txt"`|`null`|
235
+
|`pyFiles`| string | HDFS path(s) to Python files to be provided to the Spark application. These will be added to the `PYTHONPATH` so they can be imported as modules. Multiple files can be included in a single string, separated by commas. <br>Example: `"hdfs:///Project/<project_name>/Resources/module1.py,hdfs:///Project/<project_name>/Resources/module2.py"`|`null`|
236
+
|`jars`| string | HDFS path(s) to JAR files to be provided to the Spark application. These will be added to the classpath. Multiple files can be included in a single string, separated by commas. <br>Example: `"hdfs:///Project/<project_name>/Resources/lib1.jar,hdfs:///Project/<project_name>/Resources/lib2.jar"`|`null`|
237
+
|`archives`| string | HDFS path(s) to archive files to be provided to the Spark application. Multiple files can be included in a single string, separated by commas. <br>Example: `"hdfs:///Project/<project_name>/Resources/archive1.zip,hdfs:///Project/<project_name>/Resources/archive2.tar.gz"`|`null`|
0 commit comments