pyspark.sql.SparkSession.addArtifact#

SparkSession.addArtifact(*path, pyfile=False, archive=False, file=False)#

Add artifact(s) to the client session. Currently only local files are supported.

New in version 3.5.0.

Changed in version 4.0.0: Supports Spark Classic.

Parameters
*pathtuple of str

Artifact’s URIs to add.

pyfilebool

Whether to add them as Python dependencies such as .py, .egg, .zip or .jar files. The pyfiles are directly inserted into the path when executing Python functions in executors.

archivebool

Whether to add them as archives such as .zip, .jar, .tar.gz, .tgz, or .tar files. The archives are unpacked on the executor side automatically.

filebool

Add a file to be downloaded with this Spark job on every node. The path passed can only be a local file for now.