pyspark.sql.SparkSession.addArtifact#
- SparkSession.addArtifact(*path, pyfile=False, archive=False, file=False)#
- Add artifact(s) to the client session. Currently only local files are supported. - New in version 3.5.0. - Parameters
- *pathtuple of str
- Artifact’s URIs to add. 
- pyfilebool
- Whether to add them as Python dependencies such as .py, .egg, .zip or .jar files. The pyfiles are directly inserted into the path when executing Python functions in executors. 
- archivebool
- Whether to add them as archives such as .zip, .jar, .tar.gz, .tgz, or .tar files. The archives are unpacked on the executor side automatically. 
- filebool
- Add a file to be downloaded with this Spark job on every node. The - pathpassed can only be a local file for now.
 
 - Notes - This is an API dedicated to Spark Connect client only. With regular Spark Session, it throws an exception.