Default Compilation Specification for Spark Extensions
spark_default_compilation_spec
Description
This is the default compilation specification used for Spark extensions, when used with compile_package_jars.
Usage
spark_default_compilation_spec(
pkg = infer_active_package_name(),
locations = NULL
)Arguments
| Arguments | Description |
|---|---|
| pkg | The package containing Spark extensions to be compiled. |
| locations | Additional locations to scan. By default, the directories /opt/scala and /usr/local/scala will be scanned. |