mirror of https://github.com/databricks/cli.git
33c446dadd
## Changes The approach to do this was: 1. Iterate over all libraries in all job tasks 2. Find references to local libraries 3. Store pointer to `compute.Library` in the matching artifact file to signal it should be uploaded This breaks down when introducing #1098 because we can no longer track unexported state across mutators. The approach in this PR performs the path matching twice; once in the matching mutator where we check if each referenced file has an artifacts section, and once during artifact upload to rewrite the library path from a local file reference to an absolute Databricks path. ## Tests Integration tests pass. |
||
---|---|---|
.. | ||
bundles | ||
artifacts_test.go | ||
deploy_then_remove_resources_test.go | ||
empty_bundle_test.go | ||
generate_job_test.go | ||
generate_pipeline_test.go | ||
helpers.go | ||
job_metadata_test.go | ||
local_state_staleness_test.go | ||
python_wheel_test.go |