Move the location of the datalake.

Parameters:
See dedicated page for more information.
moveDataLake is a scripted action. Embedded code is accessible and customizable through this tab.
See dedicated page for more information.
Purpose: moveDataLake bulk-updates file paths used by pipelines when a data lake (or any shared storage) is reorganized. It scans selected pipelines, finds path strings in the configured scope (inputs/outputs), and rewrites those paths by replacing a “before” substring with an “after” substring. It can run in a dry-run mode to preview changes and can backup files before writing.
Bulk path refactor. Rewrites many pipeline file references in one shot (great for environment moves like /data/raw/... → /mnt/datalake/raw/..., or switching a mount prefix).
Targeted scope. You choose whether to update absolute, relative, or all paths and whether to touch inputs, outputs, or both.
Safe by design.
Actionable report. The Data tab returns a compact table per pipeline with a status of what happened.
Use moveDataLake when you need to:
/bronze/, rename /records/ → /after_Records/).You provide the list of pipelines to process.
The incoming table must include a column named path that points to pipeline files (or folders, if supported by your setup).
You tell the action what to change.
Provide the substring to find (“before”) and the substring to write (“after”). Replacement is a straightforward string substitution (no regex). Scope it to absolute / relative / all paths and to inputs / outputs / both.
It analyzes & (optionally) writes.
You review the result.
The Data tab summarizes outcomes per pipeline and the Log explains what happened internally.
path column. A simple InlineTable works well./records/ over rec).path — the pipeline file path to scan/update. (If your environment supports folders, a folder path may cause recursive processing of contained pipelines.)Tip: Feed one row per pipeline you want to update.
A compact audit table with at least:
path — the processed pipeline file.
_updateStatus — outcome for that path. Typical values include:
Updated — at least one reference changed and saved.NoChange — nothing matched your “before” substring under the chosen scope.Missing — the file could not be found at the provided path.DryRun — changes were detected but not written (because Dry run was on).BackedUp (or included in the log) — file was copied before changes.Error — the action failed to read/write the file; check the Log for details.Exact status names can vary by version; always cross-check the Log for full context.
/mnt/..., s3://..., gs://...).records/out.csv).Prepare your pipeline list.
Create (or load) a table with one row per pipeline and a path column.
Open moveDataLake and configure at a high level.
Run & review.
path and _updateStatus.Apply for real.
Turn Dry run off, keep Backup on (recommended), run again, then publish your updated pipelines if your workflow requires it.
Missing statuspath column doesn’t point to a pipeline file that moveDataLake can reach. Verify the path, workspace, and permissions.
No changes after a run
Unintended rewrites
Permission or write errors
Post-run issues when executing pipelines
