You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
[SPARK-52663][SDP] Introduce name field to pipeline spec
### What changes were proposed in this pull request?
The Declarative Pipelines SPIP included a "name" field in the pipeline spec, but we left that out in the earlier implementation. This adds it in.
The name field is required. This matches behavior for similar systems, like dbt.
### Why are the changes needed?
See above.
### Does this PR introduce _any_ user-facing change?
Yes, but only to unreleased code.
### How was this patch tested?
Updated existing tests, and added tests for proper error when the name is missing.
### Was this patch authored or co-authored using generative AI tooling?
Closes#51353 from sryza/pipeline-name.
Authored-by: Sandy Ryza <sandy.ryza@databricks.com>
Signed-off-by: Gengliang Wang <gengliang@apache.org>
Copy file name to clipboardExpand all lines: docs/declarative-pipelines-programming-guide.md
+2-1Lines changed: 2 additions & 1 deletion
Original file line number
Diff line number
Diff line change
@@ -75,6 +75,7 @@ A YAML-formatted pipeline spec file contains the top-level configuration for the
75
75
An example pipeline spec file:
76
76
77
77
```yaml
78
+
name: my_pipeline
78
79
definitions:
79
80
- glob:
80
81
include: transformations/**/*.py
@@ -99,7 +100,7 @@ The `spark-pipelines` command line interface (CLI) is the primary way to execute
99
100
100
101
### `spark-pipelines init`
101
102
102
-
`spark-pipelines init`generates a simple pipeline project, including a spec file and example definitions.
103
+
`spark-pipelines init --name my_pipeline`generates a simple pipeline project, inside a directory named "my_pipeline", including a spec file and example definitions.
0 commit comments