Skip to content

Commit a49ffde

Browse files
sryzadongjoon-hyun
authored andcommitted
[SPARK-54376][SDP] Mark most pipeline configuration options as internal
### What changes were proposed in this pull request? Marks all declarative pipelines configuration options as internal, except for `spark.sql.pipelines.maxFlowRetryAttempts`. ### Why are the changes needed? When implementing Declarative Pipelines, we made several quantities configurable. However, documented configurations are essentially public APIs, and it's too early to commit yet to supporting all of these. We should mark most of them internal except where we think users will really need them. ### Does this PR introduce _any_ user-facing change? Yes, to unreleased software. ### How was this patch tested? ### Was this patch authored or co-authored using generative AI tooling? Closes #53090 from sryza/internal-configs. Authored-by: Sandy Ryza <sandy.ryza@databricks.com> Signed-off-by: Dongjoon Hyun <dongjoon@apache.org> (cherry picked from commit 1db267e) Signed-off-by: Dongjoon Hyun <dongjoon@apache.org>
1 parent aed0547 commit a49ffde

File tree

1 file changed

+6
-0
lines changed
  • sql/catalyst/src/main/scala/org/apache/spark/sql/internal

1 file changed

+6
-0
lines changed

sql/catalyst/src/main/scala/org/apache/spark/sql/internal/SQLConf.scala

Lines changed: 6 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -6439,6 +6439,7 @@ object SQLConf {
64396439

64406440
val PIPELINES_STREAM_STATE_POLLING_INTERVAL = {
64416441
buildConf("spark.sql.pipelines.execution.streamstate.pollingInterval")
6442+
.internal()
64426443
.doc(
64436444
"Interval in seconds at which the stream state is polled for changes. This is used to " +
64446445
"check if the stream has failed and needs to be restarted."
@@ -6450,6 +6451,7 @@ object SQLConf {
64506451

64516452
val PIPELINES_WATCHDOG_MIN_RETRY_TIME_IN_SECONDS = {
64526453
buildConf("spark.sql.pipelines.execution.watchdog.minRetryTime")
6454+
.internal()
64536455
.doc(
64546456
"Initial duration in seconds between the time when we notice a flow has failed and " +
64556457
"when we try to restart the flow. The interval between flow restarts doubles with " +
@@ -6464,6 +6466,7 @@ object SQLConf {
64646466

64656467
val PIPELINES_WATCHDOG_MAX_RETRY_TIME_IN_SECONDS = {
64666468
buildConf("spark.sql.pipelines.execution.watchdog.maxRetryTime")
6469+
.internal()
64676470
.doc(
64686471
"Maximum time interval in seconds at which flows will be restarted."
64696472
)
@@ -6474,6 +6477,7 @@ object SQLConf {
64746477

64756478
val PIPELINES_MAX_CONCURRENT_FLOWS = {
64766479
buildConf("spark.sql.pipelines.execution.maxConcurrentFlows")
6480+
.internal()
64776481
.doc(
64786482
"Max number of flows to execute at once. Used to tune performance for triggered " +
64796483
"pipelines. Has no effect on continuous pipelines."
@@ -6486,6 +6490,7 @@ object SQLConf {
64866490

64876491
val PIPELINES_TIMEOUT_MS_FOR_TERMINATION_JOIN_AND_LOCK = {
64886492
buildConf("spark.sql.pipelines.timeoutMsForTerminationJoinAndLock")
6493+
.internal()
64896494
.doc("Timeout in milliseconds to grab a lock for stopping update - default is 1hr.")
64906495
.version("4.1.0")
64916496
.timeConf(TimeUnit.MILLISECONDS)
@@ -6503,6 +6508,7 @@ object SQLConf {
65036508

65046509
val PIPELINES_EVENT_QUEUE_CAPACITY = {
65056510
buildConf("spark.sql.pipelines.event.queue.capacity")
6511+
.internal()
65066512
.doc("Capacity of the event queue used in pipelined execution. When the queue is full, " +
65076513
"non-terminal FlowProgressEvents will be dropped.")
65086514
.version("4.1.0")

0 commit comments

Comments
 (0)