Skip to content

Commit 1db267e

Browse files
sryzadongjoon-hyun
authored andcommitted
[SPARK-54376][SDP] Mark most pipeline configuration options as internal
### What changes were proposed in this pull request? Marks all declarative pipelines configuration options as internal, except for `spark.sql.pipelines.maxFlowRetryAttempts`. ### Why are the changes needed? When implementing Declarative Pipelines, we made several quantities configurable. However, documented configurations are essentially public APIs, and it's too early to commit yet to supporting all of these. We should mark most of them internal except where we think users will really need them. ### Does this PR introduce _any_ user-facing change? Yes, to unreleased software. ### How was this patch tested? ### Was this patch authored or co-authored using generative AI tooling? Closes #53090 from sryza/internal-configs. Authored-by: Sandy Ryza <sandy.ryza@databricks.com> Signed-off-by: Dongjoon Hyun <dongjoon@apache.org>
1 parent 5222f1a commit 1db267e

File tree

1 file changed

+6
-0
lines changed
  • sql/catalyst/src/main/scala/org/apache/spark/sql/internal

1 file changed

+6
-0
lines changed

sql/catalyst/src/main/scala/org/apache/spark/sql/internal/SQLConf.scala

Lines changed: 6 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -6511,6 +6511,7 @@ object SQLConf {
65116511

65126512
val PIPELINES_STREAM_STATE_POLLING_INTERVAL = {
65136513
buildConf("spark.sql.pipelines.execution.streamstate.pollingInterval")
6514+
.internal()
65146515
.doc(
65156516
"Interval in seconds at which the stream state is polled for changes. This is used to " +
65166517
"check if the stream has failed and needs to be restarted."
@@ -6522,6 +6523,7 @@ object SQLConf {
65226523

65236524
val PIPELINES_WATCHDOG_MIN_RETRY_TIME_IN_SECONDS = {
65246525
buildConf("spark.sql.pipelines.execution.watchdog.minRetryTime")
6526+
.internal()
65256527
.doc(
65266528
"Initial duration in seconds between the time when we notice a flow has failed and " +
65276529
"when we try to restart the flow. The interval between flow restarts doubles with " +
@@ -6536,6 +6538,7 @@ object SQLConf {
65366538

65376539
val PIPELINES_WATCHDOG_MAX_RETRY_TIME_IN_SECONDS = {
65386540
buildConf("spark.sql.pipelines.execution.watchdog.maxRetryTime")
6541+
.internal()
65396542
.doc(
65406543
"Maximum time interval in seconds at which flows will be restarted."
65416544
)
@@ -6546,6 +6549,7 @@ object SQLConf {
65466549

65476550
val PIPELINES_MAX_CONCURRENT_FLOWS = {
65486551
buildConf("spark.sql.pipelines.execution.maxConcurrentFlows")
6552+
.internal()
65496553
.doc(
65506554
"Max number of flows to execute at once. Used to tune performance for triggered " +
65516555
"pipelines. Has no effect on continuous pipelines."
@@ -6558,6 +6562,7 @@ object SQLConf {
65586562

65596563
val PIPELINES_TIMEOUT_MS_FOR_TERMINATION_JOIN_AND_LOCK = {
65606564
buildConf("spark.sql.pipelines.timeoutMsForTerminationJoinAndLock")
6565+
.internal()
65616566
.doc("Timeout in milliseconds to grab a lock for stopping update - default is 1hr.")
65626567
.version("4.1.0")
65636568
.timeConf(TimeUnit.MILLISECONDS)
@@ -6575,6 +6580,7 @@ object SQLConf {
65756580

65766581
val PIPELINES_EVENT_QUEUE_CAPACITY = {
65776582
buildConf("spark.sql.pipelines.event.queue.capacity")
6583+
.internal()
65786584
.doc("Capacity of the event queue used in pipelined execution. When the queue is full, " +
65796585
"non-terminal FlowProgressEvents will be dropped.")
65806586
.version("4.1.0")

0 commit comments

Comments
 (0)