-
Notifications
You must be signed in to change notification settings - Fork 9.1k
HADOOP-19097. S3A: Set fs.s3a.connection.establish.timeout to 30s #6601
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
HADOOP-19097. S3A: Set fs.s3a.connection.establish.timeout to 30s #6601
Conversation
This is consistent with the java value. Change-Id: Ib24f4057f778206d6d59230de02037c5ada209f4
💔 -1 overall
This message was automatically generated. |
@HarshitGupta11 @mukund-thakur @ahmarsuhail reviews, please? targeting 3.4.1 |
Change looks good, until we remove fs.s3a settings from core-default, i wonder if we should do one round of comparison b/ s3a/Contants and core-default (as a followup jira)? |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
+1, LGTM.
@virajjasani I did a scan. Now, the clever thing would be to have a test suite which compared values so there is never a regression. e.g assertDurationEqual(conf, CONNECTION_TIMEOUT, CONNECTION_TIMEOUT_DEFAULT) |
…eout` to 30s if missing ### What changes were proposed in this pull request? This PR aims to handle HADOOP-19097 from Apache Spark side. We can remove this when Apache Hadoop `3.4.1` releases. - apache/hadoop#6601 ### Why are the changes needed? Apache Hadoop shows a warning to its default configuration. This default value issue is fixed at Apache Spark 3.4.1. ``` 24/03/25 14:46:21 WARN ConfigurationHelper: Option fs.s3a.connection.establish.timeout is too low (5,000 ms). Setting to 15,000 ms instead ``` This change will suppress Apache Hadoop default warning in the consistent way with the future Hadoop releases. ### Does this PR introduce _any_ user-facing change? No. ### How was this patch tested? Pass the CIs. Manually. **BUILD** ``` $ dev/make-distribution.sh -Phadoop-cloud ``` **BEFORE** ``` scala> spark.range(10).write.mode("overwrite").orc("s3a://express-1-zone--***--x-s3/orc/") ... 24/03/25 15:50:46 WARN ConfigurationHelper: Option fs.s3a.connection.establish.timeout is too low (5,000 ms). Setting to 15,000 ms instead ``` **AFTER** ``` scala> spark.range(10).write.mode("overwrite").orc("s3a://express-1-zone--***--x-s3/orc/") ...(ConfigurationHelper warning is gone)... ``` ### Was this patch authored or co-authored using generative AI tooling? No. Closes #45710 from dongjoon-hyun/SPARK-47552. Authored-by: Dongjoon Hyun <[email protected]> Signed-off-by: Dongjoon Hyun <[email protected]>
…eout` to 30s if missing ### What changes were proposed in this pull request? This PR aims to handle HADOOP-19097 from Apache Spark side. We can remove this when Apache Hadoop `3.4.1` releases. - apache/hadoop#6601 ### Why are the changes needed? Apache Hadoop shows a warning to its default configuration. This default value issue is fixed at Apache Spark 3.4.1. ``` 24/03/25 14:46:21 WARN ConfigurationHelper: Option fs.s3a.connection.establish.timeout is too low (5,000 ms). Setting to 15,000 ms instead ``` This change will suppress Apache Hadoop default warning in the consistent way with the future Hadoop releases. ### Does this PR introduce _any_ user-facing change? No. ### How was this patch tested? Pass the CIs. Manually. **BUILD** ``` $ dev/make-distribution.sh -Phadoop-cloud ``` **BEFORE** ``` scala> spark.range(10).write.mode("overwrite").orc("s3a://express-1-zone--***--x-s3/orc/") ... 24/03/25 15:50:46 WARN ConfigurationHelper: Option fs.s3a.connection.establish.timeout is too low (5,000 ms). Setting to 15,000 ms instead ``` **AFTER** ``` scala> spark.range(10).write.mode("overwrite").orc("s3a://express-1-zone--***--x-s3/orc/") ...(ConfigurationHelper warning is gone)... ``` ### Was this patch authored or co-authored using generative AI tooling? No. Closes apache#45710 from dongjoon-hyun/SPARK-47552. Authored-by: Dongjoon Hyun <[email protected]> Signed-off-by: Dongjoon Hyun <[email protected]>
This is consistent with the java value.
I had thought of cutting all the fs.s3a settings from core-default, but think we maybe need to review our public docs before doing that. having to look at Constants.java shouldn't be the default way to learn about an option.
How was this patch tested?
commented out my timeout from my auth-keys file (so it wasn't stamping on this default) and running the tests. The used ripgrep to look for the "is too low" message; only found in test cases where we explicitly created the problem.
For code changes:
LICENSE
,LICENSE-binary
,NOTICE-binary
files?