Details
Description
HADOOP-12950 added a timeout on shutdowns to avoid problems with hanging shutdowns. But the timeout is too short for applications where a large flush of data is needed on shutdown.
A key example of this is Spark apps which save their history to object stores, where the file close() call triggers an upload of the final local cached block of data (could be 32+MB), and then execute the final mutipart commit.
Proposed
- make the default sleep time 30s, not 10s
- make it configurable with a time duration property (with minimum time of 1s.?)
Attachments
Attachments
Issue Links
- is caused by
-
HADOOP-12950 ShutdownHookManager should have a timeout for each of the Registered shutdown hook
- Resolved
- is related to
-
HADOOP-16341 ShutDownHookManager: Regressed performance on Hook removals after HADOOP-15679
- Resolved
-
HADOOP-16461 Regression: FileSystem cache lock parses XML within the lock
- Resolved
-
SPARK-24523 InterruptedException when closing SparkContext
- Resolved
-
SPARK-25183 Spark HiveServer2 registers shutdown hook with JVM, not ShutdownHookManager; race conditions can arise
- Resolved