Details
Description
With https://issues.apache.org/jira/browse/SPARK-18107, we will disable the underlying replace(overwrite) and instead do delete in spark side and only do copy in hive side to bypass the performance issue - https://issues.apache.org/jira/browse/HIVE-11940
Conditionally, if the table location and partition location do not belong to the same [[FileSystem]], We should not disable hive overwrite. Otherwise, hive will use the [[FileSystem]] instance belong to the table location to copy files, which will fail [FileSystem#checkPath]
see https://github.com/apache/hive/blob/rel/release-2.3.7/ql/src/java/org/apache/hadoop/hive/ql/metadata/Hive.java#L1648-L1659
Attachments
Issue Links
- is related to
-
SPARK-31675 Fail to insert data to a table with remote location which causes by hive encryption check
- In Progress
- links to