Details
-
Bug
-
Status: Resolved
-
Major
-
Resolution: Fixed
-
None
-
None
Description
There are quite some new optimizations in Spark 2.0.0. We need to bump up Spark to 2.0.0 to benefit those performance improvements.
To update Spark version to 2.0.0, the following changes are required:
- Spark API updates:
- SparkShuffler#call return Iterator instead of Iterable
- SparkListener -> JavaSparkListener
- InputMetrics constructor doesn’t accept readMethod
- Method remoteBlocksFetched and localBlocksFetched in ShuffleReadMetrics return long type instead of integer
- Dependency upgrade:
- Jackson: 2.4.2 -> 2.6.5
- Netty version: 4.0.23.Final -> 4.0.29.Final
- Scala binary version: 2.10 -> 2.11
- Scala version: 2.10.4 -> 2.11.8
Attachments
Attachments
Issue Links
- is duplicated by
-
HIVE-14777 Add support of Spark-2.0.0 in Hive-2.X.X
- Resolved
- is related to
-
SPARK-12154 Upgrade to Jersey 2
- Resolved
-
SPARK-17563 Add org/apache/spark/JavaSparkListener to make Spark-2.0.0 work with Hive-2.X.X
- Resolved
-
HIVE-14825 Figure out the minimum set of required jars for Hive on Spark after bumping up to Spark 2.0.0
- Resolved
- relates to
-
HIVE-14919 Improve the performance of Hive on Spark 2.0.0
- Open
- links to