Details
-
Bug
-
Status: Closed
-
Major
-
Resolution: Fixed
-
None
-
None
Description
Phoenix's MapReduce integration lives in PhoenixInputFormat. It implements getSplits by calculating a QueryPlan for the provided SELECT query, and each split gets a mapper. As part of this QueryPlan generation, we grab all RegionLocations from .META
In PhoenixInputFormat:getQueryPlan:
// Initialize the query plan so it sets up the parallel scans
queryPlan.iterator(MapReduceParallelScanGrouper.getInstance());
In MapReduceParallelScanGrouper.getRegionBoundaries()
return context.getConnection().getQueryServices().getAllTableRegions(tableName);
This is fine.
Unfortunately, each mapper Task spawned by the job will go through this same exercise. It will pass a MapReduceParallelScanGrouper to queryPlan.iterator(), which I believe is eventually causing getRegionBoundaries to get called when the scans are initialized in the result iterator.
Since HBase 1.x and up got rid of .META prefetching and caching within the HBase client, that means that not only will each Job make potentially thousands of calls to .META, potentially thousands of Tasks will each make potentially thousands of calls to .META.
We should get a QueryPlan and setup the scans without having to read all RegionLocations, either by using the mapper's internal knowledge of its split key range, or by serializing the query plan from the client and sending it to the mapper tasks for use there.
Note that MapReduce tasks over snapshots are not affected by this, because region locations are stored in the snapshot manifest.
Attachments
Attachments
Issue Links
- relates to
-
PHOENIX-5362 Mappers should use the queryPlan from the driver rather than regenerating the plan
- Open
- links to