Details
-
New Feature
-
Status: Closed
-
Major
-
Resolution: Fixed
-
None
-
None
-
Reviewed
Description
In most of the Hadoop clusters, as more and more data is stored for longer time, the demand for storage is outstripping the compute. Hadoop needs a cost effective and easy to manage solution to meet this demand for storage. Current solution is:
- Delete the old unused data. This comes at operational cost of identifying unnecessary data and deleting them manually.
- Add more nodes to the clusters. This adds along with storage capacity unnecessary compute capacity to the cluster.
Hadoop needs a solution to decouple growing storage capacity from compute capacity. Nodes with higher density and less expensive storage with low compute power are becoming available and can be used as cold storage in the clusters. Based on policy the data from hot storage can be moved to cold storage. Adding more nodes to the cold storage can grow the storage independent of the compute capacity in the cluster.
Attachments
Attachments
Issue Links
- is depended upon by
-
HDFS-7197 Enhancements to Mover
- Open
- is related to
-
HDFS-13678 StorageType is incompatible when rolling upgrade to 2.6/2.6+ versions
- Open
-
HDFS-7147 Update archival storage user documentation
- Closed
-
HDFS-7076 Allow users to define custom storage policies
- Patch Available
- relates to
-
HDFS-7184 Allow data migration tool to run as a daemon
- Resolved
-
HDFS-7285 Erasure Coding Support inside HDFS
- Resolved
-
HDFS-2832 Enable support for heterogeneous storages in HDFS - DN as a collection of storages
- Closed
-
HDFS-5682 Heterogeneous Storage phase 2 - APIs to expose Storage Types
- Closed
-
HBASE-6572 Tiered HFile storage
- Closed
- requires
-
HDFS-6702 DFSClient should create blocks using StorageType
- Closed
-
HDFS-6685 Balancer should preserve storage type of replicas
- Closed
-
HDFS-6700 BlockPlacementPolicy shoud choose storage but not datanode for deletion
- Closed