Details
-
Bug
-
Status: Closed
-
Major
-
Resolution: Fixed
-
2.1.0-beta
-
None
-
Reviewed
Description
I had 6 nodes in a cluster with 2 NMs stopped. Then I put a host into YARN's yarn.resourcemanager.nodes.exclude-path. After running yarn rmadmin -refreshNodes, RM's JMX correctly showed decommissioned node count:
"NumActiveNMs" : 3, "NumDecommissionedNMs" : 1, "NumLostNMs" : 2, "NumUnhealthyNMs" : 0, "NumRebootedNMs" : 0
After restarting RM, the counts were shown as below in JMX.
"NumActiveNMs" : 3, "NumDecommissionedNMs" : 0, "NumLostNMs" : 0, "NumUnhealthyNMs" : 0, "NumRebootedNMs" : 0
Notice that the lost and decommissioned NM counts are both 0.
Attachments
Attachments
Issue Links
- relates to
-
YARN-2567 Add a percentage-node threshold for RM to wait for new allocations after restart/failover
- Open
-
AMBARI-2940 After restarting YARN, the number of lost nodes is incorrect
- Resolved