122 lines
5.4 KiB
Diff
122 lines
5.4 KiB
Diff
|
|
From acefcbdbb4891aa0b3f1b7afe500b8fdef440806 Mon Sep 17 00:00:00 2001
|
|||
|
|
From: huangbin <huangbin58@huawei.com>
|
|||
|
|
Date: Fri, 8 Nov 2024 16:42:52 +0800
|
|||
|
|
Subject: [PATCH] configure-group-in-json
|
|||
|
|
|
|||
|
|
---
|
|||
|
|
anteater/model/detector/slow_node_detector.py | 28 +++++++++++--------
|
|||
|
|
config/module/slow_node_detection.job.json | 9 ++++--
|
|||
|
|
2 files changed, 24 insertions(+), 13 deletions(-)
|
|||
|
|
|
|||
|
|
diff --git a/anteater/model/detector/slow_node_detector.py b/anteater/model/detector/slow_node_detector.py
|
|||
|
|
index 15a6cee..d5d10e1 100644
|
|||
|
|
--- a/anteater/model/detector/slow_node_detector.py
|
|||
|
|
+++ b/anteater/model/detector/slow_node_detector.py
|
|||
|
|
@@ -55,6 +55,9 @@ class SlowNodeDetector(Detector):
|
|||
|
|
hccl_domain = json.load(f_out)
|
|||
|
|
except Exception:
|
|||
|
|
logger.error(f"Read hccl domain info fail!")
|
|||
|
|
+ if not hccl_domain:
|
|||
|
|
+ # 增加手动设置hccl_domain
|
|||
|
|
+ hccl_domain = params.get("hccl_domain", {})
|
|||
|
|
if os.path.exists(rank_table_path):
|
|||
|
|
try:
|
|||
|
|
with open(rank_table_path, 'r', encoding='utf-8') as f_out:
|
|||
|
|
@@ -106,15 +109,15 @@ class SlowNodeDetector(Detector):
|
|||
|
|
# 获取machine_ids,
|
|||
|
|
machines_to_devices = self.get_machines_to_devices(start, end, kpis)
|
|||
|
|
npu_id2host_id, hosts_ids = self.npu_id2host_id(machines_to_devices)
|
|||
|
|
-
|
|||
|
|
group_dataloader = GroupDataLoader(self.hccl_domain, self.rank_table, machines_to_devices)
|
|||
|
|
group_ranks: list = group_dataloader.get_group_ranks()
|
|||
|
|
all_results = []
|
|||
|
|
for kpi in kpis:
|
|||
|
|
- for ranks in group_ranks:
|
|||
|
|
+ for index, ranks in enumerate(group_ranks):
|
|||
|
|
+ logger.info(f"Groups-{index}, metric: {kpi.metric}, start detection.")
|
|||
|
|
machine_ids: dict = group_dataloader.rank_table_loader.get_group_nodes_by_ranks(ranks)
|
|||
|
|
host_ids: list = self.get_host_ids_by_npu_ids(machine_ids, npu_id2host_id, hosts_ids)
|
|||
|
|
- group_result = self.group_detect_single_kpi(kpi, machine_ids, host_ids)
|
|||
|
|
+ group_result = self.group_detect_single_kpi(kpi, machine_ids, host_ids, ranks)
|
|||
|
|
all_results.extend(group_result)
|
|||
|
|
|
|||
|
|
response, all_anomaly_nodes = self.gen_final_alarm(kpis, all_results)
|
|||
|
|
@@ -148,19 +151,26 @@ class SlowNodeDetector(Detector):
|
|||
|
|
|
|||
|
|
return response, all_anomaly_nodes
|
|||
|
|
|
|||
|
|
- def group_detect_single_kpi(self, kpi: KPI, machine_ids: dict, host_ids: list) -> list:
|
|||
|
|
+ def group_detect_single_kpi(self, kpi: KPI, machine_ids: dict, host_ids: list, ranks) -> list:
|
|||
|
|
"""Detects kpi based on signal time series anomaly detection model"""
|
|||
|
|
# 普罗会一次性抓到所有的数据,需要根据machine_id, device_id去对数据作分组
|
|||
|
|
- # get数据
|
|||
|
|
metric_name: str = kpi.metric
|
|||
|
|
|
|||
|
|
all_machines_ts = []
|
|||
|
|
for machine_id in machine_ids:
|
|||
|
|
single_machine_ts_list = self.get_kpi_ts_list(metric_name, machine_id, kpi.params)
|
|||
|
|
- all_machines_ts.extend(single_machine_ts_list)
|
|||
|
|
+ if single_machine_ts_list:
|
|||
|
|
+ # 根据ranks匹配组内device的指标
|
|||
|
|
+ local_ranks = [int(rank) % 8 for rank in ranks]
|
|||
|
|
+ for single_machine_ts in single_machine_ts_list:
|
|||
|
|
+ ts_id = int(single_machine_ts.labels.get("id", -1))
|
|||
|
|
+ if ts_id in local_ranks:
|
|||
|
|
+ all_machines_ts.append(single_machine_ts)
|
|||
|
|
+
|
|||
|
|
for host_id in host_ids:
|
|||
|
|
single_machine_ts_list = self.get_kpi_ts_list(metric_name, host_id, kpi.params)
|
|||
|
|
all_machines_ts.extend(single_machine_ts_list)
|
|||
|
|
+ logger.info(f"Metric-{metric_name} single group has data {len(all_machines_ts)}. ranks: {ranks}")
|
|||
|
|
|
|||
|
|
anomaly_devices = []
|
|||
|
|
anomaly_locations = {}
|
|||
|
|
@@ -195,7 +205,7 @@ class SlowNodeDetector(Detector):
|
|||
|
|
logger.info(
|
|||
|
|
f"space_nodes_compare result: {self.output_anomaly_devices(metric_name, space_anomaly_locations)}.")
|
|||
|
|
else:
|
|||
|
|
- logger.info(f"Skip space nodes compare, due to nodes number{len(all_machines_ts)} is smaller than 4.")
|
|||
|
|
+ logger.info(f"Skip space nodes compare, due to nodes number {len(all_machines_ts)} is smaller than 4.")
|
|||
|
|
else:
|
|||
|
|
logger.info(f"Skip space nodes compare.")
|
|||
|
|
|
|||
|
|
@@ -287,10 +297,6 @@ class SlowNodeDetector(Detector):
|
|||
|
|
look_back = self.config.params.get("look_back", 10)
|
|||
|
|
metric_type = kpi_params.get("metric_type", "device")
|
|||
|
|
start, end = dt.last(minutes=look_back)
|
|||
|
|
- # from datetime import timedelta
|
|||
|
|
- # # 6:40
|
|||
|
|
- # start = start - timedelta(hours=4.5)
|
|||
|
|
- # end = end - timedelta(hours=4.5)
|
|||
|
|
|
|||
|
|
if metric_type == "device":
|
|||
|
|
# npu device
|
|||
|
|
diff --git a/config/module/slow_node_detection.job.json b/config/module/slow_node_detection.job.json
|
|||
|
|
index 91ff621..27a6d53 100644
|
|||
|
|
--- a/config/module/slow_node_detection.job.json
|
|||
|
|
+++ b/config/module/slow_node_detection.job.json
|
|||
|
|
@@ -17,7 +17,12 @@
|
|||
|
|
"obs_size": 5,
|
|||
|
|
"outlier_ratio_th": 0.6,
|
|||
|
|
"hccl_domain_json": "./hccl_domain.json",
|
|||
|
|
- "rank_table_json": "./hccl_domain.json"
|
|||
|
|
+ "hccl_domain":{
|
|||
|
|
+ "pp": 2,
|
|||
|
|
+ "dp": 1,
|
|||
|
|
+ "tp": 1
|
|||
|
|
+ },
|
|||
|
|
+ "rank_table_json": "./rank_table.json"
|
|||
|
|
}
|
|||
|
|
},
|
|||
|
|
"kpis": [
|
|||
|
|
@@ -349,4 +354,4 @@
|
|||
|
|
"metric": "gala_gopher_container_cpu_usage_seconds_total"
|
|||
|
|
}
|
|||
|
|
]
|
|||
|
|
-}
|
|||
|
|
\ No newline at end of file
|
|||
|
|
+}
|
|||
|
|
--
|
|||
|
|
2.33.0
|
|||
|
|
|