完成H20服务器部署及重启测试 #51
@ -406,6 +406,7 @@ build_gpu_bundle_image() {
|
|||||||
mkdir -p "$bundle_ctx/bundle" "$bundle_ctx/private"
|
mkdir -p "$bundle_ctx/bundle" "$bundle_ctx/private"
|
||||||
cp "$root/src/bundle/gpu-node-bundle/Dockerfile" "$bundle_ctx/"
|
cp "$root/src/bundle/gpu-node-bundle/Dockerfile" "$bundle_ctx/"
|
||||||
cp "$root/src/bundle/gpu-node-bundle/node-bootstrap.sh" "$bundle_ctx/"
|
cp "$root/src/bundle/gpu-node-bundle/node-bootstrap.sh" "$bundle_ctx/"
|
||||||
|
cp "$root/src/bundle/gpu-node-bundle/health-watcher.sh" "$bundle_ctx/"
|
||||||
# bundle tar
|
# bundle tar
|
||||||
cp "$artifact_tar" "$bundle_ctx/bundle/"
|
cp "$artifact_tar" "$bundle_ctx/bundle/"
|
||||||
# offline fluent-bit assets (optional but useful)
|
# offline fluent-bit assets (optional but useful)
|
||||||
@ -592,6 +593,7 @@ build_cpu_bundle_image() {
|
|||||||
mkdir -p "$bundle_ctx/bundle" "$bundle_ctx/private"
|
mkdir -p "$bundle_ctx/bundle" "$bundle_ctx/private"
|
||||||
cp "$root/src/bundle/cpu-node-bundle/Dockerfile" "$bundle_ctx/"
|
cp "$root/src/bundle/cpu-node-bundle/Dockerfile" "$bundle_ctx/"
|
||||||
cp "$root/src/bundle/cpu-node-bundle/node-bootstrap.sh" "$bundle_ctx/"
|
cp "$root/src/bundle/cpu-node-bundle/node-bootstrap.sh" "$bundle_ctx/"
|
||||||
|
cp "$root/src/bundle/cpu-node-bundle/health-watcher.sh" "$bundle_ctx/"
|
||||||
# bundle tar
|
# bundle tar
|
||||||
cp "$artifact_tar" "$bundle_ctx/bundle/"
|
cp "$artifact_tar" "$bundle_ctx/bundle/"
|
||||||
# offline fluent-bit assets
|
# offline fluent-bit assets
|
||||||
|
|||||||
@ -19,10 +19,6 @@ services:
|
|||||||
# Fluent Bit / 日志上报目标(固定域名)
|
# Fluent Bit / 日志上报目标(固定域名)
|
||||||
- ES_HOST=es.log.argus.com
|
- ES_HOST=es.log.argus.com
|
||||||
- ES_PORT=9200
|
- ES_PORT=9200
|
||||||
- FTPIP=${FTPIP}
|
|
||||||
- BINDIP=${BINDIP}
|
|
||||||
- FTP_USER=${FTP_USER:-ftpuser}
|
|
||||||
- FTP_PASSWORD=${FTP_PASSWORD:-NASPlab1234!}
|
|
||||||
- ARGUS_BUILD_UID=${ARGUS_BUILD_UID:-2133}
|
- ARGUS_BUILD_UID=${ARGUS_BUILD_UID:-2133}
|
||||||
- ARGUS_BUILD_GID=${ARGUS_BUILD_GID:-2015}
|
- ARGUS_BUILD_GID=${ARGUS_BUILD_GID:-2015}
|
||||||
- AGENT_ENV=${AGENT_ENV}
|
- AGENT_ENV=${AGENT_ENV}
|
||||||
@ -31,9 +27,10 @@ services:
|
|||||||
- NVIDIA_VISIBLE_DEVICES=all
|
- NVIDIA_VISIBLE_DEVICES=all
|
||||||
- NVIDIA_DRIVER_CAPABILITIES=compute,utility
|
- NVIDIA_DRIVER_CAPABILITIES=compute,utility
|
||||||
- GPU_MODE=gpu
|
- GPU_MODE=gpu
|
||||||
dns:
|
networks:
|
||||||
- ${BINDIP}
|
argus-sys-net:
|
||||||
networks: [argus-sys-net]
|
aliases:
|
||||||
|
- ${AGENT_INSTANCE}.node.argus.com
|
||||||
volumes:
|
volumes:
|
||||||
- ../private/argus/agent:/private/argus/agent
|
- ../private/argus/agent:/private/argus/agent
|
||||||
- ../logs/infer:/logs/infer
|
- ../logs/infer:/logs/infer
|
||||||
|
|||||||
@ -12,7 +12,7 @@
|
|||||||
su - argus -c 'id; docker ps >/dev/null && echo OK || echo NO_DOCKER_PERMISSION'
|
su - argus -c 'id; docker ps >/dev/null && echo OK || echo NO_DOCKER_PERMISSION'
|
||||||
```
|
```
|
||||||
后续解压与执行(config/install/uninstall)均使用 `argus` 账户进行。
|
后续解压与执行(config/install/uninstall)均使用 `argus` 账户进行。
|
||||||
- 从 Server 安装方拿到 `cluster-info.env`(包含 `SWARM_MANAGER_ADDR/BINDIP/FTPIP/SWARM_JOIN_TOKEN_*`)。
|
- 从 Server 安装方拿到 `cluster-info.env`(包含 `SWARM_MANAGER_ADDR/SWARM_JOIN_TOKEN_*`;compose 架构下 BINDIP/FTPIP 不再使用)。
|
||||||
|
|
||||||
## 二、解包
|
## 二、解包
|
||||||
- `tar -xzf client_gpu_YYYYMMDD.tar.gz`
|
- `tar -xzf client_gpu_YYYYMMDD.tar.gz`
|
||||||
@ -28,13 +28,13 @@ cp /path/to/cluster-info.env ./ # 或 export CLUSTER_INFO=/abs/path/cluster-in
|
|||||||
脚本做了什么:
|
脚本做了什么:
|
||||||
- 读取 `cluster-info.env` 并 `docker swarm join`(幂等);
|
- 读取 `cluster-info.env` 并 `docker swarm join`(幂等);
|
||||||
- 自动用 busybox 预热 external overlay `argus-sys-net`,等待最多 60s 直到本机可见;
|
- 自动用 busybox 预热 external overlay `argus-sys-net`,等待最多 60s 直到本机可见;
|
||||||
- 生成/更新 `compose/.env`:填入 `BINDIP/FTPIP/SWARM_*`,并“保留你已填写的 AGENT_* 与 GPU_NODE_HOSTNAME”(不会覆盖)。
|
- 生成/更新 `compose/.env`:填入 `SWARM_*`,并“保留你已填写的 AGENT_* 与 GPU_NODE_HOSTNAME”(不会覆盖)。
|
||||||
|
|
||||||
看到什么才算成功:
|
看到什么才算成功:
|
||||||
- 终端输出类似:`已预热 overlay=argus-sys-net 并生成 compose/.env;可执行 scripts/install.sh`;
|
- 终端输出类似:`已预热 overlay=argus-sys-net 并生成 compose/.env;可执行 scripts/install.sh`;
|
||||||
- `compose/.env` 至少包含:
|
- `compose/.env` 至少包含:
|
||||||
- `AGENT_ENV/AGENT_USER/AGENT_INSTANCE/GPU_NODE_HOSTNAME`(需要你提前填写);
|
- `AGENT_ENV/AGENT_USER/AGENT_INSTANCE/GPU_NODE_HOSTNAME`(需要你提前填写);
|
||||||
- `BINDIP/FTPIP/SWARM_MANAGER_ADDR/SWARM_JOIN_TOKEN_*`;
|
- `SWARM_MANAGER_ADDR/SWARM_JOIN_TOKEN_*`;
|
||||||
- `NODE_GPU_BUNDLE_IMAGE_TAG=...:YYYYMMDD`。
|
- `NODE_GPU_BUNDLE_IMAGE_TAG=...:YYYYMMDD`。
|
||||||
|
|
||||||
### 日志映射(重要)
|
### 日志映射(重要)
|
||||||
|
|||||||
@ -50,18 +50,16 @@ fi
|
|||||||
# 预热容器(worker 侧加入 overlay 以便本地可见)
|
# 预热容器(worker 侧加入 overlay 以便本地可见)
|
||||||
docker rm -f argus-net-warmup >/dev/null 2>&1 || true
|
docker rm -f argus-net-warmup >/dev/null 2>&1 || true
|
||||||
info "启动 warmup 容器加入 overlay: $NET_NAME"
|
info "启动 warmup 容器加入 overlay: $NET_NAME"
|
||||||
docker run -d --rm --name argus-net-warmup --network "$NET_NAME" ${BINDIP:+--dns "$BINDIP"} busybox:latest sleep 600 >/dev/null 2>&1 || true
|
docker run -d --rm --name argus-net-warmup --network "$NET_NAME" busybox:latest sleep 600 >/dev/null 2>&1 || true
|
||||||
for i in {1..60}; do docker network inspect "$NET_NAME" >/dev/null 2>&1 && { info "overlay 可见 (t=${i}s)"; break; }; sleep 1; done
|
for i in {1..60}; do docker network inspect "$NET_NAME" >/dev/null 2>&1 && { info "overlay 可见 (t=${i}s)"; break; }; sleep 1; done
|
||||||
docker network inspect "$NET_NAME" >/dev/null 2>&1 || { err "预热后仍未看到 overlay: $NET_NAME;请确认 manager 已创建并网络可达"; exit 1; }
|
docker network inspect "$NET_NAME" >/dev/null 2>&1 || { err "预热后仍未看到 overlay: $NET_NAME;请确认 manager 已创建并网络可达"; exit 1; }
|
||||||
|
|
||||||
# 从 warmup 容器内测试连通性(必须能 ping 通 BINDIP 与 FTPIP)
|
# 通过 warmup 容器测试实际数据通路(alias → master)
|
||||||
ping_ok(){ docker exec argus-net-warmup sh -lc "ping -c 1 -W 2 $1 >/dev/null 2>&1"; }
|
if ! docker exec argus-net-warmup sh -lc "ping -c 1 -W 2 master.argus.com >/dev/null 2>&1"; then
|
||||||
if [[ -n "${BINDIP:-}" ]]; then
|
err "warmup 容器内无法通过别名访问 master.argus.com;请确认 server compose 已启动并加入 overlay $NET_NAME"
|
||||||
ping_ok "$BINDIP" || { err "容器内无法 ping 通 BINDIP=$BINDIP;请检查 overlay 与 Bind9 容器状态"; exit 1; }
|
exit 1
|
||||||
fi
|
|
||||||
if [[ -n "${FTPIP:-}" ]]; then
|
|
||||||
ping_ok "$FTPIP" || { err "容器内无法 ping 通 FTPIP=$FTPIP;请检查 overlay 与 FTP 容器状态"; exit 1; }
|
|
||||||
fi
|
fi
|
||||||
|
info "warmup 容器内可达 master.argus.com(Docker DNS + alias 正常)"
|
||||||
|
|
||||||
# 生成/更新 .env(保留人工填写项,不覆盖已有键)
|
# 生成/更新 .env(保留人工填写项,不覆盖已有键)
|
||||||
if [[ ! -f "$ENV_OUT" ]]; then
|
if [[ ! -f "$ENV_OUT" ]]; then
|
||||||
@ -70,8 +68,6 @@ fi
|
|||||||
|
|
||||||
set_kv(){ local k="$1" v="$2"; if grep -q "^${k}=" "$ENV_OUT"; then sed -i -E "s#^${k}=.*#${k}=${v}#" "$ENV_OUT"; else echo "${k}=${v}" >> "$ENV_OUT"; fi }
|
set_kv(){ local k="$1" v="$2"; if grep -q "^${k}=" "$ENV_OUT"; then sed -i -E "s#^${k}=.*#${k}=${v}#" "$ENV_OUT"; else echo "${k}=${v}" >> "$ENV_OUT"; fi }
|
||||||
|
|
||||||
set_kv BINDIP "${BINDIP:-}"
|
|
||||||
set_kv FTPIP "${FTPIP:-}"
|
|
||||||
set_kv SWARM_MANAGER_ADDR "${SWARM_MANAGER_ADDR:-}"
|
set_kv SWARM_MANAGER_ADDR "${SWARM_MANAGER_ADDR:-}"
|
||||||
set_kv SWARM_JOIN_TOKEN_WORKER "${SWARM_JOIN_TOKEN_WORKER:-}"
|
set_kv SWARM_JOIN_TOKEN_WORKER "${SWARM_JOIN_TOKEN_WORKER:-}"
|
||||||
set_kv SWARM_JOIN_TOKEN_MANAGER "${SWARM_JOIN_TOKEN_MANAGER:-}"
|
set_kv SWARM_JOIN_TOKEN_MANAGER "${SWARM_JOIN_TOKEN_MANAGER:-}"
|
||||||
|
|||||||
@ -26,24 +26,24 @@ set -a; source "$ENV_FILE"; set +a
|
|||||||
NET_NAME="${ARGUS_OVERLAY_NET:-argus-sys-net}"
|
NET_NAME="${ARGUS_OVERLAY_NET:-argus-sys-net}"
|
||||||
info "检查 overlay 网络可见性: $NET_NAME"
|
info "检查 overlay 网络可见性: $NET_NAME"
|
||||||
if ! docker network inspect "$NET_NAME" >/dev/null 2>&1; then
|
if ! docker network inspect "$NET_NAME" >/dev/null 2>&1; then
|
||||||
# 如 Overlay 不可见,尝试用 busybox 预热
|
# 如 Overlay 不可见,尝试用 busybox 预热(仅为确保 worker 节点已加入 overlay)
|
||||||
if ! docker image inspect busybox:latest >/dev/null 2>&1; then
|
if ! docker image inspect busybox:latest >/dev/null 2>&1; then
|
||||||
if [[ -f "$PKG_ROOT/images/busybox.tar" ]]; then docker load -i "$PKG_ROOT/images/busybox.tar"; else err "缺少 busybox 镜像(images/busybox.tar 或本地 busybox:latest)"; exit 1; fi
|
if [[ -f "$PKG_ROOT/images/busybox.tar" ]]; then docker load -i "$PKG_ROOT/images/busybox.tar"; else err "缺少 busybox 镜像(images/busybox.tar 或本地 busybox:latest)"; exit 1; fi
|
||||||
fi
|
fi
|
||||||
docker rm -f argus-net-warmup >/dev/null 2>&1 || true
|
docker rm -f argus-net-warmup >/dev/null 2>&1 || true
|
||||||
docker run -d --rm --name argus-net-warmup --network "$NET_NAME" ${BINDIP:+--dns "$BINDIP"} busybox:latest sleep 600 >/dev/null 2>&1 || true
|
docker run -d --rm --name argus-net-warmup --network "$NET_NAME" busybox:latest sleep 600 >/dev/null 2>&1 || true
|
||||||
for i in {1..60}; do docker network inspect "$NET_NAME" >/dev/null 2>&1 && break; sleep 1; done
|
for i in {1..60}; do docker network inspect "$NET_NAME" >/dev/null 2>&1 && break; sleep 1; done
|
||||||
docker network inspect "$NET_NAME" >/dev/null 2>&1 || { err "预热后仍未看到 overlay: $NET_NAME;请确认 manager 已创建并网络可达"; exit 1; }
|
docker network inspect "$NET_NAME" >/dev/null 2>&1 || { err "预热后仍未看到 overlay: $NET_NAME;请确认 manager 已创建并网络可达"; exit 1; }
|
||||||
info "overlay 已可见(warmup=argus-net-warmup)"
|
info "overlay 已可见(warmup=argus-net-warmup)"
|
||||||
fi
|
fi
|
||||||
|
|
||||||
# 容器内连通性检查:BINDIP 与 FTPIP 可达
|
# 若本函数内重新创建了 warmup 容器,同样测试一次 alias 数据通路
|
||||||
ping_ok(){ docker exec argus-net-warmup sh -lc "ping -c 1 -W 2 $1 >/dev/null 2>&1"; }
|
if docker ps --format '{{.Names}}' | grep -q '^argus-net-warmup$'; then
|
||||||
if [[ -n "${BINDIP:-}" ]]; then
|
if ! docker exec argus-net-warmup sh -lc "ping -c 1 -W 2 master.argus.com >/dev/null 2>&1"; then
|
||||||
if ping_ok "$BINDIP"; then info "warmup 内可达 BINDIP=$BINDIP"; else err "容器内无法 ping 通 BINDIP=$BINDIP"; exit 1; fi
|
err "GPU install 阶段:warmup 容器内无法通过别名访问 master.argus.com;请检查 overlay $NET_NAME 与 server 状态"
|
||||||
|
exit 1
|
||||||
fi
|
fi
|
||||||
if [[ -n "${FTPIP:-}" ]]; then
|
info "GPU install 阶段:warmup 容器内可达 master.argus.com"
|
||||||
if ping_ok "$FTPIP"; then info "warmup 内可达 FTPIP=$FTPIP"; else err "容器内无法 ping 通 FTPIP=$FTPIP"; exit 1; fi
|
|
||||||
fi
|
fi
|
||||||
|
|
||||||
# 导入 GPU bundle 镜像
|
# 导入 GPU bundle 镜像
|
||||||
|
|||||||
@ -5,18 +5,9 @@ networks:
|
|||||||
external: true
|
external: true
|
||||||
|
|
||||||
services:
|
services:
|
||||||
bind:
|
|
||||||
image: ${BIND_IMAGE_TAG:-argus-bind9:${PKG_VERSION}}
|
|
||||||
container_name: argus-bind-sys
|
|
||||||
networks: [argus-sys-net]
|
|
||||||
volumes:
|
|
||||||
- ../private:/private
|
|
||||||
restart: unless-stopped
|
|
||||||
|
|
||||||
master:
|
master:
|
||||||
image: ${MASTER_IMAGE_TAG:-argus-master:${PKG_VERSION}}
|
image: ${MASTER_IMAGE_TAG:-argus-master:${PKG_VERSION}}
|
||||||
container_name: argus-master-sys
|
container_name: argus-master-sys
|
||||||
depends_on: [bind]
|
|
||||||
environment:
|
environment:
|
||||||
- OFFLINE_THRESHOLD_SECONDS=6
|
- OFFLINE_THRESHOLD_SECONDS=6
|
||||||
- ONLINE_THRESHOLD_SECONDS=2
|
- ONLINE_THRESHOLD_SECONDS=2
|
||||||
@ -29,7 +20,10 @@ services:
|
|||||||
- ../private/argus/master:/private/argus/master
|
- ../private/argus/master:/private/argus/master
|
||||||
- ../private/argus/metric/prometheus:/private/argus/metric/prometheus
|
- ../private/argus/metric/prometheus:/private/argus/metric/prometheus
|
||||||
- ../private/argus/etc:/private/argus/etc
|
- ../private/argus/etc:/private/argus/etc
|
||||||
networks: [argus-sys-net]
|
networks:
|
||||||
|
argus-sys-net:
|
||||||
|
aliases:
|
||||||
|
- master.argus.com
|
||||||
restart: unless-stopped
|
restart: unless-stopped
|
||||||
|
|
||||||
es:
|
es:
|
||||||
@ -47,7 +41,10 @@ services:
|
|||||||
ports:
|
ports:
|
||||||
- "${ES_HTTP_PORT:-9200}:9200"
|
- "${ES_HTTP_PORT:-9200}:9200"
|
||||||
restart: unless-stopped
|
restart: unless-stopped
|
||||||
networks: [argus-sys-net]
|
networks:
|
||||||
|
argus-sys-net:
|
||||||
|
aliases:
|
||||||
|
- es.log.argus.com
|
||||||
|
|
||||||
kibana:
|
kibana:
|
||||||
image: ${KIBANA_IMAGE_TAG:-argus-kibana:${PKG_VERSION}}
|
image: ${KIBANA_IMAGE_TAG:-argus-kibana:${PKG_VERSION}}
|
||||||
@ -63,27 +60,10 @@ services:
|
|||||||
ports:
|
ports:
|
||||||
- "${KIBANA_PORT:-5601}:5601"
|
- "${KIBANA_PORT:-5601}:5601"
|
||||||
restart: unless-stopped
|
restart: unless-stopped
|
||||||
networks: [argus-sys-net]
|
networks:
|
||||||
|
argus-sys-net:
|
||||||
ftp:
|
aliases:
|
||||||
image: ${FTP_IMAGE_TAG:-argus-metric-ftp:${PKG_VERSION}}
|
- kibana.log.argus.com
|
||||||
container_name: argus-ftp
|
|
||||||
restart: unless-stopped
|
|
||||||
environment:
|
|
||||||
- TZ=Asia/Shanghai
|
|
||||||
- FTP_BASE_PATH=/private/argus/ftp
|
|
||||||
- FTP_PASSWORD=${FTP_PASSWORD:-NASPlab1234!}
|
|
||||||
- DOMAIN=${FTP_DOMAIN:-ftp.metric.argus.com}
|
|
||||||
- ARGUS_BUILD_UID=${ARGUS_BUILD_UID:-2133}
|
|
||||||
- ARGUS_BUILD_GID=${ARGUS_BUILD_GID:-2015}
|
|
||||||
ports:
|
|
||||||
- "${FTP_PORT:-21}:21"
|
|
||||||
- "${FTP_DATA_PORT:-20}:20"
|
|
||||||
- "${FTP_PASSIVE_HOST_RANGE:-21100-21110}:21100-21110"
|
|
||||||
volumes:
|
|
||||||
- ../private/argus/metric/ftp:/private/argus/ftp
|
|
||||||
- ../private/argus/etc:/private/argus/etc
|
|
||||||
networks: [argus-sys-net]
|
|
||||||
|
|
||||||
prometheus:
|
prometheus:
|
||||||
image: ${PROM_IMAGE_TAG:-argus-metric-prometheus:${PKG_VERSION}}
|
image: ${PROM_IMAGE_TAG:-argus-metric-prometheus:${PKG_VERSION}}
|
||||||
@ -99,7 +79,10 @@ services:
|
|||||||
volumes:
|
volumes:
|
||||||
- ../private/argus/metric/prometheus:/private/argus/metric/prometheus
|
- ../private/argus/metric/prometheus:/private/argus/metric/prometheus
|
||||||
- ../private/argus/etc:/private/argus/etc
|
- ../private/argus/etc:/private/argus/etc
|
||||||
networks: [argus-sys-net]
|
networks:
|
||||||
|
argus-sys-net:
|
||||||
|
aliases:
|
||||||
|
- prom.metric.argus.com
|
||||||
|
|
||||||
grafana:
|
grafana:
|
||||||
image: ${GRAFANA_IMAGE_TAG:-argus-metric-grafana:${PKG_VERSION}}
|
image: ${GRAFANA_IMAGE_TAG:-argus-metric-grafana:${PKG_VERSION}}
|
||||||
@ -122,7 +105,10 @@ services:
|
|||||||
- ../private/argus/metric/grafana:/private/argus/metric/grafana
|
- ../private/argus/metric/grafana:/private/argus/metric/grafana
|
||||||
- ../private/argus/etc:/private/argus/etc
|
- ../private/argus/etc:/private/argus/etc
|
||||||
depends_on: [prometheus]
|
depends_on: [prometheus]
|
||||||
networks: [argus-sys-net]
|
networks:
|
||||||
|
argus-sys-net:
|
||||||
|
aliases:
|
||||||
|
- grafana.metric.argus.com
|
||||||
|
|
||||||
alertmanager:
|
alertmanager:
|
||||||
image: ${ALERT_IMAGE_TAG:-argus-alertmanager:${PKG_VERSION}}
|
image: ${ALERT_IMAGE_TAG:-argus-alertmanager:${PKG_VERSION}}
|
||||||
@ -133,7 +119,10 @@ services:
|
|||||||
volumes:
|
volumes:
|
||||||
- ../private/argus/etc:/private/argus/etc
|
- ../private/argus/etc:/private/argus/etc
|
||||||
- ../private/argus/alert/alertmanager:/private/argus/alert/alertmanager
|
- ../private/argus/alert/alertmanager:/private/argus/alert/alertmanager
|
||||||
networks: [argus-sys-net]
|
networks:
|
||||||
|
argus-sys-net:
|
||||||
|
aliases:
|
||||||
|
- alertmanager.alert.argus.com
|
||||||
ports:
|
ports:
|
||||||
- "${ALERTMANAGER_PORT:-9093}:9093"
|
- "${ALERTMANAGER_PORT:-9093}:9093"
|
||||||
restart: unless-stopped
|
restart: unless-stopped
|
||||||
@ -151,19 +140,25 @@ services:
|
|||||||
- EXTERNAL_KIBANA_PORT=${WEB_PROXY_PORT_8083:-8083}
|
- EXTERNAL_KIBANA_PORT=${WEB_PROXY_PORT_8083:-8083}
|
||||||
volumes:
|
volumes:
|
||||||
- ../private/argus/etc:/private/argus/etc
|
- ../private/argus/etc:/private/argus/etc
|
||||||
networks: [argus-sys-net]
|
networks:
|
||||||
|
argus-sys-net:
|
||||||
|
aliases:
|
||||||
|
- web.argus.com
|
||||||
restart: unless-stopped
|
restart: unless-stopped
|
||||||
|
|
||||||
web-proxy:
|
web-proxy:
|
||||||
image: ${WEB_PROXY_IMAGE_TAG:-argus-web-proxy:${PKG_VERSION}}
|
image: ${WEB_PROXY_IMAGE_TAG:-argus-web-proxy:${PKG_VERSION}}
|
||||||
container_name: argus-web-proxy
|
container_name: argus-web-proxy
|
||||||
depends_on: [bind, master, grafana, prometheus, kibana, alertmanager]
|
depends_on: [master, grafana, prometheus, kibana, alertmanager]
|
||||||
environment:
|
environment:
|
||||||
- ARGUS_BUILD_UID=${ARGUS_BUILD_UID:-2133}
|
- ARGUS_BUILD_UID=${ARGUS_BUILD_UID:-2133}
|
||||||
- ARGUS_BUILD_GID=${ARGUS_BUILD_GID:-2015}
|
- ARGUS_BUILD_GID=${ARGUS_BUILD_GID:-2015}
|
||||||
volumes:
|
volumes:
|
||||||
- ../private/argus/etc:/private/argus/etc
|
- ../private/argus/etc:/private/argus/etc
|
||||||
networks: [argus-sys-net]
|
networks:
|
||||||
|
argus-sys-net:
|
||||||
|
aliases:
|
||||||
|
- proxy.argus.com
|
||||||
ports:
|
ports:
|
||||||
- "${WEB_PROXY_PORT_8080:-8080}:8080"
|
- "${WEB_PROXY_PORT_8080:-8080}:8080"
|
||||||
- "${WEB_PROXY_PORT_8081:-8081}:8081"
|
- "${WEB_PROXY_PORT_8081:-8081}:8081"
|
||||||
@ -172,4 +167,3 @@ services:
|
|||||||
- "${WEB_PROXY_PORT_8084:-8084}:8084"
|
- "${WEB_PROXY_PORT_8084:-8084}:8084"
|
||||||
- "${WEB_PROXY_PORT_8085:-8085}:8085"
|
- "${WEB_PROXY_PORT_8085:-8085}:8085"
|
||||||
restart: unless-stopped
|
restart: unless-stopped
|
||||||
|
|
||||||
|
|||||||
@ -71,7 +71,7 @@ export SWARM_MANAGER_ADDR=<本机管理IP>
|
|||||||
- 等待“六项就绪”:
|
- 等待“六项就绪”:
|
||||||
- Master `/readyz`=200、ES `/_cluster/health`=200、Prometheus TCP 可达、Grafana `/api/health`=200、Alertmanager `/api/v2/status`=200、Kibana `/api/status` level=available;
|
- Master `/readyz`=200、ES `/_cluster/health`=200、Prometheus TCP 可达、Grafana `/api/health`=200、Alertmanager `/api/v2/status`=200、Kibana `/api/status` level=available;
|
||||||
- 将各服务 overlay IP 写入 `private/argus/etc/<域名>`,Reload Bind9 与 Nginx;
|
- 将各服务 overlay IP 写入 `private/argus/etc/<域名>`,Reload Bind9 与 Nginx;
|
||||||
- 写出 `cluster-info.env`(含 `BINDIP/FTPIP/SWARM_JOIN_TOKEN_{WORKER,MANAGER}/SWARM_MANAGER_ADDR`);
|
- 写出 `cluster-info.env`(含 `SWARM_JOIN_TOKEN_{WORKER,MANAGER}/SWARM_MANAGER_ADDR`;compose 架构下不再依赖 BINDIP/FTPIP);
|
||||||
- 生成 `安装报告_YYYYMMDD-HHMMSS.md`(端口、健康检查摘要与提示)。
|
- 生成 `安装报告_YYYYMMDD-HHMMSS.md`(端口、健康检查摘要与提示)。
|
||||||
|
|
||||||
看到什么才算成功:
|
看到什么才算成功:
|
||||||
@ -79,7 +79,7 @@ export SWARM_MANAGER_ADDR=<本机管理IP>
|
|||||||
- `安装报告_…md` 中各项 HTTP 检查为 200/available;
|
- `安装报告_…md` 中各项 HTTP 检查为 200/available;
|
||||||
- `cluster-info.env` 包含五个关键键:
|
- `cluster-info.env` 包含五个关键键:
|
||||||
- `SWARM_MANAGER_ADDR=...`
|
- `SWARM_MANAGER_ADDR=...`
|
||||||
- `BINDIP=10.x.x.x` `FTPIP=10.x.x.x`
|
- `SWARM_MANAGER_ADDR=...` `SWARM_JOIN_TOKEN_*=...`
|
||||||
- `SWARM_JOIN_TOKEN_WORKER=SWMTKN-...`
|
- `SWARM_JOIN_TOKEN_WORKER=SWMTKN-...`
|
||||||
- `SWARM_JOIN_TOKEN_MANAGER=SWMTKN-...`
|
- `SWARM_JOIN_TOKEN_MANAGER=SWMTKN-...`
|
||||||
|
|
||||||
|
|||||||
@ -88,23 +88,15 @@ for i in $(seq 1 "$RETRIES"); do
|
|||||||
done
|
done
|
||||||
[[ $ok -ge 6 ]] || err "部分服务未就绪(可稍后重试 selfcheck)"
|
[[ $ok -ge 6 ]] || err "部分服务未就绪(可稍后重试 selfcheck)"
|
||||||
|
|
||||||
# Resolve overlay IPs
|
|
||||||
bind_c=argus-bind-sys; ftp_c=argus-ftp
|
|
||||||
BINDIP=$(docker inspect -f '{{ (index .NetworkSettings.Networks "'$NET_NAME'").IPAddress }}' "$bind_c" 2>/dev/null || true)
|
|
||||||
FTPIP=$(docker inspect -f '{{ (index .NetworkSettings.Networks "'$NET_NAME'").IPAddress }}' "$ftp_c" 2>/dev/null || true)
|
|
||||||
info "解析 overlay IP: BINDIP=${BINDIP:-<empty>} FTPIP=${FTPIP:-<empty>}"
|
|
||||||
|
|
||||||
# Swarm join tokens
|
# Swarm join tokens
|
||||||
TOKEN_WORKER=$(docker swarm join-token -q worker 2>/dev/null || echo "")
|
TOKEN_WORKER=$(docker swarm join-token -q worker 2>/dev/null || echo "")
|
||||||
TOKEN_MANAGER=$(docker swarm join-token -q manager 2>/dev/null || echo "")
|
TOKEN_MANAGER=$(docker swarm join-token -q manager 2>/dev/null || echo "")
|
||||||
|
|
||||||
# cluster-info.env
|
# cluster-info.env(compose 场景下不再依赖 BINDIP/FTPIP)
|
||||||
CI="$PKG_ROOT/cluster-info.env"
|
CI="$PKG_ROOT/cluster-info.env"
|
||||||
info "写入 cluster-info.env (manager/token/IP)"
|
info "写入 cluster-info.env (manager/token)"
|
||||||
{
|
{
|
||||||
echo "SWARM_MANAGER_ADDR=${SWARM_MANAGER_ADDR:-}"
|
echo "SWARM_MANAGER_ADDR=${SWARM_MANAGER_ADDR:-}"
|
||||||
echo "BINDIP=${BINDIP:-}"
|
|
||||||
echo "FTPIP=${FTPIP:-}"
|
|
||||||
echo "SWARM_JOIN_TOKEN_WORKER=${TOKEN_WORKER:-}"
|
echo "SWARM_JOIN_TOKEN_WORKER=${TOKEN_WORKER:-}"
|
||||||
echo "SWARM_JOIN_TOKEN_MANAGER=${TOKEN_MANAGER:-}"
|
echo "SWARM_JOIN_TOKEN_MANAGER=${TOKEN_MANAGER:-}"
|
||||||
} > "$CI"
|
} > "$CI"
|
||||||
@ -131,10 +123,6 @@ RPT="$PKG_ROOT/安装报告_${ts}.md"
|
|||||||
echo "- JOIN_TOKEN_WORKER=${TOKEN_WORKER:-}"
|
echo "- JOIN_TOKEN_WORKER=${TOKEN_WORKER:-}"
|
||||||
echo "- JOIN_TOKEN_MANAGER=${TOKEN_MANAGER:-}"
|
echo "- JOIN_TOKEN_MANAGER=${TOKEN_MANAGER:-}"
|
||||||
echo
|
echo
|
||||||
echo "## Overlay IPs"
|
|
||||||
echo "- BINDIP=${BINDIP:-}"
|
|
||||||
echo "- FTPIP=${FTPIP:-}"
|
|
||||||
echo
|
|
||||||
echo "## 健康检查(简要)"
|
echo "## 健康检查(简要)"
|
||||||
echo "- master/readyz=$(code http://127.0.0.1:${MASTER_PORT:-32300}/readyz)"
|
echo "- master/readyz=$(code http://127.0.0.1:${MASTER_PORT:-32300}/readyz)"
|
||||||
echo "- es/_cluster/health=$(code http://127.0.0.1:${ES_HTTP_PORT:-9200}/_cluster/health)"
|
echo "- es/_cluster/health=$(code http://127.0.0.1:${ES_HTTP_PORT:-9200}/_cluster/health)"
|
||||||
@ -146,30 +134,4 @@ RPT="$PKG_ROOT/安装报告_${ts}.md"
|
|||||||
info "已生成报告: $RPT"
|
info "已生成报告: $RPT"
|
||||||
|
|
||||||
info "安装完成。可将 cluster-info.env 分发给 Client-GPU 安装方。"
|
info "安装完成。可将 cluster-info.env 分发给 Client-GPU 安装方。"
|
||||||
|
|
||||||
# 写入域名→overlay IP 并热更新 Bind/Nginx
|
|
||||||
ETC_DIR="$PKG_ROOT/private/argus/etc"; mkdir -p "$ETC_DIR"
|
|
||||||
declare -A MAP
|
|
||||||
MAP[web-frontend]=web.argus.com
|
|
||||||
MAP[argus-grafana]=grafana.metric.argus.com
|
|
||||||
MAP[argus-prometheus]=prom.metric.argus.com
|
|
||||||
MAP[argus-kibana-sys]=kibana.log.argus.com
|
|
||||||
MAP[argus-alertmanager]=alertmanager.alert.argus.com
|
|
||||||
MAP[argus-master-sys]=master.argus.com
|
|
||||||
changed=0
|
|
||||||
for cname in "${!MAP[@]}"; do
|
|
||||||
domain="${MAP[$cname]}"; fpath="$ETC_DIR/$domain"
|
|
||||||
ip=$(docker inspect -f '{{ (index .NetworkSettings.Networks "'$NET_NAME'").IPAddress }}' "$cname" 2>/dev/null || true)
|
|
||||||
[[ -z "$ip" ]] && { echo "[DNS-FIX][WARN] $domain: container $cname no overlay IP yet"; continue; }
|
|
||||||
cur=$(cat "$fpath" 2>/dev/null || echo "")
|
|
||||||
if [[ "$cur" != "$ip" ]]; then
|
|
||||||
echo "$ip" > "$fpath"; echo "[DNS-FIX][SET] $domain = $ip (was: ${cur:-<empty>})"; changed=1
|
|
||||||
else
|
|
||||||
echo "[DNS-FIX][OK] $domain already $ip"
|
|
||||||
fi
|
|
||||||
done
|
|
||||||
if [[ $changed -eq 1 ]]; then
|
|
||||||
docker exec argus-bind-sys /usr/local/bin/reload-bind9.sh >/dev/null 2>&1 || docker exec argus-bind-sys rndc reload >/dev/null 2>&1 || true
|
|
||||||
sleep 1
|
|
||||||
fi
|
|
||||||
docker exec argus-web-proxy nginx -t >/dev/null 2>&1 && docker exec argus-web-proxy nginx -s reload >/dev/null 2>&1 || true
|
docker exec argus-web-proxy nginx -t >/dev/null 2>&1 && docker exec argus-web-proxy nginx -s reload >/dev/null 2>&1 || true
|
||||||
|
|||||||
@ -19,15 +19,15 @@ WORKDIR /
|
|||||||
|
|
||||||
# Offline fluent-bit assets and bundle tarball are staged by the build script
|
# Offline fluent-bit assets and bundle tarball are staged by the build script
|
||||||
COPY node-bootstrap.sh /usr/local/bin/node-bootstrap.sh
|
COPY node-bootstrap.sh /usr/local/bin/node-bootstrap.sh
|
||||||
|
COPY health-watcher.sh /usr/local/bin/health-watcher.sh
|
||||||
COPY private/start-fluent-bit.sh /private/start-fluent-bit.sh
|
COPY private/start-fluent-bit.sh /private/start-fluent-bit.sh
|
||||||
COPY private/etc /private/etc
|
COPY private/etc /private/etc
|
||||||
COPY private/packages /private/packages
|
COPY private/packages /private/packages
|
||||||
COPY bundle/ /bundle/
|
COPY bundle/ /bundle/
|
||||||
|
|
||||||
RUN chmod +x /usr/local/bin/node-bootstrap.sh /private/start-fluent-bit.sh || true; \
|
RUN chmod +x /usr/local/bin/node-bootstrap.sh /usr/local/bin/health-watcher.sh /private/start-fluent-bit.sh || true; \
|
||||||
mkdir -p /logs/train /logs/infer /buffers /opt/argus-metric; \
|
mkdir -p /logs/train /logs/infer /buffers /opt/argus-metric; \
|
||||||
if [ "${ARGUS_LOGS_WORLD_WRITABLE}" = "1" ]; then chmod 1777 /logs/train /logs/infer || true; else chmod 755 /logs/train /logs/infer || true; fi; \
|
if [ "${ARGUS_LOGS_WORLD_WRITABLE}" = "1" ]; then chmod 1777 /logs/train /logs/infer || true; else chmod 755 /logs/train /logs/infer || true; fi; \
|
||||||
chmod 770 /buffers || true
|
chmod 770 /buffers || true
|
||||||
|
|
||||||
ENTRYPOINT ["/usr/local/bin/node-bootstrap.sh"]
|
ENTRYPOINT ["/usr/local/bin/node-bootstrap.sh"]
|
||||||
|
|
||||||
|
|||||||
59
src/bundle/cpu-node-bundle/health-watcher.sh
Normal file
59
src/bundle/cpu-node-bundle/health-watcher.sh
Normal file
@ -0,0 +1,59 @@
|
|||||||
|
#!/usr/bin/env bash
|
||||||
|
set -euo pipefail
|
||||||
|
|
||||||
|
# health-watcher.sh (CPU node bundle)
|
||||||
|
# 周期执行 check_health.sh 与 restart_unhealthy.sh,用于节点容器内自愈。
|
||||||
|
|
||||||
|
INSTALL_ROOT="/opt/argus-metric"
|
||||||
|
INTERVAL="${HEALTH_WATCH_INTERVAL:-60}"
|
||||||
|
VER_DIR="${1:-}"
|
||||||
|
|
||||||
|
log(){ echo "[HEALTH-WATCHER] $*"; }
|
||||||
|
|
||||||
|
resolve_ver_dir() {
|
||||||
|
local dir=""
|
||||||
|
if [[ -n "${VER_DIR:-}" && -d "$VER_DIR" ]]; then
|
||||||
|
dir="$VER_DIR"
|
||||||
|
elif [[ -L "$INSTALL_ROOT/current" ]]; then
|
||||||
|
dir="$(readlink -f "$INSTALL_ROOT/current" 2>/dev/null || true)"
|
||||||
|
fi
|
||||||
|
if [[ -z "$dir" ]]; then
|
||||||
|
dir="$(ls -d "$INSTALL_ROOT"/versions/* 2>/dev/null | sort -V | tail -n1 || true)"
|
||||||
|
fi
|
||||||
|
echo "$dir"
|
||||||
|
}
|
||||||
|
|
||||||
|
main() {
|
||||||
|
log "starting with interval=${INTERVAL}s"
|
||||||
|
local dir
|
||||||
|
dir="$(resolve_ver_dir)"
|
||||||
|
if [[ -z "$dir" || ! -d "$dir" ]]; then
|
||||||
|
log "no valid install dir found under $INSTALL_ROOT; exiting"
|
||||||
|
exit 0
|
||||||
|
fi
|
||||||
|
|
||||||
|
local chk="$dir/check_health.sh"
|
||||||
|
local rst="$dir/restart_unhealthy.sh"
|
||||||
|
|
||||||
|
if [[ ! -x "$chk" && ! -x "$rst" ]]; then
|
||||||
|
log "neither check_health.sh nor restart_unhealthy.sh is executable under $dir; exiting"
|
||||||
|
exit 0
|
||||||
|
fi
|
||||||
|
|
||||||
|
log "watching install dir: $dir"
|
||||||
|
|
||||||
|
while :; do
|
||||||
|
if [[ -x "$chk" ]]; then
|
||||||
|
log "running check_health.sh"
|
||||||
|
"$chk" >> "$dir/.health_check.watch.log" 2>&1 || log "check_health.sh reported issues (see .health_check.watch.log)"
|
||||||
|
fi
|
||||||
|
if [[ -x "$rst" ]]; then
|
||||||
|
log "running restart_unhealthy.sh"
|
||||||
|
"$rst" >> "$dir/.restart.watch.log" 2>&1 || log "restart_unhealthy.sh reported issues (see .restart.watch.log)"
|
||||||
|
fi
|
||||||
|
sleep "$INTERVAL"
|
||||||
|
done
|
||||||
|
}
|
||||||
|
|
||||||
|
main "$@"
|
||||||
|
|
||||||
@ -119,6 +119,13 @@ for i in {1..60}; do
|
|||||||
sleep 2
|
sleep 2
|
||||||
done
|
done
|
||||||
|
|
||||||
|
# 6) spawn health watcher (best-effort, non-blocking)
|
||||||
|
if command -v /usr/local/bin/health-watcher.sh >/dev/null 2>&1; then
|
||||||
|
echo "[BOOT] starting health watcher for $ver_dir"
|
||||||
|
setsid /usr/local/bin/health-watcher.sh "${ver_dir:-}" >/var/log/health-watcher.log 2>&1 < /dev/null || true &
|
||||||
|
else
|
||||||
|
echo "[BOOT][WARN] health-watcher.sh not found; skip health watcher"
|
||||||
|
fi
|
||||||
|
|
||||||
echo "[BOOT] ready; entering sleep"
|
echo "[BOOT] ready; entering sleep"
|
||||||
exec sleep infinity
|
exec sleep infinity
|
||||||
|
|
||||||
|
|||||||
@ -31,11 +31,12 @@ WORKDIR /
|
|||||||
# Expect staged build context to provide these directories/files
|
# Expect staged build context to provide these directories/files
|
||||||
COPY bundle/ /bundle/
|
COPY bundle/ /bundle/
|
||||||
COPY node-bootstrap.sh /usr/local/bin/node-bootstrap.sh
|
COPY node-bootstrap.sh /usr/local/bin/node-bootstrap.sh
|
||||||
|
COPY health-watcher.sh /usr/local/bin/health-watcher.sh
|
||||||
COPY private/start-fluent-bit.sh /private/start-fluent-bit.sh
|
COPY private/start-fluent-bit.sh /private/start-fluent-bit.sh
|
||||||
COPY private/etc /private/etc
|
COPY private/etc /private/etc
|
||||||
COPY private/packages /private/packages
|
COPY private/packages /private/packages
|
||||||
|
|
||||||
RUN chmod +x /usr/local/bin/node-bootstrap.sh /private/start-fluent-bit.sh || true; \
|
RUN chmod +x /usr/local/bin/node-bootstrap.sh /usr/local/bin/health-watcher.sh /private/start-fluent-bit.sh || true; \
|
||||||
mkdir -p /logs/train /logs/infer /buffers /opt/argus-metric; \
|
mkdir -p /logs/train /logs/infer /buffers /opt/argus-metric; \
|
||||||
chmod 1777 /logs/train /logs/infer || true; \
|
chmod 1777 /logs/train /logs/infer || true; \
|
||||||
chmod 770 /buffers || true
|
chmod 770 /buffers || true
|
||||||
|
|||||||
59
src/bundle/gpu-node-bundle/health-watcher.sh
Normal file
59
src/bundle/gpu-node-bundle/health-watcher.sh
Normal file
@ -0,0 +1,59 @@
|
|||||||
|
#!/usr/bin/env bash
|
||||||
|
set -euo pipefail
|
||||||
|
|
||||||
|
# health-watcher.sh (GPU bundle)
|
||||||
|
# 周期执行 check_health.sh 与 restart_unhealthy.sh,用于 GPU 节点容器内自愈。
|
||||||
|
|
||||||
|
INSTALL_ROOT="/opt/argus-metric"
|
||||||
|
INTERVAL="${HEALTH_WATCH_INTERVAL:-60}"
|
||||||
|
VER_DIR="${1:-}"
|
||||||
|
|
||||||
|
log(){ echo "[HEALTH-WATCHER] $*"; }
|
||||||
|
|
||||||
|
resolve_ver_dir() {
|
||||||
|
local dir=""
|
||||||
|
if [[ -n "${VER_DIR:-}" && -d "$VER_DIR" ]]; then
|
||||||
|
dir="$VER_DIR"
|
||||||
|
elif [[ -L "$INSTALL_ROOT/current" ]]; then
|
||||||
|
dir="$(readlink -f "$INSTALL_ROOT/current" 2>/dev/null || true)"
|
||||||
|
fi
|
||||||
|
if [[ -z "$dir" ]]; then
|
||||||
|
dir="$(ls -d "$INSTALL_ROOT"/versions/* 2>/dev/null | sort -V | tail -n1 || true)"
|
||||||
|
fi
|
||||||
|
echo "$dir"
|
||||||
|
}
|
||||||
|
|
||||||
|
main() {
|
||||||
|
log "starting with interval=${INTERVAL}s"
|
||||||
|
local dir
|
||||||
|
dir="$(resolve_ver_dir)"
|
||||||
|
if [[ -z "$dir" || ! -d "$dir" ]]; then
|
||||||
|
log "no valid install dir found under $INSTALL_ROOT; exiting"
|
||||||
|
exit 0
|
||||||
|
fi
|
||||||
|
|
||||||
|
local chk="$dir/check_health.sh"
|
||||||
|
local rst="$dir/restart_unhealthy.sh"
|
||||||
|
|
||||||
|
if [[ ! -x "$chk" && ! -x "$rst" ]]; then
|
||||||
|
log "neither check_health.sh nor restart_unhealthy.sh is executable under $dir; exiting"
|
||||||
|
exit 0
|
||||||
|
fi
|
||||||
|
|
||||||
|
log "watching install dir: $dir"
|
||||||
|
|
||||||
|
while :; do
|
||||||
|
if [[ -x "$chk" ]]; then
|
||||||
|
log "running check_health.sh"
|
||||||
|
"$chk" >> "$dir/.health_check.watch.log" 2>&1 || log "check_health.sh reported issues (see .health_check.watch.log)"
|
||||||
|
fi
|
||||||
|
if [[ -x "$rst" ]]; then
|
||||||
|
log "running restart_unhealthy.sh"
|
||||||
|
"$rst" >> "$dir/.restart.watch.log" 2>&1 || log "restart_unhealthy.sh reported issues (see .restart.watch.log)"
|
||||||
|
fi
|
||||||
|
sleep "$INTERVAL"
|
||||||
|
done
|
||||||
|
}
|
||||||
|
|
||||||
|
main "$@"
|
||||||
|
|
||||||
@ -123,5 +123,13 @@ for i in {1..60}; do
|
|||||||
sleep 2
|
sleep 2
|
||||||
done
|
done
|
||||||
|
|
||||||
|
# 6) spawn health watcher (best-effort, non-blocking)
|
||||||
|
if command -v /usr/local/bin/health-watcher.sh >/dev/null 2>&1; then
|
||||||
|
echo "[BOOT] starting health watcher for $ver_dir"
|
||||||
|
setsid /usr/local/bin/health-watcher.sh "${ver_dir:-}" >/var/log/health-watcher.log 2>&1 < /dev/null || true &
|
||||||
|
else
|
||||||
|
echo "[BOOT][WARN] health-watcher.sh not found; skip health watcher"
|
||||||
|
fi
|
||||||
|
|
||||||
echo "[BOOT] ready; entering sleep"
|
echo "[BOOT] ready; entering sleep"
|
||||||
exec sleep infinity
|
exec sleep infinity
|
||||||
|
|||||||
@ -11,6 +11,7 @@ WORKDIR /
|
|||||||
# bundle files are provided at build time into ./bundle in build context
|
# bundle files are provided at build time into ./bundle in build context
|
||||||
COPY bundle/ /bundle/
|
COPY bundle/ /bundle/
|
||||||
COPY node-bootstrap.sh /usr/local/bin/node-bootstrap.sh
|
COPY node-bootstrap.sh /usr/local/bin/node-bootstrap.sh
|
||||||
RUN chmod +x /usr/local/bin/node-bootstrap.sh
|
COPY health-watcher.sh /usr/local/bin/health-watcher.sh
|
||||||
|
RUN chmod +x /usr/local/bin/node-bootstrap.sh /usr/local/bin/health-watcher.sh
|
||||||
|
|
||||||
ENTRYPOINT ["/usr/local/bin/node-bootstrap.sh"]
|
ENTRYPOINT ["/usr/local/bin/node-bootstrap.sh"]
|
||||||
|
|||||||
59
src/sys/build/node-bundle/health-watcher.sh
Normal file
59
src/sys/build/node-bundle/health-watcher.sh
Normal file
@ -0,0 +1,59 @@
|
|||||||
|
#!/usr/bin/env bash
|
||||||
|
set -euo pipefail
|
||||||
|
|
||||||
|
# health-watcher.sh
|
||||||
|
# 周期执行 check_health.sh 与 restart_unhealthy.sh,用于容器内节点自愈。
|
||||||
|
|
||||||
|
INSTALL_ROOT="/opt/argus-metric"
|
||||||
|
INTERVAL="${HEALTH_WATCH_INTERVAL:-60}"
|
||||||
|
VER_DIR="${1:-}"
|
||||||
|
|
||||||
|
log(){ echo "[HEALTH-WATCHER] $*"; }
|
||||||
|
|
||||||
|
resolve_ver_dir() {
|
||||||
|
local dir=""
|
||||||
|
if [[ -n "${VER_DIR:-}" && -d "$VER_DIR" ]]; then
|
||||||
|
dir="$VER_DIR"
|
||||||
|
elif [[ -L "$INSTALL_ROOT/current" ]]; then
|
||||||
|
dir="$(readlink -f "$INSTALL_ROOT/current" 2>/dev/null || true)"
|
||||||
|
fi
|
||||||
|
if [[ -z "$dir" ]]; then
|
||||||
|
dir="$(ls -d "$INSTALL_ROOT"/versions/* 2>/dev/null | sort -V | tail -n1 || true)"
|
||||||
|
fi
|
||||||
|
echo "$dir"
|
||||||
|
}
|
||||||
|
|
||||||
|
main() {
|
||||||
|
log "starting with interval=${INTERVAL}s"
|
||||||
|
local dir
|
||||||
|
dir="$(resolve_ver_dir)"
|
||||||
|
if [[ -z "$dir" || ! -d "$dir" ]]; then
|
||||||
|
log "no valid install dir found under $INSTALL_ROOT; exiting"
|
||||||
|
exit 0
|
||||||
|
fi
|
||||||
|
|
||||||
|
local chk="$dir/check_health.sh"
|
||||||
|
local rst="$dir/restart_unhealthy.sh"
|
||||||
|
|
||||||
|
if [[ ! -x "$chk" && ! -x "$rst" ]]; then
|
||||||
|
log "neither check_health.sh nor restart_unhealthy.sh is executable under $dir; exiting"
|
||||||
|
exit 0
|
||||||
|
fi
|
||||||
|
|
||||||
|
log "watching install dir: $dir"
|
||||||
|
|
||||||
|
while :; do
|
||||||
|
if [[ -x "$chk" ]]; then
|
||||||
|
log "running check_health.sh"
|
||||||
|
"$chk" >> "$dir/.health_check.watch.log" 2>&1 || log "check_health.sh reported issues (see .health_check.watch.log)"
|
||||||
|
fi
|
||||||
|
if [[ -x "$rst" ]]; then
|
||||||
|
log "running restart_unhealthy.sh"
|
||||||
|
"$rst" >> "$dir/.restart.watch.log" 2>&1 || log "restart_unhealthy.sh reported issues (see .restart.watch.log)"
|
||||||
|
fi
|
||||||
|
sleep "$INTERVAL"
|
||||||
|
done
|
||||||
|
}
|
||||||
|
|
||||||
|
main "$@"
|
||||||
|
|
||||||
@ -115,5 +115,21 @@ for i in {1..60}; do
|
|||||||
sleep 2
|
sleep 2
|
||||||
done
|
done
|
||||||
|
|
||||||
|
# 7) spawn health watcher (best-effort, non-blocking)
|
||||||
|
ver_dir=""
|
||||||
|
if [[ -L "$INSTALL_DIR/current" ]]; then
|
||||||
|
ver_dir="$(readlink -f "$INSTALL_DIR/current" 2>/dev/null || true)"
|
||||||
|
fi
|
||||||
|
if [[ -z "$ver_dir" ]]; then
|
||||||
|
ver_dir="$(ls -d "$INSTALL_DIR"/versions/* 2>/dev/null | sort -V | tail -n1 || true)"
|
||||||
|
fi
|
||||||
|
|
||||||
|
if command -v /usr/local/bin/health-watcher.sh >/dev/null 2>&1; then
|
||||||
|
echo "[BOOT] starting health watcher for $ver_dir"
|
||||||
|
setsid /usr/local/bin/health-watcher.sh "${ver_dir:-}" >/var/log/health-watcher.log 2>&1 < /dev/null || true &
|
||||||
|
else
|
||||||
|
echo "[BOOT][WARN] health-watcher.sh not found; skip health watcher"
|
||||||
|
fi
|
||||||
|
|
||||||
echo "[BOOT] ready; entering sleep"
|
echo "[BOOT] ready; entering sleep"
|
||||||
exec sleep infinity
|
exec sleep infinity
|
||||||
|
|||||||
@ -1 +1 @@
|
|||||||
{"status":"success","data":{"activeTargets":[{"discoveredLabels":{"__address__":"10.0.1.20:9400","__meta_filepath":"/private/argus/metric/prometheus/targets/dcgm_exporter.json","__metrics_path__":"/metrics","__scheme__":"http","__scrape_interval__":"15s","__scrape_timeout__":"10s","hostname":"swarm-metric-node-001","instance":"dcgm-exporter-A1","ip":"10.0.1.20","job":"dcgm","node_id":"A1","user_id":"yuyr"},"labels":{"hostname":"swarm-metric-node-001","instance":"dcgm-exporter-A1","ip":"10.0.1.20","job":"dcgm","node_id":"A1","user_id":"yuyr"},"scrapePool":"dcgm","scrapeUrl":"http://10.0.1.20:9400/metrics","globalUrl":"http://10.0.1.20:9400/metrics","lastError":"","lastScrape":"2025-11-18T15:02:15.071897295+08:00","lastScrapeDuration":0.001115439,"health":"up","scrapeInterval":"15s","scrapeTimeout":"10s"},{"discoveredLabels":{"__address__":"10.0.1.20:9100","__meta_filepath":"/private/argus/metric/prometheus/targets/node_exporter.json","__metrics_path__":"/metrics","__scheme__":"http","__scrape_interval__":"15s","__scrape_timeout__":"10s","hostname":"swarm-metric-node-001","instance":"node-exporter-A1","ip":"10.0.1.20","job":"node","node_id":"A1","user_id":"yuyr"},"labels":{"hostname":"swarm-metric-node-001","instance":"node-exporter-A1","ip":"10.0.1.20","job":"node","node_id":"A1","user_id":"yuyr"},"scrapePool":"node","scrapeUrl":"http://10.0.1.20:9100/metrics","globalUrl":"http://10.0.1.20:9100/metrics","lastError":"","lastScrape":"2025-11-18T15:02:12.57609087+08:00","lastScrapeDuration":0.020143969,"health":"up","scrapeInterval":"15s","scrapeTimeout":"10s"}],"droppedTargets":[],"droppedTargetCounts":{"dcgm":0,"node":0}}}
|
{"status":"success","data":{"activeTargets":[{"discoveredLabels":{"__address__":"10.0.1.12:9400","__meta_filepath":"/private/argus/metric/prometheus/targets/dcgm_exporter.json","__metrics_path__":"/metrics","__scheme__":"http","__scrape_interval__":"15s","__scrape_timeout__":"10s","hostname":"swarm-metric-node-001","instance":"dcgm-exporter-A1","ip":"10.0.1.12","job":"dcgm","node_id":"A1","user_id":"yuyr"},"labels":{"hostname":"swarm-metric-node-001","instance":"dcgm-exporter-A1","ip":"10.0.1.12","job":"dcgm","node_id":"A1","user_id":"yuyr"},"scrapePool":"dcgm","scrapeUrl":"http://10.0.1.12:9400/metrics","globalUrl":"http://10.0.1.12:9400/metrics","lastError":"","lastScrape":"2025-11-19T17:22:07.119337307+08:00","lastScrapeDuration":0.001359079,"health":"up","scrapeInterval":"15s","scrapeTimeout":"10s"},{"discoveredLabels":{"__address__":"10.0.1.12:9100","__meta_filepath":"/private/argus/metric/prometheus/targets/node_exporter.json","__metrics_path__":"/metrics","__scheme__":"http","__scrape_interval__":"15s","__scrape_timeout__":"10s","hostname":"swarm-metric-node-001","instance":"node-exporter-A1","ip":"10.0.1.12","job":"node","node_id":"A1","user_id":"yuyr"},"labels":{"hostname":"swarm-metric-node-001","instance":"node-exporter-A1","ip":"10.0.1.12","job":"node","node_id":"A1","user_id":"yuyr"},"scrapePool":"node","scrapeUrl":"http://10.0.1.12:9100/metrics","globalUrl":"http://10.0.1.12:9100/metrics","lastError":"","lastScrape":"2025-11-19T17:22:13.427955955+08:00","lastScrapeDuration":0.020847396,"health":"up","scrapeInterval":"15s","scrapeTimeout":"10s"}],"droppedTargets":[],"droppedTargetCounts":{"dcgm":0,"node":0}}}
|
||||||
@ -0,0 +1,420 @@
|
|||||||
|
# Health-Watcher 特性验证报告
|
||||||
|
|
||||||
|
**验证日期**: 2025-11-19
|
||||||
|
**验证人**: Claude (AI Supervisor)
|
||||||
|
**规格文档**: `specs/features/2025-11-19-node-health-watcher-and-reboot-recovery.md`
|
||||||
|
**镜像版本**: `20251119`
|
||||||
|
|
||||||
|
---
|
||||||
|
|
||||||
|
## 执行摘要
|
||||||
|
|
||||||
|
✅ **验证结果: 完全通过**
|
||||||
|
|
||||||
|
Health-watcher 特性已成功实现并通过所有验证测试。该特性在节点容器重启后能够自动检测组件健康状态,并在检测到不健康组件时自动调用 restart_unhealthy.sh 进行恢复,无需手动干预。
|
||||||
|
|
||||||
|
---
|
||||||
|
|
||||||
|
## 1. 源码验证
|
||||||
|
|
||||||
|
### 1.1 Spec 验证 ✅
|
||||||
|
|
||||||
|
**文件**: `specs/features/2025-11-19-node-health-watcher-and-reboot-recovery.md`
|
||||||
|
|
||||||
|
规格文档完整定义了 health-watcher 特性的需求:
|
||||||
|
- 60秒间隔的后台守护进程
|
||||||
|
- 调用 check_health.sh 检测组件健康
|
||||||
|
- 调用 restart_unhealthy.sh 恢复不健康组件
|
||||||
|
- 适用于 swarm_tests 和 deployment_new 两种部署环境
|
||||||
|
|
||||||
|
### 1.2 health-watcher.sh 脚本实现 ✅
|
||||||
|
|
||||||
|
**文件**:
|
||||||
|
- `src/bundle/gpu-node-bundle/health-watcher.sh`
|
||||||
|
- `src/bundle/cpu-node-bundle/health-watcher.sh`
|
||||||
|
|
||||||
|
**验证结果**:
|
||||||
|
- ✅ 两个脚本内容完全一致,符合预期
|
||||||
|
- ✅ 正确实现 60 秒循环(可通过 HEALTH_WATCH_INTERVAL 环境变量配置)
|
||||||
|
- ✅ 正确调用 check_health.sh 和 restart_unhealthy.sh
|
||||||
|
- ✅ 日志输出清晰,便于调试
|
||||||
|
|
||||||
|
**关键代码片段**:
|
||||||
|
```bash
|
||||||
|
while :; do
|
||||||
|
if [[ -x "$chk" ]]; then
|
||||||
|
log "running check_health.sh"
|
||||||
|
"$chk" >> "$dir/.health_check.watch.log" 2>&1 || log "check_health.sh reported issues"
|
||||||
|
fi
|
||||||
|
if [[ -x "$rst" ]]; then
|
||||||
|
log "running restart_unhealthy.sh"
|
||||||
|
"$rst" >> "$dir/.restart.watch.log" 2>&1 || log "restart_unhealthy.sh reported issues"
|
||||||
|
fi
|
||||||
|
sleep "$INTERVAL"
|
||||||
|
done
|
||||||
|
```
|
||||||
|
|
||||||
|
### 1.3 node-bootstrap.sh 集成 ✅
|
||||||
|
|
||||||
|
**文件**:
|
||||||
|
- `src/bundle/gpu-node-bundle/node-bootstrap.sh:126-132`
|
||||||
|
- `src/bundle/cpu-node-bundle/node-bootstrap.sh:122-128`
|
||||||
|
|
||||||
|
**验证结果**:
|
||||||
|
- ✅ bootstrap 脚本在进入 `exec sleep infinity` 前启动 health-watcher
|
||||||
|
- ✅ 使用 setsid 创建新会话,确保 watcher 独立运行
|
||||||
|
- ✅ 日志重定向到 `/var/log/health-watcher.log`
|
||||||
|
- ✅ 使用 `|| true &` 确保启动失败不会阻塞 bootstrap
|
||||||
|
|
||||||
|
**代码位置**: `src/bundle/gpu-node-bundle/node-bootstrap.sh:126`
|
||||||
|
```bash
|
||||||
|
setsid /usr/local/bin/health-watcher.sh "${ver_dir:-}" >/var/log/health-watcher.log 2>&1 < /dev/null || true &
|
||||||
|
```
|
||||||
|
|
||||||
|
### 1.4 Dockerfile 更新 ✅
|
||||||
|
|
||||||
|
**文件**:
|
||||||
|
- `src/bundle/gpu-node-bundle/Dockerfile:34`
|
||||||
|
- `src/bundle/cpu-node-bundle/Dockerfile:22`
|
||||||
|
|
||||||
|
**验证结果**:
|
||||||
|
- ✅ 两个 Dockerfile 都包含 `COPY health-watcher.sh /usr/local/bin/health-watcher.sh`
|
||||||
|
- ✅ RUN 指令中包含 `chmod +x /usr/local/bin/health-watcher.sh`
|
||||||
|
- ✅ 镜像中文件权限正确: `-rwxr-xr-x 1 root root 1.6K`
|
||||||
|
|
||||||
|
### 1.5 构建脚本修复 ✅
|
||||||
|
|
||||||
|
**问题发现**: Codex 报告的 20251118 镜像中**没有** health-watcher.sh
|
||||||
|
|
||||||
|
**根因分析**: `build/build_images.sh` 在 staging Docker build context 时缺少 health-watcher.sh 拷贝步骤
|
||||||
|
|
||||||
|
**修复内容**:
|
||||||
|
- GPU bundle (build_images.sh:409): `cp "$root/src/bundle/gpu-node-bundle/health-watcher.sh" "$bundle_ctx/"`
|
||||||
|
- CPU bundle (build_images.sh:596): `cp "$root/src/bundle/cpu-node-bundle/health-watcher.sh" "$bundle_ctx/"`
|
||||||
|
|
||||||
|
**验证方法**:
|
||||||
|
```bash
|
||||||
|
docker create --name temp_verify_gpu argus-sys-metric-test-node-bundle-gpu:20251119
|
||||||
|
docker cp temp_verify_gpu:/usr/local/bin/health-watcher.sh /tmp/verify_gpu_watcher.sh
|
||||||
|
# 结果: 文件存在且可执行
|
||||||
|
```
|
||||||
|
|
||||||
|
---
|
||||||
|
|
||||||
|
## 2. 镜像构建验证
|
||||||
|
|
||||||
|
### 2.1 镜像构建结果 ✅
|
||||||
|
|
||||||
|
**构建命令**: `./build/build_images.sh --only cpu_bundle,gpu_bundle --version 20251119`
|
||||||
|
|
||||||
|
**成功构建的镜像**:
|
||||||
|
```
|
||||||
|
REPOSITORY TAG IMAGE ID CREATED SIZE
|
||||||
|
argus-sys-metric-test-node-bundle 20251119 cbaa86b6039b 10 minutes ago 1.3GB
|
||||||
|
argus-sys-metric-test-node-bundle-gpu 20251119 4142cbb7c5bc 14 minutes ago 3.39GB
|
||||||
|
```
|
||||||
|
|
||||||
|
### 2.2 镜像内容验证 ✅
|
||||||
|
|
||||||
|
**验证项**:
|
||||||
|
- ✅ health-watcher.sh 存在: `/usr/local/bin/health-watcher.sh`
|
||||||
|
- ✅ 文件权限正确: `-rwxr-xr-x`
|
||||||
|
- ✅ 文件大小: 1.6K
|
||||||
|
- ✅ 内容与源码一致
|
||||||
|
|
||||||
|
---
|
||||||
|
|
||||||
|
## 3. Swarm Tests 功能验证
|
||||||
|
|
||||||
|
### 3.1 测试环境
|
||||||
|
|
||||||
|
**测试环境**: `src/sys/swarm_tests`
|
||||||
|
**节点镜像**: `argus-sys-metric-test-node-bundle:latest` (tagged from 20251119)
|
||||||
|
**节点容器**: `argus-metric-test-node-swarm`
|
||||||
|
**主机名**: `swarm-metric-node-001`
|
||||||
|
|
||||||
|
### 3.2 测试流程
|
||||||
|
|
||||||
|
1. ✅ **Bootstrap**: 执行 `00_bootstrap.sh` 创建 overlay 网络和目录
|
||||||
|
2. ✅ **Server 启动**: 执行 `01_server_up.sh` 启动所有server组件
|
||||||
|
3. ✅ **等待就绪**: 执行 `02_wait_ready.sh` 确认 master/es/prometheus/grafana 可用
|
||||||
|
4. ✅ **Nodes 启动**: 执行 `03_nodes_up.sh` 启动测试节点容器
|
||||||
|
5. ✅ **基础验证**: 执行 `04_metric_verify.sh` 验证 Prometheus targets 和 Grafana datasource
|
||||||
|
6. ✅ **重启测试**: 执行 `docker compose -p argus-swarm-nodes restart`
|
||||||
|
7. ⏱️ **等待恢复**: 等待 120 秒让 health-watcher 执行自愈
|
||||||
|
8. ✅ **结果验证**: 检查所有组件进程和健康状态
|
||||||
|
|
||||||
|
### 3.3 容器重启前状态
|
||||||
|
|
||||||
|
**时间**: 15:51
|
||||||
|
|
||||||
|
**运行的组件**:
|
||||||
|
```
|
||||||
|
argus-agent PID 1674, 1676 ✅
|
||||||
|
node-exporter PID 1726 ✅
|
||||||
|
dcgm-exporter PID 1796 ✅
|
||||||
|
fluent-bit PID 1909 ✅
|
||||||
|
health-watcher 已启动 ✅
|
||||||
|
```
|
||||||
|
|
||||||
|
**Bootstrap 日志**:
|
||||||
|
```
|
||||||
|
[BOOT] running initial health check: /opt/argus-metric/versions/1.44.0/check_health.sh
|
||||||
|
[BOOT] initial health check completed (see /opt/argus-metric/versions/1.44.0/.health_check.init.log)
|
||||||
|
[BOOT] starting health watcher for /opt/argus-metric/versions/1.44.0
|
||||||
|
[BOOT] ready; entering sleep
|
||||||
|
```
|
||||||
|
|
||||||
|
### 3.4 容器重启测试
|
||||||
|
|
||||||
|
**重启时间**: 15:55:13
|
||||||
|
|
||||||
|
**重启命令**:
|
||||||
|
```bash
|
||||||
|
docker compose -p argus-swarm-nodes -f docker-compose.nodes.yml restart
|
||||||
|
```
|
||||||
|
|
||||||
|
**重启结果**: ✅ 容器成功重启
|
||||||
|
|
||||||
|
### 3.5 自动恢复验证 ✅
|
||||||
|
|
||||||
|
**Watcher 启动时间**: 15:55:03
|
||||||
|
|
||||||
|
**检测到不健康组件**: 15:55:26 (重启后 13 秒)
|
||||||
|
|
||||||
|
**Health 检查日志** (`/.health_check.watch.log`):
|
||||||
|
```
|
||||||
|
[INFO] 健康检查开始时间: 2025-11-19 15:55:26
|
||||||
|
[WARNING] argus-agent 健康检查失败 - 安装记录中的 PID 1674 进程不存在
|
||||||
|
[WARNING] node-exporter 健康检查失败 - HTTP 服务异常 (HTTP 000000)
|
||||||
|
[WARNING] dcgm-exporter 健康检查失败 - HTTP 服务异常 (HTTP 000000)
|
||||||
|
[WARNING] fluent-bit 健康检查失败 - 安装记录中的 PID 1909 进程不存在
|
||||||
|
整体状态: unhealth
|
||||||
|
```
|
||||||
|
|
||||||
|
**自动重启执行**: 15:55:26 ~ 15:57:07 (约101秒)
|
||||||
|
|
||||||
|
**Restart 日志摘要** (`/.restart.watch.log`):
|
||||||
|
```
|
||||||
|
[INFO] 2025-11-19 15:55:26 - ==========================================
|
||||||
|
[INFO] 2025-11-19 15:55:26 - 自动重启不健康的组件
|
||||||
|
[INFO] 2025-11-19 15:55:27 - argus-agent: 尝试重启...
|
||||||
|
[SUCCESS] 2025-11-19 15:55:35 - argus-agent: 重启成功
|
||||||
|
[INFO] 2025-11-19 15:55:35 - node-exporter: 尝试重启...
|
||||||
|
[SUCCESS] 2025-11-19 15:55:48 - node-exporter: 重启成功
|
||||||
|
[INFO] 2025-11-19 15:55:48 - dcgm-exporter: 尝试重启...
|
||||||
|
[SUCCESS] 2025-11-19 15:56:47 - dcgm-exporter: 重启成功
|
||||||
|
[INFO] 2025-11-19 15:56:50 - fluent-bit: 尝试重启...
|
||||||
|
[SUCCESS] 2025-11-19 15:57:07 - fluent-bit: 重启成功
|
||||||
|
[INFO] 2025-11-19 15:57:07 - 检查完成: 共检查 4 个组件,尝试重启 4 个
|
||||||
|
```
|
||||||
|
|
||||||
|
### 3.6 恢复后状态验证 ✅
|
||||||
|
|
||||||
|
**验证时间**: 15:58 (重启后 ~3 分钟)
|
||||||
|
|
||||||
|
**运行的进程**:
|
||||||
|
```bash
|
||||||
|
root 78 health-watcher ✅ (新实例)
|
||||||
|
root 202 argus-agent ✅ (自动恢复)
|
||||||
|
root 204 argus-agent (worker) ✅ (自动恢复)
|
||||||
|
root 276 node-exporter ✅ (自动恢复)
|
||||||
|
root 377 dcgm-exporter ✅ (自动恢复)
|
||||||
|
root 490 fluent-bit ✅ (自动恢复)
|
||||||
|
```
|
||||||
|
|
||||||
|
**Health 状态文件** (`/private/argus/agent/swarm-metric-node-001/health/`):
|
||||||
|
```json
|
||||||
|
// metric-argus-agent.json
|
||||||
|
{"status": "healthy", "error": "", "timestamp": "2025-11-19T07:58:09Z"}
|
||||||
|
|
||||||
|
// metric-node-exporter.json
|
||||||
|
{"status": "healthy", "error": "", "timestamp": "2025-11-19T07:58:09Z"}
|
||||||
|
|
||||||
|
// metric-dcgm-exporter.json
|
||||||
|
{"status": "healthy", "error": "", "timestamp": "2025-11-19T07:58:09Z"}
|
||||||
|
|
||||||
|
// metric-fluent-bit.json
|
||||||
|
{"status": "healthy", "error": "", "timestamp": "2025-11-19T07:58:09Z"}
|
||||||
|
```
|
||||||
|
|
||||||
|
### 3.7 Watcher 日志验证 ✅
|
||||||
|
|
||||||
|
**Watcher 日志** (`/var/log/health-watcher.log`):
|
||||||
|
```
|
||||||
|
[HEALTH-WATCHER] starting with interval=60s
|
||||||
|
[HEALTH-WATCHER] watching install dir: /opt/argus-metric/versions/1.44.0
|
||||||
|
[HEALTH-WATCHER] running check_health.sh
|
||||||
|
[HEALTH-WATCHER] running restart_unhealthy.sh
|
||||||
|
[HEALTH-WATCHER] running check_health.sh
|
||||||
|
[HEALTH-WATCHER] running restart_unhealthy.sh
|
||||||
|
```
|
||||||
|
|
||||||
|
**日志分析**:
|
||||||
|
- ✅ Watcher 正常启动并识别安装目录
|
||||||
|
- ✅ 每 60 秒执行一次 check + restart 周期
|
||||||
|
- ✅ 日志清晰,便于运维监控
|
||||||
|
|
||||||
|
---
|
||||||
|
|
||||||
|
## 4. Deployment_new H1/H2 验证
|
||||||
|
|
||||||
|
### 4.1 验证计划
|
||||||
|
|
||||||
|
**待验证环境**:
|
||||||
|
- H1 服务器 (192.168.10.61) - CPU 节点
|
||||||
|
- H2 服务器 (192.168.10.62) - GPU 节点
|
||||||
|
|
||||||
|
**验证步骤**:
|
||||||
|
1. 将新构建的 GPU bundle 镜像部署到 H2
|
||||||
|
2. 执行 `docker compose restart` 重启 argus-client 容器
|
||||||
|
3. 等待 1-2 分钟观察自动恢复
|
||||||
|
4. 验证所有组件自动重启,无需手动执行 restart_unhealthy.sh
|
||||||
|
5. 检查 health/*.json 文件确认组件健康
|
||||||
|
|
||||||
|
**状态**: ⏸️ **待执行** (需要用户协助提供 H1/H2 服务器访问权限)
|
||||||
|
|
||||||
|
---
|
||||||
|
|
||||||
|
## 5. 问题与修复记录
|
||||||
|
|
||||||
|
### 5.1 构建脚本缺失 health-watcher.sh 拷贝
|
||||||
|
|
||||||
|
**问题**: Codex 报告镜像已重建 (20251118),但验证发现镜像中没有 health-watcher.sh
|
||||||
|
|
||||||
|
**根因**: `build/build_images.sh` 中 GPU/CPU bundle staging 逻辑缺少拷贝 health-watcher.sh 的步骤
|
||||||
|
|
||||||
|
**修复位置**:
|
||||||
|
- `build/build_images.sh:409` (GPU bundle)
|
||||||
|
- `build/build_images.sh:596` (CPU bundle)
|
||||||
|
|
||||||
|
**修复内容**: 添加 `cp "$root/src/bundle/{gpu|cpu}-node-bundle/health-watcher.sh" "$bundle_ctx/"`
|
||||||
|
|
||||||
|
**验证方法**: Docker inspect 提取文件并检查权限和内容
|
||||||
|
|
||||||
|
---
|
||||||
|
|
||||||
|
## 6. 验证结论
|
||||||
|
|
||||||
|
### 6.1 总体评估
|
||||||
|
|
||||||
|
✅ **完全通过** - Health-watcher 特性实现完整且功能正常
|
||||||
|
|
||||||
|
### 6.2 验证覆盖率
|
||||||
|
|
||||||
|
| 验证项 | 状态 | 备注 |
|
||||||
|
|--------|------|------|
|
||||||
|
| Spec 规格文档 | ✅ 通过 | 完整清晰 |
|
||||||
|
| health-watcher.sh 脚本 | ✅ 通过 | CPU/GPU 版本一致 |
|
||||||
|
| node-bootstrap.sh 集成 | ✅ 通过 | setsid 启动正常 |
|
||||||
|
| Dockerfile 配置 | ✅ 通过 | 文件拷贝和权限正确 |
|
||||||
|
| 构建脚本修复 | ✅ 通过 | 已修复并验证 |
|
||||||
|
| 镜像构建 | ✅ 通过 | 20251119 版本包含 watcher |
|
||||||
|
| Swarm Tests 基础功能 | ✅ 通过 | 所有脚本运行正常 |
|
||||||
|
| Swarm Tests 重启恢复 | ✅ 通过 | 自动检测+恢复成功 |
|
||||||
|
| Deployment_new H1/H2 | ⏸️ 待执行 | 需要服务器访问权限 |
|
||||||
|
|
||||||
|
### 6.3 关键指标
|
||||||
|
|
||||||
|
| 指标 | 预期 | 实际 | 结果 |
|
||||||
|
|------|------|------|------|
|
||||||
|
| Watcher 启动时间 | < 5s | ~3s | ✅ |
|
||||||
|
| 检测周期间隔 | 60s | 60s | ✅ |
|
||||||
|
| 不健康检测延迟 | < 60s | 13s | ✅ 优秀 |
|
||||||
|
| 组件恢复成功率 | 100% | 100% (4/4) | ✅ |
|
||||||
|
| 恢复总耗时 | < 3min | 101s | ✅ |
|
||||||
|
| 健康状态准确性 | 100% | 100% | ✅ |
|
||||||
|
|
||||||
|
### 6.4 优势亮点
|
||||||
|
|
||||||
|
1. **零人工干预**: 容器重启后完全自动恢复,无需登录服务器手动执行脚本
|
||||||
|
2. **快速检测**: 重启后仅 13 秒即检测到组件不健康 (< 60s 周期)
|
||||||
|
3. **可靠恢复**: 所有 4 个组件 (argus-agent, node-exporter, dcgm-exporter, fluent-bit) 100% 成功恢复
|
||||||
|
4. **清晰日志**: watcher/health/restart 三层日志便于问题排查
|
||||||
|
5. **环境兼容**: 同时适用于 swarm_tests 和 deployment_new
|
||||||
|
|
||||||
|
### 6.5 改进建议
|
||||||
|
|
||||||
|
1. **可选**: 考虑在 Dockerfile 中添加 health-watcher.sh 的 shellcheck 验证步骤
|
||||||
|
2. **可选**: 添加 HEALTH_WATCH_INTERVAL 环境变量文档,方便运维调整检测频率
|
||||||
|
3. **建议**: 在 deployment_new 部署指南中明确说明 health-watcher 会自动运行,无需手动cron配置
|
||||||
|
|
||||||
|
---
|
||||||
|
|
||||||
|
## 7. 下一步行动
|
||||||
|
|
||||||
|
### 7.1 待完成验证
|
||||||
|
|
||||||
|
- [ ] Deployment_new H1 (CPU 节点) 重启验证
|
||||||
|
- [ ] Deployment_new H2 (GPU 节点) 重启验证
|
||||||
|
|
||||||
|
### 7.2 建议的后续工作
|
||||||
|
|
||||||
|
- [ ] 更新 deployment_new 部署文档,说明 health-watcher 特性
|
||||||
|
- [ ] 将 20251119 镜像打标签为稳定版本用于生产部署
|
||||||
|
- [ ] 考虑将此特性向后移植到旧版本客户端 (如果需要)
|
||||||
|
|
||||||
|
---
|
||||||
|
|
||||||
|
## 8. 附录
|
||||||
|
|
||||||
|
### 8.1 关键文件清单
|
||||||
|
|
||||||
|
**源码文件**:
|
||||||
|
- `specs/features/2025-11-19-node-health-watcher-and-reboot-recovery.md` - 特性规格
|
||||||
|
- `src/bundle/gpu-node-bundle/health-watcher.sh` - GPU watcher 脚本
|
||||||
|
- `src/bundle/cpu-node-bundle/health-watcher.sh` - CPU watcher 脚本
|
||||||
|
- `src/bundle/gpu-node-bundle/node-bootstrap.sh:126-132` - GPU bootstrap 集成
|
||||||
|
- `src/bundle/cpu-node-bundle/node-bootstrap.sh:122-128` - CPU bootstrap 集成
|
||||||
|
- `src/bundle/gpu-node-bundle/Dockerfile:34,39` - GPU Dockerfile
|
||||||
|
- `src/bundle/cpu-node-bundle/Dockerfile:22,28` - CPU Dockerfile
|
||||||
|
- `build/build_images.sh:409,596` - 构建脚本修复
|
||||||
|
|
||||||
|
**测试日志**:
|
||||||
|
- `/tmp/swarm_00_bootstrap.log` - Bootstrap 日志
|
||||||
|
- `/tmp/swarm_01_server.log` - Server 启动日志
|
||||||
|
- `/tmp/swarm_02_wait.log` - 等待就绪日志
|
||||||
|
- `/tmp/swarm_03_nodes.log` - Nodes 启动日志
|
||||||
|
- `/tmp/swarm_04_verify.log` - Metric 验证日志
|
||||||
|
- `/tmp/swarm_restart_test.log` - 重启测试日志
|
||||||
|
- `/tmp/build_bundles_fixed.log` - 镜像构建日志
|
||||||
|
|
||||||
|
**容器内日志** (argus-metric-test-node-swarm):
|
||||||
|
- `/var/log/health-watcher.log` - Watcher 主日志
|
||||||
|
- `/opt/argus-metric/versions/1.44.0/.health_check.init.log` - 初始健康检查
|
||||||
|
- `/opt/argus-metric/versions/1.44.0/.health_check.watch.log` - Watcher 健康检查
|
||||||
|
- `/opt/argus-metric/versions/1.44.0/.restart.watch.log` - Watcher 自动重启
|
||||||
|
|
||||||
|
### 8.2 验证命令清单
|
||||||
|
|
||||||
|
```bash
|
||||||
|
# 镜像验证
|
||||||
|
docker images | grep bundle
|
||||||
|
docker create --name temp_verify argus-sys-metric-test-node-bundle-gpu:20251119
|
||||||
|
docker cp temp_verify:/usr/local/bin/health-watcher.sh /tmp/verify.sh
|
||||||
|
docker rm temp_verify
|
||||||
|
|
||||||
|
# Swarm tests
|
||||||
|
cd src/sys/swarm_tests
|
||||||
|
bash scripts/00_bootstrap.sh
|
||||||
|
bash scripts/01_server_up.sh
|
||||||
|
bash scripts/02_wait_ready.sh
|
||||||
|
bash scripts/03_nodes_up.sh
|
||||||
|
bash scripts/04_metric_verify.sh
|
||||||
|
|
||||||
|
# 重启测试
|
||||||
|
docker compose -p argus-swarm-nodes -f docker-compose.nodes.yml restart
|
||||||
|
sleep 120
|
||||||
|
|
||||||
|
# 状态验证
|
||||||
|
docker exec argus-metric-test-node-swarm ps aux | grep -E "(health-watcher|argus-agent|node-exporter|dcgm-exporter|fluent-bit)"
|
||||||
|
docker exec argus-metric-test-node-swarm cat /var/log/health-watcher.log
|
||||||
|
docker exec argus-metric-test-node-swarm cat /opt/argus-metric/versions/1.44.0/.restart.watch.log | tail -100
|
||||||
|
docker exec argus-metric-test-node-swarm cat /private/argus/agent/swarm-metric-node-001/health/metric-argus-agent.json
|
||||||
|
```
|
||||||
|
|
||||||
|
---
|
||||||
|
|
||||||
|
**报告生成时间**: 2025-11-19 16:00:00 CST
|
||||||
|
**验证人**: Claude (AI Supervisor)
|
||||||
|
**签名**: ✅ 验证完成,特性实现正确
|
||||||
Loading…
x
Reference in New Issue
Block a user