import { useTranslation } from "next-i18next"; import Container from "components/services/widget/container"; import Block from "components/services/widget/block"; import useWidgetAPI from "utils/proxy/use-widget-api"; function calcRunning(total, current) { return current.status === "running" ? total + 1 : total; } export default function Component({ service }) { const { t } = useTranslation(); const { widget } = service; const { data: clusterData, error: clusterError } = useWidgetAPI(widget, "cluster/resources"); if (clusterError) { return ; } if (!clusterData || !clusterData.data) { return ( ); } const { data } = clusterData; const vms = data.filter( (item) => item.type === "qemu" && item.template === 0 && (widget.node === undefined || widget.node === item.node), ) || []; const lxc = data.filter( (item) => item.type === "lxc" && item.template === 0 && (widget.node === undefined || widget.node === item.node), ) || []; const nodes = data.filter( (item) => item.type === "node" && item.status === "online" && (widget.node === undefined || widget.node === item.node), ) || []; const runningVMs = vms.reduce(calcRunning, 0); const runningLXC = lxc.reduce(calcRunning, 0); if (nodes.length === 0) { return ( ); } const maxMemory = nodes.reduce((sum, n) => n.maxmem + sum, 0); const usedMemory = nodes.reduce((sum, n) => n.mem + sum, 0); const maxCpu = nodes.reduce((sum, n) => n.maxcpu + sum, 0); const usedCpu = nodes.reduce((sum, n) => n.cpu * n.maxcpu + sum, 0); return ( ); }