Compare commits
6 Commits
4cc5a1f207
...
master
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
bd955e8067
|
||
|
|
b1e9b1e0b5
|
||
|
|
4ac1ba1f24
|
||
|
|
b57db23a35
|
||
|
|
ebcc08cc8f
|
||
|
|
1d06e1c180
|
12
flake.lock
generated
12
flake.lock
generated
@@ -3,11 +3,11 @@
|
||||
"debBundler": {
|
||||
"flake": false,
|
||||
"locked": {
|
||||
"lastModified": 1725149456,
|
||||
"narHash": "sha256-rRrSD7itoPm+VIT4bIzSupQ7jw+H4eOjxRiRA89Kxb4=",
|
||||
"lastModified": 1746317543,
|
||||
"narHash": "sha256-1Xph5g1Lazzkc9XuY1nOkG5Fn7+lmSdldAC91boDawY=",
|
||||
"owner": "illustris",
|
||||
"repo": "flake",
|
||||
"rev": "257a6c986cb9a67c4d6d0e0363507cab7f958b63",
|
||||
"rev": "e86bd104d76d22b2ba36fede405e7bff290ef489",
|
||||
"type": "github"
|
||||
},
|
||||
"original": {
|
||||
@@ -18,11 +18,11 @@
|
||||
},
|
||||
"nixpkgs": {
|
||||
"locked": {
|
||||
"lastModified": 1725103162,
|
||||
"narHash": "sha256-Ym04C5+qovuQDYL/rKWSR+WESseQBbNAe5DsXNx5trY=",
|
||||
"lastModified": 1746663147,
|
||||
"narHash": "sha256-Ua0drDHawlzNqJnclTJGf87dBmaO/tn7iZ+TCkTRpRc=",
|
||||
"owner": "nixos",
|
||||
"repo": "nixpkgs",
|
||||
"rev": "12228ff1752d7b7624a54e9c1af4b222b3c1073b",
|
||||
"rev": "dda3dcd3fe03e991015e9a74b22d35950f264a54",
|
||||
"type": "github"
|
||||
},
|
||||
"original": {
|
||||
|
||||
@@ -14,7 +14,7 @@ rec {
|
||||
packages.x86_64-linux = with nixpkgs.legacyPackages.x86_64-linux; rec {
|
||||
pvemon = python3Packages.buildPythonApplication {
|
||||
pname = "pvemon";
|
||||
version = "1.3.1";
|
||||
version = "1.3.3";
|
||||
src = ./src;
|
||||
propagatedBuildInputs = with python3Packages; [
|
||||
pexpect
|
||||
|
||||
@@ -34,6 +34,7 @@ pool_cache = {
|
||||
DEFAULT_PORT = 9116
|
||||
DEFAULT_INTERVAL = 10
|
||||
DEFAULT_PREFIX = "pve"
|
||||
DEFAULT_HOST = "0.0.0.0"
|
||||
|
||||
gauge_settings = [
|
||||
('kvm_cpu', 'CPU time for VM', ['id', 'mode']),
|
||||
@@ -353,6 +354,7 @@ class PVECollector(object):
|
||||
def main():
|
||||
parser = argparse.ArgumentParser(description='PVE metrics exporter for Prometheus')
|
||||
parser.add_argument('--port', type=int, default=DEFAULT_PORT, help='Port for the exporter to listen on')
|
||||
parser.add_argument('--host', type=str, default=DEFAULT_HOST, help='Host address to bind the exporter to')
|
||||
parser.add_argument('--interval', type=int, default=DEFAULT_INTERVAL, help='THIS OPTION DOES NOTHING')
|
||||
parser.add_argument('--collect-running-vms', type=str, default='true', help='Enable or disable collecting running VMs metric (true/false)')
|
||||
parser.add_argument('--collect-storage', type=str, default='true', help='Enable or disable collecting storage info (true/false)')
|
||||
@@ -388,7 +390,7 @@ def main():
|
||||
return
|
||||
else:
|
||||
REGISTRY.register(PVECollector())
|
||||
start_http_server(cli_args.port)
|
||||
start_http_server(cli_args.port, addr=cli_args.host)
|
||||
|
||||
while True:
|
||||
time.sleep(100)
|
||||
|
||||
@@ -87,15 +87,53 @@ def parse_storage_cfg(file_path='/etc/pve/storage.cfg'):
|
||||
|
||||
def get_storage_size(storage):
|
||||
try:
|
||||
if storage["type"] in ["dir", "nfs", "cephfs", "zfspool"]:
|
||||
if storage["type"] == "zfspool":
|
||||
path = storage["mountpoint"]
|
||||
else:
|
||||
if "pool" not in storage:
|
||||
logging.debug(f"ZFS pool {storage['name']} has no pool name configured")
|
||||
return None
|
||||
|
||||
# Extract the pool name (could be in format like rpool/data)
|
||||
pool_name = storage["pool"].split("/")[0]
|
||||
|
||||
# Use zpool command to get accurate size information
|
||||
import subprocess
|
||||
try:
|
||||
result = subprocess.run(
|
||||
["zpool", "list", pool_name, "-p"],
|
||||
capture_output=True,
|
||||
text=True,
|
||||
check=True
|
||||
)
|
||||
|
||||
# Parse the output
|
||||
lines = result.stdout.strip().split("\n")
|
||||
if len(lines) < 2:
|
||||
logging.warn(f"Unexpected zpool list output format for {pool_name}")
|
||||
return None
|
||||
|
||||
# Extract values from the second line (the data line)
|
||||
values = lines[1].split()
|
||||
if len(values) < 4:
|
||||
logging.warn(f"Insufficient data in zpool list output for {pool_name}")
|
||||
return None
|
||||
|
||||
# Values are: NAME SIZE ALLOC FREE ...
|
||||
# We need the SIZE and FREE values (index 1 and 3)
|
||||
total_size = int(values[1])
|
||||
free_space = int(values[3])
|
||||
|
||||
return {
|
||||
"total": total_size,
|
||||
"free": free_space
|
||||
}
|
||||
except (subprocess.SubprocessError, ValueError, IndexError) as e:
|
||||
logging.warn(f"Error running zpool list for {pool_name}: {e}")
|
||||
return None
|
||||
|
||||
elif storage["type"] in ["dir", "nfs", "cephfs"]:
|
||||
# For non-ZFS storage, use statvfs
|
||||
path = storage["path"]
|
||||
# Get filesystem statistics
|
||||
stats = os.statvfs(path)
|
||||
# Calculate total size and free space in bytes
|
||||
# TODO: find an alternative way to calculate total_size for ZFS
|
||||
total_size = stats.f_frsize * stats.f_blocks
|
||||
free_space = stats.f_frsize * stats.f_bavail
|
||||
return {
|
||||
@@ -126,7 +164,12 @@ def collect_storage_metrics():
|
||||
|
||||
storage_pools = parse_storage_cfg()
|
||||
for storage in storage_pools:
|
||||
info_dict["node_storage"].add_metric([], storage)
|
||||
# Convert any non-string values to strings for InfoMetricFamily
|
||||
storage_info = {}
|
||||
for key, value in storage.items():
|
||||
storage_info[key] = str(value) if not isinstance(value, str) else value
|
||||
|
||||
info_dict["node_storage"].add_metric([], storage_info)
|
||||
size = get_storage_size(storage)
|
||||
if size != None:
|
||||
gauge_dict["node_storage_size"].add_metric([storage["name"], storage["type"]], size["total"])
|
||||
|
||||
@@ -2,7 +2,7 @@ from setuptools import setup, find_packages
|
||||
|
||||
setup(
|
||||
name='pvemon',
|
||||
version = "1.3.1",
|
||||
version = "1.3.3",
|
||||
packages=find_packages(),
|
||||
entry_points={
|
||||
'console_scripts': [
|
||||
|
||||
Reference in New Issue
Block a user