This is an exporter that exposes information gathered from Proxmox VE node for use by the Prometheus monitoring system.
Requires Python 3.9 or better.
python3 -m pip install prometheus-pve-exporter
pve_exporter --help
docker pull prompve/prometheus-pve-exporter
Example: Display usage message:
docker run -it --rm prompve/prometheus-pve-exporter --help
Example: Run the image with a mounted configuration file and published port:
docker run --init --name prometheus-pve-exporter -d -p 127.0.0.1:9221:9221 -v /path/to/pve.yml:/etc/pve.yml prompve/prometheus-pve-exporter
Prometheus PVE Exporter will now be reachable at http://localhost:9221/.
usage: pve_exporter [-h] [--collector.status | --no-collector.status] [--collector.version | --no-collector.version] [--collector.node | --no-collector.node] [--collector.cluster | --no-collector.cluster] [--collector.resources | --no-collector.resources] [--collector.config | --no-collector.config] [--server.keyfile SERVER_KEYFILE] [--server.certfile SERVER_CERTFILE] [config] [port] [address] positional arguments: config Path to configuration file (pve.yml) port Port on which the exporter is listening (9221) address Address to which the exporter will bind options: -h, --help show this help message and exit --server.keyfile SERVER_KEYFILE SSL key for server --server.certfile SERVER_CERTFILE SSL certificate for server cluster collectors: cluster collectors are run if the url parameter cluster=1 is set and skipped if the url parameter cluster=0 is set on a scrape url. --collector.status, --no-collector.status Exposes Node/VM/CT-Status --collector.version, --no-collector.version Exposes PVE version info --collector.node, --no-collector.node Exposes PVE node info --collector.cluster, --no-collector.cluster Exposes PVE cluster info --collector.resources, --no-collector.resources Exposes PVE resources info node collectors: node collectors are run if the url parameter node=1 is set and skipped if the url parameter node=0 is set on a scrape url. --collector.config, --no-collector.config Exposes PVE onboot status
Use [::] for the address argument in order to bind to both IPv6 and IPv4 sockets on dual stacked machines.
Visit http://localhost:9221/pve?target=1.2.3.4&cluster=1&node=1 where 1.2.3.4
is the IP of the Proxmox VE node to get metrics from. Specify the module
request parameter, to choose which module to use from the config file.
The target
request parameter defaults to localhost
. Hence if
pve_exporter
is deployed directly on the proxmox host, target
can be omitted.
Use the --collector.X / --no-collector.X flags to enable disable selected collectors.
Note that that the config collector results in one API call per guest VM/CT. It is therefore recommended to disable this collector using the --no-collector.config flag on big deployments.
See the wiki for more examples and docs.
Here's an example of the metrics exported.
# HELP pve_up Node/VM/CT-Status is online/running # TYPE pve_up gauge pve_up{id="node/proxmox"} 1.0 pve_up{id="qemu/100"} 1.0 # HELP pve_disk_size_bytes Size of storage device # TYPE pve_disk_size_bytes gauge pve_disk_size_bytes{id="qemu/100"} 6.8719476736e+010 pve_disk_size_bytes{id="node/proxmox"} 3.1044079616e+010 pve_disk_size_bytes{id="storage/proxmox/local"} 3.1044079616e+010 pve_disk_size_bytes{id="storage/proxmox/local-lvm"} 6.9243764736e+010 pve_disk_size_bytes{id="storage/proxmox/vms"} 1.934882766848e+012 # HELP pve_disk_usage_bytes Disk usage in bytes # TYPE pve_disk_usage_bytes gauge pve_disk_usage_bytes{id="qemu/100"} 0.0 pve_disk_usage_bytes{id="node/proxmox"} 1.7571426304e+010 pve_disk_usage_bytes{id="storage/proxmox/local"} 1.7571426304e+010 pve_disk_usage_bytes{id="storage/proxmox/local-lvm"} 6.619703908e+09 pve_disk_usage_bytes{id="storage/proxmox/vms"} 8.32870981632e+011 # HELP pve_memory_size_bytes Size of memory # TYPE pve_memory_size_bytes gauge pve_memory_size_bytes{id="qemu/100"} 1.7179869184e+010 pve_memory_size_bytes{id="node/proxmox"} 6.739961856e+010 # HELP pve_memory_usage_bytes Memory usage in bytes # TYPE pve_memory_usage_bytes gauge pve_memory_usage_bytes{id="qemu/100"} 1.6573280275e+010 pve_memory_usage_bytes{id="node/proxmox"} 5.3907812352e+010 # HELP pve_network_transmit_bytes Number of bytes transmitted over the network # TYPE pve_network_transmit_bytes gauge pve_network_transmit_bytes{id="qemu/100"} 7.75070828e+09 # HELP pve_network_receive_bytes Number of bytes received over the network # TYPE pve_network_receive_bytes gauge pve_network_receive_bytes{id="qemu/100"} 1.529756162e+09 # HELP pve_disk_write_bytes Number of bytes written to storage # TYPE pve_disk_write_bytes gauge pve_disk_write_bytes{id="qemu/100"} 1.50048127488e+011 # HELP pve_disk_read_bytes Number of bytes read from storage # TYPE pve_disk_read_bytes gauge pve_disk_read_bytes{id="qemu/100"} 7.473739264e+09 # HELP pve_cpu_usage_ratio CPU usage (value between 0.0 and pve_cpu_usage_limit) # TYPE pve_cpu_usage_ratio gauge pve_cpu_usage_ratio{id="qemu/100"} 0.105009724408557 pve_cpu_usage_ratio{id="node/proxmox"} 0.984243806697115 # HELP pve_cpu_usage_limit Maximum allowed CPU usage # TYPE pve_cpu_usage_limit gauge pve_cpu_usage_limit{id="qemu/100"} 1.0 pve_cpu_usage_limit{id="node/proxmox"} 4.0 # HELP pve_uptime_seconds Number of seconds since the last boot # TYPE pve_uptime_seconds gauge pve_uptime_seconds{id="qemu/100"} 315039.0 pve_uptime_seconds{id="node/proxmox"} 315069.0 # HELP pve_storage_shared Whether or not the storage is shared among cluster nodes # TYPE pve_storage_shared gauge pve_storage_shared{id="storage/proxmox/local"} 0.0 pve_storage_shared{id="storage/proxmox/local-lvm"} 0.0 pve_storage_shared{id="storage/proxmox/vms"} 0.0 # HELP pve_guest_info VM/CT info # TYPE pve_guest_info gauge pve_guest_info{id="qemu/100",name="samplevm1",node="proxmox",type="qemu"} 1.0 # HELP pve_storage_info Storage info # TYPE pve_storage_info gauge pve_storage_info{id="storage/proxmox/local",node="proxmox",storage="local"} 1.0 pve_storage_info{id="storage/proxmox/local-lvm",node="proxmox",storage="local-lvm"} 1.0 pve_storage_info{id="storage/proxmox/vms",node="proxmox",storage="vms"} 1.0 # HELP pve_node_info Node info # TYPE pve_node_info gauge pve_node_info{id="node/proxmox",level="",name="proxmox",nodeid="0"} 1.0 # HELP pve_onboot_status Proxmox vm config onboot value # TYPE pve_onboot_status gauge pve_onboot_status{id="qemu/201",node="proxmox",type="qemu"} 1.0 # HELP pve_version_info Proxmox VE version info # TYPE pve_version_info gauge pve_version_info{release="7.1",repoid="6fe299a0",version="7.1-5"} 1.0
Using pve.yml config file
Example pve.yml
for password authentication:
default:
user: prometheus@pve
password: sEcr3T!
# Optional: set to false to skip SSL/TLS verification
verify_ssl: true
Example pve.yml
for token authentication:
default:
user: prometheus@pve
token_name: "your-token-id"
token_value: "..."
Using environment variables:
If the PVE_USER
environment variable exists, then configuration is taken from
the environment instead of from the pve.yml
config file. The following
environment variables are respected:
PVE_USER
: user name
Required for password authentication:
PVE_PASSWORD
: user password
Required for token authentication:
PVE_TOKEN_NAME
: token namePVE_TOKEN_VALUE
: token value
Optional:
PVE_VERIFY_SSL
: Eithertrue
orfalse
, whether or not to verify PVE tls certificate. Defaults totrue
.PVE_MODULE
: Name of the configuration module. Defaults todefault
.
The configuration is passed directly into proxmoxer.ProxmoxAPI().
Note on verify_ssl and certificate trust store:
When operating PVE with self-signed certificates, then it is necessary to
either import the certificate into the local trust store (see this SE answer
for Debian/Ubuntu) or add verify_ssl: false
to the config dict as a sibling
to the credentials. Note that PVE supports Let's Encrypt out ouf the box. In
many cases setting up trusted certificates is the better option than operating
with self-signed certs.
For security reasons it is essential to add a user with read-only access (PVEAuditor role) for the purpose of metrics collection.
Refer to the Proxmox Documentation for the several ways of creating a user. Once created, assign the user the / path permission.
The PVE exporter can be deployed either directly on a Proxmox VE node or onto a separate machine.
Example config for PVE exporter running on PVE node:
scrape_configs:
- job_name: 'pve'
static_configs:
- targets:
- 192.168.1.2:9221 # Proxmox VE node with PVE exporter.
- 192.168.1.3:9221 # Proxmox VE node with PVE exporter.
metrics_path: /pve
params:
module: [default]
cluster: 1
node: 1
Example config for PVE exporter running on Prometheus host:
scrape_configs:
- job_name: 'pve'
static_configs:
- targets:
- 192.168.1.2 # Proxmox VE node.
- 192.168.1.3 # Proxmox VE node.
metrics_path: /pve
params:
module: [default]
cluster: 1
node: 1
relabel_configs:
- source_labels: [__address__]
target_label: __param_target
- source_labels: [__param_target]
target_label: instance
- target_label: __address__
replacement: 127.0.0.1:9221 # PVE exporter.
Note on scraping large clusters:
It is adviced to setup separate jobs to collect cluster
metrics and
node
metrics in larger deployments. Scraping any node in a cluster with the
url params set to cluster=1&node=0
results in the same set of metrics. Hence
cluster metrics can be scraped efficiently from a single node or from a subset
of cluster nodes (e.g., a different node selected on every scrape via
round-robin DNS).
Node metrics can only be scraped from a given node. In order to compile a
complete set of node metrics it is necessary to scrape every node in a cluster
with url params set to cluster=0&node=1
.