|Repos Using This
|Packages Using This
|Most Recent Commit
|8 days ago
|June 30, 2023
|extract internal monitoring data from application logs for collection in a timeseries database
|2 months ago
|October 28, 2023
|An aggregating proxy to enable HA prometheus
|Kube Rbac Proxy
|4 days ago
|October 20, 2023
|Kubernetes RBAC authorizing HTTP proxy for a single upstream.
|a year ago
|Sample Docker Swarm cluster stack of tools
|3 years ago
|September 29, 2020
|Self-deployable tool for hosting hyper:// websites
|a month ago
|March 27, 2023
|A reverse proxy designed for Prometheus exporters
|a day ago
|March 16, 2021
|S3 Reverse Proxy with GET, PUT and DELETE methods and authentication (OpenID Connect and Basic Auth)
|Prom Label Proxy
|19 days ago
|June 15, 2023
|A proxy that enforces a given label in a given PromQL query.
|a month ago
|A transparent TCP to SOCKSv5/HTTP proxy on Linux written in Rust.
|a month ago
|基于 Python Asyncio + Redis 实现的代理池
pronounced "promski" or präm-sē
Promxy is a prometheus proxy that makes many shards of prometheus appear as a single API endpoint to the user. This significantly simplifies operations and use of prometheus at scale (when you have more than one prometheus host). Promxy delivers this unified access endpoint without requiring any sidecars, custom-builds, or other changes to your prometheus infrastructure.
Short version: Prometheus itself provides no real HA/clustering support. As such the best-practice is to run multiple (e.g N) hosts with the same config. Similarly prometheus has no real built-in query federation, which means that you end up with N sources in grafana which is (1) confusing to grafana users and (2) has no support for aggregation across the sources. Promxy enables an HA prometheus setup by "merging" the data from the duplicate hosts (so if there is a gap in one, promxy will fill with the other). In addition Promxy provides a single datasource for all promql queries -- meaning your grafana can have a single source and you can have globally aggregated promql queries.
Release binaries are available on the releases page.
If you are interested in hacking on promxy (or just running your own build), you can clone and build:
git clone [email protected]:jacksontj/promxy.git
cd promxy/cmd/promxy && go build
An example configuration file is available in the repo.
With that configuration modified and ready, all that is left is to run promxy:
ServerGroup is a set of prometheus hosts configured the same. This is a common best practice
for prometheus infrastructure as prometheus itself doesn't support any HA/clustering. This
allows promxy to merge data from multiple hosts in the
ServerGroup (all until it becomes a priority).
This allows promxy to "fill" in the holes in timeseries, such as the ones created when upgrading
prometheus or rebooting the host
Promxy uses the
/v1 API of prometheus under-the-hood, meaning that promxy simply
requires that API to be present. Promxy has been used with as early as prom 1.7
and as recent as 2.13. If you run into issues with any prometheus version with the
API please open up an issue.
Promxy is currently using a fork based on prometheus 2.24. This version isn't supremely important, but it is relevant for promql features (e.g. subqueries) and sd config options.
None. Promxy is simply an aggregating proxy that sends requests to prometheus -- meaning it requires no changes to your existing prometheus install.
Yes! Promxy simply aggregates other prometheus API endpoints together so you can definitely layer promxy. Similarly you can mix prometheus API endpoints, for example you could have prometheus, promxy, and VictoriaMetrics all as downstreams of a promxy host -- since they all have prometheus compatible APIs.
Promxy's goal is to be the same performance as the slowest prometheus server it has to talk to. If you have a query that is significantly slower through promxy than on prometheus direct please open up an issue so we can get that taken care of.
Note: if you are running prometheus <2.2 you may notice "slow" performance when running queries that access large amounts of data. This is due to inefficient json marshaling in prometheus. You can workaround this by configuring promxy to use the remote_read API.
Promxy currently does a complete scatter-gather to all configured server groups. There are plans to reduce scatter-gather queries but in practice the current "scatter-gather always" implementation hasn't been a bottleneck.
Promxy is simply an aggregating proxy in front of your prometheus infrastructure. As such, you can use promxy to create alerting/recording rules which will execute across your entire prometheus infrastructure. For example, if you wanted to know that the global error rate was <10% this would be impossible on the individual prometheus hosts (without federation, or re-scraping) but trivial in promxy.
Note: recording rules in regular prometheus write to their local tsdb. Promxy has no local tsdb, so if you wish to use recording rules (or see the metrics from alerting rules) a remote_write endpoint must be defined in the promxy config (which is where it will send those metrics).
The default behavior in the event of a servergroup being down is to return an error. If all nodes in a servergroup are down the resulting data can be inaccurate (missing data, etc.) -- so we'd rather by default return an error than an inaccurate value (since alerting etc. might rely on it, we don't want to hide a problem).
Now with that said if you'd like to make some or all servergroups "optional" (meaning the errors will be ignored and we'll serve the response anyways) you can do this using the ignore_error option on the servergroup.
Feedback is greatly appreciated. If you find a bug, have a feature request, or just have a general question feel free to open up an issue!