haproxystats 0.5.1

Creator: railscoder56

Last updated:

Add to Cart

Description:

haproxystats 0.5.1

A HAProxy statistics collection program


Contents

Introduction
How haproxystats works

haproxystats-pull
haproxystats-process
Dispatchers
Statistics for HAProxy

HAProxy process


Queuing system
Statistics for haproxystats


Configuration

DEFAULT section
paths section
pull section
process section
graphite section
frontend-groups, backend-groups, and server-groups sections
local-store section


Systemd integration
Graceful shutdown
Puppet module
Ansible Playbook
Nagios checks
Starting the programs
Development

How to setup a development environment


Installation
How to make a release
Contributors
Licensing
Acknowledgement
Contacts



Introduction
haproxystats is a statistics collector for HAProxy load balancer which
processes various statistics and pushes them to graphing systems (Graphite).
It is designed to satisfy the following requirements:

Fast and configurable processing of HAProxy statistics
Perform aggregation when HAProxy runs in multiprocess (nbproc > 1)
Pull statistics at very low intervals (10secs)
Flexible dispatching of statistics to different systems (Graphite, kafka)

The main design characteristic is the split between pulling the statistics and
processing them. This provides the ability to pull data as frequently
as possible without worrying about the impact on processing time. It also
reduces the risk of losing data in case of trouble during the processing phase.
It runs locally on each load balancer node, offering a decentralized setup for
the processing phase, but it can be easily extended in the future to have a
centralized setup for the processing phase. In that centralized setup it will
be possible to perform aggregation on a cluster level as well.
Until then users can deploy carbon-c-relay for aggregation.
Because of this design haproxystats comes with two programs:
haproxystats-pull and haproxystats-process. The former pulls
statistics from HAProxy via stats socket and it uses the asyncio framework
from Python to achieve high concurrency and low footprint. The latter
processes the statistics and pushes them to various destinations. It utilizes
Pandas for data analysis and the multiprocess framework from Python.
haproxystats requires Python 3.4, docopt and Pandas to be available in the
system.


How haproxystats works

haproxystats-pull sends info and stat commands to all haproxy processes
in order to collect statistics for the daemon and for all
frontends/backends/servers. Data returned from each process and for each
command is stored in individual files which are saved under one directory. The
time (seconds since the epoch) of retrieval is used to name that directory.
haproxystats-process watches for changes on the parent directory and when a
directory is created it adds its full path to the queue. Multiple workers pick
up items (directories) from the queue and process statistics from those
directories.

haproxystats-pull
haproxystats-pull leverages the asyncio framework from Python by utilizing
coroutines to multiplex I/O access over several stats socket, which are
simple UNIX and TCP sockets.
The actual task of storing the data to the file system is off-loaded to a very
light pool of threads in order to avoid blocking the coroutines during the
disk IO phase.
haproxystats-pull manages the incoming directory and makes sure directories
are created with correct names. It also suspends the collection when the number
of directories under the incoming directory exceeds a threshold. This avoids
filling up the disk when haproxystats-process is unavailable for sometime.
This an example of directory structure:
incoming
├── 1457298067
│   ├── admin1.sock_info
│   ├── admin1.sock_stat
│   ├── admin2.sock_info
│   ├── admin2.sock_stat
│   ├── admin3.sock_info
│   ├── admin3.sock_stat
│   ├── admin4.sock_info
│   └── admin4.sock_stat
└── 1457298072
├── admin1.sock_info
├── admin1.sock_stat
├── admin2.sock_info
├── admin2.sock_stat
├── admin3.sock_info
├── admin3.sock_stat
├── admin4.sock_info
└── admin4.sock_stat


haproxystats-process
haproxystats-process is a multiprocess program. The parent process uses the
Linux kernel’s inotify API to watch for changes in incoming directory.
It receives an event when a directory is either created or moved in incoming
directory. The event contains the absolute path name of that directory. It
maintains an internal queue in which it puts directory names. Multiple child
processes pick directory names from the queue and process the data.
Its worker dispatches statistics to various destinations. The directories are
removed from incoming directory when all statistics are successfully
processed.
When haproxystats-process starts it scans the incoming directory
for new directories and processes them instantly, so you don’t lose statistics
if haproxystats-process is unavailable for sometime.


Dispatchers
haproxystats-process currently supports 2 different dispatchers.

Graphite

Pushes statistics to a Graphite system via a local or remote carbon-relay.
The recommended method is to use carbon-c-relay. It is very fast and capable
of handling millions of metrics per second. This dispatcher utilizes an internal
queue to store metrics which are failed to be sent to Graphite.
An example of graphite namespace:
<loadbalancers>.<lb-01>.haproxy.frontend.<frontendname>.
<loadbalancers>.<lb-01>.haproxy.backend.<backendname>.
<loadbalancers>.<lb-01>.haproxy.backend.<backendname>.server.<servername>
<loadbalancers>.<lb-01>.haproxy.server.<servername>.
<loadbalancers>.<lb-01>.haproxy.daemon.
<loadbalancers>.<lb-01>.haproxy.haproxystats.<metric names>.

local-store

Stores statistics in the local disk. Use it only for debugging purposes.


Statistics for HAProxy
In addition the statistics that are exposed by HAProxy, haproxystats provides
the following statistics.

HAProxy process
HAProxy exposes Idle_pct and haproxystats-process converts it to CPU
utilization without removing Idle_pct metric. This avoids the usage of
scale(-1) and offset(100) functions on graphite:
CpuUsagePct CPU utilization in percentage
The following metrics are calculated only when HAProxy is configured with more
than 1 processes (nbproc > 1):
25PercentileCpuUsagePct 25th percentile of CpuUsagePct across all processes
50PercentileCpuUsagePct 50th percentile -//-
75PercentileCpuUsagePct 75th percentile -//-
95PercentileCpuUsagePct 95th percentile -//-
99PercentileCpuUsagePct 99th percentile -//-
StdCpuUsagePct standard deviation -//-



Queuing system
The incoming directory together with the inotify API provides a simple
queueing system which is used as a communication channel between
haproxystats-pull and haproxystats-process programs.
There isn’t any feedback mechanism in place, thus haproxystats-pull monitors
the number of directories before it pulls data from HAProxy and suspends its
job when the number of directories exceeds a threshold.
See queue-size parameter of pull section.


Statistics for haproxystats
haproxystats provides statistics for the time it takes to process,
calculate and send HAProxy metrics. By default provides the following list
of metric names with values in seconds:
loadbalancers.lb-01.haproxy.haproxystats.WallClockTimeHAProxy
loadbalancers.lb-01.haproxy.haproxystats.WallClockTimeFrontends
loadbalancers.lb-01.haproxy.haproxystats.WallClockTimeBackends
loadbalancers.lb-01.haproxy.haproxystats.WallClockTimeServers
loadbalancers.lb-01.haproxy.haproxystats.WallClockTimeAllStats
It also provides the number of metrics which are send to graphite:
loadbalancers.lb-01.haproxy.haproxystats.MetricsHAProxy
loadbalancers.lb-01.haproxy.haproxystats.MetricsFrontend
loadbalancers.lb-01.haproxy.haproxystats.MetricsBackend
loadbalancers.lb-01.haproxy.haproxystats.MetricsServer



Configuration
haproxystats uses the popular INI format for its configuration file.
This is an example configuration file (/etc/haproxystats.conf):
[DEFAULT]
loglevel = info
retries = 2
timeout = 1
interval = 2

[paths]
base-dir = /var/lib/haproxystats

[pull]
loglevel = info
socket-dir = /run/haproxy
retries = 1
timeout = 0.1
interval = 0.5
pull-timeout = 2
pull-interval = 10
dst-dir = ${paths:base-dir}/incoming
tmp-dst-dir = ${paths:base-dir}/incoming.tmp
workers = 8
queue-size = 360

[process]
src-dir = ${paths:base-dir}/incoming
workers = 4
per-process-metrics = false

[graphite]
server = 127.0.0.1
port = 3002
retries = 3
interval = 1.8
connect-timeout = 1.0
write-timeout = 1.0
delay = 10
backoff = 2
namespace = loadbalancers
prefix-hostname = true
fqdn = true
queue-size = 1000000

#[local-store]
#dir = ${paths:base-dir}/local-store
All the above settings are optional as haproxystats comes with default values
for all of them. Thus, both programs can be started without supplying any
configuration.

DEFAULT section
Settings in this section can be overwritten in other sections.

loglevel Defaults to info

Log level to use, possible values are: debug, info, warning, error, critical

retries Defaults to 2

Number of times to retry a connection after a failure. Used by haproxystats-pull
and haproxystats-process when they open a connection to a UNIX/TCP socket and
Graphite respectively.

timeout Defaults to 1 (seconds)

Time to wait for establishing a connection. Used by haproxystats-pull and
haproxystats-process when they open a connection to a UNIX/TCP socket and Graphite
respectively.

interval Defaults to 2

Time to wait before trying to open a connection. Used by haproxystats-pull and
haproxystats-process when they retry a connection to a UNIX/TCP socket and Graphite
respectively.


paths section

base-dir Defaults to /var/lib/haproxystats

The directory to use as the base of the directory structure.


pull section

socket-dir Unset by default

A directory with HAProxy socket files.

servers Unset by default

A list of servers to pull statistics from. You define a server by passing a URL,
here some examples:
tcp://127.0.0.1:5555
tcp://foo.bar.com:4444
tcp://[fe80::3f2f:46b3:ef0c:a420]:4444
unix:///run/haproxy.sock
Only TCP and UNIX schemes are supported and the port for TCP servers must
be set. For UNIX scheme you can only pass a file and not a directory, but
socket-dir option can be set as well, so you can use a directory and UNIX
socket files at the same time. You can use comma as separator to pass multiple
servers:
servers = unix:///run/haproxy.sock,tcp://127.0.0.1:555,tcp://127.0.0.1:556

buffer-limit Defaults to 6291456 (bytes)

At most size bytes are read and returned from the sockets. Setting too low and
it will slow down the retrieval of statistics.
Only values greater than or equal to 1 are accepted.

retries Defaults to 1

Number of times to reconnect to UNIX/TCP socket after a failure.

timeout Defaults to 0.1 (seconds)

Time to wait for establishing a connection to UNIX/TCP socket. There is no need to
set it higher than few ms as haproxy accepts a connection within 1-2ms.

interval Defaults to 0.5 (seconds)

Time to wait before trying to reconnect to UNIX/TCP socket after a failure. Tune it
based on the duration of the reload process of haproxy. haproxy reloads within
few ms but in some environments with hundreds different SSL certificates it can
take a bit more.

pull-interval Defaults to 10 (seconds)

How often to pull statistics from HAProxy. A value of 1 second can overload
the haproxy processes in environments with thousands backends/servers.

pull-timeout Defaults to 2 (seconds)

Total time to wait for the pull process to finish. Should be always less than
pull-interval.

dst-dir Defaults /var/lib/haproxystats/incoming

A directory to store statistics retrieved by HAProxy.

tmp-dst-dir Defaults /var/lib/haproxystats/incoming.tmp

A directory to use as temporary storage location before directories are moved
to dst-dir. haproxystats-pull stores statistics for each process under
that directory and only when data from all haproxy processes are successfully
retrieved they are moved to dst-dir. Make sure dst-dir and
tmp-dst-dir are on the same file system, so the move of the directories
become a rename which is a quick and atomic operation.

workers Defaults to 8

Number of threads to use for writing statistics to disk. These are very
light threads and don’t consume a lot of resources. Shouldn’t be set higher
than the number of haproxy processes.

queue-size Defaults to 360

Suspend the pulling of statistics when the number of directories in dst-dir
exceeds this limit.


process section

src-dir Defaults /var/lib/haproxystats/incoming

A directory to watch for changes. It should point to the same directory as
the dst-dir option from pull section.

workers Defaults to 4

Number of workers to use for processing statistics. These are real processes
which can consume a fair bit of CPU.

frontend-metrics Unset by default

A list of frontend metric names separated by space to process. By default all
statistics are processed and this overwrites the default selection.
haproxystats-process emits an error and refuses to start if metrics aren’t
valid HAProxy metrics. Check the list of valid metrics in Chapter 9.1 of
management documentation of HAProxy.

backend-metrics Unset by default

A list of backend metric names separated by space to process. By default all
statistics are processed and this overwrites the default selection.
haproxystats-process emits an error and refuses to start if metrics aren’t
valid HAProxy metrics. Check the list of valid metrics in Chapter 9.1 of
management documentation of HAProxy.

server-metrics Unset by default

A list of server metric names separated by space to process. By default all
statistics are processed and this overwrites the default selection.
haproxystats-process emits an error and refuses to start if metrics aren’t
valid HAProxy metrics. Check the list of valid metrics in Chapter 9.1 of
management documentation of HAProxy.

aggr-server-metrics Defaults to false

Aggregates server’s statistics across all backends.

exclude-frontends Unset by default

A file which contains one frontend name per line for which processing is
skipped.

exclude-backends Unset by default

A file which contains one backend name per line for which processing is
skipped.

per-process-metrics Defaults to false

HAProxy daemon provides statistics and by default haproxystat-process
aggregates those statistics when HAProxy runs in multiprocess mode
(nbproc > 1).
Set this to true to get those statistics also per process as well.
This is quite useful for monitoring purposes where someone wants to monitor
sessions per process in order to see if traffic is evenly distributed to all
processes by the kernel.
It is also useful in setups where configuration for frontends and backends is
unevenly spread across all processes, for instance processes 1-4 manage SSL
frontends and processes 5-7 manage noSSL frontends.
This adds another path in Graphite under haproxy space:
loadbalancers.lb-01.haproxy.daemon.process.<process_num>.<metric>

calculate-percentages Defaults to false

Calculates percentages for a selection of metrics for HAProxy daemon. When
per-process-metrics is set to true the calculation happens also per
HAProxy process. This adds the following metric names:
ConnPercentage
ConnRatePercentage
SslRatePercentage
SslConnPercentage
Those metrics can be used for alerting when the current usage on connections
is very close the configured limit.

liveness-check-interval Defaults to 10 (seconds)

How often to check if all workers are alive and trigger a termination if at
least one is dead.


graphite section
This dispatcher is enabled by default and it can’t be disabled.

server Defaults to 127.0.0.1

Graphite server to connect to.

port Defaults to 3002

Graphite port to connect to.

retries Defaults to 3

Number of times to reconnect to Graphite after a failure.

interval Defaults to 1.8 (seconds)

Time to wait before trying to reconnect to Graphite after a failure.

connect-timeout Defaults to 1 (seconds)

Time to wait for establishing a connection to Graphite relay.

write-timeout Defaults to 1 (seconds)

Time to wait on sending data to Graphite relay.

delay Defaults to 10 (seconds)

How long to wait before trying to connect again after number of retries has
exceeded the threshold set in retries. During the delay period metrics are
stored in the queue of the dispatcher, see queue-size.

backoff Defaults to 2

A simple exponential backoff to apply for each retry.

namespace Defaults to loadbalancers

A top level graphite namespace.

prefix-hostname Defaults to true

Insert the hostname of the load balancer in the Graphite namespace, example:
loadbalancers.lb-01.haproxy.

fqdn Defaults to true

Use FQDN or short name in the graphite namespace

queue-size Defaults to 1000000

haproxystats-process uses a queue to store metrics which failed to be sent due
to a connection error/timeout. This is a First In First Out queueing system.
When the queue reaches the limit, the oldest items are removed to free space.

group-namespace Unset by default.

group graphite metrics by patterns. When a frontend, backend or server matches a
given pattern, the metric will be prefixed by this namespace, plus a
configurable group name which must be specified in the frontend-groups,
backend-groups or server-groups sections. These sections consist of
group names and their corresponding regular expression that will be matched
against frontend, backend or server names (depending on the section).
For example:
Let’s assume our metrics look something like:
loadbalancers.lb-01.haproxy.frontend.foo-001.<metric>
loadbalancers.lb-01.haproxy.frontend.foo-002.<metric>
...
loadbalancers.lb-01.haproxy.frontend.bar-001.<metric>
loadbalancers.lb-01.haproxy.frontend.bar-002.<metric>
...
And we want them to be grouped to like this:
loadbalancers.lb-01.haproxy.flavor.abc.frontend.foo-001.<metric>
loadbalancers.lb-01.haproxy.flavor.abc.frontend.foo-002.<metric>
...
loadbalancers.lb-01.haproxy.flavor.xyz.frontend.bar-001.<metric>
loadbalancers.lb-01.haproxy.flavor.xyz.frontend.bar-002.<metric>
...
The configuration should contain these settings:
[graphite]
group-namespace = flavor

[frontend-groups]
abc = ^foo-
xyz = ^bar-
Note that if the group-namespace setting is specified, then at least one of
frontend-groups, backend-groups or server-groups sections must be
specified as well.
Also note that if frontend, backend or server names contain dots, these will be
converted to underscores for graphite – because dots are graphite’s namespace
separator. The patterns will have to take this into account.

group-namespace-double-writes Unset by default.

Boolean; required only if group-namespace is specified. If True, send to
graphite the original metric as well as the grouped metrics. If False, send
only the grouped metrics. (See group-namespace.)


frontend-groups, backend-groups, and server-groups sections
Specify the patterns to match against frontend, backend and/or server names, to
group graphite metrics and give them a variable prefix. See group-namespace.
These sections are optional, unless group-namespace is set.


local-store section
This dispatcher isn’t enabled by default.
The primarily use of local-store dispatcher is to debug/troubleshoot possible
problems with the processing or/and with Graphite. There isn’t any clean-up
process in place, thus you need remove the files after they are created.
Don’t leave it enabled for more than 1 hour as it can easily fill up the disk
in environments with hundreds frontends/backends and thousands servers.

dir Defaults to /var/lib/haproxystats/local-store

A directory to stores statistics after they have been processed. The current
format is compatible with Graphite.



Systemd integration
haproxystats-pull and haproxystats-process are simple programs which are not
daemonized and they output logging messages to stdout. This is by design as it
simplifies the code. The daemonization and logging is off-loaded to systemd
which has everything we need for that job.
Under contrib/systend directory there are service files for both programs.
These are functional systemd Unit files which are used in production.
The order in which these 2 programs start doesn’t matter and there isn’t any
soft or hard dependency between them.
Furthermore, these programs don’t need to run as root. It highly recommended to
create a dedicated user to run them. You need to add that user to the group of
haproxy and adjust socket configuration of haproxy to allow write for the
group, see below an example configuration:
stats socket /run/haproxy/sock1 user haproxy group haproxy mode 660 level admin process 1
stats socket /run/haproxy/sock2 user haproxy group haproxy mode 660 level admin process 2
stats socket /run/haproxy/sock3 user haproxy group haproxy mode 660 level admin process 3
systemd Unit files use haproxystats user which has to be created prior running
haproxystats programs.


Graceful shutdown
In an effort to reduce the loss of statistics both programs support graceful
shutdown. When SIGHUP or SIGTERM signals are sent they perform a clean exit.
When a signal is sent to haproxystats-process it may take some time for the
program to exit, as it waits for all workers to empty the queue.


Puppet module
A puppet module is available under contrib directory which provides classes for
configuring both programs.
Because haproxystats-process is CPU bound program, CPU Affinity is configured
using systemd. By default it pins the workers to the last CPUs.
You should take care of pinning haproxy processes to other CPUs in order to
avoid haproxystats-process stealing CPU cycles from haproxy. In production
servers you usually pin the first 80% of CPUs to haproxy processes and you
leave the rest of CPUs for other processes. The default template of puppet
module enforces this logic.
haproxystats-pull is a single threaded program which doesn’t use a lot of CPU
cycles and by default is assigned to the last CPU.


Ansible Playbook
A Ansible playbook is available under contrib directory. For installation
instruction of the playbook please read Installation chapter of this document.


Nagios checks
Several nagios checks are provided for monitoring purposes, they can be found
under contrib/nagios directory.

check_haproxystats_process_number_of_procs.sh

Monitor the number of processes of haproxystats-process program. Systemd
monitors only the parent process and this check helps to detect cases where
some worker(s) die unexpectedly

check_haproxystats_process.sh

A wrapper around systemctl tool to detect a dead parent process.

check_haproxystats_pull.sh

A wrapper around systemctl tool to a check if haproxystats-pull is running.

check_haproxystats_queue_size.py

Checks the size of the incoming directory queue which is consumed by
haproxystats-process and alert when exceeds a threshold.


Starting the programs
haproxystats-pull -f ./haproxystats.conf
haproxystats-process -f ./haproxystats.conf
Usage:
% haproxystats-pull -h
Pulls statistics from HAProxy daemon over UNIX socket(s)

Usage:
haproxystats-pull [-f <file>] [-p | -P]

Options:
-f, --file <file> configuration file with settings
[default: /etc/haproxystats.conf]
-p, --print show default settings
-P, --print-conf show configuration
-h, --help show this screen
-v, --version show version


% haproxystats-process -h
Processes statistics from HAProxy and pushes them to Graphite

Usage:
haproxystats-process [-f <file>] [-d <dir>] [-p | -P]

Options:
-f, --file <file> configuration file with settings
[default: /etc/haproxystats.conf]
-d, --dir <dir> directory with additional configuration files
-p, --print show default settings
-P, --print-conf show configuration
-h, --help show this screen
-v, --version show version


Development
I would love to hear what other people think about haproxystats and provide
feedback. Please post your comments, bug reports and wishes on my issues page.

How to setup a development environment
Install HAProxy:
% sudo apt-get install haproxy
Use a basic HAProxy configuration in multiprocess mode:
global
log 127.0.0.1 len 2048 local2
chroot /var/lib/haproxy
stats socket /run/haproxy/admin1.sock mode 666 level admin process 1
stats socket /run/haproxy/admin2.sock mode 666 level admin process 2
stats socket /run/haproxy/admin3.sock mode 666 level admin process 3
stats socket /run/haproxy/admin4.sock mode 666 level admin process 4
# allow read/write access to anyone----------^
stats timeout 30s
user haproxy
group haproxy
daemon
nbproc 4
cpu-map 1 0
cpu-map 2 1
cpu-map 3 1
cpu-map 4 0

defaults
log global
mode http
timeout connect 5000
timeout client 50000
timeout server 50000

frontend frontend_proc1
bind 0.0.0.0:81 process 1
default_backend backend_proc1

frontend frontend_proc2
bind 0.0.0.0:82 process 2
default_backend backend_proc1

frontend frontend1_proc34
bind :83 process 3
bind :83 process 4
default_backend backend1_proc34

backend backend_proc1
bind-process 1
default-server inter 1000s
option httpchk GET / HTTP/1.1\r\nHost:\ .com\r\nUser-Agent:\ HAProxy
server member1_proc1 10.189.224.169:80 weight 100 check fall 2 rise 3
server member2_proc1 10.196.70.109:80 weight 100 check fall 2 rise 3
server bck_all_srv1 10.196.70.109:88 weight 100 check fall 2 rise 3

backend backend1_proc34
bind-process 3,4
default-server inter 1000s
option httpchk GET / HTTP/1.1\r\nHost:\ .com\r\nUser-Agent:\ HAProxy
server bck1_proc34_srv1 10.196.70.109:80 check fall 2 inter 5s rise 3
server bck1_proc34_srv2 10.196.70.109:80 check fall 2 inter 5s rise 3
server bck_all_srv1 10.196.70.109:80 check fall 2 inter 5s rise 3

backend backend_proc2
bind-process 2
default-server inter 1000s
option httpchk GET / HTTP/1.1\r\nHost:\ .com\r\nUser-Agent:\ HAProxy
server bck_proc2_srv1_proc2 127.0.0.1:8001 check fall 2 inter 5s rise 3
server bck_proc2_srv2_proc2 127.0.0.1:8002 check fall 2 inter 5s rise 3
server bck_proc2_srv3_proc2 127.0.0.1:8003 check fall 2 inter 5s rise 3
server bck_proc2_srv4_proc2 127.0.0.1:8004 check fall 2 inter 5s rise 3
Start HAProxy and check it is up:
sudo systemctl start haproxy.service;systemctl status -l haproxy.service
Create a python virtual environment using virtualenvwrapper tool:
mkvirtualenv --python=`which python3` haproxystats-dev
Do not exit the haproxystats-dev virtual environment.
Clone the project, if you are planning to contribute then you should fork it on
GitHub and clone that project instead:
mkdir ~/repo;cd ~/repo
git clone https://github.com/unixsurfer/haproxystats
Install necessary libraries:
cd haproxystats
pip install -U pbr setuptools
pip install -r ./requirements.txt
Start a TCP server which acts a Graphite relay and listens on 127.0.0.1:39991:
python3 ./contrib/tcp_server.py
Install haproxystats:
python setup.py install
Create necessary directory structure:
mkdir -p ./var/var/lib/haproxystats
mkdir -p ./var/etc
mkdir -p ./var/etc/haproxystats.d
Adjust the following configuration and save it in ./var/etc/haproxystats.conf:
[DEFAULT]
loglevel = debug
retries = 2
timeout = 1
interval = 2

[paths]
base-dir = /home/<username>/repo/haproxystats/var/var/lib/haproxystats

[pull]
socket-dir = /run/haproxy
retries = 1
timeout = 0.1
interval = 0.5
pull-timeout = 10
pull-interval = 10
dst-dir = ${paths:base-dir}/incoming
tmp-dst-dir = ${paths:base-dir}/incoming.tmp
workers = 8

[process]
src-dir = ${paths:base-dir}/incoming
workers = 2
calculate-percentages = true
per-process-metrics = true

[graphite]
server = 127.0.0.1
port = 39991
retries = 3
interval = 0.8
timeout = 0.9
delay = 10
backoff = 2
namespace = loadbalancers
prefix_hostname = true
fqdn = true
queue-size = 1000

[local-store]
dir = ${paths:base-dir}/local-store
Start haproxystats-pull and haproxystats-process on 2 different terminals:
haproxystats-pull -f var/etc/haproxystats.conf
haproxystats-process -f var/etc/haproxystats.conf
Exit from haproxystats-dev virtual environment:
deactivate
Start hacking and don’t forget to make a Pull Request



Installation
Use pip:
pip install haproxystats
From Source:
sudo python setup.py install
Build (source) RPMs:
python setup.py clean --all; python setup.py bdist_rpm
Build a source archive for manual installation:
python setup.py sdist
Use Ansible Playbook:
To deploy haproxystats By Ansible Playbook go to contrib/ansible-playbook
directory:
cd contrib/ansible-playbook
Then enter your haproxy server IP address in hosts file:
vi hosts
After that set information of your environment in variable file:
vi group_vars/all
Now for run Ansible Playbook use this command:
ansible-playbook -i hosts main-playbook.yml
When Ansible Playbook run successful completely, you can take control haproxystats-pull and haproxystats-process by systemd:
systemctl start haproxystats-pull.service

systemctl start haproxystats-process.service


How to make a release

Bump version in haproxystats/__init__.py
Commit above change with:
git commit -av -m'RELEASE 0.1.3 version'

Create a signed tag, pbr will use this for the version number:
git tag -s 0.1.3 -m 'bump release'

Create the source distribution archive (the archive will be placed in the
dist directory):
python setup.py sdist

pbr updates ChangeLog file and we want to squeeze this change to the
previous commit, thus run:
git commit -av --amend

Move current tag to the last commit:
git tag -fs 0.1.3 -m 'bump release'

Push changes:
git push;git push --tags

Upload to Python Package Index:
twine upload -s -p dist/*




Contributors
The following people have contributed to project with feedback and code reviews

Károly Nagy https://github.com/charlesnagy
Dan Achim https://github.com/danakim



Licensing
Apache 2.0


Acknowledgement
This program was originally developed for Booking.com. With approval
from Booking.com, the code was generalised and published as Open Source
on github, for which the author would like to express his gratitude.


Contacts
Project website: https://github.com/unixsurfer/haproxystats
Author: Pavlos Parissis <pavlos.parissis@gmail.com>

License

For personal and professional use. You cannot resell or redistribute these repositories in their original state.

Customer Reviews

There are no reviews.