This module provides a full nagios monitoring solution, for setting up both servers and clients. There can be any number of each, though there is typically one server and many clients. The main limitation is that all server and client nodes need to have all of their nagios related puppet configuration on the same puppetmaster.
The server part is still very Red Hat Enterprise Linux specific. The client part is much more generic and has been tested on both RHEL and Gentoo.
Please note that this module is not for the faint of heart. Even I (the author) have my head hurt each time I have to make modificiations to it... but it is worth it, as it allows having monitoring automatically enabled by default on all nodes as well as for all standard services detected on them.
Version 2 of the module has updated all default checks to become classes, in order to work better with hiera's class parameters automatic lookup. The drawback is that overrides which were set from manifests possibly need to be moved to hieradata.
When upgrading from version 1, if you see checks changing arguments or getting added or removed, look at the check's parameters and migrate your existing overrides, typically :
# Override the critical value for check foo
nagios::check::foo::args: '-c 10%'
# Disable the bar check completely
nagios::check::bar::ensure: 'absent'
- Stored configurations enabled on the puppetmaster (mandatory)
apache_httpd
andphp
modules used for the nagios server node (optional)
The apache_httpd
and php
modules are required for the server part, though
optionally since it is also possible to use an existing non-puppet managed web
server or different puppet modules. For a new empty node, configuring it as a
nagios server will be much quicker if those modules can be used.
Stored configurations are essential on the puppetmaster for the module to work at all, since it relies on having all nodes create their own exported nagios host and service resources, which the nagios server node then realizes to build its configuration.
For RHEL, any packages which might be required but are not part of the official repositories or EPEL can be found on http://dl.marmotte.net/rpms/
When SELinux is enforcing, the selinux::audit2allow
definition is required
to allow some basic nagios/nrpe accesses, though it can also be disabled.
Nagios server instance (node specific, typically inside a node section) :
class { '::nagios::server':
apache_httpd_ssl => false,
apache_httpd_conf_content => template('my/httpd-nagios.conf.erb'),
apache_httpd_htpasswd_source => 'puppet:///modules/my/htpasswd',
cgi_authorized_for_system_information => '*',
cgi_authorized_for_configuration_information => '*',
cgi_authorized_for_system_commands => '*',
cgi_authorized_for_all_services => '*',
cgi_authorized_for_all_hosts => '*',
cgi_authorized_for_all_service_commands => '*',
cgi_authorized_for_all_host_commands => '*',
cgi_default_statusmap_layout => '3',
}
Nagios client instances (typically from site.pp
) :
class { '::nagios::client':
nrpe_allowed_hosts => '127.0.0.1,192.168.1.1',
}
Nagios client specific overrides. See client.pp
and check/*.pp
for all of
the variables which can be manipulated this way. The following :
nagios::client::config { 'host_address': value => $::ipaddress_eth2 }
Will result in having $::nagios_host_name
get $ipaddress_eth2
as its value
for the entire configuration of the client where it is applied.
Nagios client check override configuration examples :
nagios::client::config { 'check_ram_ensure': value => 'absent' }
nagios::client::config { 'check_cpu_args': value => '-w 50 -c 20' }
Configuring a default check (must be done from a scope where nagios::client
can inherit it) :
Nagios::Check::Swap { ensure => 'absent' }
if $::domain == 'example.com' {
Nagios::Check::Cpu { notification_period => 'workhours' }
}
To enable nagiosgraph on the server :
class { '::nagios::nagiosgraph':
# This is the default
perflog => '/var/log/nagios/service_perfdata.log',
# To enable the mouseover graphs
nagios_service_action_url => '/nagiosgraph/cgi-bin/show.cgi?host=$HOSTNAME$&service=$SERVICEDESC$\' onMouseOver=\'showGraphPopup(this)\' onMouseOut=\'hideGraphPopup()\' rel=\'/nagiosgraph/cgi-bin/showgraph.cgi?host=$HOSTNAME$&service=$SERVICEDESC$',
ssi => true,
}
# This is what needs to be changed/added for nagios::server
class { '::nagios::server':
process_performance_data => '1',
service_perfdata_file => '/var/log/nagios/service_perfdata.log',
service_perfdata_file_template => '$LASTSERVICECHECK$||$HOSTNAME$||$SERVICEDESC$||$SERVICEOUTPUT$||$SERVICEPERFDATA$',
service_perfdata_file_processing_interval => '30',
service_perfdata_file_processing_command => 'process-service-perfdata-nagiosgraph',
}
To enable nagiosgraph for the client's services in the server web interface :
class { '::nagios::client':
service_use => 'generic-service,nagiosgraph-service',
}
To override the parameters of a default template using hiera :
```yaml
---
# Remove default warning notifications for services
nagios::server::template_generic_service:
notification_options: 'u,c,r'
Debug any startup or configuration problems on the server with :
nagios -v /etc/nagios/nagios.cfg
A lot can be configured semi-dynamically for nagios::client
(ideally using
hiera's automatic class parameter lookup) :
class { '::nagios::client':
host_notification_period => $::domain ? {
/\.dev$/ => 'workhours',
default => '24x7',
}
# You will need to use the type "nagios_hostgroup" on the server for
# all of the possible domain values to create the hostgroups.
host_hostgroups => $::domain,
}
Client overrides should be done in this order :
- For all services on some nodes : Using
nagios::client service_*
parameters - For all services on a node : Using the
nagios::client host_*
parameters - For specific services on all nodes : Using
Nagios::Check::Foo { paramname => 'value' }
(the scope must be the same or lower than wherenagios::client
is called from) - For specific services on a node : Using
nagios::client::config { 'check_*': value => 'foo' }
overrides
Note : nagios::client::config
can also be used to override just about
anything, though you must take into account that since it's done using custom
facts, you'll need two puppet runs on the client node then one on the server
node for the change to be applied : One for the fact to be created, then
another for the client configuration to take the new fact into account, then
the server run to update the nagios configuration. This might take a little
while depending on how often puppet is run on the nodes.
For the mysql_health
based checks to work, you will need to create the MySQL
user on your servers, allowed for localhost since we use nrpe for execution.
Example :
# This could go in site.pp, the fact is present only if mysqld is found
if $::nagios_mysqld {
mysql_user { 'nagios@localhost':
ensure => 'present',
password_hash => mysql_password('mysupersecretpassword'),
}
mysql_grant { 'nagios@localhost/*.*':
user => 'nagios@localhost',
table => '*.*',
privileges => [ 'REPLICATION CLIENT' ],
require => Mysql_user['nagios@localhost'],
}
}
# In hieradata
nagios::check::mysql_health::args: '--username nagios --password mysupersecretpassword'
The single mysql_health
script has many different 'modes', which are all
enabled by default. Because hyphens shouldn't be used in puppet variable names,
we use underscores instead in their names.
You can either selectively disable some :
# Disable some checks (modes)
nagios::check::mysql_health::modes_disabled:
- 'slave_io_running'
- 'slave_lag'
- 'slave_sql_running'
Or selectively enable some :
# Enable only the following checks (modes)
nagios::check::mysql_health::modes_enabled:
- 'connection_time'
- 'open_files'
- 'uptime'
Then for each mode, you can also pass some arguments, typically to change the warning and critical values as needed :
# Tweak some check values
nagios::check::mysql_health::args_connection_time: '--warning 5 --critical 10'
The postgres
checks are very similar to the mysql_health
ones. The single
postgres
script has many 'actions' ('modes'), which are enabled by default.
You can either selectively disable some :
# Disable some checks (modes)
nagios::check::postgres::modes_disabled:
- 'new_version_pg'
- 'new_version_tnm'
- 'pgb_pool_cl_active'
Or selectively enable some :
# Enable only the following checks (modes)
nagios::check::postgres::modes_enabled:
- 'locks'
- 'logfile'
- 'query_time'
Then for each mode, you can also pass some arguments, typically to change the warning and critical values as needed :
Note: The following pgbouncer related checks (modes) are disabled by default if no pgbouncer installation is detected:
pgb_pool_cl_active
pgb_pool_cl_waiting
pgb_pool_sv_active
pgb_pool_sv_idle
pgb_pool_sv_used
pgb_pool_sv_tested
pgb_pool_sv_login
pgb_pool_maxwait
pgbouncer_backends
pgbouncer_checksum
# Tweak some check values
nagios::check::postgres::args_query_time: '--warning=20s --critical=2m'
It is also possible to specify some custom query checks :
nagios::check::postgres::custom_queries:
custom_query_1:
query: 'SELECT SUBSTRING(version(), 12, 5)'
valtype: 'string'
warning: '9.4.4'
custom_query_2:
query: "SELECT SUBSTRING(version(), 12, 1) AS result"
warning: 9
valtype: 'integer'
reverse: true
For more info please refer to the check_postgres
nagios plugin
documentation : https://bucardo.org/check_postgres/check_postgres.pl.html
The mongodb
checks are very similar to the mysql_health
ones. The single
mongodb
script has many 'actions' ('modes'), which are enabled by default.
They may be enabled and disabled individually, or in groups of relevant
checks, for instance all replication checks at once.
You will need to create the monitoring user and set the information :
nagios::check::mongodb::user: 'nagios'
nagios::check::mongodb::pass: 'mysupersecretpassword'
db.createUser(
{
user: "nagios",
pwd: "mysupersecretpassword",
roles:
[
"readAnyDatabase",
"clusterMonitor",
{ role: "readWrite", db: "local" },
{ role: "readWrite", db: "nagios" }
]
}
)
You can completely disable MongoDB monitoring for some nodes :
nagios::check::mongodb::ensure: 'absent'
You can selectively disable some :
# Disable some checks (modes)
nagios::check::mongodb::modes_disabled:
- 'oplog'
- 'queries_per_second'
- 'queues'
Or selectively enable some :
# Enable only the following checks (modes)
nagios::check::mongodb::modes_enabled:
- 'connect'
- 'page_faults'
Or disable entire groups of non-relevant checks :
# Disable non relevant checks
nagios::check::mongodb::mmapv1: false
nagios::check::mongodb::v2: false
nagios::check::mongodb::replication: false
nagios::check::mongodb::sharding: false
For an arbiter, you can disable all non-relevant checks :
nagios::check::mongodb::arbiter: true
Then for each mode, you can also pass some arguments, typically to change the warning and critical values as needed :
# Tweak some check values
nagios::check::mongodb::args_connect: '-W 2 -C 4'
nagios::check::mongodb::args_connections: '-W 70 -C 80'
nagios::check::mongodb::args_memory: '-W 8 -C 16'
nagios::check::mongodb::args_opcounters: '-W 10000 -C 50000'
nagios::check::mongodb::args_replication_lag: '-W 15 -C 30'
For more info please refer to the nagios-plugin-mongodb
documentation :
https://github.com/mzupan/nagios-plugin-mongodb
The zookeeper
checks are very similar to the mysql_health
ones. The single
zookeeper
script has many 'actions' ('keys'), which are enabled by default.
You can either selectively disable some :
# Disable some checks (keys)
nagios::check::zookeeper::keys_disabled:
- 'zk_max_latency'
- 'zk_outstanding_requests'
Or selectively enable some :
# Enable only the following checks (keys)
nagios::check::zookeeper::keys_enabled:
- 'zk_avg_latency'
- 'zk_open_file_descriptor_count'
Then for each key, you can also pass some arguments, typically to change the warning and critical values as needed :
# Tweak some check values
nagios::check::zookeeper::zk_avg_latency: '--warning=1 --critical=10'
When nagios::check::zookeeper::leader
is set to true
, the following
additional checks are enabled by default:
zk_pending_syncs
zk_synced_followers
For more info please refer to the check_zookeeper
nagios plugin
documentation: https://github.com/andreisavu/zookeeper-monitoring
If you want to have your nagios notifications in Slack, enable the slack plugin
by setting $plugin_slack
parameter to true
.
Additional parameters:
$plugin_slack_domain
(mandatory): your team's Slack domain$plugin_slack_token
(mandatory): the token from your Nagios services page
Sample Slack contact and commands configuration:
nagios_contact { 'slack':
alias => 'Slack',
service_notification_period => '24x7',
host_notification_period => '24x7',
service_notification_options => 'w,u,c,r',
host_notification_options => 'd,r',
service_notification_commands => 'notify-service-by-slack',
host_notification_commands => 'notify-host-by-slack',
}
nagios_command { 'notify-service-by-slack':
command_line => '$USER1$/slack_nagios -field slack_channel=#exads-monitoring-nagios -field HOSTALIAS="$HOSTNAME$" -field SERVICEDESC="$SERVICEDESC$" -field SERVICESTATE="$SERVICESTATE$" -field SERVICEOUTPUT="$SERVICEOUTPUT$" -field NOTIFICATIONTYPE="$NOTIFICATIONTYPE$"',
}
nagios_command { 'notify-host-by-slack':
command_line => '$USER1$/slack_nagios -field slack_channel=#exads-monitoring-nagios -field HOSTALIAS="$HOSTNAME$" -field HOSTSTATE="$HOSTSTATE$" -field HOSTOUTPUT="$HOSTOUTPUT$" -field NOTIFICATIONTYPE="$NOTIFICATIONTYPE$"',
}
Then the slack
contact should be added to the $admins_members
parameter of
the main nagios::server
class.
For the rabbitmq
based checks to work, you will need to create a nagios user
on your rabbit servers
Example :
# This could go in site.pp, the fact is present only if rabbitmq-server is found
if $::nagios_rabbitmq {
rabbitmq_user { 'nagios':
password => 'mysupersecretpassword',
tags => ['monitoring'],
}
rabbitmq_user_permissions { 'nagios@/':
read_permission => '.*',
}
}
# In hieradata
nagios::check::rabbitmq::user: 'nagios'
nagios::check::rabbitmq::pass: 'mysupersecretpassword'
The single rabbitmq
script has many different 'modes', which are all
enabled by default.
You can either selectively disable some :
# Disable some checks (modes)
nagios::check::rabbitmq::modes_disabled:
- 'aliveness'
- 'cluster_status'
Or selectively enable some :
# Enable only the following checks (modes)
nagios::check::rabbitmq::modes_enabled:
- 'cluster_status'
- 'connection_count'
- 'queues_count'
- 'mem_usage'
Then for each mode, you can also pass some arguments, typically to change the warning and critical values as needed :
# Tweak some check values
nagios::check::rabbitmq::connection_count: '-C 100 -W 50'
The single redis
script has many different 'modes', which are all
enabled by default.
You can either selectively disable some :
# Disable some checks (modes)
nagios::check::redis::modes_disabled:
- 'connected_slaves'
- 'blocked_clients'
Or selectively enable some :
# Enable only the following checks (modes)
nagios::check::redis::modes_enabled:
- 'hitrate'
- 'response_time'
- 'rejected_connections'
- 'uptime_in_seconds'
Then for each mode, you can also pass some arguments, typically to change the warning and critical values as needed :
# Tweak some check values
nagios::check::redis::args_response_time: '0.005,0.010' # Warning,Critical
nagios::check::redis::args_hitrate: '70,100' # Warning,Critical
The redis_sentinel
check is enable by default. This check will validate the number of healthy redis slaves and sentinels
You can define the master:
nagios::check::redis_sentinel::master: 'MyAwesomeMaster'
You can also pass some arguments, typically to change the warning and critical values as needed :
# Tweak some check values
nagios::check::redis_sentinel::args: '-c 0,2 -w 0,2' # Slaves,Sentinels
If you want to monitor multiple redis databases on a single host you must use the definition nagios::check::redis::mdbs
nagios::check::redis_mdbs { 'db_name' :
fqdn => 'host.domain.foo',
port => 'port_num',
modes => {
'connected_clients' => '100,200', # Mode => 'Warning,Critical'
'evicted_keys' => '10,20',
'rejected_connections' => '20,50',
},
}
On the other hand, if you want monitor multiple sentinels on a single host you must use the definiton nagios::check::redis_sentinel_mmasters
nagios::check::redis_sentinel_mmasters { 'sentinel_master':
port => 'port_num',
fqdn => 'host.domain.foo',
}
Note: In these kinds of scenarios the plugins will run on the Nagios Server. The nrpe agent won't be used to perform these checks.
RHEL IDM manages few services. In this module we only monitor the following:
- IDM status via
/usr/sbin/ipactl status
- IDM replication
- KRB status
IDM general status monitoring is enabled by default if the file /usr/sbin/ipactl
is found on the server.
You can enable the others with the following yaml setup:
# IDM Replication
nagios::check::ipa_replication::bind_dn: 'uid=nagios_user,cn=users,cn=accounts,dc=dummy,dc=domain,dc=com'
nagios::check::ipa_replication::bind_pass: 'mysupersecretpassword'
# KRB status
nagios::check::krb5::keytab: '/path/to/auth/keytab/file'
nagios::check::krb5::principal: 'nagios_user'
nagios::check::krb5::realm: 'DUMMY.DOMAIN.COM'
In order to get the keytab file please read the documentation of the ktutils
command
The elastic search monitoring has many different 'modes', which are all enabled by default.
You can either selectively disable some :
# Disable some checks (modes)
nagios::check::elasticsearch::modes_disabled:
- 'split_brain'
- 'unassigned_shards'
Or selectively enable some :
# Enable only the following checks (modes)
nagios::check::elasticsearch::modes_enabled:
- 'cluster_status'
- 'nodes'
- 'jvm_usage'
Then for each mode, you can also pass some arguments, typically to change the warning and critical values as needed :
# Tweak some check values
nagios::check::elasticsearch::args_jvm_usage: '-N 10.0.0.1 -C 90 -W 80'
nagios::check::elasticsearch::args_nodes: '-E 5' # Expected nodes in cluester
The OpenSearch monitoring integration offers several 'modes' for comprehensive health and performance checks of your OpenSearch clusters. By default, all checks are enabled.
You have the flexibility to customize which checks are active either by disabling specific modes or enabling a select few:
If you prefer to exclude certain checks, specify them in the modes_disabled
configuration. For example, to disable JVM usage and disk usage checks:
# Disable some checks (modes)
nagios::check::opensearch::modes_disabled:
- 'jvm_usage'
- 'disk_usage'
Alternatively, you can enable only specific checks by listing them in the modes_enabled
configuration. This is useful for focusing monitoring on particular areas of interest or concern:
# Enable only the following checks (modes)
nagios::check::opensearch::modes_enabled:
- 'cluster_status'
- 'nodes'
- 'unassigned_shards'
For each mode, you can pass arguments to adjust warning and critical threshold values or specify other options such as node names or expected node counts. This customization allows you to tailor monitoring to your cluster's specific needs and operational parameters:
# Example setting critical and warning for jvm_usage and 5 expected nodes
nagios::check::opensearch::mode_args:
jvm_usage: '-C 90 -W 80'
nodes: '-E 5'
Available Checks (Modes):
cluster_status
: Checks the overall health status of the OpenSearch cluster.nodes
: Monitors the number of active nodes against expected counts.unassigned_shards
: Alerts on unassigned shards within the cluster.disk_usage
: Monitors disk space usage on cluster nodes.jvm_usage
: Assesses JVM memory usage across nodes.thread_pool_queues
: Checks the sizes of thread pool queues to identify high load conditions.no_replica_indices
: Check for indices without replicas.node_uptime
: Check if node uptime is less than 10 minutes.check_disk_space_for_resharding
: Check disk space for index resharding.
The Fluent Bit monitoring uses the /api/v1/health
endpoint to determine the health status of Fluent Bit.
The health check is performed by a custom script named check_fluentbit_health.sh
,
which is located by default at /usr/lib64/nagios/plugins/
.
This script checks the health of the Fluent Bit service by accessing the API endpoint /api/v1/health
,
which provides health status directly from the Fluent Bit service.
To utilize the health check functionality, Fluent Bit must be configured to enable its built-in HTTP server,
which exposes the /api/v1/health
endpoint.
For more information please check Health Check for Fluent Bit
By default, localhost and port 2020 are used for the check but you can customize it from hiera:
# Fluent Bit health check parameters
nagios::check::fluentbit::args: '-H your-fluentbit-host -p your-port-number'
Kafka monitoring checks producing to and consuming from specific Kafka topic, which can be set from hiera:
nagios::check::kafka::topic: 'nagios'
By default, random partition of the above topic is used on each run, but this
behavior may be changed using -p
parameter of the check script.
nagios::check::kafka::args: '-p 1'
By default localhost:9092
is used as a broker list, but it may be controlled:
nagios::check::kafka::brokers:
- '1.1.1.1:9093'
- '1.1.1.2:9093'
There is also ISR Kafka check, which stands for 'In-Sync Replicas'. It checks for under-replicated partitions.
The accepted parameters are address or host of the zookeeper server and zookeeper chroot:
nagios::check::kafka_isr::zookeeper_ipaddr:
- '1.1.1.1'
- '1.1.1.2'
nagios::check::kafka_isr::zookeeper_chroot: 'my-cluster'
HAProxy-stats check the state of Frontend, Servers and Backend.
If you need more options just change the args as documented in the plugin page:
nagios::check::haproxy_stats::args: '-s /var/lib/haproxy/stats -P statistics -m'
https://github.com/tatref/nagios-scripts/
Consul check which connects to Consul health API in order to gather node's state.
The default data center to check is '', but it may be controlled via hiera:
nagios::check::consul::datacenter: 'ovh'
If using ACLs, also consider passing the token with read permissions on node level:
nagios::check::consul::token: 'foo'
Check status of system services for Linux, FreeBSD, OSX, and AIX.
Add the following on the client manifest:
nagios::check::service { 'foo_service': }
Check the Syncthing status via de API You'll need to they API key from your node.
nagios::check::syncthing::api_key: 'l0r2m1psumd0l0rs1t4m3tC0ns3ct3tur'
You can enable and disable modes as with the previous checks with
the modes_enabled
and modes_disabled
parameters
# Enable only the following checks (modes)
nagios::check::syncthing::modes_enabled:
- 'alive'
- 'devices'
- 'folders_status'
# Disable only the following checks (modes)
nagios::check::syncthing::modes_disabled:
- 'last_scans'
- 'folders_status'
check_ups provided by nagios-plugins-ups from https://www.nagios-plugins.org/
If you need more options just change the args as documented in the plugin page:
nagios::check::ups::args: '-H 127.0.0.1 -u nutdev1'
If you want to define a custom service (non-NRPE) without modifying module code:
nagios::server::commands:
check_dns_addr:
command_line: "$USER1$/check_dns -H $ARG1$ $ARG2$"
nagios::client::services:
"check_command_tcp_port_8888_%{::fqdn}":
check_command: 'check_tcp!8888'
service_description: 'TCP port 8888'
"check_dns_hostname_%{::fqdn}":
check_command: 'check_dns_addr!$HOSTNAME$!-a $HOSTADDRESS$'
service_description: 'DNS Hostname'
If you want to monitor a custom service (via NRPE) without modifying module code, use the following hieradata definitions:
# define server command that uses your custom plugin
# make sure the plugin exists at ${module_name}/templates/plugins/check_command
nagios::server::commands:
check_nrpe_command:
command_line: "%{::nagios::params::nrpe_command} %{::nagios::params::nrpe_options} -c check_command"
# define NRPE plugin to be installed on a client
nagios::client::nrpe_plugins:
check_command:
ensure: 'present'
# define NRPE file to be delivered to a client
nagios::client::nrpe_files:
check_command:
ensure: 'present'
plugin: 'check_command'
args: '-w 600 -c 900'
sudo: true
# finally, define a service that uses our new custom NRPE plugin
nagios::client::services:
"service_name_%{::fqdn}":
check_command: 'check_nrpe_command'
service_description: 'service description'
contact_groups: 'all'
Having multiple client services/nrpe_files/nrpe_plugins definitions (e.g. multiple Hiera roles), you might want to change Hiera merge behaviour, e.g.:
lookup_options:
nagios::client::nrpe_plugins:
merge:
strategy: deep
merge_hash_arrays: true
nagios::client::nrpe_files:
merge:
strategy: deep
merge_hash_arrays: true
nagios::client::services:
merge:
strategy: deep
merge_hash_arrays: true
knockout_prefix: '--'
## Host/Service Escalation rules support
In order to define a Host/Service escalation rule use the hierdata template below:
```yaml
# The easiest way to escalate service is to use hostgroups
nagios::server::hostescalation:
orca-hostescalation:
hostgroup_name: 'all'
contact_groups: 'oncall,backup-oncall'
first_notification: '6'
last_notification: '0'
notification_interval: '15'
escalation_options: 'd,u,r'
escalation_period: '24x7'
# The easiest way to escalate service is to use servicegroups
nagios::server::serviceescalation:
orca-serviceescalation:
servicegroup_name: 'escalation'
contact_groups: 'oncall,backup-oncall'
first_notification: '6'
last_notification: '0'
notification_interval: '15'
escalation_options: 'w,u,c,r'
escalation_period: '24x7'
If you decommission a Nagios-monitored host a couple of manual steps are required to clean up.
# On the Puppet Master
puppet node deactivate <my_host>
# On the Nagios server
puppet agent -t
service nagios reload|restart
See Issue #21 on why the service restart is required.