Manually configure metrics collection on a *nix host for Splunk App for Infrastructure
To configure data collection, you must log in to an account with permissions to use sudo for root access. Don't log in as the root user.
Manually install the collectd agent to collect system metrics on a *nix host instead of using the data collection script in the Splunk App for Infrastructure (SAI) if at least one of these things is true for your environment:
- You're installing collectd on a closed network.
- You already have collectd on the host from which you want to collect data.
- You don't have trusted URLs to download the required packages and dependencies from.
For more information about collectd, see About using collectd.
The following steps show you how to set up collectd to forward metrics data for these *nix host metrics to SAI:
- CPU
- Memory
- File system usage
- System load
- Disk performance
- Interface traffic
- Process monitoring
- Docker containers
You can collect additional metrics with collectd, but SAI provides visualizations for these metrics.
If you manually configure metrics collection, you also need to manually configure log collection. For more information, see Manually configure log collection on a *nix host for Splunk App for Infrastructure.
You can also configure collectd to forward metrics data to a local universal forwarder. For more information, see Send collectd data to a universal forwarder.
Prerequisites
Before configuring metrics collection manually, confirm your system is compatible and has the required dependencies. For more information, see *nix data collection requirements.
Steps
Follow these steps to collect metrics for a *nix host with collectd.
1. Install collectd
If you haven't already installed collectd on your host, install one of these versions of collectd, depending on what's compatible with your operating system:
- 5.7.x
- 5.8.x
- 5.9.x
- 5.10.x
If you have an earlier version of collectd, update to a compatible version now.
For a full list of collectd install commands for supported operating systems, see collectd package sources, install commands, and locations.
2. Install the libcurl package
If you have not already installed the libcurl package on your system, install it now. For Linux systems, the version of libcurl you have to install depends on the Linux OS version you're running.
To install libcurl3 on a Debian 7, 8, or 9 system or on a Ubuntu 14 or 16 system, enter:
$ sudo apt-get install libcurl3
To install libcurl4 on a Debian 10 or on a Ubuntu 18, 04, 18.10 or 19 system, enter:
$ sudo apt-get install libcurl4
To install libcurl on a CentOS, Redhat, or Fedora system, enter:
$ sudo yum install libcurl
To install libcurl on a SUSE or openSUSE system, enter:
$ sudo zypper install libcurl4
To install libcurl on a Solaris system, enter:
$ pkgadd -d http://get.opencsw.org/now $ /opt/csw/bin/pkgutil -U $ /opt/csw/bin/pkgutil -i libcurl4_feature $ /usr/sbin/pkgchk -L CSWlibcurl4-feature # list files
3. (Optional) Install the libyajl package
If you monitor Docker containers on a Linux or Mac OS X host, you must have the libyajl version 2 package on your system. If you don't already have the package, install it now.
To install libyajl on a Debian or Ubuntu system, enter:
$ apt-get install libyajl2
To install libyajl on a CentOS, Redhat, or Fedora system, enter:
$ yum install yajl
To install libyajl on a SUSE or openSUSE system, enter:
$ zypper install libyajl2
To install libyajl on a Mac OS X system, enter:
$ brew install yajl
4. Copy the plug-ins to collectd's plug-in directory
For information about plug-in locations, see collectd package sources, install commands, and locations. You cannot monitor Docker containers on Solaris systems that you deployed without an orchestration tool.
The write_splunk
collectd plug-in is a replacement for the write_http
plug-in that directs metrics data to the Splunk HTTP Event Collector (HEC).
write_splunk
creates five dimensions for the corresponding entity when you integrate a system:
host
ip
os
os_version
kernel_version
You cannot delete the dimensions the plug-in creates.
If you want to monitor process metrics, copy the processmon.so
plug-in. The processmon.so
plugin adds the following dimensions:
- process_name: The name of the process.
- pid: The process ID.
- user: The user associated with the process.
You can split processmon metrics by dimension. Suppose you want to view the process by cpu usage. In SAI, you can split processmon.cpu.percent
metrics by process_name
.
If you monitor Docker containers on Linux systems, copy the docker.so
plug-in.Use the docker.so
plug-in to monitor Docker containers you didn't deploy with an orchestration tool such as Docker Swarm, Kubernetes, or OpenShift. If you want to monitor Docker containers you deployed with Kubernetes or OpenShift, see these topics:
- Collect Kubernetes metrics and logs with Splunk App for Infrastructure
- Collect OpenShift metrics and logs with Splunk App for Infrastructure
If you're monitoring a Linux system, the plug-in locations depend on which version of libcurl you're using. See the following commands for each operating system and plug-in.
Debian/Ubuntu with libcurl3 and collectd 5.7.x or 5.8.x
Collectd plug-in | Install commands |
---|---|
write_splunk
|
$ wget https://<hostname>:8000/en-US/static/app/splunk_app_infrastructure/unix_agent/unix-agent.tgz $ tar xvzf unix-agent.tgz $ cp unix-agent/plugin_5_7_5_8/write_splunk.so <plug-in_directory> |
processmon
|
$ wget https://<hostname>:8000/en-US/static/app/splunk_app_infrastructure/unix_agent/unix-agent.tgz $ tar xvzf unix-agent.tgz $ cp unix-agent/plugin_5_7_5_8/processmon.so <plug-in_directory> |
docker
|
$ wget https://<hostname>:8000/en-US/static/app/splunk_app_infrastructure/unix_agent/unix-agent.tgz $ tar xvzf unix-agent.tgz $ cp unix-agent/plugin_5_7_5_8/docker.so <plug-in_directory> |
Debian/Ubuntu with libcurl4 and collectd 5.7.x or 5.8.x
Collectd plug-in | Install commands |
---|---|
write_splunk
|
$ wget https://<hostname>:8000/en-US/static/app/splunk_app_infrastructure/unix_agent/unix-agent.tgz $ tar xvzf unix-agent.tgz $ cp unix-agent/plugin_5_7_5_8/deb_libcurl4/write_splunk.so <plug-in_directory> |
processmon
|
$ wget https://<hostname>:8000/en-US/static/app/splunk_app_infrastructure/unix_agent/unix-agent.tgz $ tar xvzf unix-agent.tgz $ cp unix-agent/plugin_5_7_5_8/deb_libcurl4/processmon.so <plug-in_directory> |
docker
|
$ wget https://<hostname>:8000/en-US/static/app/splunk_app_infrastructure/unix_agent/unix-agent.tgz $ tar xvzf unix-agent.tgz $ cp unix-agent/plugin_5_7_5_8/deb_libcurl4/docker.so <plug-in_directory> |
Debian/Ubuntu with collectd 5.9.x or 5.10.x
Collectd plug-in | Install commands |
---|---|
write_splunk
|
$ wget https://<hostname>:8000/en-US/static/app/splunk_app_infrastructure/unix_agent/unix-agent.tgz $ tar xvzf unix-agent.tgz $ cp unix-agent/plugin_5_9_5_10/write_splunk.so <plug-in_directory> |
processmon
|
$ wget https://<hostname>:8000/en-US/static/app/splunk_app_infrastructure/unix_agentunix-agent.tgz $ tar xvzf unix-agent.tgz $ cp unix-agent/plugin_5_9_5_10/processmon.so <plug-in_directory> |
docker
|
$ wget https://<hostname>:8000/en-US/static/app/splunk_app_infrastructure/unix_agent/unix-agent.tgz $ tar xvzf unix-agent.tgz $ cp unix-agent/plugin_5_9_5_10/docker.so <plug-in_directory> |
RHEL, CentOS, Fedora, SUSE, or openSUSE with collectd 5.7.x or 5.8.x
Collectd plug-in | Install commands |
---|---|
write_splunk
|
$ wget https://<hostname>:8000/en-US/static/app/splunk_app_infrastructure/unix_agent/unix-agent.tgz $ tar xvzf unix-agent.tgz $ cp unix-agent/plugin_5_7_5_8/write_splunk.so <plug-in_directory> |
processmon
|
$ wget https://<hostname>:8000/en-US/static/app/splunk_app_infrastructure/unix_agent/unix-agent.tgz $ tar xvzf unix-agent.tgz $ cp unix-agent/plugin_5_7_5_8/processmon.so <plug-in_directory> |
docker
|
$ wget https://<hostname>:8000/en-US/static/app/splunk_app_infrastructure/unix_agent/unix-agent.tgz $ tar xvzf unix-agent.tgz $ cp unix-agent/plugin_5_7_5_8/docker.so <plug-in_directory> |
RHEL, CentOS, Fedora, SUSE, or openSUSE with collectd 5.9.x or 5.10.x
Collectd plug-in | Install commands |
---|---|
write_splunk
|
$ wget https://<hostname>:8000/en-US/static/app/splunk_app_infrastructure/unix_agent/unix-agent.tgz $ tar xvzf unix-agent.tgz $ cp unix-agent/plugin_5_9_5_10/write_splunk.so <plug-in_directory> |
processmon
|
$ wget https://<hostname>:8000/en-US/static/app/splunk_app_infrastructure/unix_agent/unix-agent.tgz $ tar xvzf unix-agent.tgz $ cp unix-agent/plugin_5_9_5_10/processmon.so <plug-in_directory> |
docker
|
$ wget https://<hostname>:8000/en-US/static/app/splunk_app_infrastructure/unix_agent/unix-agent.tgz $ tar xvzf unix-agent.tgz $ cp unix-agent/plugin_5_9_5_10/docker.so <plug-in_directory> |
Solaris with collectd 5.7.x or 5.8.x
Collectd plug-in | Install commands |
---|---|
write_splunk
|
$ wget https://<hostname>:8000/en-US/static/app/splunk_app_infrastructure/unix_agent/unix-agent.tgz $ tar xvzf unix-agent.tgz $ cp unix-agent/plugin_5_7_5_8/write_splunk-solaris.so "/opt/csw/lib/collectd/write_splunk.so" |
Solaris with collectd 5.9.x or 5.10.x
Collectd plug-in | Install commands |
---|---|
write_splunk
|
$ wget https://<hostname>:8000/en-US/static/app/splunk_app_infrastructure/unix_agent/unix-agent.tgz $ tar xvzf unix-agent.tgz $ cp unix-agent/plugin_5_9_5_10/write_splunk-solaris.so "/opt/csw/lib/collectd/write_splunk.so" |
Mac OS X with collectd 5.7.x or 5.8.x
Collectd plug-in | Install commands |
---|---|
write_splunk
|
$ curl -ksL -o osx-agent.tgz http://waitomo-next.sv.splunk.com:/static/app/splunk_app_infrastructure/osx_agent/osx-agent.tgz $ tar -xzf osx-agent.tgz $ cp osx-agent/plugin_5_7_5_8/write_splunk.so <plug-in_directory> |
docker
|
$ curl -ksL -o osx-agent.tgz http://waitomo-next.sv.splunk.com:/static/app/splunk_app_infrastructure/osx_agent/osx-agent.tgz $ tar -xzf osx-agent.tgz $ cp osx-agent/plugin_5_7_5_8/docker.so <plug-in_directory> |
Mac OS X with collectd 5.9.x or 5.10.x
Collectd plug-in | Install commands |
---|---|
write_splunk
|
$ curl -ksL -o osx-agent.tgz http://waitomo-next.sv.splunk.com:/static/app/splunk_app_infrastructure/osx_agent/osx-agent.tgz $ tar -xzf osx-agent.tgz $ cp osx-agent/plugin_5_9_5_10/write_splunk.so <plug-in_directory> |
docker
|
$ curl -ksL -o osx-agent.tgz http://waitomo-next.sv.splunk.com:/static/app/splunk_app_infrastructure/osx_agent/osx-agent.tgz $ tar -xzf osx-agent.tgz $ cp osx-agent/plugin_5_9_5_10/docker.so <plug-in_directory> |
5. Configure collectd.conf to send data to the Splunk App for Infrastructure
To configure collectd.conf
, you must add the <Plugin write_splunk>
plug-in, add plug-ins for every other metric you want to monitor, and modify the Hostname
field. For information about required plug-in locations, see collectd package sources, install commands, and locations.
- Add a LoadPlugin for each plug-in you want to use.
<LoadPlugin "write_splunk"> FlushInterval 30 </LoadPlugin> LoadPlugin cpu LoadPlugin uptime LoadPlugin memory LoadPlugin df LoadPlugin load LoadPlugin disk LoadPlugin interface LoadPlugin docker LoadPlugin processmon
- Add metrics configuration stanzas for each metric you want to collect from the host. The following stanzas are default stanzas the data collection script configures. There isn't a stanza for the
uptime
metric. A stanza for theprocessmon
plug-in is optional. Include aprocessmon
stanza to specify allow lists and deny lists and report IO metrics for monitored processes. The followingprocessmon
stanza is just an example that includes the settings you can configure.Plug-in Supported OS Stanza write_splunk - Linux
- Solaris
- Mac OS X
<Plugin write_splunk> server "<receiving_server>" port "<HEC PORT>" token "<HEC TOKEN>" ssl true verifyssl false Dimension "entity_type:nix_host" Dimension "key2:value2" </Plugin>
server
: The IP or hostname of the Splunk deployment to which you are sending data. If you are sending data to a distributed deployment, the IPs or hostnames of the indexer. If you deploy a load balancer, the IP or hostname of the load balancer.port
: The HEC port.token
: The HEC token.
CPU - Linux
- Solaris
- Mac OS X
<Plugin cpu> ReportByCpu false ReportByState true ValuesPercentage true </Plugin>
Memory - Linux
- Solaris
- Mac OS X
<Plugin memory> ValuesAbsolute false ValuesPercentage true </Plugin>
DF - Linux
- Solaris
- Mac OS X
<Plugin df> FSType "ext2" FSType "ext3" FSType "ext4" FSType "XFS" FSType "rootfs" FSType "overlay" FSType "hfs" FSType "apfs" FSType "zfs" FSType "ufs" ReportByDevice true ValuesAbsolute false ValuesPercentage true IgnoreSelected false </Plugin>
Load - Linux
- Solaris
- Mac OS X
<Plugin load> ReportRelative true </Plugin>
Disk - Linux
- Solaris
- Mac OS X
<Plugin disk> Disk "" IgnoreSelected true UdevNameAttr "DEVNAME" </Plugin>
Interface - Linux
- Solaris
- Mac OS X
<Plugin interface> IgnoreSelected true </Plugin>
Docker - Linux
- Mac OS X
<Plugin docker> dockersock "/var/run/docker.sock" apiversion "v1.20" </Plugin>
By default, collectd fails if you're running more than 100 Docker containers. To monitor more 100 or more Docker containers, add the
ReadBufferSize
parameter to the Docker plug-in. The max value is32000
.Process monitoring - Linux
<Plugin processmon> ReadIo true whitelist "process1.*" whitelist "process2.*" blacklist "process3.*" </Plugin>
This plug-in is optional. If you don't configure aprocessmon
stanza, the plug-in monitors every process and doesn't collect IO metrics.
If you both deny and allow a process, the plug-in denies the process.
The plug-in uses POSIX Extended Regular Expression syntax for the regular expression you enter to allow or deny processes.
The plug-in uses thecomm
field in/proc/[pid]/stat
for process names. For more information, see the Linux Programmer's Manual. - Update the
Hostname
field with the IP or hostname of the system that's running collectd. TheHostname
must be unique to the system because it is used to identify the entity in SAI.
Optionally, you can also add dimensions as Dimension "key:value"
to the write_splunk plug-in.
6. Start collectd
Start collectd on Linux systems:
$ sudo service collectd restart
Start collectd on Solaris systems:
$ sudo svcadm enable cswcollectd
Start collectd on Mac OS X systems:
$ sudo brew services restart collectd
Example metrics collection configuration file
Here is an example collectd.conf
file that includes every plug-in ITSI entity integrations use.
# # Config file for collectd(1). # Please read collectd.conf(5) for a list of options. # http://collectd.org/ # ############################################################################## # Global # #----------------------------------------------------------------------------# # Global settings for the daemon. # ############################################################################## Hostname "collectd.server.sample" FQDNLookup false #BaseDir "/var/lib/collectd" #PIDFile "/var/run/collectd.pid" #PluginDir "/usr/lib64/collectd" #TypesDB "/usr/share/collectd/types.db" #----------------------------------------------------------------------------# # When enabled, plugins are loaded automatically with the default options # # when an appropriate <Plugin ...> block is encountered. # # Disabled by default. # #----------------------------------------------------------------------------# #AutoLoadPlugin false #----------------------------------------------------------------------------# # When enabled, internal statistics are collected, using "collectd" as the # # plugin name. # # Disabled by default. # #----------------------------------------------------------------------------# #CollectInternalStats false #----------------------------------------------------------------------------# # Interval at which to query values. This may be overwritten on a per-plugin # # base by using the 'Interval' option of the LoadPlugin block: # # <LoadPlugin foo> # # Interval 60 # # </LoadPlugin> # #----------------------------------------------------------------------------# Interval 60 #MaxReadInterval 86400 #Timeout 2 #ReadThreads 5 #WriteThreads 5 # Limit the size of the write queue. Default is no limit. Setting up a limit is # recommended for servers handling a high volume of traffic. WriteQueueLimitHigh 1000000 WriteQueueLimitLow 800000 ############################################################################## # Logging # #----------------------------------------------------------------------------# # Plugins which provide logging functions should be loaded first, so log # # messages generated when loading or configuring other plugins can be # # accessed. # ############################################################################## LoadPlugin syslog LoadPlugin logfile <LoadPlugin "write_splunk"> FlushInterval 30 </LoadPlugin> ############################################################################## # LoadPlugin section # #----------------------------------------------------------------------------# # Lines beginning with a single `#' belong to plugins which have been built # # but are disabled by default. # # # # Lines beginning with `##' belong to plugins which have not been built due # # to missing dependencies or because they have been deactivated explicitly. # ############################################################################## #LoadPlugin csv LoadPlugin cpu LoadPlugin uptime LoadPlugin memory LoadPlugin df LoadPlugin load LoadPlugin disk LoadPlugin interface LoadPlugin docker LoadPlugin processmon ############################################################################## # Plugin configuration # #----------------------------------------------------------------------------# # In this section configuration stubs for each plugin are provided. A desc- # # ription of those options is available in the collectd.conf(5) manual page. # ############################################################################## <Plugin logfile> LogLevel info File "/etc/collectd/collectd.log" Timestamp true PrintSeverity true </Plugin> <Plugin syslog> LogLevel info </Plugin> <Plugin cpu> ReportByCpu false ReportByState true ValuesPercentage true </Plugin> <Plugin memory> ValuesAbsolute false ValuesPercentage true </Plugin> <Plugin df> FSType "ext2" FSType "ext3" FSType "ext4" FSType "XFS" FSType "rootfs" FSType "overlay" FSType "hfs" FSType "apfs" FSType "zfs" FSType "ufs" ReportByDevice true ValuesAbsolute false ValuesPercentage true IgnoreSelected false </Plugin> <Plugin load> ReportRelative true </Plugin> <Plugin disk> Disk "" IgnoreSelected true UdevNameAttr "DEVNAME" </Plugin> <Plugin interface> IgnoreSelected true </Plugin> <Plugin docker> dockersock "/var/run/docker.sock" apiversion "v1.20" </Plugin> <Plugin processmon> ReadIo true whitelist "collectd" whitelist "bash" blacklist "splunkd" </Plugin> <Plugin write_splunk> server "<splunk infrastructure app server>" port "<HEC PORT>" token "<HEC TOKEN>" ssl true verifyssl false Dimension "entity_type:nix_host" </Plugin>
Manually configure log collection on a *nix host for Splunk App for Infrastructure | Send collectd data to a local universal forwarder |
This documentation applies to the following versions of Splunk® App for Infrastructure (Legacy): 2.0.2, 2.0.3, 2.0.4, 2.1.0, 2.1.1 Cloud only
Feedback submitted, thanks!