inputs.conf
The following are the spec and example files for inputs.conf
.
inputs.conf.spec
Version 8.0.3
OVERVIEW
This file contains possible settings you can use to configure inputs, distributed inputs such as forwarders, and file system monitoring in inputs.conf. Each stanza controls different search commands settings. There is a inputs.conf file in the $SPLUNK_HOME/etc/system/default/ directory. Never change or copy the configuration files in the default directory. The files in the default directory must remain intact and in their original location. To set custom configurations, create a new file with the name inputs.conf in the $SPLUNK_HOME/etc/system/local/ directory. Then add the specific settings that you want to customize to the local configuration file. For examples, see inputs.conf.example. You must restart the Splunk instance to enable configuration changes. To learn more about configuration files (including file precedence) see the documentation located at http://docs.splunk.com/Documentation/Splunk/latest/Admin/Aboutconfigurationfiles
GLOBAL SETTINGS
Use the [default] stanza to define any global settings. * You can also define global settings outside of any stanza, at the top of the file. * Each conf file should have at most one default stanza. If there are multiple default stanzas, settings are combined. In the case of multiple definitions of the same setting, the last definition in the file wins. * If an setting is defined at both the global level and in a specific stanza, the value in the specific stanza takes precedence. GENERAL SETTINGS: The following settings are valid for all input types (except file system change monitor, which is described in a separate section in this file). You must first enter a stanza header in square brackets, specifying the input type. See further down in this file for examples. Then, use any of the following settings. To specify global settings for Windows Event Log inputs, place them in the [WinEventLog] global stanza as well as the [default] stanza. host = <string> * Sets the host key/field to a static value for this input stanza. * The input uses this field during parsing and indexing. It also uses this field at search time. * As a convenience, the input prepends the chosen string with 'host::'. * If set to '$decideOnStartup', sets the field to the hostname of executing machine. This occurs on each splunkd startup. * If you run multiple instances of the software on the same machine (hardware or virtual machine), choose unique values for 'host' to differentiate your data, ex. myhost-sh-1 or myhost-idx-2. * Do not put the <string> value in quotes. Use host=foo, not host="foo". * If you set 'host' to "$decideOnStartup", you can further control how splunkd derives the hostname by using the 'hostnameOption' setting in server.conf. * For example, if you want splunkd to use the fully qualified domain name for the machine, set "host = $decideOnStartup" in inputs.conf and "hostnameOption = fullyqualifiedname" in server.conf. * More information on hostname options can be found in the server.conf specification file. * If you remove the 'host' setting from $SPLUNK_HOME/etc/system/local/inputs.conf or remove $SPLUNK_HOME/etc/system/local/inputs.conf, the setting changes to "$decideOnStartup". Apps that need a resolved host value should use the 'host_resolved' property in the response for the REST 'GET' call of the input source. This property is set to the hostname of the local Splunk instance. It is a read only property that is not written to inputs.conf. * Default: "$decideOnStartup", but at installation time, the setup logic adds the local hostname, as determined by DNS, to the $SPLUNK_HOME/etc/system/local/inputs.conf default stanza, which is the effective default value. index = <string> * Sets the index to store events from this input. * Primarily used to specify the index to store events that come in through this input stanza. * Default: "main" (or whatever you have set as your default index). source = <string> * Sets the source key/field for events from this input. * Detail: Sets the source key initial value. The key is used during parsing/indexing, in particular to set the source field during indexing. It is also the source field used at search time. * As a convenience, the chosen string is prepended with 'source::'. * Avoid overriding the source key. The input layer provides a more accurate string to aid in problem analysis and investigation, recording the file from which the data was retrieved. Consider using source types, tagging, and search wildcards before overriding this value. * Do not put the <string> value in quotes: Use source=foo, not source="foo". * Default: the input file path. sourcetype = <string> * Sets the sourcetype key/field for events from this input. * Explicitly declares the source type for this input instead of letting it be determined through automated methods. This is important for search and for applying the relevant configuration for this data type during parsing and indexing. * Sets the sourcetype key initial value. The key is used during parsing or indexing to set the source type field during indexing. It is also the source type field used at search time. * As a convenience, the chosen string is prepended with 'sourcetype::'. * Do not put the <string> value in quotes: Use sourcetype=foo, not sourcetype="foo". * If not set, the indexer analyzes the data and chooses a source type. * No default. queue = [parsingQueue|indexQueue] * Sets the queue where the input processor should deposit the events it reads. * Set to "parsingQueue" to apply the props.conf file and other parsing rules to your data. For more information about the props.conf file and rules timestamps and linebreaks, see the props.conf fileand the online documentation at http://docs.splunk.com/Documentation. * Set to "indexQueue" to send your data directly into the index. * Default: parsingQueue Pipeline Key defaulting. * Pipeline keys in general can be defaulted in inputs stanzas. * The list of user-available, modifiable pipeline keys is described in transforms.conf.spec. See transforms.conf.spec for further information on these keys. * The currently-defined keys which are available literally in inputs stanzas are as follows: queue = <value> _raw = <value> _meta = <value> _time = <value> * Inputs have special support for mapping host, source, sourcetype, and index to their metadata names such as host -> Metadata:Host * Defaulting these values is not recommended, and is generally only useful as a workaround to other product issues. * Defaulting these keys in most cases will override the default behavior of input processors, but this behavior is not guaranteed in all cases. * Values defaulted here, as with all values provided by inputs, can be altered by transforms at parse time. This section contains options for routing data using inputs.conf rather than outputs.conf. NOTE: concerning routing via inputs.conf: This is a simplified set of routing options you can use as data comes in. For more flexible options or details on configuring required or optional settings, see outputs.conf.spec. _TCP_ROUTING = <tcpout_group_name>,<tcpout_group_name>,<tcpout_group_name>, ... * A comma-separated list of tcpout group names. * This setting lets you selectively forward data to specific indexer(s). * Specify the tcpout group that the forwarder should use when forwarding the data. The tcpout group names are defined in outputs.conf with [tcpout:<tcpout_group_name>]. * To forward data to all tcpout group names that have been defined in outputs.conf, set to '*' (asterisk). * To forward data from the "_internal" index, you must explicitly set '_TCP_ROUTING' to either "*" or a specific splunktcp target group. * Default: The groups specified in 'defaultGroup' in [tcpout] stanza in outputs.conf file _SYSLOG_ROUTING = <syslog_group_name>,<syslog_group_name>,<syslog_group_name>, ... * A comma-separated list of syslog group names. * Using this, you can selectively forward the data to specific destinations as syslog events. * Specify the syslog group to use when forwarding the data. The syslog group names are defined in outputs.conf with [syslog:<syslog_group_name>]. * The destination host must be configured in outputs.conf, using "server=[<ip>|<servername>]:<port>". * Default: The groups present in "defaultGroup" in the [syslog] stanza in outputs.conf file _INDEX_AND_FORWARD_ROUTING = <string> * Only has effect if you use the 'selectiveIndexing' feature in outputs.conf. * If set for any input stanza, should cause all data coming from that input stanza to be labeled with this setting. * When 'selectiveIndexing' is in use on a forwarder: * data without this label will not be indexed by that forwarder. * data with this label will be indexed in addition to any forwarding. * This setting does not actually cause data to be forwarded or not forwarded in any way, nor does it control where the data is forwarded in multiple-forward path cases. * Default: not present.
Blacklist
[blacklist:<path>] * Protects files on the file system from being indexed or previewed. * The input treats a file as blacklisted if the file starts with any of the defined blacklisted <paths>. * Blacklisting of a file with the specified path occurs even if a monitor stanza defines a whitelist that matches the file path. * The preview endpoint returns an error when asked to preview a blacklisted file. * The oneshot endpoint and command also returns an error. * When a blacklisted file is monitored (monitor:// or batch://), the 'filestatus' endpoint shows an error. * For fschange with the 'sendFullEvent' option enabled, contents of blacklisted files are not indexed.
Valid input types follow, along with their input-specific settings:
MONITOR:
[monitor://<path>] * Configures a file monitor input to watch all files in <path>. * <path> can be an entire directory or a single file. * You must specify the input type and then the path, so put three slashes in your path if you are starting at the root on *nix systems (to include the slash that indicates an absolute path). Additional settings: host_regex = <regular expression> * If specified, <regular expression> extracts host from the path to the file for each input file. * Detail: This feature examines the source key; if source is set explicitly in the stanza, that string is matched, not the original filename. * Specifically, the first group of the regex is used as the host. * If the regex fails to match, the default 'host' setting is used. * If 'host_regex' and 'host_segment' are both set, the input ignores 'host_regex'. * No default. host_segment = <integer> * If set to N, Splunk software sets the Nth "/"-separated segment of the path as 'host'. * For example, if host_segment=3, the third segment is used. * If the value is not an integer or is less than 1, the default 'host' setting is used. * On Windows machines, the drive letter and colon before the backslash count as one segment. * For example, if you set host_segment=3 and the monitor path is D:\logs\servers\host01, Splunk software sets the host as "servers" because that is the third segment. * No default. whitelist = <regular expression> * If set, files from this input are monitored only if their path matches the specified regex. * Takes precedence over the deprecated '_whitelist' setting, which functions the same way. * No default. blacklist = <regular expression> * If set, files from this input are NOT monitored if their path matches the specified regex. * Takes precedence over the deprecated '_blacklist' setting, which functions the same way. * If a file matches the regexes in both the blacklist and whitelist settings, the file is NOT monitored. Blacklists take precedence over whitelists. * No default. Note concerning wildcards and monitor: * You can use wildcards to specify your input path for monitored inputs. Use "..." for recursive directory matching and "*" for wildcard matching in a single directory segment. * "..." recurses through directories. This means that /foo/.../bar matches foo/1/bar, foo/1/2/bar, etc. * You can use multiple "..." specifications in a single input path. For example: /foo/.../bar/... * The asterisk (*) matches anything in a single path segment; unlike "...", it does not recurse. For example, /foo/*/bar matches the files /foo/1/bar, /foo/2/bar, etc. However, it does not match /foo/bar or /foo/1/2/bar. A second example: /foo/m*r/bar matches /foo/mr/bar, /foo/mir/bar, /foo/moor/bar, etc. It does not match /foo/mi/or/bar. * You can combine "*" and "..." as needed: foo/.../bar/* matches any file in the bar directory within the specified path. crcSalt = <string> * Use this setting to force the input to consume files that have matching CRCs (cyclic redundancy checks). * By default, the input only performs CRC checks against the first 256 bytes of a file. This behavior prevents the input from indexing the same file twice, even though you might have renamed it, as with rolling log files, for example. Because the CRC is based on only the first few lines of the file, it is possible for legitimately different files to have matching CRCs, particularly if they have identical headers. * If set, <string> is added to the CRC. * If set to the literal string "<SOURCE>" (including the angle brackets), the full directory path to the source file is added to the CRC. This ensures that each file being monitored has a unique CRC. When 'crcSalt' is invoked, it is usually set to <SOURCE>. * Be cautious about using this setting with rolling log files; it could lead to the log file being re-indexed after it has rolled. * In many situations, 'initCrcLength' can be used to achieve the same goals. * Default: empty string. initCrcLength = <integer> * How much of a file, in bytes, that the input reads before trying to identify whether it is a file that has already been seen. You might want to adjust this if you have many files with common headers (comment headers, long CSV headers, etc) and recurring filenames. * Cannot be less than 256 or more than 1048576. * CAUTION: Improper use of this setting causes data to be re-indexed. You might want to consult with Splunk Support before adjusting this value - the default is fine for most installations. * Default: 256 (bytes). ignoreOlderThan = <non-negative integer>[s|m|h|d] * The monitor input compares the modification time on files it encounters with the current time. If the time elapsed since the modification time is greater than the value in this setting, Splunk software puts the file on the ignore list. * Files on the ignore list are not checked again until the Splunk platform restarts, or the file monitoring subsystem is reconfigured. This is true even if the file becomes newer again at a later time. * Reconfigurations occur when changes are made to monitor or batch inputs through Splunk Web or the command line. * Use 'ignoreOlderThan' to increase file monitoring performance when monitoring a directory hierarchy that contains many older, unchanging files, and when removing or blacklisting those files from the monitoring location is not a reasonable option. * Do NOT select a time that files you want to read could reach in age, even temporarily. Take potential downtime into consideration! * Suggested value: 14d, which means 2 weeks * For example, a time window in significant numbers of days or small numbers of weeks are probably reasonable choices. * If you need a time window in small numbers of days or hours, there are other approaches to consider for performant monitoring beyond the scope of this setting. * NOTE: Most modern Windows file access APIs do not update file modification time while the file is open and being actively written to. Windows delays updating modification time until the file is closed. Therefore you might have to choose a larger time window on Windows hosts where files may be open for long time periods. * Value must be: <number><unit>. For example, "7d" indicates one week. * Valid units are "d" (days), "h" (hours), "m" (minutes), and "s" (seconds). * No default, meaning there is no threshold and no files are ignored for modification time reasons followTail = [0|1] * Whether or not the input should skip past current data in a monitored file for a given input stanza. This lets you skip over data in files, and immediately begin indexing current data. * If you set to "1", monitoring starts at the end of the file (like *nix 'tail -f'). The input does not read any data that exists in the file when it is first encountered. The input only reads data that arrives after the first encounter time. * If you set to "0", monitoring starts at the beginning of the file. * This is an advanced setting. Contact Splunk Support before using it. * Best practice for using this setting: * Enable this setting and start the Splunk instance. * Wait enough time for the input to identify the related files. * Disable the setting and restart the instance. * Do not leave 'followTail' enabled in an ongoing fashion. * Do not use 'followTail' for rolling log files (log files that get renamed as they age) or files whose names or paths vary. * Default: 0 alwaysOpenFile = [0|1] * Opens a file to check whether it has already been indexed, by skipping the modification time/size checks. * Only useful for files that do not update modification time or size. * Only known to be needed when monitoring files on Windows, mostly for Internet Information Server logs. * Configuring this setting to "1" can increase load and slow indexing. Use it only as a last resort. * Default: 0. time_before_close = <integer> * The amount of time, in seconds, that the file monitor must wait for modifications before closing a file after reaching an End-of-File (EOF) marker. * Tells the input not to close files that have been updated in the past 'time_before_close' seconds. * Default: 3 multiline_event_extra_waittime = <boolean> * By default, the file monitor sends an event delimiter when: * It reaches EOF of a file it monitors and * The last character it reads is a newline. * In some cases, it takes time for all lines of a multiple-line event to arrive. * Set to "true" to delay sending an event delimiter until the time that the file monitor closes the file, as defined by the 'time_before_close' setting, to allow all event lines to arrive. * Default: false recursive = <boolean> * Whether or not the input monitors subdirectories that it finds within a monitored directory. * If you set this setting to "false", the input does not monitor sub-directories * Default: true followSymlink = <boolean> * Whether or not to follow any symbolic links within a monitored directory. * If you set this setting to "false", the input ignores symbolic links that it finds within a monitored directory. * If you set the setting to "true", the input follows symbolic links and monitors files at the symbolic link destination. * Additionally, any whitelists or blacklists that the input stanza defines also apply to files at the symbolic link destination. * Default: true. _whitelist = ... * DEPRECATED. * This setting is valid unless the 'whitelist' setting also exists. _blacklist = ... * DEPRECATED. * This setting is valid unless the 'blacklist' setting also exists.
BATCH ("Upload a file" in Splunk Web):
Use the 'batch' input for large archives of historic data. If you want to continuously monitor a directory or index small archives, use 'monitor' (see above). 'batch' reads in the file and indexes it, and then deletes the file on disk. [batch://<path>] * A one-time, destructive input of files in <path>. * This stanza must include the 'move_policy = sinkhole' setting. * This input reads and indexes the files, then DELETES THEM IMMEDIATELY. * For continuous, non-destructive inputs of files, use 'monitor' instead. Additional settings: move_policy = sinkhole * This setting is required. You *must* include "move_policy = sinkhole" when you define batch inputs. * This setting causes the input to load the file destructively. * CAUTION: Do not use the 'batch' input type for files you do not want to delete after indexing. * The 'move_policy' setting exists for historical reasons, but remains as a safeguard. As an administrator, you must explicitly declare that you want the data in the monitored directory (and its sub-directories) to be deleted after being read and indexed. host_regex = see MONITOR, above. host_segment = see MONITOR, above. crcSalt = see MONITOR, above. log_on_completion = <boolean> * When set to false, this setting prevents Splunk software from logging to splunkd.log when it indexes files with this input. * Default: true 'batch' inputs do not use the following setting: source = <string> followSymlink = <boolean> * Works similarly to the same setting for monitor, but does not delete files after following a symbolic link out of the monitored directory. The following settings work identically as for [monitor::] stanzas, documented above host_regex = <regular expression> host_segment = <integer> crcSalt = <string> recursive = <boolean> whitelist = <regular expression> blacklist = <regular expression> initCrcLength = <integer>
TCP:
[tcp://<remote server>:<port>] * Configures the input to listen on a specific TCP network port. * If a <remote server> makes a connection to this instance, the input uses this stanza to configure itself. * If you do not specify <remote server>, this stanza matches all connections on the specified port. * Generates events with source set to "tcp:<port>", for example: tcp:514 * If you do not specify a sourcetype, generates events with sourcetype set to "tcp-raw". Additional settings: connection_host = [ip|dns|none] * "ip" sets the host to the IP address of the system sending the data. * "dns" sets the host to the reverse DNS entry for the IP address of the system sending the data. * "none" leaves the host as specified in inputs.conf, typically the Splunk system hostname. * Default: "dns". queueSize = <integer>[KB|MB|GB] * The maximum size of the in-memory input queue. * Default: 500KB. persistentQueueSize = <integer>[KB|MB|GB|TB] * Maximum size of the persistent queue file. * Persistent queues can help prevent loss of transient data. For information on persistent queues and how the 'queueSize' and 'persistentQueueSize' settings interact, search the online documentation for "persistent queues". * If you set this to a value other than 0, then 'persistentQueueSize' must be larger than either the in-memory queue size (as defined by the 'queueSize' setting in inputs.conf or 'maxSize' settings in [queue] stanzas in server.conf). * Default: 0 (no persistent queue). requireHeader = <boolean> * Whether or not to require a header be present at the beginning of every stream. * This header can be used to override indexing settings. * Default: false. listenOnIPv6 = [no|yes|only] * Whether or not the input listens on IPv4, IPv6, or both * Set to 'yes' to listen on both IPv4 and IPv6 protocols. * Set to 'only' to listen on only the IPv6 protocol. * Default: The setting in the [general] stanza of the server.conf file acceptFrom = <network_acl> ... * Lists a set of networks or addresses from which to accept connections. * Separate multiple rules with commas or spaces. * Each rule can be in one of the following formats: 1. A single IPv4 or IPv6 address (examples: "10.1.2.3", "fe80::4a3") 2. A Classless Inter-Domain Routing (CIDR) block of addresses (examples: "10/8", "192.168.1/24", "fe80:1234/32") 3. A DNS name, possibly with a "*" used as a wildcard (examples: "myhost.example.com", "*.splunk.com") 4. "*", which matches anything * You can also prefix an entry with '!' to cause the rule to reject the connection. The input applies rules in order, and uses the first one that matches. For example, "!10.1/16, *" allows connections from everywhere except the 10.1.*.* network. * Default: "*" (accept from anywhere) rawTcpDoneTimeout = <seconds> * The amount of time, in seconds, that a network connection can remain idle before Splunk software declares that the last event over that connection has been received. * If a connection over this port remains idle for more than 'rawTcpDoneTimeout' seconds after receiving data, it adds a Done-key. This declares that the last event has been completely received. * Default: 10 [tcp:<port>] * Configures the input listen on the specified TCP network port. * This stanza is similar to [tcp://<remote server>:<port>], but listens for connections to the specified port from any host. * Generates events with a source of tcp:<port>. * If you do not specify a sourcetype, generates events with a source type of tcp-raw. * This stanza supports the following settings: connection_host = [ip|dns|none] queueSize = <integer>[KB|MB|GB] persistentQueueSize = <integer>[KB|MB|GB|TB] requireHeader = <boolean> listenOnIPv6 = [no|yes|only] acceptFrom = <network_acl> ... rawTcpDoneTimeout = <integer>
Data distribution:
Global settings for splunktcp. Used on the receiving side for data forwarded from a forwarder. [splunktcp] route = [has_key|absent_key:<key>:<queueName>;...] * Settings for the light forwarder. * The receiver sets these parameters automatically -- you do not need to set them yourself. * The property route is composed of rules delimited by ';' (semicolon). * The receiver checks each incoming data payload through the cooked TCP port against the route rules. * If a matching rule is found, the receiver sends the payload to the specified <queueName>. * If no matching rule is found, the receiver sends the payload to the default queue specified by any queue= for this stanza. If no queue= key is set in the stanza or globally, the receiver sends the events to the parsingQueue. enableS2SHeartbeat = <boolean> * Specifies the global keepalive setting for all splunktcp ports. * This option is used to detect forwarders which might have become unavailable due to network, firewall, or other problems. * The receiver monitors each connection for presence of a heartbeat, and if the heartbeat is not seen for 's2sHeartbeatTimeout' seconds, it closes the connection. * Default: true (heartbeat monitoring enabled) s2sHeartbeatTimeout = <seconds> * The amount of time, in seconds, that a receiver waits for heartbeats from forwarders that connect to this instance. * The receiver closes a forwarder connection if it does not receive a heartbeat for 's2sHeartbeatTimeout' seconds. * Default: 600 (10 minutes) inputShutdownTimeout = <seconds> * The amount of time, in seconds, that a receiver waits before shutting down inbound TCP connections after it receives a signal to shut down. * Used during shutdown to minimize data loss when forwarders are connected to a receiver. * During shutdown, the TCP input processor waits for 'inputShutdownTimeout' seconds and then closes any remaining open connections. * If all connections close before the end of the timeout period, shutdown proceeds immediately, without waiting for the timeout. stopAcceptorAfterQBlock = <seconds> * Specifies the time, in seconds, to wait before closing the splunktcp port. * If the receiver is unable to insert received data into the configured queue for more than the specified number of seconds, it closes the splunktcp port. * This action prevents forwarders from establishing new connections to this receiver. * Forwarders that have an existing connection will notice the port is closed upon test-connections and move to other receivers. * Once the queue unblocks, and TCP Input can continue processing data, the receiver starts listening on the port again. * This setting should not be adjusted lightly as extreme values can interact poorly with other defaults. * Default: 300 (5 minutes) listenOnIPv6 = no|yes|only * Select whether this receiver listens on IPv4, IPv6, or both protocols. * Set this to 'yes' to listen on both IPv4 and IPv6 protocols. * Set to 'only' to listen on only the IPv6 protocol. * If not present, the input uses the setting in the [general] stanza of server.conf. acceptFrom = <network_acl> ... * Lists a set of networks or IP addresses from which to accept connections. * Specify multiple rules with commas or spaces. * Each rule can be in the following forms: 1. A single IPv4 or IPv6 address (examples: "10.1.2.3", "fe80::4a3") 2. A CIDR block of addresses (examples: "10/8", "fe80:1234/32") 3. A DNS name, possibly with a "*" used as a wildcard (examples: "myhost.example.com", "*.splunk.com") 4. "*", which matches anything. * You can also prefix an entry with '!' to cause the rule to reject the connection. * The input applies rules in order, and uses the first one that matches. For example, "!10.1/16, *" allows connections from everywhere except the 10.1.*.* network. * Default: "*" (accept from anywhere) negotiateProtocolLevel = <unsigned integer> * If set, lets forwarders that connect to this receiver (or specific port) send data using only up to the specified feature level of the Splunk forwarder protocol. * If set to a value that is lower than the default, denies the use of newer forwarder protocol features during connection negotiation. This might impact indexer efficiency. * Default (if 'negotiateNewProtocol' is "true"): 1 * Default (if 'negotiateNewProtocol' is not "true"): 0 negotiateNewProtocol = <boolean> * Controls the default configuration of the 'negotiateProtocolLevel' setting. * DEPRECATED. * Use the 'negotiateProtocolLevel' instead. * Default: true. concurrentChannelLimit = <unsigned integer> * The number of unique channel codes that are available for forwarders to use to communicate with an indexer. * Each forwarder that connects to this indexer may use up to 'concurrentChannelLimit' unique channel codes. * In other words, each forwarder may have up to 'concurrentChannelLimit' channels in flight concurrently. * The receiver closes a forwarder connection if a forwarder attempts to exceed this value. * This setting only applies when the new forwarder protocol is in use. * Default: 300. Forwarder-specific settings for splunktcp. [splunktcp://[<remote server>]:<port>] * Receivers use this input stanza. * This is the same as the [tcp://] stanza, except the remote server is assumed to be a Splunk instance, most likely a forwarder. * <remote server> is optional. If you specify it, the receiver only listen for data from <remote server>. * Use of <remote server is not recommended. Use the 'acceptFrom' setting, which supersedes this setting. connection_host = [ip|dns|none] * For splunktcp, the 'host' or 'connection_host' is be used if the remote Splunk instance does not set a host, or if the host is set to "<host>::<localhost>". * "ip" sets the host to the IP address of the system sending the data. * "dns" sets the host to the reverse DNS entry for IP address of the system sending the data. * "none" leaves the host as specified in inputs.conf, typically the Splunk system hostname. * Default: "ip". compressed = <boolean> * Whether or not the receiver communicates with the forwarder in compressed format. * Applies to non-Secure Sockets Layer (SSL) receiving only. There is no compression setting required for SSL. * If set to "true", the receiver communicates with the forwarder in compressed format. * If set to "true", there is no longer a requirement to also set "compressed = true" in the outputs.conf file on the forwarder. * Default: false. enableS2SHeartbeat = <boolean> * Specifies the keepalive setting for the splunktcp port. * This option is used to detect forwarders which might have become unavailable due to network, firewall, or other problems. * The receiver monitors the connection for presence of a heartbeat, and if it does not see the heartbeat in 's2sHeartbeatTimeout' seconds, it closes the connection. * This overrides the default value specified at the global [splunktcp] stanza. * Default: true (heartbeat monitoring enabled). s2sHeartbeatTimeout = <integer> * The amount of time, in seconds, that a receiver waits for heartbeats from forwarders that connect to this instance. * The receiver closes the forwarder connection if it does not see a heartbeat for 's2sHeartbeatTimeout' seconds. * This overrides the default value specified at the global [splunktcp] stanza. * Default: 600 (10 minutes). queueSize = <integer>[KB|MB|GB] * The maximum size of the in-memory input queue. * Default: 500KB. negotiateProtocolLevel = <unsigned integer> * See the description for this setting in the [splunktcp] stanza. negotiateNewProtocol = <boolean> * See the description for this setting in the [splunktcp] stanza. concurrentChannelLimit = <unsigned integer> * See the description for this setting in the [splunktcp] stanza. [splunktcp:<port>] * This input stanza is the same as [splunktcp://[<remote server>]:<port>], but accepts connections from any server. * See the online documentation for [splunktcp://[<remote server>]:<port>] for more information on the following supported settings: connection_host = [ip|dns|none] compressed = <boolean> enableS2SHeartbeat = <boolean> s2sHeartbeatTimeout = <integer> queueSize = <integer>[KB|MB|GB] negotiateProtocolLevel = <unsigned integer> negotiateNewProtocol = <boolean> concurrentChannelLimit = <unsigned integer> Access control settings. [splunktcptoken://<token name>] * Use this stanza to specify forwarders from which to accept data. * You must configure a token on the receiver, then configure the same token on forwarders. * The receiver discards data from forwarders that do not have the token configured. * This setting is enabled for all receiving ports. * This setting is optional. token = <string> * Value of token. SSL settings for data distribution: [splunktcp-ssl:<port>] * Use this stanza type if you are receiving encrypted, parsed data from a forwarder. * Set <port> to the port on which the forwarder sends the encrypted data. * Forwarder settings are set in outputs.conf on the forwarder. * Compression for SSL is enabled by default. On the forwarder you can still specify compression with the 'useClientSSLCompression' setting in outputs.conf. * The 'compressed' setting is used for non-SSL connections. However, if you still specify 'compressed' for SSL, ensure that the 'compressed' setting is the same as on the forwarder, as splunktcp protocol expects the same 'compressed' setting from forwarders. connection_host = [ip|dns|none] * For splunktcp, the host or connection_host is used if the remote Splunk instance does not set a host, or if the host is set to "<host>::<localhost>". * "ip" sets the host to the IP address of the system sending the data. * "dns" sets the host to the reverse DNS entry for IP address of the system sending the data. * "none" leaves the host as specified in inputs.conf, typically the Splunk system hostname. * Default: "ip". compressed = <boolean> * See the description for this setting in the [splunktcp:<port>] stanza. enableS2SHeartbeat = <boolean> * See the description for this setting in the [splunktcp:<port>] stanza. s2sHeartbeatTimeout = <seconds> * See the description for this setting in the [splunktcp:<port>] stanza. listenOnIPv6 = [no|yes|only] * Select whether this receiver listens on IPv4, IPv6, or both protocols. * Set to "yes" to listen on both IPv4 and IPv6 protocols. * Set to "only" to listen on only the IPv6 protocol. * Default: The setting in the [general] stanza of the server.conf file. acceptFrom = <network_acl> ... * Lists a set of networks or IP addresses from which to accept connections. * Specify multiple rules with commas or spaces. * Each rule can be in the following forms: 1. A single IPv4 or IPv6 address (examples: "10.1.2.3", "fe80::4a3") 2. A CIDR block of addresses (examples: "10/8", "fe80:1234/32") 3. A DNS name, possibly with a "*" used as a wildcard (examples: "myhost.example.com", "*.splunk.com") 4. "*", which matches anything. * You can also prefix an entry with '!' to cause the rule to reject the connection. The input applies rules in order, and uses the first one that matches. For example, "!10.1/16, *" allows connections from everywhere except the 10.1.*.* network. * Default: "*" (accept from anywhere) negotiateProtocolLevel = <unsigned integer> * See the description for this setting in the [splunktcp] stanza. negotiateNewProtocol = <boolean> * See the description for this setting in the [splunktcp] stanza. concurrentChannelLimit = <unsigned integer> * See the description for this setting in the [splunktcp] stanza. To specify global ssl settings, that are applicable for all ports, add the settings to the SSL stanza. Specify any ssl setting that deviates from the global setting here. For a detailed description of each ssl setting, refer to the [SSL] stanza. serverCert = <path> sslPassword = <password> requireClientCert = <boolean> sslVersions = <string> cipherSuite = <cipher suite string> ecdhCurves = <comma separated list of ec curves> dhFile = <path> allowSslRenegotiation = <boolean> sslQuietShutdown = <boolean> sslCommonNameToCheck = <commonName1>, <commonName2>, ... sslAltNameToCheck = <alternateName1>, <alternateName2>, ... useSSLCompression = <boolean> [tcp-ssl:<port>] * Use this stanza type if you are receiving encrypted, unparsed data from a forwarder or third-party system. * Set <port> to the port on which the forwarder/third-party system is sending unparsed, encrypted data. * To create multiple SSL inputs, you can add the following attributes to each [tcp-ssl:<port>] input stanza. If you do not configure a certificate in the port, the certificate information is pulled from the default [SSL] stanza: * serverCert = <path_to_cert> * sslRootCAPath = <path_to_cert> This attribute should only be added if you have not configured your sslRootPath in server.conf. * sslPassword = <password> listenOnIPv6 = [no|yes|only] * Select whether the receiver listens on IPv4, IPv6, or both protocols. * Set to "yes" to listen on both IPv4 and IPv6 protocols. * Set to "only" to listen on only the IPv6 protocol. * If not present, the receiver uses the setting in the [general] stanza of server.conf. acceptFrom = <network_acl> ... * Lists a set of networks or IP addresses from which to accept connections. * Specify multiple rules with commas or spaces. * Each rule can be in the following forms: 1. A single IPv4 or IPv6 address (examples: "10.1.2.3", "fe80::4a3") 2. A CIDR block of addresses (examples: "10/8", "fe80:1234/32") 3. A DNS name, possibly with a '*' used as a wildcard (examples: "myhost.example.com", "*.splunk.com") 4. A single '*', which matches anything. * You can also prefix an entry with '!' to cause the rule to reject the connection. The input applies rules in order, and uses the first one that matches. For example, "!10.1/16, *" allows connections from everywhere except the 10.1.*.* network. * Default: "*" (accept from anywhere) [SSL] * Set the following specifications for receiving Secure Sockets Layer (SSL) communication underneath this stanza name. serverCert = <path> * The full path to the server certificate Privacy-Enhanced Mail (PEM) format file. * PEM is the most common text-based storage format for SSL certificate files. * No default. sslPassword = <string> * The server certificate password, if it exists. * Initially set to plain-text password. * Upon first use, the input encrypts and rewrites the password to $SPLUNK_HOME/etc/system/local/inputs.conf. password = <string> * DEPRECATED. * Do not use this setting. Use the 'sslPassword' setting instead. rootCA = <path> * DEPRECATED. * Do not use this setting. Use 'server.conf/[sslConfig]/sslRootCAPath' instead. * Used only if 'sslRootCAPath' is not set. * The <path> must refer to a PEM format file containing one or more root CA certificates concatenated together. requireClientCert = <boolean> * Determines whether a client must present an SSL certificate to authenticate. * Full path to the root CA (Certificate Authority) certificate store. * Default: false (if using self-signed and third-party certificates) * Default: true (if using the default certificates, overrides the existing "false" setting) sslVersions = <string> * A comma-separated list of SSL versions to support. * The versions available are "ssl3", "tls1.0", "tls1.1", and "tls1.2" * The special version "*" selects all supported versions. The version "tls" selects all versions that begin with "tls". * To remove a version from the list, prefix it with "-". * SSLv2 is always disabled. Specifying "-ssl2" in the version list has no effect. * When configured in Federal Information Processing Standard (FIPS) mode, the "ssl3" version is always disabled, regardless of this configuration. * The default can vary. See the 'sslVersions' setting in $SPLUNK_HOME/etc/system/default/inputs.conf for the current default. supportSSLV3Only = <boolean> * DEPRECATED. * SSLv2 is now always disabled. * Use the 'sslVersions' setting to set the list of supported SSL versions. cipherSuite = <cipher suite string> * If set, uses the specified cipher string for the input processors. * Must specify 'dhFile' to enable any Diffie-Hellman ciphers. * The default can vary. See the 'cipherSuite' setting in $SPLUNK_HOME/etc/system/default/inputs.conf for the current default. ecdhCurveName = <string> * DEPRECATED. * Use the 'ecdhCurves' setting instead. * This setting specifies the Elliptic Curve Diffie-Hellman (ECDH) curve to use for ECDH key negotiation. * Splunk software only supports named curves that have been specified by their SHORT name. * The list of valid named curves by their short and long names can be obtained by running this CLI command: $SPLUNK_HOME/bin/splunk cmd openssl ecparam -list_curves * Default: empty string. ecdhCurves = <comma-separated list> * A list of ECDH curves to use for ECDH key negotiation. * The curves should be specified in the order of preference. * The client sends these curves as a part of an SSL Client Hello. * The server supports only the curves specified in the list. * Splunk software only supports named curves that have been specified by their SHORT names. * The list of valid named curves by their short and long names can be obtained by running this CLI command: $SPLUNK_HOME/bin/splunk cmd openssl ecparam -list_curves * Example setting: "ecdhCurves = prime256v1,secp384r1,secp521r1" * The default can vary. See the 'ecdhCurves' setting in $SPLUNK_HOME/etc/system/default/inputs.conf for the current default. dhFile = <path> * Full path to the Diffie-Hellman parameter file. * DH group size should be no less than 2048 bits. * This file is required in order to enable any Diffie-Hellman ciphers. * No default. dhfile = <path> * DEPRECATED. * Use the 'dhFile' setting instead. allowSslRenegotiation = <boolean> * Whether or not to let SSL clients renegotiate their connections. * In the SSL protocol, a client might request renegotiation of the connection settings from time to time. * Setting this to false causes the server to reject all renegotiation attempts, which breaks the connection. * This limits the amount of CPU a single TCP connection can use, but it can cause connectivity problems, especially for long-lived connections. * Default: true. sslQuietShutdown = <boolean> * Enables quiet shutdown mode in SSL. * Default: false. sslCommonNameToCheck = <comma-separated list> * Checks the common name of the client certificate against this list of names. * If there is no match, assumes that the Splunk instance is not authenticated against this server. * This setting is optional. * For this setting to work, you must also set 'requireClientCert' to "true". * Default: empty string (no common name checking). sslAltNameToCheck = <comma-separated list> * Checks the alternate name of the client certificate against this list of names. * If there is no match, assumes that the Splunk instance is not authenticated against this server. * This setting is optional. * For this setting to work, you must also set 'requireClientCert' to "true". * Default: empty string (no alternate name checking). useSSLCompression = <boolean> * If set to "true", the server allows forwarders to negotiate SSL-layer data compression. * Default: The value of 'server.conf/[sslConfig]/allowSslCompression'
UDP:
[udp://<remote server>:<port>] * Similar to the [tcp://] stanza, except that this stanza causes the Splunk instance to listen on a UDP port. * Only one stanza per port number is currently supported. * Configures the instance to listen on a specific port. * If you specify <remote server>, the specified port only accepts data from that host. * If <remote server> is empty - [udp://<port>] - the port accepts data sent from any host. * The use of <remote server> is not recommended. Use the 'acceptFrom' setting, which supersedes this setting. * Generates events with source set to udp:portnumber, for example: udp:514 * If you do not specify a sourcetype, generates events with sourcetype set to udp:portnumber. Additional settings: connection_host = [ip|dns|none] * "ip" sets the host to the IP address of the system sending the data. * "dns" sets the host to the reverse DNS entry for IP address of the system sending the data. * "none" leaves the host as specified in inputs.conf, typically the Splunk system hostname. * If the input is configured with a 'sourcetype' that has a transform that overrides the 'host' field e.g. 'sourcetype=syslog', that takes precedence over the host specified here. * Default: "ip" _rcvbuf = <integer> * The receive buffer, in bytes, for the UDP port. * If you set the value to 0 or a negative number, the input ignores the value. * If the default value is too large for an OS, the instance tries to set the value to 1572864/2. If that value is also too large, the instance retries with 1572864/(2*2). It continues to retry by halving the value until it succeeds. * Default: 1572864. no_priority_stripping = <boolean> * Whether or not the input strips <priority> syslog fields from events it receives over the syslog input. * If you set this setting to true, the instance does NOT strip the <priority> syslog field from received events. * NOTE: Do NOT set this setting if you want to strip <priority>. * Default: false. no_appending_timestamp = <boolean> * Whether or not to append a timestamp and host to received events. * If you set this setting to true, the instance does NOT append a timestamp and host to received events. * NOTE: Do NOT set this setting if you want to append timestamp and host to received events. * Default: false. queueSize = <integer>[KB|MB|GB] * Maximum size of the in-memory input queue. * Default: 500KB. persistentQueueSize = <integer>[KB|MB|GB|TB] * Maximum size of the persistent queue file. * Persistent queues can help prevent loss of transient data. For information on persistent queues and how the 'queueSize' and 'persistentQueueSize' settings interact, search the online documentation for "persistent queues".. * If you set this to a value other than 0, then 'persistentQueueSize' must be larger than either the in-memory queue size (as defined by the 'queueSize' setting in inputs.conf or 'maxSize' settings in [queue] stanzas in server.conf). * Default: 0 (no persistent queue). listenOnIPv6 = <no | yes | only> * Select whether the instance listens on the IPv4, IPv6, or both protocols. * Set this to 'yes' to listen on both IPv4 and IPv6 protocols. * Set to 'only' to listen on only the IPv6 protocol. * If not present, the input uses the setting in the [general] stanza of server.conf. acceptFrom = <network_acl> ... * Lists a set of networks or IP addresses from which to accept connections. * Specify multiple rules with commas or spaces. * Each rule can be in the following forms: 1. A single IPv4 or IPv6 address (examples: "10.1.2.3", "fe80::4a3") 2. A CIDR block of addresses (examples: "10/8", "fe80:1234/32") 3. A DNS name, possibly with a "*"" used as a wildcard (examples: "myhost.example.com", "*.splunk.com") 4. "*", which matches anything. * You can also prefix an entry with '!' to cause the rule to reject the connection. The input applies rules in order, and uses the first one that matches. For example, "!10.1/16, *" allows connections from everywhere except the 10.1.*.* network. * Default: "*" (accept from anywhere) [udp:<port>] * This input stanza is the same as [udp://<remote server>:<port>], but does not have a <remote server> restriction. * See the documentation for [udp://<remote server>:<port>] to configure supported settings: connection_host = [ip|dns|none] _rcvbuf = <integer> no_priority_stripping = <boolean> no_appending_timestamp = <boolean> queueSize = <integer>[KB|MB|GB] persistentQueueSize = <integer>[KB|MB|GB|TB] listenOnIPv6 = <no | yes | only> acceptFrom = <network_acl> ...
FIFO (First In, First Out queue):
[fifo://<path>] * This stanza configures the monitoring of a FIFO at the specified path. queueSize = <integer>[KB|MB|GB] * Maximum size of the in-memory input queue. * Default: 500KB. persistentQueueSize = <integer>[KB|MB|GB|TB] * Maximum size of the persistent queue file. * Persistent queues can help prevent loss of transient data. For information on persistent queues and how the 'queueSize' and 'persistentQueueSize' settings interact, search the online documentation for "persistent queues".. * If you set this to a value other than 0, then 'persistentQueueSize' must be larger than either the in-memory queue size (as defined by the 'queueSize' setting in inputs.conf or 'maxSize' settings in [queue] stanzas in server.conf). * Default: 0 (no persistent queue).
Scripted Input:
[script://<cmd>] * Runs <cmd> at a configured interval (see below) and indexes the output that <cmd> returns. * The <cmd> must reside in one of the following directories: * $SPLUNK_HOME/etc/system/bin/ * $SPLUNK_HOME/etc/apps/$YOUR_APP/bin/ * $SPLUNK_HOME/bin/scripts/ * The path to <cmd> can be an absolute path, make use of an environment variable such as $SPLUNK_HOME, or use the special pattern of an initial '.' as the first directory to indicate a location inside the current app. * The '.' specification must be followed by a platform-specific directory separator. * For example, on UNIX: [script://./bin/my_script.sh] Or on Windows: [script://.\bin\my_program.exe] This '.' pattern is strongly recommended for app developers, and necessary for operation in search head pooling environments. * <cmd> can also be a path to a file that ends with a ".path" suffix. A file with this suffix is a special type of pointer file that points to a command to be run. Although the pointer file is bound by the same location restrictions mentioned above, the command referenced inside it can reside anywhere on the file system. The .path file must contain exactly one line: the path to the command to run, optionally followed by command-line arguments. The file can contain additional empty lines and lines that begin with ''. The input ignores these lines. interval = [<decimal>|<cron schedule>] * How often, in seconds, to run the specified command, or a valid "cron" schedule. * If you specify the interval as a number, it may have a fractional component; for example, 3.14 * To specify a cron schedule, use the following format: * "<minute> <hour> <day of month> <month> <day of week>" * Cron special characters are acceptable. You can use combinations of "*", ",", "/", and "-" to specify wildcards, separate values, specify ranges of values, and step values. * The cron implementation for data inputs does not currently support names of months or days. * The special value 0 forces this scripted input to be run continuously. As soon as the script exits, the input restarts it. * The special value -1 causes the scripted input to run once on start-up. * NOTE: when you specify a cron schedule, the input does not run the script on start-up. * Default: 60.0 passAuth = <username> * User to run the script as. * If you provide a username, the instance generates an auth token for that user and passes it to the script through stdin. * No default. python.version = {default|python|python2|python3} * For Python scripts only, selects which Python version to use. * Set to either "default" or "python" to use the system-wide default Python version. * Optional. * Default: Not set; uses the system-wide Python version. queueSize = <integer>[KB|MB|GB] * Maximum size of the in-memory input queue. * Default: 500KB. persistentQueueSize = <integer>[KB|MB|GB|TB] * Maximum size of the persistent queue file. * Persistent queues can help prevent loss of transient data. For information on persistent queues and how the 'queueSize' and 'persistentQueueSize' settings interact, search the online documentation for "persistent queues".. * If you set this to a value other than 0, then 'persistentQueueSize' must be larger than either the in-memory queue size (as defined by the 'queueSize' setting in inputs.conf or 'maxSize' settings in [queue] stanzas in server.conf). * Default: 0 (no persistent queue). index = <string> * The index where the scripted input sends the data. * NOTE: The script passes this parameter as a command-line argument to <cmd> in the format: -index <index name>. If the script does not need the index info, it can ignore this argument. * If you do not specify an index, the script uses the default index. send_index_as_argument_for_path = <boolean> * Whether or not to pass the index as an argument when specified for stanzas that begin with 'script://' * When this setting is "true", the script passes the argument as '-index <index name>'. * To avoid passing the index as a command line argument, set this to "false". * Default: true. start_by_shell = <boolean> * Whether or not to run the specified command through the operating system shell or command prompt. * If you set this setting to "true", the host operating system runs the specified command through the OS shell ("/bin/sh -c" on *NIX, "cmd.exe /c" on Windows.) * If you set the setting to "false", the input runs the program directly without attempting to expand shell metacharacters. * You might want to explicitly set the setting to "false" for scripts that you know do not need UNIX shell metacharacter expansion. This is a Splunk best practice. * Default: true (on *NIX hosts) * Default: false (on Windows hosts).
File system change monitor (fschange monitor)
The file system change monitor has been deprecated as of Splunk Enterprise version 5.0 and might be removed in a future version of the product. You cannot simultaneously monitor a directory with both the 'fschange' and 'monitor' stanza types. [fschange:<path>] * Monitors changes (such as additions, updates, and deletions) to this directory and any of its sub-directories. * <path> is the direct path. Do not preface it with '//' like with other inputs. * Sends an event for every change. Additional settings: NOTE: The 'fschange' stanza type does not use the same settings as other input types. It uses only the following settings: index = <string> * The index where the input sends the data. * Default: _audit (if you do not set 'signedaudit' or set 'signedaudit' to "false") * Default: the default index (in all other cases) signedaudit = <boolean> * Whether or not to send cryptographically signed add/update/delete events. * If this setting is "true", the input does the following to events that it generates: * Puts the events in the _audit index. * Sets the event sourcetype to 'audittrail' * If this setting is "false", the input: * Places events in the default index. * Sets the sourcetype to whatever you specify (or "fs_notification" by default). * You must set 'signedaudit' to "false" if you want to set the index for fschange events. * You must also enable auditing in audit.conf. * Default: false. filters = <comma-separated list> * The fschange input applies each filter, left to right, for each file or directory found during the monitor poll cycle. * See the "File System Monitoring Filters" section below for help on how to define a fschange filter. recurse = <boolean> * Whether or not the fschange input should look through all sub-directories for changes to files in a directory. * If this setting is "true", the input recurses through sub-directories within the directory specified in [fschange]. * Default: true. followLinks = <boolean> * Whether or not the fschange input should follow any symbolic links it encounters. * If you set this setting to "true", the input follows symbolic links. * CAUTION: Do not set this setting to "true" unless you can confirm that doing so will not create a file system loop (For example, in Directory A, symbolic link B points back to Directory A.) * Default: false. pollPeriod = <integer> * How often, in seconds, to check a directory for changes. * Default: 3600 (1 hour). hashMaxSize = <integer> * Tells the fschange input to calculate a SHA256 hash for every file that is this size or smaller, in bytes. * The input uses this hash as an additional method for detecting changes to the file/directory. * Default: -1 (disabled). fullEvent = <boolean> * Whether or not to send the full event if the input detects an add or update change. * Set to true to send the full event if an add or update change is detected. * Further qualified by the 'sendEventMaxSize' setting. * Default: false. sendEventMaxSize = <integer> * The maximum size, in bytes, that an fschange event can be for the input to send the full event to be indexed. * Limits the size of event data that the fschange input sends. * This limits the size of indexed file data. * Default: -1 (unlimited). sourcetype = <string> * Sets the source type for events from this input. * The input automatically prepends "sourcetype=" to <string>. * Default: "audittrail" (if you set the 'signedaudit' setting to "true".) * Default: "fs_notification" (if you set the 'signedaudit' setting to "false".) host = <string> * Sets the host name for events from this input. * Default: whatever host sent the event. filesPerDelay = <integer> * The number of files that the fschange input processes between processing delays, as specified by the 'delayInMills' setting. * After a delay of 'delayInMills' milliseconds, the fschange input processes 'filesPerDelay' files, then waits 'delayInMills' milliseconds again before repeating this process. * This setting helps throttle file system monitoring so it consumes less CPU. * Default: 10. delayInMills = <integer> * The delay, in milliseconds, that the fschange input waits between processing 'filesPerDelay' files. * After a delay of 'delayInMills' milliseconds, the fschange input processes 'filesPerDelay' files, then waits 'delayInMills' milliseconds again before repeating this process. * This setting helps throttle file system monitoring so it consumes less CPU. * Default: 100.
File system monitoring filters:
[filter:<filtertype>:<filtername>] * Defines a filter of type <filtertype> and names it <filtername>. * <filtertype>: * Filter types are either 'blacklist' or 'whitelist.' * A whitelist filter processes all file names that match the regular expression list that you define within the stanza. * A blacklist filter skips all file names that match the regular expression list. * <filtername> * The fschange input uses filter names that you specify with the 'filters' setting for a given fschange stanza. * You can specify multiple filters buy separating them with commas. regex<integer> = <regular expression> * Blacklist and whitelist filters can include a set of regular expressions. * The name of each regular expression MUST be 'regex<integer>', where <integer> starts at 1 and increments by 1. * The input applies each regular expression in numeric order: regex1=<regular expression> regex2=<regular expression> ...
http: (HTTP Event Collector)
Global settings for the HTTP Event Collector (HEC) Input. [http] port = <positive integer> * The event collector data endpoint server port. * Default: 8088. disabled = [0|1] * Whether or not the event collector input is active. * Set this setting to "1" to disable the input, and "0" to enable it. * Default: 1 (disabled). outputgroup = <string> * The name of the output group that the event collector forwards data to. * Default: empty string. useDeploymentServer = [0|1] * Whether or not the HTTP event collector input should write its configuration to a deployment server repository. * When you enable this setting, the input writes its configuration to the directory that you specify with the 'repositoryLocation' setting in the serverclass.conf file. * You must copy the full contents of the splunk_httpinput app directory to this directory for the configuration to work. * When enabled, only the tokens defined in the splunk_httpinput app in this repository are viewable and editable through the API and Splunk Web. * When disabled, the input writes its configuration to $SPLUNK_HOME/etc/apps by default. * Default: 0 (disabled). index = <string> * The default index to use. * Default: the "default" index. sourcetype = <string> * The default source type for the events that the input generates. * If you do not specify a sourcetype, the input does not set a sourcetype for events it generates. enableSSL = [0|1] * Whether or not the HTTP Event Collector uses SSL. * HEC shares SSL settings with the Splunk management server and cannot have SSL enabled when the Splunk management server has SSL disabled. * Default: 1 (enabled). dedicatedIoThreads = <number> * The number of dedicated input/output threads in the event collector input. * Default: 0 (The input uses a single thread). replyHeader.<name> = <string> * Adds a static header to all HTTP responses that this server generates. * For example, "replyHeader.My-Header = value" causes the response header "My-Header: value" to be included in the reply to every HTTP request made to the event collector endpoint server. * No default. maxSockets = <integer> * The number of HTTP connections that the HTTP event collector input accepts simultaneously. * Set this setting to constrain resource usage. * If you set this setting to 0, the input automatically sets it to one third of the maximum allowable open files on the host. * If this value is less than 50, the input sets it to 50. If this value is greater than 400000, the input sets it to 400000. * If set to a negative value, the input does not enforce a limit on connections. * Default: 0. maxThreads = <integer> * The number of threads that can be used by active HTTP transactions. * Set this to constrain resource usage. * If you set this setting to 0, the input automatically sets the limit to one third of the maximum allowable threads on the host. * If this value is less than 20, the input sets it to 20. If this value is greater than 150000, the input sets it to 150000. * If the 'maxSockets' setting has a positive value and 'maxThreads' is greater than 'maxSockets', then the input sets 'maxThreads' to be equal to 'maxSockets'. * If set to a negative value, the input does not enforce a limit on threads. * Default: 0. keepAliveIdleTimeout = <integer> * How long, in seconds, that the HTTP Event Collector input lets a keep-alive connection remain idle before forcibly disconnecting it. * If this value is less than 7200, the input sets it to 7200. * Default: 7200. busyKeepAliveIdleTimeout = <integer> * How long, in seconds, that the HTTP Event Collector lets a keep-alive connection remain idle while in a busy state before forcibly disconnecting it. * CAUTION: Setting this to a value that is too large can result in file descriptor exhaustion due to idling connections. * If this value is less than 12, the input sets it to 12. * Default: 12. serverCert = <path> * The full path to the server certificate PEM format file. * The same file may also contain a private key. * Splunk software automatically generates certificates when it first starts. * You may replace the auto-generated certificate with your own certificate. * Default: $SPLUNK_HOME/etc/auth/server.pem. sslKeysfile = <filename> * DEPRECATED. * Use the 'serverCert' setting instead. * The file that contains the SSL keys. Splunk software looks for this file in the directory specified by 'caPath'. * Default: server.pem. sslPassword = <string> * The server certificate password. * Initially set to a plain-text password. * Upon first use, Splunk software encrypts and rewrites the password. * Default: "password". sslKeysfilePassword = <string> * DEPRECATED. * Use the 'sslPassword' setting instead. caCertFile = <string> * DEPRECATED. * Use the 'server.conf:[sslConfig]/sslRootCAPath' setting instead. * Used only if you do not set the 'sslRootCAPath' setting. * Specifies the file name (relative to 'caPath') of the CA (Certificate Authority) certificate PEM format file that contains one or more certificates concatenated together. * Default: cacert.pem. caPath = <string> * DEPRECATED. * Use absolute paths for all certificate files. * If certificate files given by other settings in this stanza are not absolute paths, then they are relative to this path. * Default: $SPLUNK_HOME/etc/auth. sslVersions = <comma-separated list> * A comma-separated list of SSL versions to support. * The versions available are "ssl3", "tls1.0", "tls1.1", and "tls1.2" * The special version "*" selects all supported versions. The version "tls" selects all versions "tls1.0" or newer. * To remove a version from the list, prefix it with "-". * SSLv2 is always disabled. Specifying "-ssl2" in the version list has no effect. * When configured in Federal Information Processing Standard (FIPS) mode, the "ssl3" version is always disabled, regardless of this configuration. * Default: "*,-ssl2". (anything newer than SSLv2) cipherSuite = <string> * The cipher string to use for the HTTP Event Collector input. * Use this setting to ensure that the server does not accept connections using weak encryption protocols. * If you set this setting, the input uses the specified cipher string for the HTTP server. * Default: The default cipher string that 'OpenSSL' provides listenOnIPv6 = [no|yes|only] * Whether or not this input listens on IPv4, IPv6, or both. * Set to "no" to make the input listen only on the IPv4 protocol. * Set to "yes" to make the input listen on both IPv4 and IPv6 protocols. * Set to "only" to make the input listen on only the IPv6 protocol. * Default: The setting in the [general] stanza of the server.conf file acceptFrom = <network_acl> ... * A list of networks or IP addresses from which to accept connections. * Specify multiple rules with commas or spaces. * Each rule can be in the following forms: 1. A single IPv4 or IPv6 address (examples: "10.1.2.3", "fe80::4a3") 2. A CIDR block of addresses (examples: "10/8", "fe80:1234/32") 3. A DNS name, possibly with a "*" used as a wildcard (examples: "myhost.example.com", "*.splunk.com") 4. "*", which matches anything. * You can also prefix an entry with "!" to cause the rule to reject the connection. The input applies rules in order, and uses the first one that matches. For example, "!10.1/16, *" allows connections from everywhere except the 10.1.*.* network. * Default: "*" (accept from anywhere) requireClientCert = <boolean> * Requires that any client connecting to the HEC port has a certificate that can be validated by the certificate authority specified in the 'caCertFile' setting. * Default: false. ecdhCurveName = <string> * DEPRECATED. * Use the 'ecdhCurves' setting instead. * This setting specifies the ECDH curve to use for ECDH key negotiation. * Splunk software only supports named curves that have been specified by their SHORT names. * The list of valid named curves by their short or long names can be obtained by executing this command: $SPLUNK_HOME/bin/splunk cmd openssl ecparam -list_curves * Default: empty string. ecdhCurves = <comma-separated list> * ECDH curves to use for ECDH key negotiation. * The curves should be specified in the order of preference. * The client sends these curves as a part of Client Hello. * The server supports only the curves specified in the list. * Splunk software only supports named curves that have been specified by their SHORT names. * The list of valid named curves by their short or long names can be obtained by executing this command: $SPLUNK_HOME/bin/splunk cmd openssl ecparam -list_curves * Example setting: ecdhCurves = prime256v1,secp384r1,secp521r1 * Default: empty string. crossOriginSharingPolicy = <origin_acl> ... * A list of the HTTP Origins for which to return Access-Control-Allow-* Cross-origin Resource Sharing (CORS) headers. * These headers tell browsers that web applications at those sites can be trusted to make requests to the REST interface. * The origin is passed as a URL without a path component (for example "https://app.example.com:8000"). * This setting can take a list of acceptable origins, separated by spaces and/or commas. * Each origin can also contain wildcards for any part. Examples: * *://app.example.com:* (either HTTP or HTTPS on any port) * https://*.example.com (any host under example.com, including example.com itself). * An address can be prefixed with a '!' to negate the match, with the first matching origin taking precedence. Example: * "!*://evil.example.com:* *://*.example.com:*" to not avoid matching one host in a domain. * "*" matches all origins. * Default: empty string. crossOriginSharingHeaders = <string> * A list of the HTTP headers to which splunkd sets "Access-Control-Allow-Headers" when replying to Cross-Origin Resource Sharing (CORS) preflight requests. * The "Access-Control-Allow-Headers" header is used in response to a CORS preflight request to tell browsers which HTTP headers can be used during the actual request. * A CORS preflight request is a CORS request that checks to see if the CORS protocol is understood and a server is aware of using specific methods and headers. * This setting can take a list of acceptable HTTP headers, separated by commas. * A single "*" can also be used to match all headers. * Default: Empty string. forceHttp10 = [auto|never|always] * Whether or not the REST HTTP server forces clients that connect to it to use the HTTP 1.0 specification for web communications. * When set to "always", the REST HTTP server does not use some HTTP 1.1 features such as persistent connections or chunked transfer encoding. * When set to "auto", it does this only if the client did not send a User-Agent header, or if the user agent is known to have bugs in its support of HTTP/1.1. * When set to "never" it always allows HTTP 1.1, even to clients it suspects might be buggy. * Default: "auto". sslCommonNameToCheck = <commonName1>, <commonName2>, ... * A list of SSL Common Names to match against certificates that incoming HTTPS connections present to this instance. * If you configure this setting and also set 'requireClientCert' to "true", splunkd limits most inbound HTTPS connections to hosts that use a cert with one of the listed common names. * The most important scenario to use this setting is distributed search. * This feature does not work with the deployment server and client communication over SSL. * This setting is optional. * Default: empty string (no common name checking.) sslAltNameToCheck = <alternateName1>, <alternateName2>, ... * If you set this setting and also set 'requireClientCert' to true, splunkd can verify certificates that have a so-called "Subject Alternate Name" that matches any of the alternate names in this list. * Subject Alternate Names are effectively extended descriptive fields in SSL certs beyond the commonName. A common practice for HTTPS certs is to use these values to store additional valid hostnames or domains where the cert should be considered valid. * Accepts a comma-separated list of Subject Alternate Names to consider valid. * Items in this list are never validated against the SSL Common Name. * This feature does not work with the deployment server and client communication over SSL. * This setting is optional. * Default: empty string (no alternate name checking.) sendStrictTransportSecurityHeader = <boolean> * Whether or not to force inbound connections to always use SSL with the "Strict-Transport-Security" header.. * If set to "true", the REST interface sends a "Strict-Transport-Security" header with all responses to requests made over SSL. * This can help prevent a client being tricked later by a Man-In-The-Middle attack to accept a non-SSL request. However, this requires a commitment that no non-SSL web hosts will ever be run on this hostname on any port. For example, if Splunk Web is in default non-SSL mode this can break the ability of the browser to connect to it. Enable with caution. * Default: false. allowSslCompression = <boolean> * Whether or not to allow data compression over SSL. * If set to "true", the server allows clients to negotiate SSL-layer data compression. * Default: true. allowSslRenegotiation = <boolean> * Whether or not to let SSL clients renegotiate their connections. * In the SSL protocol, a client may request renegotiation of the connection settings from time to time. * Setting this to false causes the server to reject all renegotiation attempts, which breaks the connection. * This limits the amount of CPU a single TCP connection can use, but it can cause connectivity problems, especially for long-lived connections. * Default: true. ackIdleCleanup = <boolean> * Whether or not to remove ACK channels that have been idle after a period of time, as defined by the 'maxIdleTime' setting. * If set to "true", the server removes the ACK channels that are idle for 'maxIdleTime' seconds. * Default: true. maxIdleTime = <integer> * The maximum amount of time, in seconds, that ACK channels can be idle before they are removed. * If 'ackIdleCleanup' is "true", the system removes ACK channels that have been idle for 'maxIdleTime' seconds. * Default: 600 (10 minutes.) channel_cookie = <string> * The name of the cookie to use when sending data with a specified channel ID. * The value of the cookie is the channel sent. For example, if you have set 'channel_cookie=foo' and sent a request with channel ID set to 'bar', then you will have a cookie in the response with the value 'foo=bar'. * If no channel ID is present in the request, then no cookie is returned. * This setting is to be used for load balancers (for example, AWS ELB) that can only provide sticky sessions on cookie values and not general header values. * If no value is set (the default), then no cookie is returned. * Default: empty string (no cookie). maxEventSize = <positive integer>[KB|MB|GB] * The maximum size of a single HEC (HTTP Event Collector) event. * HEC disregards and triggers a parsing error for events whose size is greater than 'maxEventSize'. * Default: 5MB.
HTTP Event Collector (HEC) - Local stanza for each token
[http://name] token = <string> * The value of the HEC token. * HEC uses this token to authenticate inbound connections. Your application or web client must present this token when attempting to connect to HEC. * No default. disabled = [0|1] * Whether or not this token is active. * Default: 0 (enabled). description = <string> * A human-readable description of this token. * Default: empty string. indexes = <string> * The indexes that events for this token can go to. * If you do not specify this value, the index list is empty, and any index can be used. * No default. index = <string> * The default index to use for this token. * Default: the default index. sourcetype = <string> * The default sourcetype to use if it is not specified in an event. * Default: empty string. outputgroup = <string> * The name of the forwarding output group to send data to. * Default: empty string. queueSize = <integer>[KB|MB|GB] * The maximum size of the in-memory input queue. * Default: 500KB. persistentQueueSize = <integer>[KB|MB|GB|TB] * Maximum size of the persistent queue file. * Persistent queues can help prevent loss of transient data. For information on persistent queues and how the 'queueSize' and 'persistentQueueSize' settings interact, search the online documentation for "persistent queues".. * If you set this to a value other than 0, then 'persistentQueueSize' must be larger than either the in-memory queue size (as defined by the 'queueSize' setting in inputs.conf or 'maxSize' settings in [queue] stanzas in server.conf). * Default: 0 (no persistent queue). connection_host = [ip|dns|proxied_ip|none] * Specifies the host if an event doesn't have a host set. * "ip" sets the host to the IP address of the system sending the data. * "dns" sets the host to the reverse DNS entry for IP address of the system sending the data. * "proxied_ip" checks whether an X-Forwarded-For header was sent (presumably by a proxy server) and if so, sets the host to that value. Otherwise, the IP address of the system sending the data is used. * "none" leaves the host as specified in the HTTP header. * No default. useACK = <boolean> * When set to "true", acknowledgment (ACK) is enabled. Events in a request are tracked until they are indexed. An events status (indexed or not) can be queried from the ACK endpoint with the ID for the request. * When set to false, acknowledgment is not enabled. * This setting can be set at the stanza level. * Default: false allowQueryStringAuth = <boolean> * Enables or disables sending authorization tokens with a query string. * This is a token level configuration. It may only be set for a particular token. * To use this feature, set to "true" and configure the client application to include the token in the query string portion of the URL they use to send data to HEC in the following format: "https://<URL>?<your=query-string>&token=<your-token>" or "https://<URL>?token=<your-token>" if the token is the first element in the query string. * If a token is sent in both the query string and an HTTP header, the token in the query string takes precedence, even if this feature is disabled. In other words, if a token is present in the query string, any token in the header for that request is not used. * NOTE: Query strings may be observed in transit and/or logged in cleartext. There is no confidentiality protection for the transmitted tokens. * Before using this in production, consult security personnel in your organization to understand and plan to mitigate the risks. * At a minimum, always use HTTPS when you enable this feature. Check your client application, proxy, and logging configurations to confirm that the token is not logged in clear text. * Give minimal access permissions to the token in HEC and restrict the use of the token only to trusted client applications. * Default: false
WINDOWS INPUTS:
* Windows platform specific input processor. *********** Splunk software on Windows ships with several Windows-only inputs. They are defined in the default inputs.conf. * Use the "disabled=" setting to enable/disable any of them. * A short summary of the inputs follows: * Perfmon: Monitors Windows performance counters, objects, and instances. * WinRegMon: Tracks and report any changes that occur in the local system Registry. * ADMon: Indexes existing Active Directory (AD) objects and listens for AD changes. * WMI: Retrieves event logs remotely and locally through the Windows Management. Instrumentation subsystem. It can also gather performance data remotely, as well as receive various system notifications. See wmi.conf.spec for information on how to configure this input. ******* The following Windows input specifications are for parsing on non-Windows platforms. *******
Performance Monitor
[perfmon://<name>] * This section explains possible settings for configuring the Windows Performance Monitor input. * Each perfmon:// stanza represents an individually configured performance monitoring input. If you configure the input through Splunk Web, then the value of "<NAME>" matches what was specified there. While you can add performance monitor inputs manually, it is a best practice to use Splunk Web to configure them, because it is easy to mistype the values for Performance Monitor objects, counters, and instances. * NOTE: The perfmon stanza is for local systems ONLY. To define performance monitor inputs for remote machines, use wmi.conf. object = <string> * A valid Performance Monitor object as defined within Performance Monitor (for example, "Process," "Server," "PhysicalDisk.") * You can specify a single valid Performance Monitor object or use a regular expression (regex) to specify multiple objects. * This setting is required, and the input does not run if the setting is not present. * No default. counters = <semicolon-separated list> * This can be a single counter, or multiple valid Performance Monitor counters. * This setting is required, and the input does not run if the setting is not present. * "*" is equivalent to all available counters for a given Performance Monitor object. * No default. instances = <semicolon-separated list> * One or more multiple valid Performance Monitor instances. * "*"" is equivalent to all available instances for a given Performance Monitor counter. * If applicable instances are available for a counter and this setting is not present, then the input logs data for all available instances (this is the same as setting "instances = *"). * If there are no applicable instances for a counter, then you can omit this setting. * No default. interval = <integer> * How often, in seconds, to poll for new data. * This setting is required, and the input does not run if the setting is not present. * The recommended setting depends on the Performance Monitor object, counter(s), and instance(s) that you define in the input, and how much performance data you need. * Objects with numerous instantaneous or per-second counters, such as "Memory", "Processor", and "PhysicalDisk" should have shorter interval times specified (anywhere from 1-3 seconds). * Less volatile counters such as "Terminal Services", "Paging File", and "Print Queue" can have longer intervals configured. * Default: 300 mode = [single|multikv] * Specifies how the performance monitor input generates events. * Set to "single" to print each event individually. * Set to "multikv" to print events in multikv (formatted multiple key-value pair) format. * Default: "single" samplingInterval = <positive integer> * How often, in milliseconds, to poll for new data. * This is an advanced setting. * Enables high-frequency performance sampling. The input collects performance data every sampling interval. It then reports averaged data and other statistics at every interval. * The minimum legal value is 100, and the maximum legal value must be less than the 'interval' setting. * If not set, high-frequency sampling does not occur. * No default (disabled). stats = <average;count;dev;min;max> * Reports statistics for high-frequency performance sampling. * This is an advanced setting. * Acceptable values are: average, count, dev, min, max. * You can specify multiple values by separating them with semicolons. * If not specified, the input does not produce high-frequency sampling statistics. * No default. (disabled) disabled = [0|1] * Specifies whether or not the input is enabled. * Set to 1 to disable the input, and 0 to enable it. * Default: 0 (enabled) showZeroValue = [0|1] * Specfies whether or not zero-value event data should be collected. * Set to 1 to capture zero value event data, and 0 to ignore such data. * Default: 0 (ignore zero value event data) useEnglishOnly = <boolean> * Controls which Windows Performance Monitor API the input uses. * If set to "true", the input uses PdhAddEnglishCounter() to add the counter string. This ensures that counters display in English regardless of the Windows machine locale. * If set to "false", the input uses PdhAddCounter() to add the counter string. * NOTE: if you set this setting to true, the 'object' setting does not accept a regular expression as a value on machines that have a non-English locale. * Default: false. useWinApiProcStats = <boolean> * Whether or not the Performance Monitor input uses process kernel mode and user mode times to calculate CPU usage for a process, rather than using the standard Performance Data Helper (PDH) APIs to calculate those values. * A problem was found in the PDH APIs that causes Performance Monitor inputs to show maximum values of 100% usage for a process on multicore Windows machines, even when the process uses more than 1 core at a time. * When you configure this setting to "true", the input uses the GetProcessTime() function in the core Windows API to calculate CPU usage for a process, for the following Performance Monitor counters, only: ** Processor Time ** User Time ** Privileged Time * This means that, if a process uses 5 of 8 cores on an 8-core machine, that the input should return a value of around 500, rather than the incorrect 100. * When you configure the setting to "false", the input uses the standard PDH APIs to calculate CPU usage for a process. On multicore systems, the maximum value that PDH APIs return is 100, regardless of the number of cores in the machine that the process uses. * Performance monitor inputs use the PDH APIs for all other Performance Monitor counters. Configuring this setting has no effect on those counters. * NOTE: If the Windows machine uses a non-English system locale, and you have set 'useWinApiProcStats' to "true" for a Performance Monitor input, then you must also set 'useEnglishOnly' to "true" for that input. * Default: false formatString = <string> * Controls the print format for double-precision statistic counters. * Do not use quotes when specifying this string. * Default: "%.20g" (without quotes). usePDHFmtNoCap100 = <boolean> * Whether or not performance counter values that are greater than 100 (for example, counter values that measure the processor load on computers with multiple processors) are reset to 100. * If set to "true", the counter values can exceed 100. * If set to "false", the input resets counter values to 100 if the processor load on multiprocessor computers exceeds 100. * Default: false Direct Access File Monitor (does not use file handles) For Windows systems only. [MonitorNoHandle://<path>] * This input intercepts file writes to the specific file. * <path> must be a fully qualified path name to a specific file. Wildcards and directories are not accepted. * This input type does not function on *nix machines. * You can specify more than one stanza of this type. disabled = [0|1] * Whether or not the input is enabled. * Default: 0 (enabled) index = <string> * Specifies the index that this input should send the data to. * This setting is optional. * Default: the default index
Windows Event Log Monitor
[WinEventLog://<name>] * This section explains possible settings for configuring the Windows Event Log monitor. * Each WinEventLog:// stanza represents an individually configured WinEventLog monitoring input. If you you configure the input through Splunk Web, the value of "<NAME>" matches what was specified there. While you can add event log monitor inputs manually, it is best practice to use Splunk Web to configure Windows event log monitor inputs because it is easy to mistype the values for event log channels. * NOTE: The WinEventLog stanza is for local systems only. To define event log monitor inputs for remote machines, use wmi.conf. start_from = <string> * How the input should chronologically read the Event Log channels. * If you set this setting to "oldest", the input reads Windows event logs from oldest to newest. * If you set this setting to "newest" the input reads Windows event logs in reverse, from newest to oldest. Once the input consumes the backlog of events, it stops. * If you set this setting to "newest", and at the same time set the "current_only" setting to 0, the combination can result in the input indexing duplicate events. * Do not set this setting to "newest" and at the same time set the "current_only" setting to 1. This results in the input not collecting any events because you instructed it to read existing events from oldest to newest and read only incoming events concurrently (A logically impossible combination.) * Default: "oldest" use_old_eventlog_api = <boolean> * Whether or not to read Event Log events with the Event Logging API. * This is an advanced setting. Contact Splunk Support before you change it. * If set to "true", the input uses the Event Logging API (instead of the Windows Event Log API) to read from the Event Log on Windows Server 2008, Windows Vista, and later installations. * Default: false (Use the API that is specific to the OS.) use_threads = <integer> * Specifies the number of threads, in addition to the default writer thread, that can be created to filter events with the blacklist/whitelist regular expression. * This is an advanced setting. Contact Splunk Support before you change it. * The maximum number of threads is 15. * Default: 0 thread_wait_time_msec = <integer> * The interval, in milliseconds, between attempts to re-read Event Log files when a read error occurs. * This is an advanced setting. Contact Splunk Support before you change it. * Default: 5000 suppress_checkpoint = <boolean> * Whether or not the Event Log strictly follows the 'checkpointInterval' setting when it saves a checkpoint. * This is an advanced setting. Contact Splunk Support before you change it. * By default, the Event Log input saves a checkpoint from between zero and 'checkpointInterval' seconds, depending on incoming event volume. If you set this setting to "true", that does not happen. * Default: false suppress_sourcename = <boolean> * Whether or not to exclude the 'sourcename' field from events. * This is an advanced setting. Contact Splunk Support before you change it. * When set to true, the input excludes the 'sourcename' field from events and thruput performance (the number of events processed per second) improves. * Default: false suppress_keywords = <boolean> * Whether or not to exclude the 'keywords' field from events. * This is an advanced setting. Contact Splunk Support before you change it. * When set to true, the input excludes the 'keywords' field from events and thruput performance (the number of events processed per second) improves. * Default: false suppress_type = <boolean> * Whether or not to exclude the 'type' field from events. * This is an advanced setting. Contact Splunk Support before you change it. * When set to true, the input excludes the 'type' field from events and thruput performance (the number of events processed per second) improves. * Default: false suppress_task = <boolean> * Whether or not to exclude the 'task' field from events. * This is an advanced setting. Contact Splunk Support before you change it. * When set to true, the input excludes the 'task' field from events and thruput performance (the number of events processed per second) improves. * Default: false suppress_opcode = <boolean> * Whether or not to exclude the 'opcode' field from events. When set to true, the input excludes the 'opcode' field from events and thruput performance (the number of events processed per second) improves. * This is an advanced setting. Contact Splunk Support before you change it. * Default: false current_only = [0|1] * Whether or not to acquire only events that arrive while the instance is running. * If you set this setting to 1, the input only acquires events that arrive while the instance runs and the input is enabled. The input does not read data which was stored in the Windows Event Log while the instance was not running. This means that there will be gaps in the data if you restart the instance or experiences downtime. * If you set the setting to 0, the input first gets all existing events already stored in the log that have higher event IDs (have arrived more recently) than the most recent events acquired. The input then monitors events that arrive in real time. * If you set this setting to 0, and at the same time set the 'start_from' setting to "newest", the combination can result in the indexing of duplicate events. * Do not set this setting to 1 and at the same time set the 'start_from' setting to "newest". This results in the input not collecting any events because you instructed it to read existing events from oldest to newest and read only incoming events concurrently (A logically impossible combination.) * Default: 0 (false, gathering stored events first before monitoring live events) batch_size = <integer> * How many Windows Event Log items to read per request. * If troubleshooting identifies that the Event Log input is a bottleneck in acquiring data, increasing this value can help. * NOTE: Splunk Support has seen cases where large values can result in a stall in the Event Log subsystem. If you increase this value significantly, monitor closely for trouble. * In local and customer acceptance testing, a value of 10 was acceptable for both throughput and reliability. * Default: 10. checkpointInterval = <integer> * How often, in seconds, that the Windows Event Log input saves a checkpoint. * Checkpoints store the eventID of acquired events. This lets the input continue monitoring at the correct event after a shutdown or outage. * Default: 0 disabled = [0|1] * Whether or not the input is enabled. * Set to 1 to disable the input, and 0 to enable it. * Default: 0 (enabled) evt_resolve_ad_obj = [0|1] * How the input should interact with Active Directory while indexing Windows Event Log events. * If you set this setting to 1, the input resolves the Active Directory Security IDentifier (SID) objects to their canonical names for a specific Windows Event Log channel. * If you enable the setting, the rate at which the input reads events on high-traffic Event Log channels can decrease. Latency can also increase during event acquisition. This is due to the overhead involved in performing AD translations. * When you set this setting to 1, you can optionally specify the domain controller name or dns name of the domain to bind to with the 'evt_dc_name' setting. The input connects to that domain controller to resolve the AD objects. * If you set this setting to 0, the input does not attempt any resolution. * Default: 0 (disabled) for all channels evt_dc_name = <string> * Which Active Directory domain controller to bind to for AD object resolution. * If you prefix a dollar sign to a value (for example, $my_domain_controller), the input interprets the value as an environment variable. If the environment variable has not been defined on the host, it is the same as if the value is blank. * This setting is optional. * This setting can be set to the NetBIOS name of the domain controller or the fully-qualified DNS name of the domain controller. Either name type can, optionally, be preceded by two backslash characters. The following examples represent correctly formatted domain controller names: * "FTW-DC-01" * "\\FTW-DC-01" * "FTW-DC-01.splunk.com" * "\\FTW-DC-01.splunk.com" * $my_domain_controller evt_dns_name = <string> * The fully-qualified DNS name of the domain that the input should bind to for AD object resolution. * This setting is optional. evt_resolve_ad_ds = [auto|PDC] * How the input should choose the domain controller to bind for AD resolution. * This setting is optional. * If set to PDC, the input only contacts the primary domain controller to resolve AD objects. * If set to auto, the input lets Windows chose the best domain controller. * If you set the 'evt_dc_name' setting, the input ignores this setting. * Default: 'auto' (let Windows determine the domain controller to use.) evt_ad_cache_disabled = [0|1] * Enables or disables the AD object cache. * Default: 0 (enabled) evt_ad_cache_exp = <integer> * The expiration time, in seconds, for AD object cache entries. * This setting is optional. * The minimum allowed value is 10 and the maximum allowed value is 31536000. * Default: 3600 (1 hour) evt_ad_cache_exp_neg = <integer> * The expiration time, in seconds, for negative AD object cache entries. * This setting is optional. * The minimum allowed value is 10 and the maximum allowed value is 31536000. * Default: 10 evt_ad_cache_max_entries = <integer> * The maximum number of AD object cache entries. * This setting is optional. * The minimum allowed value is 10 and the maximum allowed value is 40000. * Default: 1000 evt_sid_cache_disabled = [0|1] * Enables or disables account Security IDentifier (SID) cache. * This setting is global. It affects all Windows Event Log stanzas. * Default: 0 evt_sid_cache_exp = <unsigned integer> * The expiration time, in seconds, for account SID cache entries. * This setting is optional. * This setting is global. It affects all Windows Event Log stanzas. * The minimum allowed value is 10 and the maximum allowed value is 31536000. * Default: 3600 evt_sid_cache_exp_neg = <unsigned integer> * The expiration time, in seconds, for negative account SID cache entries. * This setting is optional. * This setting is global. It affects all Windows Event Log stanzas. * The minimum allowed value is 10 and the maximum allowed value is 31536000. * Default: 10 evt_sid_cache_max_entries = <unsigned integer> * The maximum number of account SID cache entries. * This setting is optional. * This setting is global. It affects all Windows Event Log stanzas. * The minimum allowed value is 10 and the maximum allowed value is 40000. * Default: 10 index = <string> * Specifies the index that this input should send the data to. * This setting is optional. * Default: The default index Event Log filtering Filtering at the input layer is desirable to reduce the total processing load in network transfer and computation on the Splunk platform nodes that acquire and processing Event Log data. whitelist = <list of eventIDs> | key=regex [key=regex] blacklist = <list of eventIDs> | key=regex [key=regex] whitelist1 = <list of eventIDs> | key=regex [key=regex] whitelist2 = <list of eventIDs> | key=regex [key=regex] whitelist3 = <list of eventIDs> | key=regex [key=regex] whitelist4 = <list of eventIDs> | key=regex [key=regex] whitelist5 = <list of eventIDs> | key=regex [key=regex] whitelist6 = <list of eventIDs> | key=regex [key=regex] whitelist7 = <list of eventIDs> | key=regex [key=regex] whitelist8 = <list of eventIDs> | key=regex [key=regex] whitelist9 = <list of eventIDs> | key=regex [key=regex] blacklist1 = <list of eventIDs> | key=regex [key=regex] blacklist2 = <list of eventIDs> | key=regex [key=regex] blacklist3 = <list of eventIDs> | key=regex [key=regex] blacklist4 = <list of eventIDs> | key=regex [key=regex] blacklist5 = <list of eventIDs> | key=regex [key=regex] blacklist6 = <list of eventIDs> | key=regex [key=regex] blacklist7 = <list of eventIDs> | key=regex [key=regex] blacklist8 = <list of eventIDs> | key=regex [key=regex] blacklist9 = <list of eventIDs> | key=regex [key=regex] * These settings are optional. * Both numbered and unnumbered whitelists and blacklists support two formats: * A comma-separated list of event IDs. * A list of key=regular expression pairs. * You cannot combine these formats. You can use either format on a specific line. * Numbered whitelist settings are permitted from 1 to 9, so whitelist1 through whitelist9 and blacklist1 through blacklist9 are supported. * If no whitelist or blacklist rules are present, the input reads all events.
Event Log whitelist and blacklist formats
* Event ID list format: * A comma-separated list of terms. * Terms may be a single event ID (e.g. 6) or range of event IDs (e.g. 100-200) * Example: 4,5,7,100-200 * This applies to events with IDs 4, 5, 7, or any event ID between 100 and 200, inclusive. * The event ID list format provides no additional functionality over the key=regex format, but can be easier to understand: List format: 4,5,7,100-200 Regex equivalent: EventCode=%^(4|5|7|1..|200)$% * key=regex format: * A whitespace-separated list of Event Log components to match, and regular expressions to match against against them. * There can be one match expression or multiple expressions per line. * The key must belong to the set of valid keys provided below. * The regex consists of a leading delimiter, the regex expression, and a trailing delimiter. Examples: %regex%, *regex*, "regex" * When multiple match expressions are present, they are treated as a logical AND. In other words, all expressions must match for the line to apply to the event. * If the value represented by the key does not exist, it is not considered a match, regardless of the regex. * Example: whitelist = EventCode=%^200$% User=%jrodman% Include events only if they have EventCode 200 and relate to User jrodman Valid keys for the key=regex format: * The following keys are equivalent to the fields that appear in the text of the acquired events: * Category, CategoryString, ComputerName, EventCode, EventType, Keywords, LogName, Message, OpCode, RecordNumber, Sid, SidType, SourceName, TaskCategory, Type, User * There are two special keys that do not appear literally in the event. * $TimeGenerated: The time that the computer generated the event * $Timestamp: The time that the event was received and recorded by the Event Log service. * The 'EventType' key is only available on Windows Server 2003 / Windows XP and earlier. * The 'Type' key is only available on Windows Server 2008 / Windows Vista and later. * For a detailed definition of these keys, see the "Monitor Windows Event Log Data" topic in the online documentation. suppress_text = [0|1] * Whether or not to include the description of the event text for a given Event Log event. * This setting is optional. * Set this setting to 1 to suppress the inclusion of the event text description. * Set this value to 0 to include the event text description. * Default: 0 renderXml = <boolean> * Whether or not the input returns the event data in XML (eXtensible Markup Language) format or in plain text. * Set this to "true" to render events in XML. * Set this to "false" to output events in plain text. * If you set this setting to "true", you should also set the 'suppress_text', 'suppress_sourcename', 'suppress_keywords', 'suppress_task', and 'suppress_opcode' settings to "true" to improve thruput performance. * Default: false
Active Directory Monitor
[admon://<name>] * This section explains possible settings for configuring the Active Directory monitor input. * Each admon:// stanza represents an individually configured Active Directory monitoring input. If you configure the input with Splunk Web, then the value of "<NAME>" matches what was specified there. While you can add Active Directory monitor inputs manually, it is best practice to use Splunk Web to configure Active Directory monitor inputs because it is easy to mistype the values for Active Directory monitor objects. targetDc = <string> * The fully qualified domain name of a valid, network-accessible Active Directory domain controller (DC). * This setting is case sensitive. Do not use 'targetdc' or 'targetDC', but rather 'targetDc'. * Default: The DC that the local host used to connect to AD. The input binds to its root Distinguished Name (DN). startingNode = <string> * Where in the Active Directory directory tree to start monitoring. * The user that you configure Splunk software to run as at installation determines where the input starts monitoring. * Default: the root of the directory tree monitorSubtree = [0|1] * Whether or not to monitor the subtree(s) of a given Active Directory tree path. * Set this to 1 to monitor subtrees of a given directory tree path and 0 to monitor only the path itself. * Default: 1 (monitor subtrees of a given directory tree path) disabled = [0|1] * Whether or not the input is enabled. * Set this to 1 to disable the input and 0 to enable it. * Default: 0 (enabled) index = <string> * The index to store incoming data into for this input. * This setting is optional. * Default: the default index printSchema = [0|1] * Whether or not to print the Active Directory schema. * Set this to 1 to print the schema and 0 to not print the schema. * Default: 1 (print the Active Directory schema) baseline = [0|1] * Whether or not to query baseline objects. * Baseline objects are objects which currently reside in Active Directory. * Baseline objects also include previously deleted objects. * Set this to 1 to query baseline objects, and 0 to not query baseline objects. * Default: 0 (do not query baseline objects)
Windows Registry Monitor
[WinRegMon://<name>] * This section explains possible settings for configuring the Windows Registry Monitor input. * Each WinRegMon:// stanza represents an individually configured WinRegMon monitoring input. * If you configure the inputs with Splunk Web, the value of "<NAME>" matches what was specified there. While you can add event log monitor inputs manually, it is best practice to use Splunk Web to configure Windows registry monitor inputs because it is easy to mistype the values for Registry hives and keys. * The WinRegMon input is for local systems only. You cannot monitor the Registry remotely. proc = <string> * The processes this input should monitor for Registry access. * If set, matches against the process name which performed the Registry access. * The input includes events from processes that match the regular expression that you specify here. * The input filters out events for processes that do not match the regular expression. * No default. hive = <string> * The Registry hive(s) that this input should monitor for Registry access. * If set, matches against the Registry key that was accessed. * The input includes events from Registry hives that match the regular expression that you specify here. * The input filters out events for Registry hives that do not match the regular expression. * No default. type = <string> * A regular expression that specifies the type(s) of Registry event(s) that you want the input to monitor. * No default. baseline = [0|1] * Whether or not the input should get a baseline of Registry events when it starts. * If you set this to 1, the input captures a baseline for the specified hive when it starts for the first time. It then monitors live events. * Default: 0 (do not capture a baseline for the specified hive first before monitoring live events) baseline_interval = <integer> * Selects how much downtime in continuous registry monitoring should trigger a new baseline for the monitored hive and/or key. * In detail: * Sets the minimum time interval, in seconds, between baselines. * At startup, a WinRegMon input does not generate a baseline if less time has passed since the last checkpoint than baseline_interval chooses. * In normal operation, checkpoints are updated frequently as data is acquired, so this will cause baselines to occur only when monitoring was not operating for a period of time. * If baseline is set to 0 (disabled), has no effect. * Default: 0 (always baseline on startup, if baseline is 1) disabled = [0|1] * Whether or not the input is enabled. * Set this to 1 to disable the input, or 0 to enable it. * Default: 0 (enabled) index = <string> * The index that this input should send the data to. * This setting is optional. * Default: the default index
Windows Host Monitoring
[WinHostMon://<name>] * This section explains possible settings for configuring the Windows host monitor input. * Gathers status information from the local Windows system components as per the type field below. * Each WinHostMon:// stanza represents an WinHostMon monitoring input. * The "<name>" component of the stanza name is used as the source field on generated events, unless an explicit source setting is added to the stanza. It does not affect what data is collected (see type setting for that). * If you configure the input in Splunk Web, the value of "<name>" matches what was specified there. * NOTE: The WinHostMon input is for local Windows systems only. You cannot monitor Windows host information remotely. type = <semicolon-separated list> * An expression that specifies the type(s) of host inputs that you want the input to monitor. * Type can be (case insensitive): Computer;Process;Processor;NetworkAdapter;Service;OperatingSystem;Disk;Driver;Roles * No default. interval = <integer> * The interval, in seconds, between when the input runs to gather Windows host information and generate events. * See 'interval' in the Scripted input section for more information. disabled = [0|1] * Whether or not the input is enabled. * Set this to 1 to disable the input, or 0 to enable it. * Default: 0 (enabled) index = <string> * The index that this input should send the data to. * This setting is optional. * Default: the default index [WinPrintMon://<name>] * This section explains possible settings for configuring the Windows print monitor input. * Each WinPrintMon:// stanza represents an WinPrintMon monitoring input. The value of "<name>" matches what was specified in Splunk Web. * NOTE: The WinPrintMon input is for local Windows systems only. * The "<name>" component of the stanza name is used as the source field on generated events, unless an explicit source setting is added to the stanza. It does not affect what data is collected (see type setting for that). type = <semicolon-separated list> * An expression that specifies the type(s) of print inputs that you want the input to monitor. * Type can be (case insensitive): Printer;Job;Driver;Port * No default. baseline = [0|1] * Whether or not to capture a baseline of print objects when the input starts for the first time. * If you set this setting to 1, the input captures a baseline of the current print objects when the input starts for the first time. * Default: 0 (do not capture a baseline) disabled = [0|1] * Whether or not the input is enabled. * Set to 1 to disable the input, or 0 to enable it. * Default: 0 (enabled) index = <string> * The index that this input should send the data to. * This setting is optional. * Default: the default index [WinNetMon://<name>] * This section explains possible settings for configuring a Network Monitor input. * Each WinNetMon:// stanza represents an individually configured network monitoring input. The value of "<name>" matches what was specified in Splunk Web. It is best practice to use Splunk Web to configure Network Monitor inputs because it is easy to mistype the values for Network Monitor objects. remoteAddress = <regular expression> * A regular expression that represents the remote IP address of a host that is involved in network communication. * This setting accepts a regular expression that matches against IP addresses only, not host names. For example: 192\.163\..* * The input includes events for remote IP addresses that match the regular expression that you specify here. * The input filters out events for remote IP addresses that do not match the regular expression. * No default (including all remote address events) process = <regular expression> * A regular expression that represents the process or application that performed a network access. * The input includes events for processes that match the regular expression that you specify here. * The input filters out events for processes that do not match the regular expression. * No default (including all processes and application events) user = <regular expression> * A regular expression that represents the Windows user name that performed a network access. * The input includes events for user names that match the regular expression that you specify here. * The input filters out events for user names that do not match the regular expression. * No default (including all user name events) addressFamily = ipv4;ipv6 * Determines the events to include by network address family. * Setting "ipv4" alone includes only IPv4 packets, while "ipv6" alone includes only IPv6 packets. * To specify both families, separate them with a semicolon. For example: ipv4;ipv6 * No default (including events with both address families) packetType = connect;accept;transport. * Determines the events to include by network packet type. * To specify multiple packet types, separate them with a semicolon. For example: connect;transport * No default (including events with any packet type) direction = inbound;outbound * Determines the events to include by network transport direction. * To specify multiple directions, separate them with a semicolon. For example: inbound;outbound * No default (including events with any direction) protocol = tcp;udp * Determines the events to include by network protocol. * To specify multiple protocols, separate them with a semicolon. For example: tcp;udp * For more information about protocols, see http://www.ietf.org/rfc/rfc1700.txt * No default (including events with all protocols) readInterval = <integer> * How often, in milliseconds, that the input should read the network kernel driver for events. * Advanced option. Use the default value unless there is a problem with input performance. * Set this to adjust the frequency of calls into the network kernel driver. * Choosing lower values (higher frequencies) can reduce network performance, while higher numbers (lower frequencies) can cause event loss. * The minimum allowed value is 10 and the maximum allowed value is 1000. * Default: 100 driverBufferSize = <integer> * The maximum number of packets that the network kernel driver retains for retrieval by the input. * Set to adjust the maximum number of network packets retained in the network driver buffer. * Advanced option. Use the default value unless there is a problem with input performance. * Configuring this setting to lower values can result in event loss, while higher values can increase the size of non-paged memory on the host. * The minimum allowed value is 128 and the maximum allowed value is 32768. * Default: 32768 userBufferSize = <integer> * The maximum size, in megabytes, of the user mode event buffer. * Controls amount of packets cached in the the user mode. * Advanced option. Use the default value unless there is a problem with input performance. * Configuring this setting to lower values can result in event loss, while higher values can increase the amount of memory that the network monitor uses. * The minimum allowed value is 20 and the maximum allowed value is 500. * Default: 20 mode = single|multikv * Specifies how the network monitor input generates events. * Set to "single" to generate one event per packet. * Set to "multikv" to generate combined events of many packets in multikv format (many packets described in a single table as one event). * Default: single multikvMaxEventCount = <integer> * The maximum number of packets to combine in multikv format when you set the 'mode' setting to "multikv". * Has no effect when 'mode' is set to "single". * Advanced option. * The minimum allowed value is 10 and the maximum allowed value is 500. * Default: 100 multikvMaxTimeMs = <integer> * The maximum amount of time, in milliseconds, to accumulate packet data to combine into a large tabular event in multikv format. * Has no effect when 'mode' is set to 'single'. * Advanced option. * The minimum allowed value is 100 and the maximum allowed value is 5000. * Default: 1000 sid_cache_disabled = 0|1 * Enables or disables account Security IDentifier (SID) cache. * This setting is global. It affects all Windows Network Monitor stanzas. * Default: 0 sid_cache_exp = <integer> * The expiration time, in seconds, for account SID cache entries. * Optional. * This setting is global. It affects all Windows Network Monitor stanzas. * The minimum allowed value is 10 and the maximum allowed value is 31536000. * Default: 3600 sid_cache_exp_neg = <integer> * The expiration time, in seconds, for negative account SID cache entries. * Optional. * This setting is global. It affects all Windows Network Monitor stanzas. * The minimum allowed value is 10 and the maximum allowed value is 31536000. * Default: 10 sid_cache_max_entries = <integer> * The maximum number of account SID cache entries. * Optional. * This setting is global. It affects all Windows Network Monitor stanzas. * The minimum allowed value is 10 and the maximum allowed value is 40000. * Default: 10 disabled = 0|1 * Whether or not the input is enabled. * Set to 1 to disable the input, and 0 to enable it. * Default: 0 (enabled) index = <string> * The index that this input should send the data to. * Optional. * Default: the default index Global settings for the powershell modinput. [powershell] io_threads = <integer> * The number of threads that Splunk software spawns to run PowerShell scripts that have been configured in inputs.conf. * If you specify a value that is less than or equal to 0, Splunk software autotunes this setting. * The default can vary. Splunk software autotunes the number of threads based on the availability of CPU resources on the machine. serialization_threads = <integer> * The number of threads that Splunk software spawns for serialization of PowerShell objects that it has collected into XML strings. * This serialization, or conversion of objects, occurs according to the Modular Input XML protocol. * If you specify a value that is less than or equal to 0, Splunk software autotunes this setting. * Default: The default can vary. Splunk software autotunes the number of threads based on available CPU resources on the machine. event_serialization_format = [ kv | json ] * The event format that Powershell objects are serialized into. * The supported event formats are "kv" and "json". * For example, given the following PowerShell object: $psObj = @{ A: "a string" B: 18 C: "a log line" } If you set 'event_serialization_format' to "kv", the Splunk platform indexes the event as follows: A="a string" B=18 C="a log line" If you set 'event_serialization_format' to "json", the Splunk platform indexes the event as follows: { "A": "a string", "B": 18, "C": "a log line" } * Default: kv [powershell://<name>] * Runs Windows PowerShell version 3 commands or scripts. script = <string> * A PowerShell command-line script or .ps1 script file that the input should run. * No default. schedule = [<positive integer>|<cron schedule>] * How often to run the specified PowerShell command or script. * You can specify a number in seconds, or provide a valid cron schedule. * Default: Runs the command or script once, at startup. Global settings for the powershell2 modinput. [powershell2] io_threads = <integer> * The number of threads that Splunk software spawns to run PowerShell scripts that have been configured in inputs.conf. * If you specify a value that is less than or equal to 0, Splunk software autotunes this setting. * The default can vary. Splunk software autotunes the number of threads based on the availability of CPU resources on the machine. event_serialization_format = [ kv | json ] * The event format that Powershell objects are serialized into. * The supported event formats are "kv" and "json". * For example, given the following PowerShell object: $psObj = @{ A: "a string" B: 18 C: "a log line" } If you set 'event_serialization_format' to "kv", the Splunk platform indexes the event as follows: A="a string" B=18 C="a log line" If you set 'event_serialization_format' to "json", the Splunk platform indexes the event as follows: { "A": "a string", "B": 18, "C": "a log line" } * Default: kv [powershell2://<name>] * Runs Windows PowerShell version 2 commands or scripts. script = <string> * A PowerShell command-line script or .ps1 script file that the input should run. * No default. schedule = <schedule> * How often to run the specified PowerShell command or script. * You can provide a valid cron schedule. * Default: Runs the command or script once, at startup.
Remote Queue Monitor
[remote_queue:<name>] * This section explains possible settings for configuring a remote queue. * Each remote_queue: stanza represents an individually configured remote queue monitoring input. remote_queue.* = <string> * With remote queues, communication between the indexer and the remote queue system might require additional configuration, specific to the type of remote queue. * You can pass configuration information to the storage system by specifying the settings through the following schema: remote_queue.<scheme>.<config-variable> = <value>. For example: remote_queue.sqs.access_key = ACCESS_KEY * This setting is optional. * No default. remote_queue.type = [sqs|kinesis] * Currently not supported. This setting is related to a feature that is still under development. * Required. * Specifies the remote queue type, either Amazon Web Services (AWS) Simple Queue Service (SQS) or Amazon Kinesis. compressed = <boolean> * See the description for TCPOUT ATTRIBUTES in outputs.conf.spec. negotiateProtocolLevel = <unsigned integer> * See the description for TCPOUT ATTRIBUTES in outputs.conf.spec. channelReapInterval = <integer> * See the description for TCPOUT ATTRIBUTES in outputs.conf.spec. channelTTL = <integer> * See the description for TCPOUT ATTRIBUTES in outputs.conf.spec. channelReapLowater = <integer> * See the description for TCPOUT ATTRIBUTES in outputs.conf.spec. concurrentChannelLimit = <unsigned integer> * See the description for [splunktcp].
SQS specific settings
remote_queue.sqs.access_key = <string> * Currently not supported. This setting is related to a feature that is still under development. * The access key to use when authenticating with the remote queue system supporting the SQS API. * If not specified, the indexer looks for these environment variables: 'AWS_ACCESS_KEY_ID' or 'AWS_ACCESS_KEY' (in that order). If the environment variables are not set and the indexer is running on Elastic Compute Cloud (EC2), the indexer attempts to use the secret key from the Identity and Access Management (IAM) role. * This setting is optional. * No default. remote_queue.sqs.secret_key = <string> * Currently not supported. This setting is related to a feature that is still under development. * The secret key to use when authenticating with the remote queue system supporting the SQS API. * If not specified, the indexer looks for these environment variables: AWS_SECRET_ACCESS_KEY or AWS_SECRET_KEY (in that order). If the environment variables are not set and the indexer is running on EC2, the indexer attempts to use the secret key from the IAM role. * This setting is optional. * No default. remote_queue.sqs.auth_region = <string> * Currently not supported. This setting is related to a feature that is still under development. * The authentication region to use when signing the requests when interacting with the remote queue system supporting the SQS API. * If not specified and the indexer is running on EC2, the auth_region is constructed automatically based on the EC2 region of the instance where the the indexer is running. * This setting is optional. * No default. remote_queue.sqs.endpoint = <URL> * Currently not supported. This setting is related to a feature that is still under development. * The URL of the remote queue system supporting the SQS API. * The scheme, http or https, can be used to enable or disable SSL connectivity with the endpoint. * If not specified, the endpoint is constructed automatically based on the auth_region as follows: https://sqs.<auth_region>.amazonaws.com * If specified, the endpoint must match the effective auth_region, which is either a value specified in 'remote_queue.sqs.auth_region' or a value constructed automatically based on the EC2 region of the running instance. * Example: https://sqs.us-west-2.amazonaws.com/ * This setting is optional. * No default. remote_queue.sqs.max_connections = <unsigned integer> * Currently not supported. This setting is related to a feature that is still under development. * The maximum number of HTTP connections to have in progress for certain queue operations. * A value of 0 means unlimited. * Default: 8 remote_queue.sqs.message_group_id = <string> * Currently not supported. This setting is related to a feature that is still under development. * The Message Group ID for Amazon Web Services Simple Queue Service (SQS) First-In, First-Out (FIFO) queues. * Setting a Message Group ID controls how messages within an AWS SQS queue are processed. * For information on SQS FIFO queues and how messages in those queues are processed, see "Recommendations for FIFO queues" in the AWS SQS Developer Guide. * If you configure this setting, Splunk software assumes that the SQS queue is a FIFO queue, and that messages in the queue should be processed first-in, first-out. * Otherwise, Splunk software assumes that the SQS queue is a standard queue. * Can be between 1-128 alphanumeric or punctuation characters. * NOTE: FIFO queues must have Content-Based Deduplication enabled. * This setting is optional. * No default. remote_queue.sqs.retry_policy = [max_count|none] * Currently not supported. This setting is related to a feature that is still under development. * The retry policy to use for remote queue operations. * A retry policy specifies whether and how to retry file operations that fail for those failures that might be intermittent. * Retry policies: + "max_count": Imposes a maximum number of times a queue operation can be retried upon intermittent failure. + "none": Do not retry file operations upon failure. * This setting is optional. * Default: "max_count" remote_queue.sqs.max_count.max_retries_per_part = <unsigned integer> * Currently not supported. This setting is related to a feature that is still under development. * When 'remote_queue.sqs.retry_policy' is set to "max_count", sets the maximum number of times a queue operation can be retried upon intermittent failure. * This setting is optional. * Default: 9 remote_queue.sqs.timeout.connect = <unsigned integer> * Currently not supported. This setting is related to a feature that is still under development. * The connection timeout, in seconds, when interacting with SQS for this queue. * This setting is optional. * Default: 5 remote_queue.sqs.timeout.read = <unsigned integer> * Currently not supported. This setting is related to a feature that is still under development. * The read timeout, in seconds, when interacting with SQS for this queue. * This setting is optional. * Default: 60 remote_queue.sqs.timeout.write = <unsigned integer> * Currently not supported. This setting is related to a feature that is still under development. * The write timeout, in seconds, when interacting with SQS for this queue. * This setting is optional. * Default: 60 remote_queue.sqs.timeout.receive_message = <unsigned integer> * The receive message wait time, in seconds, when interacting with SQS for this queue. * When set to greater than 0, enables "long polling." If there are no messages immediately available, the queue waits at most 'remote_queue.sqs.timeout.receive_message' seconds for a message to become available. * When 0, disables long polling. * When not set, uses the value configured for the queue via the AWS SQS console. * Maximum value: 20 * This setting is optional. * Default: 20 remote_queue.sqs.timeout.visibility = <unsigned integer> * Currently not supported. This setting is related to a feature that is still under development. * The default "visibility timeout," in seconds, to use when explicitly changing the visibility of specific messages in the queue. * NOTE: Changing the value of 'remote_queue.sqs.timeout.visibility' does not change the implicit visibility timeout configured for the queue in the AWS SQS console. * This setting is optional. * Default: 60 remote_queue.sqs.buffer.visibility = <unsigned integer> * Currently not supported. This setting is related to a feature that is still under development. * The default time, in seconds, before 'remote_queue.sqs.timeout.visibility' at which visibility of specific messages in the queue needs to be changed. * This setting is optional. * Default: 15 remote_queue.sqs.executor_max_workers_count = <positive integer> * Currently not supported. This setting is related to a feature that is still under development. * The maximum number of worker threads that can be used by indexer per pipeline set to execute SQS tasks. * A value of 0 is equivalent to 1. * Default: 8 remote_queue.sqs.min_pending_messages = <unsigned integer> * Currently not supported. This setting is related to a feature that is still under development. * The default "minimum number of pending messages" to use before receiving messages off remote queue. Messages are only received when the sum of internal queue message count and pending object GET (from large messages storage) count is below the set value. * This setting is optional. * Default: 10 remote_queue.sqs.large_message_store.endpoint = <string> * Currently not supported. This setting is related to a feature that is still under development. * The URL of the remote storage system supporting the S3 API. * The scheme, http or https, can be used to enable or disable SSL connectivity with the endpoint. * If not specified, the endpoint is constructed automatically based on the auth_region as follows: https://s3-<auth_region>.amazonaws.com * If specified, the endpoint must match the effective auth_region, which is either a value specified via 'remote_queue.sqs.auth_region' or a value constructed automatically based on the EC2 region of the running instance. * Example: https://s3-us-west-2.amazonaws.com/ * This setting is optional. * No default. remote_queue.sqs.large_message_store.path = <string> * Currently not supported. This setting is related to a feature that is still under development. * The remote storage location where messages that are larger than the underlying queue maximum message size will reside. * The format for this attribute is: <scheme>://<remote-location-specifier> * The "scheme" identifies a supported external storage system type. * The "remote-location-specifier" is an external system-specific string for identifying a location inside the storage system. * These external systems are supported: - Object stores that support the AWS S3 protocol. These use the scheme "s3". For example, "path=s3://mybucket/some/path". * If not specified, messages exceeding the underlying queue's maximum message size are dropped. * This setting is optional. * No default. remote_queue.large_message_store.supports_versioning = <boolean> * Currently not supported. This setting is related to a feature that is still under development. * Specifies whether or not the remote storage supports versioning. * Versioning is a means of keeping multiple variants of an object in the same bucket on the remote storage. * This setting is optional. * Default: false
Kinesis specific settings
remote_queue.kinesis.access_key = <string> * Currently not supported. This setting is related to a feature that is still under development. * Specifies the access key to use when authenticating with the remote queue system supporting the Kinesis API. * If not specified, the forwarder will look for these environment variables: AWS_ACCESS_KEY_ID or AWS_ACCESS_KEY (in that order). If the environment variables are not set and the forwarder is running on EC2, the forwarder attempts to use the secret key from the IAM role. * This setting is optional. * No default. remote_queue.kinesis.secret_key = <string> * Currently not supported. This setting is related to a feature that is still under development. * Specifies the secret key to use when authenticating with the remote queue system supporting the Kinesis API. * If not specified, the forwarder will look for these environment variables: AWS_SECRET_ACCESS_KEY or AWS_SECRET_KEY (in that order). If the environment variables are not set and the forwarder is running on EC2, the forwarder attempts to use the secret key from the IAM role. * This setting is optional. * No default. remote_queue.kinesis.auth_region = <string> * Currently not supported. This setting is related to a feature that is still under development. * The authentication region to use when signing the requests when interacting with the remote queue system supporting the Kinesis API. * If not specified and the forwarder is running on EC2, the auth_region will be constructed automatically based on the EC2 region of the instance where the the forwarder is running. * This setting is optional. * No default. remote_queue.kinesis.endpoint = <URL> * Currently not supported. This setting is related to a feature that is still under development. * The URL of the remote queue system supporting the Kinesis API. * The scheme, http or https, can be used to enable or disable SSL connectivity with the endpoint. * If not specified, the endpoint is constructed automatically based on the auth_region as follows: https://kinesis.<auth_region>.amazonaws.com * If specified, the endpoint must match the effective auth_region, which is either a value specified via 'remote_queue.kinesis.auth_region' or a value constructed automatically based on the EC2 region of the running instance. * Example: https://kinesis.us-west-2.amazonaws.com/ * This setting is optional. * No default. remote_queue.kinesis.retry_policy = [max_count|none] * The retry policy to use for remote queue operations. * A retry policy specifies whether and how to retry file operations that fail for those failures that might be intermittent. * Retry policies: + "max_count": Imposes a maximum number of times a queue operation will be retried upon intermittent failure. + "none": Do not retry file operations upon failure. * This setting is optional. * Default: "max_count" remote_queue.kinesis.max_count.max_retries_per_part = <unsigned integer> * When 'remote_queue.kinesis.retry_policy' is "max_count", sets the maximum number of times a queue operation is retried upon intermittent failure. * This setting is optional. * Default: 9 remote_queue.kinesis.timeout.connect = <unsigned integer> * Currently not supported. This setting is related to a feature that is still under development. * The connection timeout, in milliseconds, when interacting with Kinesis for this queue. * This setting is optional. * Default: 5000 remote_queue.kinesis.timeout.read = <unsigned integer> * Currently not supported. This setting is related to a feature that is still under development. * The read timeout, in milliseconds, when interacting with Kinesis for this queue. * This setting is optional. * Default: 60000 remote_queue.kinesis.timeout.write = <unsigned integer> * Currently not supported. This setting is related to a feature that is still under development. * The write timeout, in milliseconds, when interacting with Kinesis for this queue. * This setting is optional. * Default: 60000 remote_queue.kinesis.executor_max_workers_count = <positive integer> * Currently not supported. This setting is related to a feature that is still under development. * The maximum number of worker threads that can be used by indexer per pipeline set to execute kinesis queue tasks. * A value of 0 is equivalent to 1. * Default: 8 remote_queue.kinesis.max_messages = <unsigned integer> * Currently not supported. This setting is related to a feature that is still under development. * The default "maximum number of messages" (that are received from remote_queue endpoint) to store in kinesis in-memory message queue. * This setting is optional. * Default: 10000 remote_queue.kinesis.min_pending_messages = <unsigned integer> * Currently not supported. This setting is related to a feature that is still under development. * The default "minimum number of pending messages" to use before receiving messages off kinesis in-memory message queue. Messages are only received when sum of internal queue message count and pending object GET (from large messages storage) count is below the set value. * This setting is optional. * Default: 50 remote_queue.kinesis.max_checkpoints = <unsigned integer> * Currently not supported. This setting is related to a feature that is still under development. * The default "maximum number of messages" (that have been received from remote_queue endpoint and completely consumed) to store in the Kinesis in-memory checkpoint queue. * This setting is optional. * Default: 100000 remote_queue.kinesis.roll_remote_buckets_interval = <unsigned integer> * Currently not supported. This setting is related to a feature that is still under development. * The default interval, in seconds, that the Kinesis remote queue input worker waits before it rolls the remote storage enabled buckets. * This setting is optional. * Default: 30 remote_queue.kinesis.large_message_store.endpoint = <string> * Currently not supported. This setting is related to a feature that is still under development. * The URL of the remote storage system supporting the S3 API. * The scheme, http or https, can be used to enable or disable SSL connectivity with the endpoint. * If not specified, the endpoint will be constructed automatically based on the auth_region as follows: https://s3-<auth_region>.amazonaws.com * If specified, the endpoint must match the effective auth_region, which is either a value specified via 'remote_queue.kinesis.auth_region' or a value constructed automatically based on the EC2 region of the running instance. * Example: https://s3-us-west-2.amazonaws.com/ * This setting is optional. * No default. remote_queue.kinesis.large_message_store.path = <string> * Currently not supported. This setting is related to a feature that is still under development. * The remote storage location where messages larger than the underlying queue maximum message size will reside. * The format for this attribute is: <scheme>://<remote-location-specifier> * The "scheme" identifies a supported external storage system type. * The "remote-location-specifier" is an external system-specific string for identifying a location inside the storage system. * These external systems are supported: - Object stores that support AWS's S3 protocol. These use the scheme "s3". For example, "path=s3://mybucket/some/path". * If not specified, messages exceeding the underlying queue maximum message size are dropped. * This setting is optional. * No default.
Modular Inputs
python.version = {default|python|python2|python3} * For Python scripts only, selects which Python version to use. * Either "default" or "python" select the system-wide default Python version. * Optional. * Default: Not set; uses the system-wide Python version. run_introspection = <boolean> * Whether or not Splunk software runs introspection on a modular input scheme when you have disabled all of its associated scripts by using the 'disabled = 1' setting. * This setting applies only for modular inputs. It takes effect only if you specify it under a default stanza of a modular input scheme. * A default stanza of a modular input scheme begins with the notation [<scheme name>] * If set to "true", Splunk software runs introspection on a modular input scheme even when you have disabled all the input scripts for the scheme. * If set to "false", Splunk software does not run introspection on a modular input scheme where you have disabled all scripts for the scheme. * If introspection does not run for a scheme, then Splunk software does not register the modular input scripts that are associated with the scheme for execution and it is disabled completely. * Use the 'disabled' setting to enable or disable individual modular input scripts. * For example, to turn introspection off for the modular input scheme "myScheme": [myScheme] run_introspection = false * Default: true
inputs.conf.example
# Version 8.0.3 # # This is an example inputs.conf. Use this file to configure data inputs. # # To use one or more of these configurations, copy the configuration block into # inputs.conf in $SPLUNK_HOME/etc/system/local/. You must restart Splunk to # enable configurations. # # To learn more about configuration files (including precedence) please see the # documentation located at # http://docs.splunk.com/Documentation/Splunk/latest/Admin/Aboutconfigurationfiles # The following configuration reads all the files in the directory /var/log. [monitor:///var/log] # The following configuration reads all the files under /var/log/httpd and # classifies them as sourcetype::access_common. # # When checking a file for new data, if the file's modification time is from # before seven days ago, the file will no longer be checked for changes # until you restart the software. [monitor:///var/log/httpd] sourcetype = access_common ignoreOlderThan = 7d # The following configuration reads all the # files under /mnt/logs. When the path is /mnt/logs/<host>/... it # sets the hostname (by file) to <host>. [monitor:///mnt/logs] host_segment = 3 # The following configuration listens on TCP port 9997 for raw # data from ANY remote server (not just a Splunk instance). The host of the # data is set to the IP address of the remote server. [tcp://:9997] # The following configuration listens on TCP port 9995 for raw # data from ANY remote server. The host of the data is set as the host name of # the remote server. All data will also be assigned the sourcetype "log4j" and # the source "tcp:9995". [tcp://:9995] connection_host = dns sourcetype = log4j source = tcp:9995 # The following configuration listens on TCP port 9995 for raw # data from 10.1.1.10. # All data is assigned the host "webhead-1", the sourcetype "access_common" and # the the source "//10.1.1.10/var/log/apache/access.log". [tcp://10.1.1.10:9995] host = webhead-1 sourcetype = access_common source = //10.1.1.10/var/log/apache/access.log # The following configuration listens on TCP port 9996 for # Splunk cooked event data from ANY splunk forwarder. # The host of the data is set to the host name of the remote server ONLY IF the # remote data has no host set, or if it is set to "localhost". [splunktcp://:9996] connection_host = dns # The following configuration listens on TCP port 9996 for # distributed search data from 10.1.1.100. The data is processed the same as # locally indexed data. [splunktcp://10.1.1.100:9996] # The following configuration listens on TCP port 514 for data # from syslog.corp.company.net. The data is assigned the sourcetype "syslog" # and the host is set to the host name of the remote server. [tcp://syslog.corp.company.net:514] sourcetype = syslog connection_host = dns # Following configuration limits the acceptance of data to forwarders # that have been configured with the token value specified in 'token' field. # NOTE: The token value is encrypted. The REST endpoint encrypts the token # while saving it. [splunktcptoken://tok1] token = $7$ifQTPTzHD/BA8VgKvVcgO1KQAtr3N1C8S/1uK3nAKIE9dd9e9g== # Set up Secure Sockets Layer (SSL): [SSL] serverCert=$SPLUNK_HOME/etc/auth/server.pem password=password rootCA=$SPLUNK_HOME/etc/auth/cacert.pem requireClientCert=false [splunktcp-ssl:9996] # Use file system change monitor: [fschange:/etc/] fullEvent=true pollPeriod=60 recurse=true sendEventMaxSize=100000 index=main # Monitor the Security Windows Event Log channel, getting the most recent # events first, then older, and finally continuing to gather newly arriving events [WinEventLog://Security] disabled = 0 start_from = newest evt_dc_name = evt_dns_name = evt_resolve_ad_ds = evt_resolve_ad_obj = 1 checkpointInterval = 5 # Monitor the ForwardedEvents Windows Event Log channel, only gathering the # events that arrive after monitoring starts, going forward in time. [WinEventLog://ForwardedEvents] disabled = 0 start_from = oldest current_only = 1 batch_size = 10 checkpointInterval = 5 [tcp://9994] queueSize=50KB persistentQueueSize=100MB # Perfmon: Windows performance monitoring examples # You must specify the names of objects, counters and instances # exactly as they are shown in the Performance Monitor application. Splunk Web # is the recommended interface to use to configure performance monitor inputs. # These stanzas gather performance data from the local system only. # Use wmi.conf for performance monitor metrics on remote systems. # Query the PhysicalDisk performance object and gather disk access data for # all physical drives installed in the system. Store this data in the # "perfmon" index. # Note: If the interval attribute is set to 0, Splunk will reset the interval # to 1. [perfmon://LocalPhysicalDisk] interval = 0 object = PhysicalDisk counters = Disk Bytes/sec; % Disk Read Time; % Disk Write Time; % Disk Time instances = * disabled = 0 index = PerfMon # Gather common memory statistics using the Memory performance object, every # 5 seconds. Store the data in the "main" index. Since none of the counters # specified have applicable instances, the instances attribute is not required. [perfmon://LocalMainMemory] interval = 5 object = Memory counters = Committed Bytes; Available Bytes; % Committed Bytes In Use disabled = 0 index = main # Gather data on USB activity levels every 10 seconds. Store this data in the # default index. [perfmon://USBChanges] interval = 10 object = USB counters = Usb Control Data Bytes/Sec instances = * disabled = 0 # Admon: Windows Active Directory monitoring examples # Monitor the default domain controller (DC) for the domain that the computer # running Splunk belongs to. Start monitoring at the root node of Active # Directory. [admon://NearestDC] targetDc = startingNode = # Monitor a specific DC, with a specific starting node. Store the events in # the "admon" Splunk index. Do not print Active Directory schema. Do not # index baseline events. [admon://DefaultTargetDC] targetDc = pri01.eng.ad.splunk.com startingNode = OU=Computers,DC=eng,DC=ad,DC=splunk,DC=com index = admon printSchema = 0 baseline = 0 # Monitor two different DCs with different starting nodes. [admon://DefaultTargetDC] targetDc = pri01.eng.ad.splunk.com startingNode = OU=Computers,DC=eng,DC=ad,DC=splunk,DC=com [admon://SecondTargetDC] targetDc = pri02.eng.ad.splunk.com startingNode = OU=Computers,DC=hr,DC=ad,DC=splunk,DC=com
indexes.conf | instance.cfg.conf |
This documentation applies to the following versions of Splunk® Enterprise: 8.0.3
Feedback submitted, thanks!