indexes.conf
The following are the spec and example files for indexes.conf
.
indexes.conf.spec
# Version 7.1.1 # # This file contains all possible options for an indexes.conf file. Use # this file to configure Splunk's indexes and their properties. # # There is an indexes.conf in $SPLUNK_HOME/etc/system/default/. To set # custom configurations, place an indexes.conf in # $SPLUNK_HOME/etc/system/local/. For examples, see indexes.conf.example. # You must restart Splunk to enable configurations. # # To learn more about configuration files (including precedence) please see # the documentation located at # http://docs.splunk.com/Documentation/Splunk/latest/Admin/Aboutconfigurationfiles # # CAUTION: You can drastically affect your Splunk installation by changing # these settings. Consult technical support # (http://www.splunk.com/page/submit_issue) if you are not sure how to # configure this file. #
GLOBAL SETTINGS
# Use the [default] stanza to define any global settings. # * You can also define global settings outside of any stanza, at the top # of the file. # * Each conf file should have at most one default stanza. If there are # multiple default stanzas, attributes are combined. In the case of # multiple definitions of the same attribute, the last definition in the # file wins. # * If an attribute is defined at both the global level and in a specific # stanza, the value in the specific stanza takes precedence. sync = <nonnegative integer> * The index processor syncs events every <integer> number of events. * Set to 0 to disable. * Highest legal value is 32767 * Defaults to 0. defaultDatabase = <index name> * If no index is specified during search, Splunk searches the default index. * The specified index displays as the default in Splunk Manager settings. * Defaults to "main". queryLanguageDefinition = <path to file> * DO NOT EDIT THIS SETTING. SERIOUSLY. * The path to the search language definition file. * Defaults to $SPLUNK_HOME/etc/searchLanguage.xml. lastChanceIndex = <index name> * Gives ability to define a last chance index for events destined for non-existent indexes. * If an event arrives whose index destination key points to an index that is not configured (such as when using index=<index name> in the input stanza or by a setting in a transform), it will route that event to the index specified by this setting. The index destination key of that event will be overwritten with the specified index name before routing. * <index name> must name an existing enabled index. Splunk will not start if this is not the case. * If this setting is not defined or is empty, it will drop such events. * If set to "default", then the default index specified by the "defaultDatabase" will be used as a last chance index. * Defaults to empty. memPoolMB = <positive integer>|auto * Determines how much memory is given to the indexer memory pool. This restricts the number of outstanding events in the indexer at any given time. * Must be greater than 0; maximum value is 1048576 (which corresponds to 1 TB) * Setting this too high can lead to splunkd memory usage going up substantially. * Setting this too low can degrade splunkd indexing performance. * Setting this to "auto" or an invalid value will cause Splunk to autotune this parameter. * Defaults to "auto". * The values derived when "auto" is seen are as follows: * System Memory Available less than ... | memPoolMB 1 GB | 64 MB 2 GB | 128 MB 8 GB | 128 MB 8 GB or higher | 512 MB * Only set this value if you are an expert user or have been advised to by Splunk Support. * CARELESSNESS IN SETTING THIS MAY LEAD TO PERMANENT BRAIN DAMAGE OR LOSS OF JOB. indexThreads = <nonnegative integer>|auto * Determines the number of threads to use for indexing. * Must be at least 1 and no more than 16. * This value should not be set higher than the number of processor cores in the box. * If splunkd is also doing parsing and aggregation, the number should be set lower than the total number of processors minus two. * Setting this to "auto" or an invalid value will cause Splunk to autotune this parameter. * Only set this value if you are an expert user or have been advised to by Splunk Support. * CARELESSNESS IN SETTING THIS MAY LEAD TO PERMANENT BRAIN DAMAGE OR LOSS OF JOB. * Defaults to "auto". rtRouterThreads = 0|1 * Set this to 1 if you expect to use non-indexed real time searches regularly. Index throughput drops rapidly if there are a handful of these running concurrently on the system. * If you are not sure what "indexed vs non-indexed" real time searches are, see README of indexed_realtime* settings in limits.conf * NOTE: This is not a boolean value, only 0 or 1 is accepted. In the future, we may allow more than a single thread, but current implementation only allows one to create a single thread per pipeline set rtRouterQueueSize = <positive integer> * Defaults to 10000 * This setting is only relevant if rtRouterThreads != 0 * This queue sits between the indexer pipeline set thread (producer) and the rtRouterThread * Changing the size of this queue may impact real time search performance selfStorageThreads = <positive integer> * Specifies the number of threads used to transfer data to customer-owned remote storage. * Defaults to 2 * The threads are created on demand when any index is configured with self storage options. assureUTF8 = true|false * Verifies that all data retrieved from the index is proper by validating all the byte strings. * This does not ensure all data will be emitted, but can be a workaround if an index is corrupted in such a way that the text inside it is no longer valid utf8. * Will degrade indexing performance when enabled (set to true). * Can only be set globally, by specifying in the [default] stanza. * Defaults to false. enableRealtimeSearch = true|false * Enables or disables real-time searches across the indexer. This cannot be overridden on an index-by-index basis. * Defaults to true. suppressBannerList = <comma-separated list of strings> * suppresses index missing warning banner messages for specified indexes * Defaults to empty maxRunningProcessGroups = <positive integer> * splunkd fires off helper child processes like splunk-optimize, recover-metadata, etc. This param limits how many child processes can be running at any given time. * This maximum applies to entire splunkd, not per index. If you have N indexes, there will be at most maxRunningProcessGroups child processes, not N*maxRunningProcessGroups * Must maintain maxRunningProcessGroupsLowPriority < maxRunningProcessGroups * This is an advanced parameter; do NOT set unless instructed by Splunk Support * Highest legal value is 4294967295 * Defaults to 8 (note: up until 5.0 it defaulted to 20) maxRunningProcessGroupsLowPriority = <positive integer> * Of the maxRunningProcessGroups (q.v.) helper child processes, at most maxRunningProcessGroupsLowPriority may be low-priority (e.g. fsck) ones. * This maximum applies to entire splunkd, not per index. If you have N indexes, there will be at most maxRunningProcessGroupsLowPriority low-priority child processes, not N*maxRunningProcessGroupsLowPriority * Must maintain maxRunningProcessGroupsLowPriority < maxRunningProcessGroups * This is an advanced parameter; do NOT set unless instructed by Splunk Support * Highest legal value is 4294967295 * Defaults to 1 bucketRebuildMemoryHint = <positive integer>[KB|MB|GB]|auto * Suggestion for the bucket rebuild process for the size (bytes) of tsidx file it will try to build. * Larger files use more memory in rebuild, but rebuild will fail if there is not enough. * Smaller files make the rebuild take longer during the final optimize step. * Note: this value is not a hard limit on either rebuild memory usage or tsidx size. * This is an advanced parameter, do NOT set this unless instructed by Splunk Support. * Defaults to "auto", which varies by the amount of physical RAM on the host * less than 2GB RAM = 67108864 (64MB) tsidx * 2GB to 8GB RAM = 134217728 (128MB) tsidx * more than 8GB RAM = 268435456 (256MB) tsidx * If not "auto", then must be 16MB-1GB. * Value may be specified using a size suffix: "16777216" or "16MB" are equivalent. * Inappropriate use of this parameter will cause splunkd to not start if rebuild is required. * Highest legal value (in bytes) is 4294967295 inPlaceUpdates = true|false * If true, metadata updates are written to the .data files directly * If false, metadata updates are written to a temporary file and then moved into place * Intended for advanced debugging of metadata issues * Setting this parameter to false (to use a temporary file) will impact indexing performance, particularly with large numbers of hosts, sources, or sourcetypes (~1 million, across all indexes.) * This is an advanced parameter; do NOT set unless instructed by Splunk Support * Defaults to true serviceInactiveIndexesPeriod = <positive integer> * Defines how frequently inactive indexes are serviced, in seconds. * An inactive index is an index that has not been written to for a period greater than the value of serviceMetaPeriod. The inactive state is not affected by whether the index is being read from. * Defaults to 60 (seconds). * Highest legal value is 4294967295 serviceOnlyAsNeeded = true|false * DEPRECATED; use 'serviceInactiveIndexesPeriod'. * Causes index service (housekeeping tasks) overhead to be incurred only after index activity. * Indexer module problems may be easier to diagnose when this optimization is disabled (set to false). * Defaults to true. serviceSubtaskTimingPeriod = <positive integer> * Subtasks of indexer service task will be timed on every Nth execution, where N = value of this parameter, in seconds. * Smaller values will give greater accuracy; larger values will lessen timer overhead. * Timer measurements will be found in metrics.log, marked "group=subtask_seconds, task=indexer_service" * Highest legal value is 4294967295 * We strongly suggest value of this parameter divide evenly into value of 'rotatePeriodInSecs' parameter. * Defaults to 30 processTrackerServiceInterval = <nonnegative integer> * Controls how often, in seconds, indexer checks status of the child OS processes it had launched to see if it can launch new processes for queued requests. * If set to 0, indexer will check child process status every second. * Highest legal value is 4294967295 * Defaults to 15 maxBucketSizeCacheEntries = <nonnegative integer> * This value is not longer needed and its value is ignored. tsidxStatsHomePath = <path on server> * An absolute path that specifies where Splunk creates namespace data with 'tscollect' command * If the directory does not exist, we attempt to create it. * Optional. If this is unspecified, we default to the 'tsidxstats' directory under $SPLUNK_DB * CAUTION: Path "$SPLUNK_DB" must be writable. hotBucketTimeRefreshInterval = <positive integer> * Controls how often each index refreshes the available hot bucket times used by the indexes REST endpoint. * Refresh will occur every N times service is performed for each index. * For busy indexes, this is a multiple of seconds. * For idle indexes, this is a multiple of the second-long-periods in which data is received. * This tunable is only intended to relax the frequency of these refreshes in * the unexpected case that it adversely affects performance in unusual production scenarios. * This time is tracked on a per-index basis, and thus can be adjusted on a per-index basis if needed. * If, for some reason, you want have the index information refreshed with every service (and accept minor performance overhead), you can use the value 1. * Defaults to 10 (services).
PER INDEX OPTIONS
# These options may be set under an [<index>] entry. # # Index names must consist of only numbers, lowercase letters, underscores, # and hyphens. They cannot begin with an underscore or hyphen, or contain # the word "kvstore". #************************************************************************** disabled = true|false * Toggles your index entry off and on. * Set to true to disable an index. * Defaults to false. deleted = true * If present, means that this index has been marked for deletion: if splunkd is running, deletion is in progress; if splunkd is stopped, deletion will re-commence on startup. * Normally absent, hence no default. * Do NOT manually set, clear, or modify value of this parameter. * Seriously: LEAVE THIS PARAMETER ALONE. homePath = <path on index server> * An absolute path that contains the hotdb and warmdb for the index. * It is recommended that you specify the path with the following syntax: homePath = $SPLUNK_DB/$_index_name/db At runtime, Splunk expands "$_index_name" to the name of the index. For example, if the index name is "newindex", homePath becomes "$SPLUNK_DB/newindex/db". * Splunkd keeps a file handle open for warmdbs at all times. * May contain a volume reference (see volume section below) in place of $SPLUNK_DB. * CAUTION: The parent path "$SPLUNK_DB/$_index_name/" must be writable. * Required. Splunk will not start if an index lacks a valid homePath. * Must restart splunkd after changing this parameter; index reload will not suffice. * We strongly recommend that you avoid the use of other environment variables in index paths, aside from the possible exception of SPLUNK_DB. * As an exception, SPLUNK_DB is explicitly managed by the provided software, so most possible downsides here do not exist. * Environment variables could be different from launch to launch of the software, causing severe problems with management of indexed data, including: * Data in the prior location will not be searchable. * The indexer may not be able to write to the new location, causing outages and/or data loss. * Writing to a new, unexpected location could lead to disk exhaustion causing additional operational problems. * Recovery from such a scenario will require manual intevention and bucket renaming, especially difficult in an index clustered environment. * In all circumstances, Splunk Diag, the diagnostic tool we use to support you, will have no way to determine the correct values for the environment variables, so cannot reliably operate. You may need to manually acquire information about your index buckets in troubleshooting scenarios. * Generally speaking, volumes provide a more appropriate way to control the storage location for indexes in a general way. coldPath = <path on index server> * An absolute path that contains the colddbs for the index. * It is recommended that you specify the path with the following syntax: coldPath = $SPLUNK_DB/$_index_name/colddb At runtime, Splunk expands "$_index_name" to the name of the index. For example, if the index name is "newindex", coldPath becomes "$SPLUNK_DB/newindex/colddb". * Cold databases are opened as needed when searching. * May contain a volume reference (see volume section below) in place of $SPLUNK_DB. * CAUTION: Path must be writable. * Required. Splunk will not start if an index lacks a valid coldPath. * Must restart splunkd after changing this parameter; index reload will not suffice. * We strongly recommend that you avoid the use of environment variables in index paths, aside from the possible exception of SPLUNK_DB. See homePath for the complete rationale. thawedPath = <path on index server> * An absolute path that contains the thawed (resurrected) databases for the index. * May NOT contain a volume reference. * CAUTION: Path must be writable. * Required. Splunk will not start if an index lacks a valid thawedPath. * Must restart splunkd after changing this parameter; index reload will not suffice. * We strongly recommend that you avoid the use of environment variables in index paths, aside from the possible exception of SPLUNK_DB. See homePath for the complete rationale. bloomHomePath = <path on index server> * Location where the bloomfilter files for the index are stored. * If specified, bloomHomePath must be defined in terms of a volume definition (see volume section below). * If bloomHomePath is not specified, bloomfilter files for index will be stored inline, inside bucket directories. * Path must be writable. * Must restart splunkd after changing this parameter; index reload will not suffice. * We strongly recommend that you avoid the use of environment variables in index paths, aside from the possible exception of SPLUNK_DB. See homePath for the complete rationale. * CAUTION: Do not set this parameter on indexes that have been configured to use remote storage with the "remotePath" parameter. createBloomfilter = true|false * Controls whether to create bloomfilter files for the index. * TRUE: bloomfilter files will be created. FALSE: not created. * Defaults to true. * CAUTION: Do not set this parameter to "false" on indexes that have been configured to use remote storage with the "remotePath" parameter. summaryHomePath = <path on index server> * An absolute path where transparent summarization results for data in this index should be stored. Must be different for each index and may be on any disk drive. * It is recommended that you specify the path with the following syntax: summaryHomePath = $SPLUNK_DB/$_index_name/summary At runtime, Splunk expands "$_index_name" to the name of the index. For example, if the index name is "newindex", summaryHomePath becomes "$SPLUNK_DB/newindex/summary". * May contain a volume reference (see volume section below) in place of $SPLUNK_DB. * Volume reference must be used if data retention based on data size is desired. * CAUTION: Path must be writable. * If not specified, Splunk creates a directory 'summary' in the same location as homePath * For example, if homePath is "/opt/splunk/var/lib/splunk/index1/db", then summaryHomePath would be "/opt/splunk/var/lib/splunk/index1/summary". * CAUTION: The parent path "/opt/splunk/var/lib/splunk/index1" must be writable. * Must restart splunkd after changing this parameter; index reload will not suffice. * We strongly recommend that you avoid the use of environment variables in index paths, aside from the possible exception of SPLUNK_DB. See homePath for the complete rationale. * Defaults to unset. tstatsHomePath = <path on index server> * Required. * Location where datamodel acceleration TSIDX data for this index should be stored * MUST be defined in terms of a volume definition (see volume section below) * Must restart splunkd after changing this parameter; index reload will not suffice. * CAUTION: Path must be writable. * Defaults to volume:_splunk_summaries/$_index_name/datamodel_summary, where $_index_name is runtime-expanded to the name of the index remotePath = <root path for remote volume, prefixed by a URI-like scheme> * Currently not supported. This setting is related to a feature that is still under development. * Optional. * Presence of this parameter means that this index uses remote storage, instead of the local file system, as the main repository for bucket storage. The index processor works with a cache manager to fetch buckets locally, as necessary, for searching and to evict them from local storage as space fills up and they are no longer needed for searching. * This setting must be defined in terms of a storageType=remote volume definition. See the volume section below. * The path portion that follows the volume reference is relative to the path specified for the volume. For example, if the path for a volume "v1" is "s3://bucket/path" and "remotePath" is "volume:v1/idx1", then the fully qualified path will be "s3://bucket/path/idx1". The rules for resolving the relative path with the absolute path specified in the volume can vary depending on the underlying storage type. * If "remotePath" is specified, the "coldPath" and "thawedPath" attributes are ignored. However, they still must be specified. maxBloomBackfillBucketAge = <nonnegative integer>[smhd]|infinite * If a (warm or cold) bloomfilter-less bucket is older than this, Splunk will not create a bloomfilter for that bucket. * When set to 0, bloomfilters are never backfilled * When set to "infinite", bloomfilters are always backfilled * NB that if createBloomfilter=false, bloomfilters are never backfilled regardless of the value of this parameter * Highest legal value in computed seconds is 2 billion, or 2000000000, which is approximately 68 years. * Defaults to 30d. enableOnlineBucketRepair = true|false * Controls asynchronous "online fsck" bucket repair, which runs concurrently with Splunk * When enabled, you do not have to wait until buckets are repaired, to start Splunk * When enabled, you might observe a slight performance degradation * Defaults to true. enableDataIntegrityControl = true|false * If set to true, hashes are computed on the rawdata slices and stored for future data integrity checks * If set to false, no hashes are computed on the rawdata slices * It has a global default value of false maxWarmDBCount = <nonnegative integer> * The maximum number of warm buckets. * Warm buckets are located in the <homePath> for the index. * If set to zero, Splunk will not retain any warm buckets (will roll them to cold as soon as it can) * Highest legal value is 4294967295 * Defaults to 300. maxTotalDataSizeMB = <nonnegative integer> * The maximum size of an index (in MB). * If an index grows larger than the maximum size, the oldest data is frozen. * This parameter only applies to hot, warm, and cold buckets. It does not apply to thawed buckets. * CAUTION: This setting takes precedence over other settings like 'frozenTimePeriodInSecs' with regard to data retention. If the index grows beyond 'maxTotalDataSizeMB' megabytes before 'frozenTimePeriodInSecs' seconds have passed, data could prematurely roll to frozen. As the default policy for rolling data to frozen is deletion, unintended data loss could occur. * Highest legal value is 4294967295 * Defaults to 500000. maxGlobalDataSizeMB = <nonnegative integer> * Currently not supported. This setting is related to a feature that is still under development. * The maximum amount of local disk space (in MB) that a remote storage enabled index can occupy, shared across all peers in the cluster. * This attribute controls the disk space that the index occupies on the peers only. It does not control the space that the index occupies on remote storage. * If the size that an index occupies across all peers exceeds the maximum size, the oldest data is frozen. * For example, assume that the attribute is set to 500 for a four-peer cluster, and each peer holds a 100 MB bucket for the index. If a new bucket of size 200 MB is then added to one of the peers, the cluster freezes the oldest bucket in the cluster, no matter which peer the bucket resides on. * This value applies to hot, warm and cold buckets. It does not apply to thawed buckets. * The maximum allowable value is 4294967295 * Defaults to 0, which means that it does not limit the space that the index can occupy on the peers. rotatePeriodInSecs = <positive integer> * Controls the service period (in seconds): how often splunkd performs certain housekeeping tasks. Among these tasks are: * Check if a new hotdb needs to be created. * Check if there are any cold DBs that should be frozen. * Check whether buckets need to be moved out of hot and cold DBs, due to respective size constraints (i.e., homePath.maxDataSizeMB and coldPath.maxDataSizeMB) * This value becomes the default value of the rotatePeriodInSecs attribute for all volumes (see rotatePeriodInSecs in the Volumes section) * Highest legal value is 4294967295 * Defaults to 60. frozenTimePeriodInSecs = <nonnegative integer> * Number of seconds after which indexed data rolls to frozen. * If you do not specify a coldToFrozenScript, data is deleted when rolled to frozen. * IMPORTANT: Every event in the DB must be older than frozenTimePeriodInSecs before it will roll. Then, the DB will be frozen the next time splunkd checks (based on rotatePeriodInSecs attribute). * Highest legal value is 4294967295 * Defaults to 188697600 (6 years). warmToColdScript = <script path> * Specifies a script to run when moving data from warm to cold. * This attribute is supported for backwards compatibility with versions older than 4.0. Migrating data across filesystems is now handled natively by splunkd. * If you specify a script here, the script becomes responsible for moving the event data, and Splunk-native data migration will not be used. * The script must accept two arguments: * First: the warm directory (bucket) to be rolled to cold. * Second: the destination in the cold path. * Searches and other activities are paused while the script is running. * Contact Splunk Support (http://www.splunk.com/page/submit_issue) if you need help configuring this setting. * The script must be in $SPLUNK_HOME/bin or a subdirectory thereof. * Defaults to empty. coldToFrozenScript = [path to script interpreter] <path to script> * Specifies a script to run when data will leave the splunk index system. * Essentially, this implements any archival tasks before the data is deleted out of its default location. * Add "$DIR" (quotes included) to this setting on Windows (see below for details). * Script Requirements: * The script must accept one argument: * An absolute path to the bucket directory to archive. * Your script should work reliably. * If your script returns success (0), Splunk will complete deleting the directory from the managed index location. * If your script return failure (non-zero), Splunk will leave the bucket in the index, and try calling your script again several minutes later. * If your script continues to return failure, this will eventually cause the index to grow to maximum configured size, or fill the disk. * Your script should complete in a reasonable amount of time. * If the script stalls indefinitely, it will occupy slots. * This script should not run for long as it would occupy resources which will affect indexing. * If the string $DIR is present in this setting, it will be expanded to the absolute path to the directory. * If $DIR is not present, the directory will be added to the end of the invocation line of the script. * This is important for Windows. * For historical reasons, the entire string is broken up by shell-pattern expansion rules. * Since windows paths frequently include spaces, and the windows shell breaks on space, the quotes are needed for the script to understand the directory. * If your script can be run directly on your platform, you can specify just the script. * Examples of this are: * .bat and .cmd files on Windows * scripts set executable on UNIX with a #! shebang line pointing to a valid interpreter. * You can also specify an explicit path to an interpreter and the script. * Example: /path/to/my/installation/of/python.exe path/to/my/script.py * Splunk ships with an example archiving script in that you SHOULD NOT USE $SPLUNK_HOME/bin called coldToFrozenExample.py * DO NOT USE the example for production use, because: * 1 - It will be overwritten on upgrade. * 2 - You should be implementing whatever requirements you need in a script of your creation. If you have no such requirements, use coldToFrozenDir * Example configuration: * If you create a script in bin/ called our_archival_script.py, you could use: UNIX: coldToFrozenScript = "$SPLUNK_HOME/bin/python" "$SPLUNK_HOME/bin/our_archival_script.py" Windows: coldToFrozenScript = "$SPLUNK_HOME/bin/python" "$SPLUNK_HOME/bin/our_archival_script.py" "$DIR" * The example script handles data created by different versions of splunk differently. Specifically data from before 4.2 and after are handled differently. See "Freezing and Thawing" below: * The script must be in $SPLUNK_HOME/bin or a subdirectory thereof. coldToFrozenDir = <path to frozen archive> * An alternative to a coldToFrozen script - simply specify a destination path for the frozen archive * Splunk will automatically put frozen buckets in this directory * For information on how buckets created by different versions are handled, see "Freezing and Thawing" below. * If both coldToFrozenDir and coldToFrozenScript are specified, coldToFrozenDir will take precedence * Must restart splunkd after changing this parameter; index reload will not suffice. * May NOT contain a volume reference. # Freezing and Thawing (this should move to web docs 4.2 and later data: * To archive: remove files except for the rawdata directory, since rawdata contains all the facts in the bucket. * To restore: run splunk rebuild <bucket_dir> on the archived bucket, then atomically move the bucket to thawed for that index 4.1 and earlier data: * To archive: gzip the .tsidx files, as they are highly compressible but cannot be recreated * To restore: unpack the tsidx files within the bucket, then atomically move the bucket to thawed for that index compressRawdata = true|false * This parameter is ignored. The splunkd process always compresses raw data. maxConcurrentOptimizes = <nonnegative integer> * The number of concurrent optimize processes that can run against the hot DB. * This number should be increased if: * There are always many small tsidx files in the hot DB. * After rolling, there are many tsidx files in warm or cold DB. * Must restart splunkd after changing this parameter; index reload will not suffice. * Highest legal value is 4294967295 * Defaults to 6 maxDataSize = <positive integer>|auto|auto_high_volume * The maximum size in MB for a hot DB to reach before a roll to warm is triggered. * Specifying "auto" or "auto_high_volume" will cause Splunk to autotune this parameter (recommended). * You should use "auto_high_volume" for high-volume indexes (such as the main index); otherwise, use "auto". A "high volume index" would typically be considered one that gets over 10GB of data per day. * Defaults to "auto", which sets the size to 750MB. * "auto_high_volume" sets the size to 10GB on 64-bit, and 1GB on 32-bit systems. * Although the maximum value you can set this is 1048576 MB, which corresponds to 1 TB, a reasonable number ranges anywhere from 100 to 50000. Before proceeding with any higher value, please seek approval of Splunk Support. * If you specify an invalid number or string, maxDataSize will be auto tuned. * NOTE: The maximum size of your warm buckets may slightly exceed 'maxDataSize', due to post-processing and timing issues with the rolling policy. rawFileSizeBytes = <positive integer> * Deprecated in version 4.2 and later. We will ignore this value. * Rawdata chunks are no longer stored in individual files. * If you really need to optimize the new rawdata chunks (highly unlikely), edit rawChunkSizeBytes rawChunkSizeBytes = <positive integer> * Target uncompressed size in bytes for individual raw slice in the rawdata journal of the index. * If 0 is specified, rawChunkSizeBytes will be set to the default value. * NOTE: rawChunkSizeBytes only specifies a target chunk size. The actual chunk size may be slightly larger by an amount proportional to an individual event size. * WARNING: This is an advanced parameter. Only change it if you are instructed to do so by Splunk Support. * Must restart splunkd after changing this parameter; index reload will not suffice. * Highest legal value is 18446744073709551615 * Defaults to 131072 (128KB). minRawFileSyncSecs = <nonnegative decimal>|disable * How frequently we force a filesystem sync while compressing journal slices. During this interval, uncompressed slices are left on disk even after they are compressed. Then we force a filesystem sync of the compressed journal and remove the accumulated uncompressed files. * If 0 is specified, we force a filesystem sync after every slice completes compressing. * Specifying "disable" disables syncing entirely: uncompressed slices are removed as soon as compression is complete * Some filesystems are very inefficient at performing sync operations, so only enable this if you are sure it is needed * Must restart splunkd after changing this parameter; index reload will not suffice. * No exponent may follow the decimal. * Highest legal value is 18446744073709551615 * Defaults to "disable". maxMemMB = <nonnegative integer> * The amount of memory to allocate for indexing. * This amount of memory will be allocated PER INDEX THREAD, or, if indexThreads is set to 0, once per index. * IMPORTANT: Calculate this number carefully. splunkd will crash if you set this number higher than the amount of memory available. * The default is recommended for all environments. * Highest legal value is 4294967295 * Defaults to 5. maxHotSpanSecs = <positive integer> * Upper bound of timespan of hot/warm buckets in seconds. * NOTE: If you set this too small, you can get an explosion of hot/warm buckets in the filesystem. * NOTE: If you set maxHotBuckets to 1, Splunk attempts to send all events to the single hot bucket and maxHotSpanSeconds will not be enforced. * If you set this parameter to less than 3600, it will be automatically reset to 3600. * This is an advanced parameter that should be set with care and understanding of the characteristics of your data. * Highest legal value is 4294967295 * Defaults to 7776000 seconds (90 days). * Note that this limit will be applied per ingestion pipeline. For more information about multiple ingestion pipelines see parallelIngestionPipelines in server.conf.spec file. * With N parallel ingestion pipelines, each ingestion pipeline will write to and manage its own set of hot buckets, without taking into account the state of hot buckets managed by other ingestion pipelines. Each ingestion pipeline will independently apply this setting only to its own set of hot buckets. * NOTE: the bucket timespan snapping behavior is removed from this setting. See the 6.5 spec file for details of this behavior. maxHotIdleSecs = <nonnegative integer> * Provides a ceiling for buckets to stay in hot status without receiving any data. * If a hot bucket receives no data for more than maxHotIdleSecs seconds, Splunk rolls it to warm. * This setting operates independently of maxHotBuckets, which can also cause hot buckets to roll. * A value of 0 turns off the idle check (equivalent to infinite idle time). * Highest legal value is 4294967295 * Defaults to 0. maxHotBuckets = <positive integer> * Maximum hot buckets that can exist per index. * When maxHotBuckets is exceeded, Splunk rolls the least recently used (LRU) hot bucket to warm. * Both normal hot buckets and quarantined hot buckets count towards this total. * This setting operates independently of maxHotIdleSecs, which can also cause hot buckets to roll. * Highest legal value is 4294967295 * Defaults to 3. * Note that this limit will be applied per ingestion pipeline. For more information about multiple ingestion pipelines see parallelIngestionPipelines in server.conf.spec file. * With N parallel ingestion pipelines the maximum number of hot buckets across all of the ingestion pipelines will be N * maxHotBuckets but maxHotBuckets for each ingestion pipeline. Each ingestion pipeline will independently write to and manage up to maxHotBuckets number of hot buckets. As a consequence of this, when multiple ingestion pipelines are used, there may be multiple (dependent on number of ingestion pipelines configured) hot buckets with events with overlapping time ranges. minHotIdleSecsBeforeForceRoll = <nonnegative integer>|auto * When there are no existing hot buckets that can fit new events because of their timestamps and the constraints on the index (refer to maxHotBuckets, maxHotSpanSecs and quarantinePastSecs), if any hot bucket has been idle (i.e. not receiving any data) for minHotIdleSecsBeforeForceRoll number of seconds, a new bucket will be created to receive these new events and the idle bucket will be rolled to warm. * If no hot bucket has been idle for minHotIdleSecsBeforeForceRoll number of seconds, or if minHotIdleSecsBeforeForceRoll has been set to zero, then a best fit bucket will be chosen for these new events from the existing set of hot buckets. * This setting operates independently of maxHotIdleSecs, which causes hot buckets to roll after they have been idle for maxHotIdleSecs number of seconds, *regardless* of whether new events can fit into the existing hot buckets or not due to an event timestamp. minHotIdleSecsBeforeForceRoll, on the other hand, controls a hot bucket roll *only* under the circumstances when the timestamp of a new event cannot fit into the existing hot buckets given the other parameter constraints on the system (parameters such as maxHotBuckets, maxHotSpanSecs and quarantinePastSecs). * auto: Specifying "auto" will cause Splunk to autotune this parameter (recommended). The value begins at 600 seconds but automatically adjusts upwards for optimal performance. Specifically, the value will increase when a hot bucket rolls due to idle time with a significantly smaller size than maxDataSize. As a consequence, the outcome may be fewer buckets, though these buckets may span wider earliest-latest time ranges of events. * 0: A value of 0 turns off the idle check (equivalent to infinite idle time). Setting this to zero means that we will never roll a hot bucket for the reason that an event cannot fit into an existing hot bucket due to the constraints of other parameters. Instead, we will find a best fitting bucket to accommodate that event. * Highest legal value is 4294967295. * NOTE: If you set this configuration, there is a chance that this could lead to frequent hot bucket rolls depending on the value. If your index contains a large number of buckets whose size-on-disk falls considerably short of the size specified in maxDataSize, and if the reason for the roll of these buckets is due to "caller=lru", then setting the parameter value to a larger value or to zero may reduce the frequency of hot bucket rolls (see AUTO above). You may check splunkd.log for a similar message below for rolls due to this setting. INFO HotBucketRoller - finished moving hot to warm bid=_internal~0~97597E05-7156-43E5-85B1-B0751462D16B idx=_internal from=hot_v1_0 to=db_1462477093_1462477093_0 size=40960 caller=lru maxHotBuckets=3, count=4 hot buckets,evicting_count=1 LRU hots * Defaults to "auto". splitByIndexKeys = <comma separated index keys> * By default, buckets are split by time ranges with each bucket having its earliest and latest time. * If one or several keys are provided, buckets will be split by the index key or a combination of the index keys if more than one is provided. The buckets will no longer be split by time ranges. * Valid values are: host, sourcetype, source, metric_name * Defaults to an empty string (no key). If not set, splitting by time span is applied. * NOTE: This setting only applies to metric indexes. quarantinePastSecs = <positive integer> * Events with timestamp of quarantinePastSecs older than "now" will be dropped into quarantine bucket. * This is a mechanism to prevent the main hot buckets from being polluted with fringe events. * Highest legal value is 4294967295 * Defaults to 77760000 (900 days). quarantineFutureSecs = <positive integer> * Events with timestamp of quarantineFutureSecs newer than "now" will be dropped into quarantine bucket. * This is a mechanism to prevent main hot buckets from being polluted with fringe events. * Highest legal value is 4294967295 * Defaults to 2592000 (30 days). maxMetaEntries = <nonnegative integer> * Sets the maximum number of unique lines in .data files in a bucket, which may help to reduce memory consumption * If exceeded, a hot bucket is rolled to prevent further increase * If your buckets are rolling due to Strings.data hitting this limit, the culprit may be the 'punct' field in your data. If you do not use punct, it may be best to simply disable this (see props.conf.spec) * NOTE: since at least 5.0.x, large strings.data from punct will be rare. * There is a delta between when maximum is exceeded and bucket is rolled. * This means a bucket may end up with epsilon more lines than specified, but this is not a major concern unless excess is significant * If set to 0, this setting is ignored (it is treated as infinite) * Highest legal value is 4294967295 syncMeta = true|false * When "true", a sync operation is called before file descriptor is closed on metadata file updates. * This functionality was introduced to improve integrity of metadata files, especially in regards to operating system crashes/machine failures. * NOTE: Do not change this parameter without the input of a Splunk support professional. * Must restart splunkd after changing this parameter; index reload will not suffice. * Defaults to true. serviceMetaPeriod = <positive integer> * Defines how frequently metadata is synced to disk, in seconds. * Defaults to 25 (seconds). * You may want to set this to a higher value if the sum of your metadata file sizes is larger than many tens of megabytes, to avoid the hit on I/O in the indexing fast path. * Highest legal value is 4294967295 partialServiceMetaPeriod = <positive integer> * Related to serviceMetaPeriod. If set, it enables metadata sync every <integer> seconds, but only for records where the sync can be done efficiently in-place, without requiring a full re-write of the metadata file. Records that require full re-write will be synced at serviceMetaPeriod. * <integer> specifies how frequently it should sync. Zero means that this feature is turned off and serviceMetaPeriod is the only time when metadata sync happens. * If the value of partialServiceMetaPeriod is greater than serviceMetaPeriod, this setting will have no effect. * By default it is turned off (zero). * This parameter is ignored if serviceOnlyAsNeeded = true (the default). * Highest legal value is 4294967295 throttleCheckPeriod = <positive integer> * Defines how frequently Splunk checks for index throttling condition, in seconds. * NOTE: Do not change this parameter without the input of a Splunk Support professional. * Highest legal value is 4294967295 * Defaults to 15 maxTimeUnreplicatedWithAcks = <nonnegative decimal> * Important if you have enabled indexer acknowledgements (ack) on forwarders and have replication enabled (via Index Clustering) * This parameter puts an upper limit on how long events can sit unacknowledged in a raw slice * To disable this, you can set to 0, but this is NOT recommended!!! * NOTE: This is an advanced parameter; make sure you understand the settings on all your forwarders before changing this. This number should not exceed ack timeout configured on any forwarders, and should indeed be set to at most half of the minimum value of that timeout. You can find this setting in outputs.conf readTimeout setting, under the tcpout stanza. * Highest legal value is 2147483647 * Defaults to 60 (seconds) maxTimeUnreplicatedNoAcks = <nonnegative decimal> * Important only if replication is enabled for this index, otherwise ignored * This parameter puts an upper limit on how long an event can sit in raw slice. * If there are any ack''d events sharing this raw slice, this parameter will not apply (maxTimeUnreplicatedWithAcks will be used instead) * Highest legal value is 2147483647 * To disable this, you can set to 0; please be careful and understand the consequences before changing this parameter * Defaults to 300 (seconds) isReadOnly = true|false * Set to true to make an index read-only. * If true, no new events can be added to the index, but the index is still searchable. * Must restart splunkd after changing this parameter; index reload will not suffice. * Defaults to false. homePath.maxDataSizeMB = <nonnegative integer> * Specifies the maximum size of homePath (which contains hot and warm buckets). * If this size is exceeded, Splunk will move buckets with the oldest value of latest time (for a given bucket) into the cold DB until homePath is below the maximum size. * If this attribute is missing or set to 0, Splunk will not constrain the size of homePath. * Highest legal value is 4294967295 * Defaults to 0. coldPath.maxDataSizeMB = <nonnegative integer> * Specifies the maximum size of coldPath (which contains cold buckets). * If this size is exceeded, Splunk will freeze buckets with the oldest value of latest time (for a given bucket) until coldPath is below the maximum size. * If this attribute is missing or set to 0, Splunk will not constrain size of coldPath * If we freeze buckets due to enforcement of this policy parameter, and coldToFrozenScript and/or coldToFrozenDir archiving parameters are also set on the index, these parameters *will* take into effect * Highest legal value is 4294967295 * Defaults to 0. disableGlobalMetadata = true|false * NOTE: This option was introduced in 4.3.3, but as of 5.0 it is obsolete and ignored if set. * It used to disable writing to the global metadata. In 5.0 global metadata was removed. repFactor = 0|auto * Valid only for indexer cluster peer nodes. * Determines whether an index gets replicated. * Value of 0 turns off replication for this index. * Value of "auto" turns on replication for this index. * This attribute must be set to the same value on all peer nodes. * Defaults to 0. minStreamGroupQueueSize = <nonnegative integer> * Minimum size of the queue that stores events in memory before committing them to a tsidx file. As Splunk operates, it continually adjusts this size internally. Splunk could decide to use a small queue size and thus generate tiny tsidx files under certain unusual circumstances, such as file system errors. The danger of a very low minimum is that it can generate very tiny tsidx files with one or very few events, making it impossible for splunk-optimize to catch up and optimize the tsidx files into reasonably sized files. * Defaults to 2000. * Only set this value if you have been advised to by Splunk Support. * Highest legal value is 4294967295 streamingTargetTsidxSyncPeriodMsec = <nonnegative integer> * Period we force sync tsidx files on streaming targets. This setting is needed for multi-site clustering where streaming targets may be primary. * if set to 0, we never sync (equivalent to infinity) journalCompression = gzip|lz4 * Select compression algorithm for rawdata journal file * Defaults to gzip enableTsidxReduction = true|false * By enabling this setting, you turn on the tsidx reduction capability. This causes the indexer to reduce the tsidx files of buckets, when the buckets reach the age specified by timePeriodInSecBeforeTsidxReduction. * CAUTION: Do not set this parameter to "true" on indexes that have been configured to use remote storage with the "remotePath" parameter. * Defaults to false. suspendHotRollByDeleteQuery = true|false * When the "delete" search command is run, all buckets containing data to be deleted are marked for updating of their metadata files. The indexer normally first rolls any hot buckets, as rolling must precede the metadata file updates. * When suspendHotRollByDeleteQuery is set to true, the rolling of hot buckets for the "delete" command is suspended. The hot buckets, although marked, do not roll immediately, but instead wait to roll in response to the same circumstances operative for any other hot buckets; for example, due to reaching a limit set by maxHotBuckets, maxDataSize, etc. When these hot buckets finally roll, their metadata files are then updated. * Defaults to false tsidxReductionCheckPeriodInSec = <positive integer> * Time period between service runs to reduce the tsidx files for any buckets that have reached the age specified by timePeriodInSecBeforeTsidxReduction. * Defaults to 600 (seconds). timePeriodInSecBeforeTsidxReduction = <positive integer> * Age at which buckets become eligible for tsidx reduction. The bucket age is the difference between the current time and the timestamp of the bucket's latest event. * Defaults to 604800 (seconds).
PER PROVIDER FAMILY OPTIONS
# A provider family is a way of collecting properties that are common to # multiple providers. There are no properties that can only be used in a # provider family, and not in a provider. If the same property is specified # in a family, and in a provider belonging to that family, then the latter # value "wins". # # All family stanzas begin with "provider-family:". For example: # [provider-family:family_name] # vix.mode=stream # vix.command = java # vix.command.arg.1 = -Xmx512m # .... #**************************************************************************
PER PROVIDER OPTIONS
# These options affect External Resource Providers. All provider stanzas # begin with "provider:". For example: # [provider:provider_name] # vix.family = hadoop # vix.env.JAVA_HOME = /path/to/java/home # vix.env.HADOOP_HOME = /path/to/hadoop/client/libraries # # Each virtual index must reference a provider. #************************************************************************** vix.family = <family> * A provider family to which this provider belongs. * The only family available by default is "hadoop". Others may be added. vix.mode = stream|report * Usually specified at the family level. * Typically should be "stream". In general, do not use "report" without consulting Splunk Support. vix.command = <command> * The command to be used to launch an external process for searches on this provider. * Usually specified at the family level. vix.command.arg.<N> = <argument> * The Nth argument to the command specified by vix.command. * Usually specified at the family level, but frequently overridden at the provider level, for example to change the jars used depending on the version of Hadoop to which a provider connects. vix.<property name> = <property value> * All such properties will be made available as "configuration properties" to search processes on this provider. * For example, if this provider is in the Hadoop family, the configuration property "mapreduce.foo = bar" can be made available to the Hadoop via the property "vix.mapreduce.foo = bar". vix.env.<env var name> = <env var variable> * Will create an environment variable available to search processes on this provider. * For example, to set the JAVA_HOME variable to "/path/java" for search processes on this provider, use "vix.env.JAVA_HOME = /path/java". #************************************************************************** # PER PROVIDER OPTIONS -- HADOOP # These options are specific to ERPs with the Hadoop family. # NOTE: Many of these properties specify behavior if the property is not # set. However, default values set in system/default/indexes.conf # take precedence over the "unset" behavior. #************************************************************************** vix.javaprops.<JVM system property name> = <value> * All such properties will be used as Java system properties. * For example, to specify a Kerberos realm (say "foo.com") as a Java system property, use the property "vix.javaprops.java.security.krb5.realm = foo.com". vix.mapred.job.tracker = <logical name or server:port> * In high-availability mode, use the logical name of the Job Tracker. * Otherwise, should be set to server:port for the single Job Tracker. * Note: this property is passed straight to Hadoop. Not all such properties are documented here. vix.fs.default.name = <logical name or hdfs://server:port> * In high-availability mode, use the logical name for a list of Name Nodes. * Otherwise, use the URL for the single Name Node. * Note: this property is passed straight to Hadoop. Not all such properties are documented here. vix.splunk.setup.onsearch = true|false * Whether to perform setup (install & bundle replication) on search. * Defaults to false. vix.splunk.setup.package = current|<path to file> * Splunk .tgz package to install and use on data nodes (in vix.splunk.home.datanode). * Uses the current install if set to value 'current' (without quotes). vix.splunk.home.datanode = <path to dir> * Path to where splunk should be installed on datanodes/tasktrackers, i.e. SPLUNK_HOME. * Required. vix.splunk.home.hdfs = <path to dir> * Scratch space for this Splunk instance on HDFS * Required. vix.splunk.search.debug = true|false * Whether to run searches against this index in debug mode. In debug mode, additional information is logged to search.log. * Optional. Defaults to false. vix.splunk.search.recordreader = <list of classes> * Comma separated list of data preprocessing classes. * Each such class must extend BaseSplunkRecordReader and return data to be consumed by Splunk as the value. vix.splunk.search.splitter = <class name> * Set to override the class used to generate splits for MR jobs. * Classes must implement com.splunk.mr.input.SplitGenerator. * Unqualified classes will be assumed to be in the package com.splunk.mr.input. * May be specified in either the provider stanza, or the virtual index stanza. * To search Parquet files, use ParquetSplitGenerator. * To search Hive files, use HiveSplitGenerator. vix.splunk.search.mr.threads = <postive integer> * Number of threads to use when reading map results from HDFS * Numbers less than 1 will be treated as 1. * Numbers greater than 50 will be treated as 50. * If not set, defaults to 10. vix.splunk.search.mr.maxsplits = <positive integer> * Maximum number of splits in an MR job. * If not set, defaults to 10000. vix.splunk.search.mr.minsplits = <positive integer> * Number of splits for first MR job associated with a given search. * If not set, defaults to 100. vix.splunk.search.mr.splits.multiplier = <decimal greater than or equal to 1.0> * Factor by which the number of splits is increased in consecutive MR jobs for a given search, up to the value of maxsplits. * If not set, defaults to 10. vix.splunk.search.mr.poll = <positive integer> * Polling period for job status, in milliseconds. * If not set, defaults to 1000 (ie. 1 second). vix.splunk.search.mr.mapper.output.replication = <positive integer> * Replication level for mapper output. * Defaults to 3. vix.splunk.search.mr.mapper.output.gzlevel = <integer between 0 and 9, inclusive> * The compression level used for the mapper output. * Defaults to 2. vix.splunk.search.mixedmode = true|false * Whether mixed mode execution is enabled. * Defaults to true. vix.splunk.search.mixedmode.maxstream = <nonnegative integer> * Max # of bytes to stream during mixed mode. * Value = 0 means there's no stream limit. * Will stop streaming after the first split that took the value over the limit. * If not set, defaults to 10 GB. vix.splunk.jars = <list of paths> * Comma delimited list of Splunk dirs/jars to add to the classpath in the Search Head and MR. vix.env.HUNK_THIRDPARTY_JARS = <list of paths> * Comma delimited list of 3rd-party dirs/jars to add to the classpath in the Search Head and MR. vix.splunk.impersonation = true|false * Enable/disable user impersonation. vix.splunk.setup.bundle.replication = <positive integer> * Set custom replication factor for bundles on HDFS. * Must be an integer between 1 and 32767. * Increasing this setting may help performance on large clusters by decreasing the average access time for a bundle across Task Nodes. * Optional. If not set, the default replication factor for the file-system will apply. vix.splunk.setup.bundle.max.inactive.wait = <positive integer> * A positive integer represent a time interval in seconds. * Defaults to 5. * While a task waits for a bundle being replicated to the same node by another task, if the bundle file is not modified for this amount of time, the task will begin its own replication attempt. vix.splunk.setup.bundle.poll.interval = <positive integer> * A positive number, representing a time interval in milliseconds. * Defaults to 100. * While a task waits for a bundle to be installed by another task on the same node, it will check once per interval whether that installation is complete. vix.splunk.setup.bundle.setup.timelimit = <positive integer> * A postive number, representing a time duration in milliseconds. * Defaults to 20,000 (i.e. 20 seconds). * A task will wait this long for a bundle to be installed before it quits. vix.splunk.setup.package.replication = true|false * Set custom replication factor for the Splunk package on HDFS. This is the package set in the property vix.splunk.setup.package. * Must be an integer between 1 and 32767. * Increasing this setting may help performance on large clusters by decreasing the average access time for the package across Task Nodes. * Optional. If not set, the default replication factor for the file-system will apply. vix.splunk.setup.package.max.inactive.wait = <positive integer> * A positive integer represent a time interval in seconds. * Defaults to 5. * While a task waits for a Splunk package being replicated to the same node by another task, if the package file is not modified for this amount of time, the task will begin its own replication attempt. vix.splunk.setup.package.poll.interval = <positive integer> * A positive number, representing a time interval in milliseconds. * Defaults to 100. * While a task waits for a Splunk package to be installed by another task on the same node, it will check once per interval whether that installation is complete. vix.splunk.setup.package.setup.timelimit = <positive integer> * A positive number, representing a time duration in milliseconds. * Defaults to 20,000 (i.e. 20 seconds). * A task will wait this long for a Splunk package to be installed before it quits. vix.splunk.setup.bundle.reap.timelimit = <positive integer> * Specific to Hunk provider * For bundles in the working directory on each data node, this property controls how old they must be before they are eligible for reaping. * Unit is milliseconds * Defaults to 24 hours, e.g. 24 * 3600 * 1000. * Values larger than 24 hours will be treated as if set to 24 hours. vix.splunk.search.column.filter = true|false * Enables/disables column filtering. When enabled, Hunk will trim columns that are not necessary to a query on the Task Node, before returning the results to the search process. * Should normally increase performance, but does have its own small overhead. * Works with these formats: CSV, Avro, Parquet, Hive. * If not set, defaults to true. # # Kerberos properties # vix.kerberos.principal = <kerberos principal name> * Specifies principal for Kerberos authentication. * Should be used with vix.kerberos.keytab and either 1) vix.javaprops.java.security.krb5.realm and vix.javaprops.java.security.krb5.kdc, or 2) security.krb5.conf vix.kerberos.keytab = <kerberos keytab path> * Specifies path to keytab for Kerberos authentication. * See usage note with vix.kerberos.principal. # # The following properties affect the SplunkMR heartbeat mechanism. If this # mechanism is turned on, the SplunkMR instance on the Search Head updates a # heartbeat file on HDFS. Any MR job spawned by report or mix-mode searches # checks the heartbeat file. If it is not updated for a certain time, it will # consider SplunkMR to be dead and kill itself. # vix.splunk.heartbeat = true|false * Turn on/off heartbeat update on search head, and checking on MR side. * If not set, defaults to true. vix.splunk.heartbeat.path = <path on HDFS> * Path to heartbeat file. * If not set, defaults to <vix.splunk.home.hdfs>/dispatch/<sid>/ vix.splunk.heartbeat.interval = <positive integer> * Frequency with which the Heartbeat will be updated on the Search Head. * Unit is millisecond. * Default value is 6 seconds (6000). * Minimum value is 1000. Smaller values will cause an exception to be thrown. vix.splunk.heartbeat.threshold = <postive integer> * The number of times the MR job will detect a missing heartbeat update before it considers SplunkMR dead and kills itself. * Default value is 10. ## The following sections are specific to data input types. # # Sequence file # vix.splunk.search.recordreader.sequence.ignore.key = true|false * When reading sequence files, if this key is enabled, events will be expected to only include a value. Otherwise, the expected representation is key+"\t"+value. * Defaults to true. # # Avro # vix.splunk.search.recordreader.avro.regex = <regex> * Regex that files must match in order to be considered avro files. * Optional. Defaults to \.avro$ # # Parquet # vix.splunk.search.splitter.parquet.simplifyresult = true|false * If enabled, field names for map and list type fields will be simplified by dropping intermediate "map" or "element" subfield names. Otherwise, a field name will match parquet schema completely. * May be specified in either the provider stanza or in the virutal index stanza. * Defaults to true. # # Hive # vix.splunk.search.splitter.hive.ppd = true|false * Enable or disable Hive ORC Predicate Push Down. * If enabled, ORC PPD will be applied whenever possible to prune unnecessary data as early as possible to optimize the search. * If not set, defaults to true. * May be specified in either the provider stanza or in the virutal index stanza. vix.splunk.search.splitter.hive.fileformat = textfile|sequencefile|rcfile|orc * Format of the Hive data files in this provider. * If not set, defaults to "textfile". * May be specified in either the provider stanza or in the virutal index stanza. vix.splunk.search.splitter.hive.dbname = <DB name> * Name of Hive database to be accessed by this provider. * Optional. If not set, defaults to "default". * May be specified in either the provider stanza or in the virutal index stanza. vix.splunk.search.splitter.hive.tablename = <table name> * Table accessed by this provider. * Required property. * May be specified in either the provider stanza or in the virutal index stanza. vix.splunk.search.splitter.hive.columnnames = <list of column names> * Comma-separated list of file names. * Required if using Hive, not using metastore. * May be specified in either the provider stanza or in the virutal index stanza. vix.splunk.search.splitter.hive.columntypes = string:float:int # COLON separated list of column types, required * Colon-separated list of column- types. * Required if using Hive, not using metastore. * May be specified in either the provider stanza or in the virutal index stanza. vix.splunk.search.splitter.hive.serde = <SerDe class> * Fully-qualified class name of SerDe. * Required if using Hive, not using metastore, and if specified in creation of Hive table. * May be specified in either the provider stanza or in the virutal index stanza. vix.splunk.search.splitter.hive.serde.properties = <list of key-value pairs> * Comma-separated list of "key=value" pairs. * Required if using Hive, not using metastore, and if specified in creation of Hive table. * May be specified in either the provider stanza or in the virutal index stanza. vix.splunk.search.splitter.hive.fileformat.inputformat = <InputFormat class> * Fully-qualified class name of an InputFormat to be used with Hive table data. * May be specified in either the provider stanza or in the virutal index stanza. vix.splunk.search.splitter.hive.rowformat.fields.terminated = <delimiter> * Will be set as the Hive SerDe property "field.delim". * Optional. * May be specified in either the provider stanza or in the virutal index stanza. vix.splunk.search.splitter.hive.rowformat.escaped = <escape char> * Will be set as the Hive SerDe property "escape.delim". * Optional. * May be specified in either the provider stanza or in the virutal index stanza. vix.splunk.search.splitter.hive.rowformat.lines.terminated = <delimiter> * Will be set as the Hive SerDe property "line.delim". * Optional. * May be specified in either the provider stanza or in the virutal index stanza. vix.splunk.search.splitter.hive.rowformat.mapkeys.terminated = <delimiter> * Will be set as the Hive SerDe property "mapkey.delim". * Optional. * May be specified in either the provider stanza or in the virutal index stanza. vix.splunk.search.splitter.hive.rowformat.collectionitems.terminated = <delimiter> * Will be set as the Hive SerDe property "colelction.delim". * Optional. * May be specified in either the provider stanza or in the virutal index stanza. # # Archiving # vix.output.buckets.max.network.bandwidth = 0|<bits per second> * Throttles network bandwidth to <bits per second> * Defaults to 0, meaning no throttling. * Set at provider level. Applied to all virtual indexes using a provider with this setting.
PER VIRTUAL INDEX OPTIONS
# These options affect virtual indexes. Like indexes, these options may # be set under an [<virtual-index>] entry. # # Virtual index names have the same constraints as normal index names. # # Each virtual index must reference a provider. I.e: # [virtual_index_name] # vix.provider = <provider_name> # # All configuration keys starting with "vix." will be passed to the # external resource provider (ERP). #************************************************************************** vix.provider = <provider_name> * Name of the external resource provider to use for this virtual index. #************************************************************************** # PER VIRTUAL INDEX OPTIONS -- HADOOP # These options are specific to ERPs with the Hadoop family. #************************************************************************** # # The vix.input.* configurations are grouped by an id. # Inputs configured via the UI always use '1' as the id. # In this spec we'll use 'x' as the id. # vix.input.x.path = <path> * Path in a hadoop filesystem (usually HDFS or S3). * May contain wildcards. * Checks the path for data recursively when ending with '...' * Can extract fields with ${field}. I.e: "/data/${server}/...", where server will be extracted. * May start with a schema. * The schema of the path specifies which hadoop filesystem implementation to use. Examples: * hdfs://foo:1234/path, will use a HDFS filesystem implementation * s3a://s3-bucket/path, will use a S3 filesystem implementation vix.input.x.accept = <regex> * Specifies a whitelist regex. * Only files within the location given by matching vix.input.x.path, whose paths match this regex, will be searched. vix.input.x.ignore = <regex> * Specifies a blacklist regex. * Searches will ignore paths matching this regex. * These matches take precedence over vix.input.x.accept matches. vix.input.x.required.fields = <comma separated list of fields> * Fields that will be kept in search results even if the field is not required by the search # Earliest time extractions - For all 'et' settings, there's an equivalent 'lt' setting. vix.input.x.et.regex = <regex> * Regex extracting earliest time from vix.input.x.path vix.input.x.et.format = <java.text.SimpleDateFormat date pattern> * Format of the extracted earliest time. * See documentation for java.text.SimpleDateFormat vix.input.x.et.offset = <seconds> * Offset in seconds to add to the extracted earliest time. vix.input.x.et.timezone = <java.util.SimpleTimeZone timezone id> * Timezone in which to interpret the extracted earliest time. * Examples: "America/Los_Angeles" or "GMT-8:00" vix.input.x.et.value = mtime|<epoch time in milliseconds> * Sets the earliest time for this virtual index. * Can be used instead of extracting times from the path via vix.input.x.et.regex * When set to "mtime", uses the file modification time as the earliest time. # Latest time extractions - See "Earliest time extractions" vix.input.x.lt.regex = <regex> * Latest time equivalent of vix.input.x.et.regex vix.input.x.lt.format = <java.text.SimpleDateFormat date pattern> * Latest time equivalent of vix.input.x.et.format vix.input.x.lt.offset = <seconds> * Latest time equivalent of vix.input.x.et.offset vix.input.x.lt.timezone = <java.util.SimpleTimeZone timezone id> * Latest time equivalent of vix.input.x.et.timezone vix.input.x.lt.value = <mod time> * Latest time equivalent of vix.input.x.et.value # # Archiving # vix.output.buckets.path = <hadoop path> * Path to a hadoop filesystem where buckets will be archived vix.output.buckets.older.than = <seconds> * Buckets must be this old before they will be archived. * A bucket's age is determined by the the earliest _time field of any event in the bucket. vix.output.buckets.from.indexes = <comma separated list of splunk indexes> * List of (non-virtual) indexes that will get archived to this (virtual) index. vix.unified.search.cutoff_sec = <seconds> * Window length before present time that configures where events are retrieved for unified search * Events from now to now-cutoff_sec will be retrieved from the splunk index and events older than cutoff_sec will be retrieved from the archive index #************************************************************************** # PER VIRTUAL INDEX OR PROVIDER OPTIONS -- HADOOP # These options can be set at either the virtual index level or provider # level, for the Hadoop ERP. # # Options set at the virtual index level take precedence over options set # at the provider level. # # Virtual index level prefix: # vix.input.<input_id>.<option_suffix> # # Provider level prefix: # vix.splunk.search.<option_suffix> #************************************************************************** # The following options are just defined by their <option_suffix> # # Record reader options # recordreader.<name>.<conf_key> = <conf_value> * Sets a configuration key for a RecordReader with <name> to <conf_value> recordreader.<name>.regex = <regex> * Regex specifying which files this RecordReader can be used for. recordreader.journal.buffer.size = <bytes> * Buffer size used by the journal record reader recordreader.csv.dialect = default|excel|excel-tab|tsv * Set the csv dialect for csv files * A csv dialect differs on delimiter_char, quote_char and escape_char. * Here is a list of how the different dialects are defined in order delim, quote, and escape: * default = , " \ * excel = , " " * excel-tab = \t " " * tsv = \t " \ # # Splitter options # splitter.<name>.<conf_key> = <conf_value> * Sets a configuration key for a split generator with <name> to <conf_value> * See comment above under "PER VIRTUAL INDEX OR PROVIDER OPTIONS". This means that the full format is: vix.input.N.splitter.<name>.<conf_key> (in a vix stanza) vix.splunk.search.splitter.<name>.<conf_key> (in a provider stanza) splitter.file.split.minsize = <bytes> * Minimum size in bytes for file splits. * Defaults to 1. splitter.file.split.maxsize = <bytes> * Maximum size in bytes for file splits. * Defaults to Long.MAX_VALUE. #************************************************************************** # Volume settings. This section describes settings that affect the volume- # optional and volume-mandatory parameters only. # # All volume stanzas begin with "volume:". For example: # [volume:volume_name] # path = /foo/bar # # These volume stanzas can then be referenced by individual index # parameters, e.g. homePath or coldPath. To refer to a volume stanza, use # the "volume:" prefix. For example, to set a cold DB to the example stanza # above, in index "hiro", use: # [hiro] # coldPath = volume:volume_name/baz # This will cause the cold DB files to be placed under /foo/bar/baz. If the # volume spec is not followed by a path # (e.g. "coldPath=volume:volume_name"), then the cold path would be # composed by appending the index name to the volume name ("/foo/bar/hiro"). # # If "path" is specified with a URI-like value (e.g., "s3://bucket/path"), # this is a remote storage volume. A remote storage volume can only be # referenced by a remotePath parameter, as described above. An Amazon S3 # remote path might look like "s3://bucket/path", whereas an NFS remote path # might look like "file:///mnt/nfs". The name of the scheme ("s3" or "file" # from these examples) is important, because it can indicate some necessary # configuration specific to the type of remote storage. To specify a # configuration under the remote storage volume stanza, you use parameters # with the pattern "remote.<scheme>.<param name>". These parameters vary # according to the type of remote storage. For example, remote storage of # type S3 might require that you specify an access key and a secret key. # You would do this through the "remote.s3.access_key" and # "remote.s3.secret_key" parameters. # # Note: thawedPath may not be defined in terms of a volume. # Thawed allocations are manually controlled by Splunk administrators, # typically in recovery or archival/review scenarios, and should not # trigger changes in space automatically used by normal index activity. #************************************************************************** storageType = local | remote * Setting this to "remote" is currently not supported. This setting is related to a feature that is still under development. * Optional. * Specifies whether the volume definition is for indexer local storage or remote storage. Only the remotePath attribute references a remote volume. * Defaults to: local. path = <path on server> * Required. * If storageType is set to its default value of "local": * The path attribute points to the location on the file system where all indexes that will use this volume reside. * This location must not overlap with the location for any other volume or index. * If storageType is set to "remote": * The path attribute points to the remote storage location where indexes reside. * The format for this attribute is: <scheme>://<remote-location-specifier> * The "scheme" identifies a supported external storage system type. * The "remote-location-specifier" is an external system-specific string for identifying a location inside the storage system. * These external systems are supported: - Object stores that support AWS's S3 protocol. These use the scheme "s3". For example, "path=s3://mybucket/some/path". - POSIX file system, potentially a remote filesystem mounted over NFS. These use the scheme "file". For example, "path=file:///mnt/cheap-storage/some/path". maxVolumeDataSizeMB = <positive integer> * Optional, ignored for storageType=remote * If set, this attribute limits the total size of all databases that reside on this volume to the maximum size specified, in MB. Note that this it will act only on those indexes which reference this volume, not on the total size of the path set in the path attribute of this volume. * If the size is exceeded, Splunk will remove buckets with the oldest value of latest time (for a given bucket) across all indexes in the volume, until the volume is below the maximum size. This is the trim operation. Note that this can cause buckets to be chilled [moved to cold] directly from a hot DB, if those buckets happen to have the least value of latest-time (LT) across all indexes in the volume. * Highest legal value is 4294967295, lowest legal value is 1. rotatePeriodInSecs = <nonnegative integer> * Optional, ignored for storageType=remote * Specifies period of trim operation for this volume. * If not set, the value of global rotatePeriodInSecs attribute is inherited. * Highest legal value is 4294967295 datatype = <event|metric> * Optional, defaults to 'event'. * Determines whether the index stores log events or metric data. * If set to 'metric', we optimize the index to store metric data which can be queried later only using the mstats operator as searching metric data is different from traditional log events. * Use 'metric' data type only for metric sourcetypes like statsd. remote.* = <String> * Optional. * With remote volumes, communication between the indexer and the external storage system may require additional configuration, specific to the type of storage system. You can pass configuration information to the storage system by specifying the settings through the following schema: remote.<scheme>.<config-variable> = <value>. For example: remote.s3.access_key = ACCESS_KEY ################################################################ ##### S3 specific settings ################################################################ remote.s3.header.<http-method-name>.<header-field-name> = <String> * Currently not supported. This setting is related to a feature that is still under development. * Optional. * Enable server-specific features, such as reduced redundancy, encryption, and so on, by passing extra HTTP headers with the REST requests. The <http-method-name> can be any valid HTTP method. For example, GET, PUT, or ALL, for setting the header field for all HTTP methods. * Example: remote.s3.header.PUT.x-amz-storage-class = REDUCED_REDUNDANCY remote.s3.access_key = <String> * Currently not supported. This setting is related to a feature that is still under development. * Optional. * Specifies the access key to use when authenticating with the remote storage system supporting the S3 API. * If not specified, the indexer will look for these environment variables: AWS_ACCESS_KEY_ID or AWS_ACCESS_KEY (in that order). * If the environment variables are not set and the indexer is running on EC2, the indexer attempts to use the access key from the IAM role. * Default: unset remote.s3.secret_key = <String> * Currently not supported. This setting is related to a feature that is still under development. * Optional. * Specifies the secret key to use when authenticating with the remote storage system supporting the S3 API. * If not specified, the indexer will look for these environment variables: AWS_SECRET_ACCESS_KEY or AWS_SECRET_KEY (in that order). * If the environment variables are not set and the indexer is running on EC2, the indexer attempts to use the secret key from the IAM role. * Default: unset remote.s3.signature_version = v2|v4 * Currently not supported. This setting is related to a feature that is still under development. * Optional. * The signature version to use when authenticating with the remote storage system supporting the S3 API. * If not specified, it defaults to v4. * For 'sse-kms' server-side encryption scheme, you must use signature_version=v4. remote.s3.auth_region = <String> * Currently not supported. This setting is related to a feature that is still under development. * Optional * The authentication region to use when signing the requests when interacting with the remote storage system supporting the S3 API. If unset, Splunk will attempt to automatically extract the value from the endpoint URL * Defaults: unset remote.s3.use_delimiter = true | false * Currently not supported. This setting is related to a feature that is still under development. * Optional. * Specifies whether a delimiter (currently "guidSplunk") should be used to list the objects that are present on the remote storage. * A delimiter groups objects that have the same delimiter value so that the listing process can be more efficient as it does not need to report similar objects. * Defaults to: true remote.s3.supports_versioning = true | false * Currently not supported. This setting is related to a feature that is still under development. * Optional. * Specifies whether the remote storage supports versioning. * Versioning is a means of keeping multiple variants of an object in the same bucket on the remote storage. * Defaults to: true remote.s3.endpoint = <URL> * Currently not supported. This setting is related to a feature that is still under development. * Optional. * The URL of the remote storage system supporting the S3 API. * The scheme, http or https, can be used to enable or disable SSL connectivity with the endpoint. * If not specified and the indexer is running on EC2, the endpoint will be constructed automatically based on the EC2 region of the instance where the indexer is running, as follows: https://s3-<region>.amazonaws.com * Example: https://s3-us-west-2.amazonaws.com remote.s3.multipart_download.part_size = <unsigned int> * Currently not supported. This setting is related to a feature that is still under development. * Optional. * Sets the download size of parts during a multipart download. * This setting uses HTTP/1.1 Range Requests (RFC 7233) to improve throughput overall and for retransmission of failed transfers. * A value of 0 disables downloading in multiple parts, i.e., files will always be downloaded as a single (large) part. * Do not change this value unless that value has been proven to improve throughput. * Minimum value: 5242880 (5 MB) * Defaults: 134217728 (128 MB) remote.s3.multipart_upload.part_size = <unsigned int> * Currently not supported. This setting is related to a feature that is still under development. * Optional. * Sets the upload size of parts during a multipart upload. * Minimum value: 5242880 (5 MB) * Defaults: 134217728 (128 MB) remote.s3.multipart_max_connections = <unsigned int> * Specifies the maximum number of HTTP connections to have in progress for either multipart download or upload. * A value of 0 means unlimited. * Default: 8 remote.s3.enable_data_integrity_checks = <bool> * If set to true, Splunk sets the data checksum in the metadata field of the HTTP header during upload operation to S3. * The checksum is used to verify the integrity of the data on uploads. * Default: false remote.s3.enable_signed_payloads = <bool> * If set to true, Splunk signs the payload during upload operation to S3. * Valid only for remote.s3.signature_version = v4 * Default: true remote.s3.retry_policy = max_count * Optional. * Sets the retry policy to use for remote file operations. * A retry policy specifies whether and how to retry file operations that fail for those failures that might be intermittent. * Retry policies: + "max_count": Imposes a maximum number of times a file operation will be retried upon intermittent failure both for individual parts of a multipart download or upload and for files as a whole. * Defaults: max_count remote.s3.max_count.max_retries_per_part = <unsigned int> * Optional. * When the remote.s3.retry_policy setting is max_count, sets the maximum number of times a file operation will be retried upon intermittent failure. * The count is maintained separately for each file part in a multipart download or upload. * Defaults: 9 remote.s3.max_count.max_retries_in_total = <unsigned int> * Optional. * When the remote.s3.retry_policy setting is max_count, sets the maximum number of times a file operation will be retried upon intermittent failure. * The count is maintained for each file as a whole. * Defaults: 128 remote.s3.timeout.connect = <unsigned int> * Currently not supported. This setting is related to a feature that is still under development. * Optional * Set the connection timeout, in milliseconds, to use when interacting with S3 for this volume * Defaults: 5000 remote.s3.timeout.read = <unsigned int> * Currently not supported. This setting is related to a feature that is still under development. * Optional * Set the read timeout, in milliseconds, to use when interacting with S3 for this volume * Defaults: 60000 remote.s3.timeout.write = <unsigned int> * Currently not supported. This setting is related to a feature that is still under development. * Optional * Set the write timeout, in milliseconds, to use when interacting with S3 for this volume * Defaults: 60000 remote.s3.sslVerifyServerCert = <bool> * Currently not supported. This setting is related to a feature that is still under development. * Optional * If this is set to true, Splunk verifies certificate presented by S3 server and checks that the common name/alternate name matches the ones specified in 'remote.s3.sslCommonNameToCheck' and 'remote.s3.sslAltNameToCheck'. * Defaults: false remote.s3.sslVersions = <versions_list> * Currently not supported. This setting is related to a feature that is still under development. * Optional * Comma-separated list of SSL versions to connect to 'remote.s3.endpoint'. * The versions available are "ssl3", "tls1.0", "tls1.1", and "tls1.2". * The special version "*" selects all supported versions. The version "tls" selects all versions tls1.0 or newer. * If a version is prefixed with "-" it is removed from the list. * SSLv2 is always disabled; "-ssl2" is accepted in the version list but does nothing. * When configured in FIPS mode, ssl3 is always disabled regardless of this configuration. * Defaults: tls1.2 remote.s3.sslCommonNameToCheck = <commonName1>, <commonName2>, .. * Currently not supported. This setting is related to a feature that is still under development. * If this value is set, and 'remote.s3.sslVerifyServerCert' is set to true, splunkd checks the common name of the certificate presented by the remote server (specified in 'remote.s3.endpoint') against this list of common names. * Defaults: unset remote.s3.sslAltNameToCheck = <alternateName1>, <alternateName2>, .. * Currently not supported. This setting is related to a feature that is still under development. * If this value is set, and 'remote.s3.sslVerifyServerCert' is set to true, splunkd checks the alternate name(s) of the certificate presented by the remote server (specified in 'remote.s3.endpoint') against this list of subject alternate names. * Defaults: unset remote.s3.sslRootCAPath = <path> * Currently not supported. This setting is related to a feature that is still under development. * Optional * Full path to the Certificate Authrity (CA) certificate PEM format file containing one or more certificates concatenated together. S3 certificate will be validated against the CAs present in this file. * Defaults: [sslConfig/caCertFile] in server.conf remote.s3.cipherSuite = <cipher suite string> * Currently not supported. This setting is related to a feature that is still under development. * Optional * If set, uses the specified cipher string for the SSL connection. * If not set, uses the default cipher string. * Must specify 'dhFile' to enable any Diffie-Hellman ciphers. * Defaults: TLSv1+HIGH:TLSv1.2+HIGH:@STRENGTH remote.s3.ecdhCurves = <comma separated list of ec curves> * Currently not supported. This setting is related to a feature that is still under development. * Optional * ECDH curves to use for ECDH key negotiation. * The curves should be specified in the order of preference. * The client sends these curves as a part of Client Hello. * We only support named curves specified by their SHORT names. (see struct ASN1_OBJECT in asn1.h) * The list of valid named curves by their short/long names can be obtained by executing this command: $SPLUNK_HOME/bin/splunk cmd openssl ecparam -list_curves * e.g. ecdhCurves = prime256v1,secp384r1,secp521r1 * Defaults: unset remote.s3.dhFile = <path> * Currently not supported. This setting is related to a feature that is still under development. * Optional * PEM format Diffie-Hellman parameter file name. * DH group size should be no less than 2048bits. * This file is required in order to enable any Diffie-Hellman ciphers. * Defaults:unset. remote.s3.encryption = sse-s3 | sse-kms | sse-c | none * Currently not supported. This setting is related to a feature that is still under development. * Optional * Specifies the scheme to use for Server-side Encryption (SSE) for data-at-rest. * sse-s3: Check http://docs.aws.amazon.com/AmazonS3/latest/dev/UsingServerSideEncryption.html * sse-kms: Check http://docs.aws.amazon.com/AmazonS3/latest/dev/UsingKMSEncryption.html * sse-c: Check http://docs.aws.amazon.com/AmazonS3/latest/dev/ServerSideEncryptionCustomerKeys.html * none: no Server-side encryption enabled. Data is stored unencrypted on the remote storage. * Defaults: none remote.s3.encryption.sse-c.key_type = kms * Currently not supported. This setting is related to a feature that is still under development. * Optional * Determines the mechanism Splunk uses to generate the key for sending over to S3 for SSE-C. * The only valid value is 'kms', indicating AWS KMS service. * One must specify required KMS settings: e.g. remote.s3.kms.key_id for Splunk to start up while using SSE-C. * Defaults: kms. remote.s3.encryption.sse-c.key_refresh_interval = <unsigned int> * Currently not supported. This setting is related to a feature that is still under development. * Optional * Specifies period in seconds at which a new key will be generated and used for encrypting any new data being uploaded to S3. * Defaults: 86400 remote.s3.kms.key_id = <String> * Currently not supported. This setting is related to a feature that is still under development. * Required if remote.s3.encryption = sse-c | sse-kms * Specifies the identifier for Customer Master Key (CMK) on KMS. It can be the unique key ID or the Amazon Resource Name (ARN) of the CMK or the alias name or ARN of an alias that refers to the CMK. * Examples: Unique key ID: 1234abcd-12ab-34cd-56ef-1234567890ab CMK ARN: arn:aws:kms:us-east-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab Alias name: alias/ExampleAlias Alias ARN: arn:aws:kms:us-east-2:111122223333:alias/ExampleAlias * Defaults: unset remote.s3.kms.access_key = <String> * Currently not supported. This setting is related to a feature that is still under development. * Optional. * Similar to 'remote.s3.access_key'. * If not specified, KMS access uses 'remote.s3.access_key'. * Default: unset remote.s3.kms.secret_key = <String> * Currently not supported. This setting is related to a feature that is still under development. * Optional. * Similar to 'remote.s3.secret_key'. * If not specified, KMS access uses 'remote.s3.secret_key'. * Default: unset remote.s3.kms.auth_region = <String> * Required if 'remote.s3.auth_region' is unset and Splunk can not automatically extract this information. * Similar to 'remote.s3.auth_region'. * If not specified, KMS access uses 'remote.s3.auth_region'. * Defaults: unset remote.s3.kms.max_concurrent_requests = <unsigned int> * Currently not supported. This setting is related to a feature that is still under development. * Optional. * Limits maximum concurrent requests to KMS from this Splunk instance. * NOTE: Can severely affect search performance if set to very low value. * Defaults: 10 remote.s3.kms.<ssl_settings> = <...> * Currently not supported. This setting is related to a feature that is still under development. * Optional. * Check the descriptions of the SSL settings for remote.s3.<ssl_settings> above. e.g. remote.s3.sslVerifyServerCert. * Valid ssl_settings are sslVerifyServerCert, sslVersions, sslRootCAPath, sslAltNameToCheck, sslCommonNameToCheck, cipherSuite, ecdhCurves and dhFile. * All of these are optional and fall back to same defaults as remote.s3.<ssl_settings>. #************************************************************************** # Archiver settings. This section describes settings that affect the archiver- # optional and archiver-mandatory parameters only. # # As the first step in the Archiver feature, Self Storage allows users to move # their data from Splunk indexes to customer-owned external storage in AWS S3 # when the data reaches the end of the retention period. Note that only the # raw data and delete marker files are transferred to the external storage. # Future development may include the support for storage hierarchies and the # automation of data rehydration. # # For example, use the following settings to configure Self Storage. # archiver.selfStorageProvider = S3 # archiver.selfStorageBucket = mybucket # archiver.selfStorageBucketFolder = folderXYZ #************************************************************************** archiver.selfStorageProvider = <string> * Specifies the storage provider for Self Storage. * Optional. Only required when using Self Storage. * The only supported provider is S3. More providers will be added in the future for other cloud vendors and other storage options. archiver.selfStorageBucket = <string> * Specifies the destination bucket for Self Storage. * Optional. Only required when using Self Storage. archiver.selfStorageBucketFolder = <string> * Specifies the folder on the destination bucket for Self Storage. * Optional. If not specified, data is uploaded to the root path in the destination bucket.
indexes.conf.example
# Version 7.1.1 # # This file contains an example indexes.conf. Use this file to configure # indexing properties. # # To use one or more of these configurations, copy the configuration block # into indexes.conf in $SPLUNK_HOME/etc/system/local/. You must restart # Splunk to enable configurations. # # To learn more about configuration files (including precedence) please see # the documentation located at # http://docs.splunk.com/Documentation/Splunk/latest/Admin/Aboutconfigurationfiles # # The following example defines a new high-volume index, called "hatch", and # sets this to be the default index for both incoming data and search. # # Note that you may want to adjust the indexes that your roles have access # to when creating indexes (in authorize.conf) defaultDatabase = hatch [hatch] homePath = $SPLUNK_DB/hatchdb/db coldPath = $SPLUNK_DB/hatchdb/colddb thawedPath = $SPLUNK_DB/hatchdb/thaweddb maxDataSize = 10000 maxHotBuckets = 10 # The following example changes the default amount of space used on a # per-index basis. [default] maxTotalDataSizeMB = 650000 maxGlobalDataSizeMB = 0 # The following example changes the time data is kept around by default. # It also sets an export script. NOTE: You must edit this script to set # export location before running it. [default] maxWarmDBCount = 200 frozenTimePeriodInSecs = 432000 rotatePeriodInSecs = 30 coldToFrozenScript = "$SPLUNK_HOME/bin/python" "$SPLUNK_HOME/bin/myColdToFrozenScript.py" # This example freezes buckets on the same schedule, but lets Splunk do the # freezing process as opposed to a script [default] maxWarmDBCount = 200 frozenTimePeriodInSecs = 432000 rotatePeriodInSecs = 30 coldToFrozenDir = "$SPLUNK_HOME/myfrozenarchive" ### This example demonstrates the use of volumes ### # volume definitions; prefixed with "volume:" [volume:hot1] path = /mnt/fast_disk maxVolumeDataSizeMB = 100000 [volume:cold1] path = /mnt/big_disk # maxVolumeDataSizeMB not specified: no data size limitation on top of the # existing ones [volume:cold2] path = /mnt/big_disk2 maxVolumeDataSizeMB = 1000000 # index definitions [idx1] homePath = volume:hot1/idx1 coldPath = volume:cold1/idx1 # thawedPath must be specified, and cannot use volume: syntax # choose a location convenient for reconstitition from archive goals # For many sites, this may never be used. thawedPath = $SPLUNK_DB/idx1/thaweddb [idx2] # note that the specific indexes must take care to avoid collisions homePath = volume:hot1/idx2 coldPath = volume:cold2/idx2 thawedPath = $SPLUNK_DB/idx2/thaweddb [idx3] homePath = volume:hot1/idx3 coldPath = volume:cold2/idx3 thawedPath = $SPLUNK_DB/idx3/thaweddb ### Indexes may be allocated space in effective groups by sharing volumes ### # perhaps we only want to keep 100GB of summary data and other # low-volume information [volume:small_indexes] path = /mnt/splunk_indexes maxVolumeDataSizeMB = 100000 # and this is our main event series, allowing 50 terabytes [volume:large_indexes] path = /mnt/splunk_indexes maxVolumeDataSizeMB = 50000000 # summary and rare_data together will be limited to 100GB [summary] homePath=volume:small_indexes/summary/db coldPath=volume:small_indexes/summary/colddb thawedPath=$SPLUNK_DB/summary/thaweddb # low-volume indexes probably don't want a lot of hot buckets maxHotBuckets = 2 # if the volume is quite low, and you have data sunset goals you may # want to have smaller buckets maxDataSize = 500 [rare_data] homePath=volume:small_indexes/rare_data/db coldPath=volume:small_indexes/rare_data/colddb thawedPath=$SPLUNK_DB/rare_data/thaweddb maxHotBuckets = 2 # main, and any other large volume indexes you add sharing large_indexes # will be together be constrained to 50TB, separately from the 100GB of # the small_indexes [main] homePath=volume:large_indexes/main/db coldPath=volume:large_indexes/main/colddb thawedPath=$SPLUNK_DB/main/thaweddb # large buckets and more hot buckets are desirable for higher volume # indexes, and ones where the variations in the timestream of events is # hard to predict. maxDataSize = auto_high_volume maxHotBuckets = 10 [idx1_large_vol] homePath=volume:large_indexes/idx1_large_vol/db coldPath=volume:large_indexes/idx1_large_vol/colddb homePath=$SPLUNK_DB/idx1_large/thaweddb # this index will exceed the default of .5TB requiring a change to maxTotalDataSizeMB maxTotalDataSizeMB = 750000 maxDataSize = auto_high_volume maxHotBuckets = 10 # but the data will only be retained for about 30 days frozenTimePeriodInSecs = 2592000 ### This example demonstrates database size constraining ### # In this example per-database constraint is combined with volumes. While a # central volume setting makes it easy to manage data size across multiple # indexes, there is a concern that bursts of data in one index may # significantly displace data from others. The homePath.maxDataSizeMB setting # can be used to assure that no index will ever take more than certain size, # therefore alleviating the concern. # global settings # will be inherited by all indexes: no database will exceed 1TB homePath.maxDataSizeMB = 1000000 # volumes [volume:caliente] path = /mnt/fast_disk maxVolumeDataSizeMB = 100000 [volume:frio] path = /mnt/big_disk maxVolumeDataSizeMB = 1000000 # and this is our main event series, allowing about 50 terabytes [volume:large_indexes] path = /mnt/splunk_indexes maxVolumeDataSizeMB = 50000000 # indexes [i1] homePath = volume:caliente/i1 # homePath.maxDataSizeMB is inherited coldPath = volume:frio/i1 # coldPath.maxDataSizeMB not specified: no limit - old-style behavior thawedPath = $SPLUNK_DB/i1/thaweddb [i2] homePath = volume:caliente/i2 # overrides the default maxDataSize homePath.maxDataSizeMB = 1000 coldPath = volume:frio/i2 # limits the cold DB's coldPath.maxDataSizeMB = 10000 thawedPath = $SPLUNK_DB/i2/thaweddb [i3] homePath = /old/style/path homePath.maxDataSizeMB = 1000 coldPath = volume:frio/i3 coldPath.maxDataSizeMB = 10000 thawedPath = $SPLUNK_DB/i3/thaweddb # main, and any other large volume indexes you add sharing large_indexes # will together be constrained to 50TB, separately from the rest of # the indexes [main] homePath=volume:large_indexes/main/db coldPath=volume:large_indexes/main/colddb thawedPath=$SPLUNK_DB/main/thaweddb # large buckets and more hot buckets are desirable for higher volume indexes maxDataSize = auto_high_volume maxHotBuckets = 10 ### This example demonstrates how to configure a volume that points to ### S3-based remote storage, and an index that uses this volume. The setting ### "storageType=remote" indicates that this is a remote-storage volume. ### The "remotePath" parameter associates the index with that volume ### and configures a top-level location for uploading buckets. ### ### Note: Remote storage volumes are currently under development and ### are not supported for production use. [volume:s3] storageType = remote path = s3://example-s3-bucket/remote_volume remote.s3.access_key = S3_ACCESS_KEY remote.s3.secret_key = S3_SECRET_KEY [foo] coldPath=$SPLUNK_DB/foo/colddb homePath=$SPLUNK_DB/foo/db thawedPath=$SPLUNK_DB/foo/thaweddb remotePath=volume:s3/$_index_name ### This example demonstrates how to configure a volume that points to ### file-based remote storage, such as NFS-based remote storage. ### The "remotePath" parameter under the [default] stanza causes all ### indexes on the system to use the remote storage. ### ### Note: Remote storage volumes are currently under development and ### are not supported for production use. [default] remotePath = volume:file/$_index_name [volume:file] storageType = remote path = file:///mount/nfs
health.conf | inputs.conf |
This documentation applies to the following versions of Splunk® Enterprise: 7.1.1
Feedback submitted, thanks!