Splunk® Data Stream Processor

Function Reference

Acrobat logo Download manual as PDF

On April 3, 2023, Splunk Data Stream Processor will reach its end of sale, and will reach its end of life on February 28, 2025. If you are an existing DSP customer, please reach out to your account team for more information.

All DSP releases prior to DSP 1.4.0 use Gravity, a Kubernetes orchestrator, which has been announced end-of-life. We have replaced Gravity with an alternative component in DSP 1.4.0. Therefore, we will no longer provide support for versions of DSP prior to DSP 1.4.0 after July 1, 2023. We advise all of our customers to upgrade to DSP 1.4.0 before July 1, 2023 in order to continue to receive full product support from Splunk.
This documentation does not apply to the most recent version of Splunk® Data Stream Processor. Click here for the latest version.
Acrobat logo Download topic as PDF

Get data from Splunk DSP Firehose

Use the Splunk DSP Firehose to get data sent from the Forwarder, Collect, Ingest, DSP HEC, and Syslog (through SC4S) API services. These services support data ingest for the following data sources:

  • Splunk forwarders
  • HTTP clients
  • Syslog data
  • The Ingest REST API
  • Amazon CloudWatch
  • Amazon S3
  • Amazon Web Services (AWS) metadata
  • Google Cloud Monitoring
  • Microsoft 365
  • Microsoft Azure Monitor

The Splunk DSP Firehose collects data from all of the supported services concurrently and outputs the combined data in a single stream, allowing you to ingest data from multiple data sources at the same time with minimal pipeline configuration. See Connecting multiple data sources to your DSP pipeline in the Connect to Data Sources and Destinations with DSP manual for more information.

Function output schema

This function outputs data pipeline events using the event schema for events or metrics schema for metrics.

Optional arguments

Description: The position in the data stream where you want to start reading data. Defaults to LATEST.
  • LATEST: Start reading data from the latest position on the data stream.
  • TRIM_HORIZON: Start reading data from the very beginning of the data stream.
Example: "LATEST"

SPL2 example

When working in the SPL View, you can write the function using the following syntax.

| from splunk_firehose("TRIM_HORIZON") |...;

Alternatively, you can use the named argument syntax to write this function.

| from splunk_firehose(initial_position: "TRIM_HORIZON") |...;

Limitations of the Splunk DSP Firehose

The maximum supported event or metric record size is 5 MB.

Last modified on 25 May, 2021
Write thru KV Store
Get data from Forwarders service

This documentation applies to the following versions of Splunk® Data Stream Processor: 1.2.0, 1.2.1-patch02, 1.2.1, 1.2.2-patch02, 1.2.4, 1.2.5, 1.3.0, 1.3.1

Was this documentation topic helpful?

You must be logged into splunk.com in order to post comments. Log in now.

Please try to keep this discussion focused on the content covered in this documentation topic. If you have a more general question about Splunk functionality or are experiencing a difficulty with Splunk, consider posting a question to Splunkbase Answers.

0 out of 1000 Characters