Splunk® Data Stream Processor

Use the Data Stream Processor

On April 3, 2023, Splunk Data Stream Processor will reach its end of sale, and will reach its end of life on February 28, 2025. If you are an existing DSP customer, please reach out to your account team for more information.
This documentation does not apply to the most recent version of Splunk® Data Stream Processor. For documentation on the most recent version, go to the latest release.

Create a Splunk DSP pipeline that processes universal forwarder data

Because the universal forwarder doesn't parse incoming data, except in certain cases, you must use the Group by and Merge Events functions to properly ingest data from the universal forwarder into your data pipeline. The Data Stream Processor provides a Splunk universal forwarder template so you do not need to construct this pipeline from scratch.

Prerequisites

Steps

  1. From the Build Pipeline page, select the Splunk universal forwarder template.
    This template creates a pipeline that reads data from Splunk Forwarders, does the appropriate processing required by the universal forwarder data source, and sends the data to the main index of the preconfigured Splunk Enterprise instance associated with the Data Stream Processor.
  2. Most of the pipeline is preconfigured for you, but you'll need to provide a regular expression delimiter to correctly stitch your events together. Click the Merge Events function.
  3. In the Delimiter text box, use a Java 8 regular expression to express a delimiter for your log files. For example, if your log file events look like the following.
    2018-12-18 15:09:00,144 log event 1
    log event 1 continue
    log event 1 continue
    log event 1 continue
    2018-12-18 15:09:01,144 log event 2

    You can use the following regular expression to correctly group the events together.

    (\\n)[0-9]{4}-[0-9]{2}-[0-9]{2}

    This results in your events being broken correctly into two events:

    2018-12-18 15:09:00,144 log event 1 log event 1 continue log event 1 continue log event 1 continue
    2018-12-18 15:09:01,144 log event 2
  4. Click Validate and Start Preview to check if your events are passing through your pipeline as-expected.
Last modified on 06 December, 2019
About Splunk Data Stream Processor regular expressions   Deserialize and send Kafka data from a DSP pipeline

This documentation applies to the following versions of Splunk® Data Stream Processor: 1.0.0


Was this topic useful?







You must be logged into splunk.com in order to post comments. Log in now.

Please try to keep this discussion focused on the content covered in this documentation topic. If you have a more general question about Splunk functionality or are experiencing a difficulty with Splunk, consider posting a question to Splunkbase Answers.

0 out of 1000 Characters