Splunk® Data Stream Processor

Connect to Data Sources and Destinations with DSP

On April 3, 2023, Splunk Data Stream Processor reached its end of sale, and will reach its end of life on February 28, 2025. If you are an existing DSP customer, please reach out to your account team for more information.

All DSP releases prior to DSP 1.4.0 use Gravity, a Kubernetes orchestrator, which has been announced end-of-life. We have replaced Gravity with an alternative component in DSP 1.4.0. Therefore, we will no longer provide support for versions of DSP prior to DSP 1.4.0 after July 1, 2023. We advise all of our customers to upgrade to DSP 1.4.0 in order to continue to receive full product support from Splunk.

Getting started with DSP data connections

As a administrator, you are responsible for creating connections in DSP to get data in from a data source or send data out to a data destination. DSP includes connectors that provide read and write support for a variety of data sources and destinations including Splunk indexes, databases, and pub/sub messaging systems. To allow DSP to access your data, you must configure a connection that contains your credentials for the data source or destination. DSP can then use these connections to access your data, and start reading from data sources or writing to data destinations.

Any credentials that you provide are transmitted securely by HTTPS, encrypted, and securely stored in a secrets manager.

Only DSP administrators are permitted to create connections. For instructions on how to create a connection, refer to the chapter corresponding to the type of data source or destination that you are using. For example:

  • To create a connection that gets data from a Splunk forwarder, see the Splunk forwarders chapter.
  • To create a connection that sends data to a Splunk index, see the Splunk indexes chapter.
  • To create a connection that gets data from multiple data sources concurrently, or send data to multiple data destinations concurrently, see the Multiple data sources and destinations chapter.

Once you've created connections to your data source and destination of choice, you can build a data pipeline that uses these connections to access your data. You can then start streaming and transforming data using DSP.

See also

For this information Refer to this documentation
How to assign the DSP admin role, and the permissions that admins have. Manage users and admins
Step-by-step tutorial that guides you through the process of creating and using a data pipeline. Tutorial
Details about the supported methods for creating a data pipeline. Navigating the Splunk Data Stream Processor
Details about the DSP functions that use connections to collect data from supported data sources. Source functions
Details about the DSP functions that use connections to send data from pipelines to supported data destinations. Sink functions
Last modified on 29 July, 2022
  Data sources and destinations

This documentation applies to the following versions of Splunk® Data Stream Processor: 1.2.0, 1.2.1-patch02, 1.2.1, 1.2.2-patch02, 1.2.4, 1.2.5, 1.3.0, 1.3.1, 1.4.0, 1.4.1, 1.4.2, 1.4.3, 1.4.4, 1.4.5, 1.4.6


Was this topic useful?







You must be logged into splunk.com in order to post comments. Log in now.

Please try to keep this discussion focused on the content covered in this documentation topic. If you have a more general question about Splunk functionality or are experiencing a difficulty with Splunk, consider posting a question to Splunkbase Answers.

0 out of 1000 Characters