All DSP releases prior to DSP 1.4.0 use Gravity, a Kubernetes orchestrator, which has been announced end-of-life. We have replaced Gravity with an alternative component in DSP 1.4.0. Therefore, we will no longer provide support for versions of DSP prior to DSP 1.4.0 after July 1, 2023. We advise all of our customers to upgrade to DSP 1.4.0 in order to continue to receive full product support from Splunk.
Increase internal partitions to improve pipeline throughput
The throughput of your pipelines is highly correlated with the parallelism of the pipeline. You can increase the parallelism of certain pipelines by increasing the number of input partitions of the internal Apache Pulsar message bus. The Splunk Data Stream Processor uses Apache Pulsar as the message bus for the following data sources: Read from Splunk Firehose, Read from Forwarders Service, and Read from the Ingest REST API.
Data loss may occur when decreasing the number of partitions later on. Therefore, if you want to increase the number of input partitions, make sure that you do not overallocate input partitions in the process. If you do need to decrease the number of partitions, contact Splunk Support.
Steps:
- From a controller node in your cluster, get a list of running Apache Pulsar broker pods.
kubectl get pods -n pulsar
- Log into a running broker pod.
kubectl exec -it pulsar-broker-0 -n pulsar /bin/bash
- (Optional) Get the current number of partitions.
pulsar-admin topics get-partitioned-topic-metadata persistent://dsp/ingest_default/input
- Use the
pulsar-admin
CLI tool to update the number of partitions.pulsar-admin topics update-partitioned-topic -p <Number-of-Partitions> persistent://dsp/ingest_default/input
- Confirm that the number of partitions has been changed by using the
pulsar-admin
CLI tool again.pulsar-admin topics get-partitioned-topic-metadata persistent://dsp/ingest_default/input
- Log in to the Data Stream Processor and restart your pipelines for changes to take effect.
To further improve pipeline throughput, you can add a batching function in your pipeline. See batch bytes or batch records.
Back up your Splunk Data Stream Processor deployment | About the Splunk App for DSP |
This documentation applies to the following versions of Splunk® Data Stream Processor: 1.4.0, 1.4.1, 1.4.2, 1.4.3, 1.4.4, 1.4.5
Feedback submitted, thanks!