Splunk® Enterprise

Monitoring Splunk Enterprise

Download manual as PDF

Download topic as PDF

Investigate feature health status changes

There are two ways to access feature health status information from the /server/health/splunkd endpoint:

  • View the splunkd health report in Splunk Web.
  • Query the server/health/splunkd endpoint.

You can also monitor feature health status changes in $SPLUNK_HOME/var/log/health.log. For more information on health.log file entries, see Configure health status logs.

View the splunkd health report

The splunkd health report lets you view the current status of features in the splunkd health status tree. You can use the report to identify features whose status indicates a problem, and investigate the cause of those problems.

  1. On the instance you want to monitor, log in to Splunk Web.
  2. In the main menu, check the color of the health report icon. The color of this icon indicates the overall status of splunkd.
  3. Click the health report icon to open the health report.
    The image shows the splunkd health report. The health report includes the splunkd status tree, and provides detailed diagnostic information about changes in feature health status.
  4. In the health status tree, click on any feature to view information about the feature's status.
  5. For features in the red or yellow state, review Root Cause and Last 50 related messages for information that can help you identify the cause of a feature's status change.

Example

This example shows how you can use the splunkd health report to investigate feature health status changes on a cluster master instance.

  1. Review the splunkd health report.
  2. Investigate root cause and related messages.
  3. Confirm the cause of feature status change.

1. Review the splunkd health report

  1. On the cluster master instance, log in to Splunk Web.
  2. Check the color of the health report icon.
    The image shows a red health report icon in the Splunk Web  main menu
    The red icon indicates that one or more features in the splunkd health status tree has a severe issue.
  3. Click the health report icon to open the health report.
    Data durability red.png
    The health report shows the data_durability feature has a severe issue.

2. Investigate root cause and related messages

In the health status tree, click the data_durability feature. The following diagnostic information appears:

Root Cause: "Replication Factor is not met."
If the number of peers in the cluster is less than the replication factor, the replication factor cannot be met. Therefore a possible cause of the feature's red status is an offline peer.

Last 50 Related Messages: Searching the related messages, you see log entries that contain streaming errors. For example:

09-04-2018 18:33:43.724 +0000 INFO CMMaster - streaming error src=CFEAE46E-A7F5-485E-99B3-9DE293772D88 tgt=88A0C501-5A3B-4BBD-9B32-BBD20C7158C5 failing=target bid=_internal~2~CFEAE46E-A7F5-485E-99B3-9DE293772D88 bytes=2442356

The streaming error suggests that bucket replication is failing because a source peer cannot communicate with a target peer. This type of error can be caused by a network interruption or an offline peer.

3. Confirm the cause of feature status change

After you use the splunkd health report to investigate the cause of the feature's status change, which suggests a peer is offline, you can use the Monitoring Console to check the status of cluster peers and confirm if the suspected cause is correct.

  1. In Splunk Web, click Settings > Monitoring Console.
  2. Click Indexing > Indexer Clustering > Indexer Clustering: Status.
    The dashboard shows that the peer node idx3 is stopped.
    Monitoring console idx3.png
  3. Click on the idx3 peer to see more information. The idx3 peer GUID matches the GUID of the target peer in the streaming error message. This confirms that the cause of the data_durability feature's red status is an offline peer.
    Idx3 GUID 7.2.png
    You can now take steps to restart the peer and have it rejoin the cluster, which returns the data_durability feature to the green state.

Query the server/health/splunkd endpoint

You can integrate proactive Splunk component monitoring with existing third-party monitoring tools and other applications.

To see the overall health status of splunkd, query the server/health/splunkd endpoint. For example:

curl -k -u admin:pass https://<host>:8089/services/server/health/splunkd

For endpoint details, see server/health/splunkd in the REST API Reference Manual.

To see the overall health of splunkd, as well as the health of each feature reporting to the status tree, and information on feature health status changes, query server/health/splunkd/details. For example:

curl -k -u admin:pass https://<host>:8089/services/server/health/splunkd/details

For endpoint details, see server/health/splunkd/details in the REST API Reference Manual.

For more information on how to use Splunk REST API endpoints, see Connecting to splunkd in the REST API User Manual.

For information on Splunk SDKs that support the Splunk REST API, see Overview of Splunk SDKs.

PREVIOUS
Set access controls for the splunkd health report
 

This documentation applies to the following versions of Splunk® Enterprise: 7.2.0, 7.2.1, 7.2.2, 7.2.3, 7.2.4, 7.2.5, 7.2.6, 7.3.0


Was this documentation topic helpful?

Enter your email address, and someone from the documentation team will respond to you:

Please provide your comments here. Ask a question or make a suggestion.

You must be logged into splunk.com in order to post comments. Log in now.

Please try to keep this discussion focused on the content covered in this documentation topic. If you have a more general question about Splunk functionality or are experiencing a difficulty with Splunk, consider posting a question to Splunkbase Answers.

0 out of 1000 Characters