All DSP releases prior to DSP 1.4.0 use Gravity, a Kubernetes orchestrator, which has been announced end-of-life. We have replaced Gravity with an alternative component in DSP 1.4.0. Therefore, we will no longer provide support for versions of DSP prior to DSP 1.4.0 after July 1, 2023. We advise all of our customers to upgrade to DSP 1.4.0 in order to continue to receive full product support from Splunk.
Get data from Amazon Metadata
Use the Amazon Metadata source function to get data from the resources and infrastructure in Amazon Web Services (AWS).
Prerequisites
Before you can use this function, you must create a connection. See Create a DSP connection to get metadata from AWS in the Connect to Data Sources and Destinations with the manual. When configuring this source function, set the connection_id
argument to the ID of that connection.
Function output schema
This function outputs data pipeline events using the event schema.
The data that is included in the body
field varies depending on the specific AWS API that the event comes from, but typically contains all the attributes of the corresponding AWS resource. For example, the body
field of an event from the ec2_key_pairs
API would include all attributes of the ec2.KeyPairInfo
resource. However, in some cases, additional related attributes are included in the body
field. See How AWS metadata is collected in the Connect to Data Sources and Destinations with DSP manual for more information.
For all events, the function also includes the following attributes in addition to the ones that are part of the original payload:
AccountID
: The ID of the AWS account associated with the event. This attribute is omitted if the account ID cannot be retrieved.Region
: The AWS region associated with the event.
The following is an example of a record from the read_from_aws_cloudwatch_metrics
function. In this example, the record contains an event from the ec2_key_pairs
API.
{ "timestamp": 1568050119000, "nanos": 0, "id": "2823738566644596", "host": "test-host-1", "source": "us-east-1:ec2:describeKeyPairs", "source_type": "aws:ec2:keypair", "body": { "KeyFingerprint":"d7:a1:98:76:fe:f6:29:3a:76:8e:15:57:d1:d6:e4:35:69:f2:b8:3e", "KeyName":"my_crypto_key", "KeyPairId":"AJRAKBM7YBYC4DUD5QPM" }, "attributes": { "AccountID": "123412341234", "Region": "ca-central-1" } }
Required arguments
- connection_id
- Syntax: string
- Description: The ID of your Amazon Metadata connection.
- Example: "576205b3-f6f5-4ab7-8ffc-a4089a95d0c4"
Optional arguments
- initial_position
- Syntax: LATEST | TRIM_HORIZON
- Description: The position in the data stream where you want to start reading data. Defaults to LATEST.
- LATEST: Start reading data from the latest position on the data stream.
- TRIM_HORIZON: Start reading data from the very beginning of the data stream.
- Example: LATEST
SPL2 example
When working in the SPL View, you can write the function by listing arguments in this exact order.
| from read_from_aws_metadata("my-connection-id", "TRIM_HORIZON") |... ;
Alternatively, you can use named arguments to declare arguments in any order. The following SPL2 example uses named arguments to specify the initial_position
argument before the connection_id
argument.
| from read_from_aws_metadata(initial_position: "TRIM_HORIZON", connection_id: "my-connection-id") |... ;
If you want to use a mix of unnamed and named arguments in your functions, you need to list all unnamed arguments in the correct order before providing the named arguments.
Limitations of the Amazon Metadata source function
The Amazon Metadata source function uses scheduled data collection jobs to ingest data. See Limitations of scheduled data collection jobs for information about limitations that apply to all scheduled data collection jobs.
Get data from Amazon Kinesis Data Stream | Get data from Amazon S3 |
This documentation applies to the following versions of Splunk® Data Stream Processor: 1.2.0, 1.2.1-patch02, 1.2.1, 1.2.2-patch02, 1.2.4, 1.2.5, 1.3.0, 1.3.1
Feedback submitted, thanks!