Splunk® User Behavior Analytics

Install and Upgrade Splunk User Behavior Analytics

System requirements for Splunk UBA

Install Splunk UBA with assistance from Splunk Professional Services.

Hardware requirements

You can install Splunk UBA on a physical server, a virtual machine, or in the cloud. Hardware requirements for UBA are the same no matter where you install UBA. UBA installations are supported in any virtual machine or cloud environment, so long as the underlying hardware and operating system requirements are met.

Install Splunk UBA on its own hardware stack. Do not install Splunk UBA on the same machines as Splunk Enterprise.

You can use Splunk Professional Services resources to assist with your UBA installation.

Verify the following hardware requirements before installing Splunk UBA:

Disk space and memory requirements for installing Splunk UBA

Every machine in your Splunk UBA deployment must meet the following requirements. Not all machines in your deployment need to have matching specifications, but they must all meet the minimum requirements.

  • 16 CPU cores. If a machine has more than 16 cores, the additional cores are not used by Splunk UBA.
  • 64GB RAM. If a machine has more than 64GB RAM, the additional RAM is used by Splunk UBA as needed.
  • Disk 1 - 100GB dedicated disk space for Splunk UBA installation on bare metal systems running RHEL or OEL. The AMI images are pre-configured with 100GB root disks for Splunk UBA installation.
  • Disk 2 - 1TB (1024GB) additional disk space for metadata storage.
  • Disk 3 - 1TB (1024GB) additional disk space for each node running Spark services.

See Where services run in Splunk UBA in the Administer Splunk User Behavior Analytics manual for more information about how to determine where Splunk UBA services are running in your deployment. See Directories created or modified on the disk for more information about Splunk UBA directories and space requirements.

Do not manually mount the disks before installing Splunk UBA. During the installation procedure, the mount-a command will properly mount the disks for you.

The following table summarizes the disk requirements per deployment.

Splunk UBA Deployment Nodes Requiring 100GB Disk Space for Disk 1 Nodes Requiring a 1TB Disk for Disk 2 Nodes Requiring a 1TB Disk for Disk 3
1 Node Node 1 Node 1 Node 1
3 Nodes All Nodes All Nodes Nodes 1, 3
5 Nodes All Nodes All Nodes Nodes 1, 4, 5
7 Nodes All Nodes All Nodes Node 7
10 Nodes All Nodes All Nodes Nodes 9, 10
20 Nodes - Classic All Nodes All Nodes Nodes 17, 18, 19, 20

For disk space and memory requirements for 20 Nodes- XL, see About the 20 Node XL option in the Plan and Scale your Splunk UBA Deployment manual.

(Optional) Plan for configuring Splunk UBA warm standby

Configure warm standby in your deployment for high availability and disaster recovery. Allocate additional servers for a warm standby solution, where you can manually failover Splunk UBA to a full backup system. The backup system must have the same number of nodes as the active system. See Configure warm standby in Splunk UBA in Administer Splunk User Behavior Analytics.

(Optional) Add additional disks for offline Splunk UBA backups

Use the backup and restore scripts located in /opt/caspida/bin/utils to migrate your Splunk UBA deployment to the next larger size on the same operating system. For example, you can migrate from 5 nodes to 7 nodes, or 10 nodes to 20 nodes. If you want to migrate from 7 nodes to 20 nodes, migrate from 7 nodes to 10 nodes first, then from 10 nodes to 20 nodes.

Add an additional disk to the Splunk UBA management node mounted as /backup for the Splunk UBA backups used to restore Splunk UBA during the migration process.

The size of the additional disk must follow these guidelines:

  • The disk size must be at least half the size of your deployment in terabytes. For example, a 10-node system requires a 5TB disk.
  • If you are creating archives, allow for an additional 50 percent of the backup disk size. For example, a 10-node system requires a 5TB disk for backups, and an additional 2.5TB if for archives, so you would need a 7.5TB disk for archived backups.

The table summarizes the minimum disk size requirements for Splunk UBA backups per deployment:

For 20 node instances, the disk sizes in this table must be installed on both node1 and node2.

Number of Splunk UBA Nodes Minimum Disk Size for Backup (without archives) Minimum Disk Size for Backup (with archives)
1 Node 1TB 1.5TB
3 Nodes 1TB 1.5TB
5 Nodes 2TB 3TB
7 Nodes 4TB 6TB
10 Nodes 5TB 7.5TB
20 Nodes - Classic 10TB 15TB
20 Nodes - XL 30TB 45TB

If you have previous backups on the same disk, be sure to also take this into account when determining available disk space. See Prepare to backup Splunk UBA in Administer Splunk User Behavior Analytics.

Supported AWS server instance types

If you run Splunk UBA on an AWS instance:

  • AWS measures CPU power on Elastic Compute Cloud (EC2) instances in virtual CPUs (vCPUs), not real CPUs.
  • Each vCPU is a hyper thread of an Intel Xeon core on most AWS instance types. See Amazon EC2 Instance Types on the AWS website.
  • As a hyper thread of a core, a vCPU acts as a core, but the physical core must schedule its workload among other workloads of other vCPUs that the physical core handles.

Installation of Splunk UBA on Amazon Web Services (AWS) servers is supported on the following instance types:

  • m4.4xlarge
  • m5.4xlarge
  • m5a.4xlarge
  • m5.8xlarge
  • m6a.4xlarge
  • m6i.4xlarge

Not all available AWS instance types are supported when using the 20 Nodes - XL option. See, Use 20 Node XL on a Splunk UBA upgrade - AWS users in the Plan and Scale your Splunk UBA Deployment manual.

The m6a.4xlarge and m6i.4xlarge instance types have less memory than the other supported instance types. This memory issue is surfaced as a warning when running UBA pre-check scripts. You can ignore this warning.

All Splunk UBA nodes in your AWS environment must use gp3 (3000 IOPS) or higher performant volumes for storage.

Disk subsystem IOPS requirements

For all new Splunk UBA deployments, the disk subsystem for each Splunk UBA server must support an average Input/Output Operations per second (IOPS) of 1200 IOPS. Existing deployments on 800 IOPS servers can be upgraded without having to upgrade the disks.

IOPS are a measurement of how much data throughput a hard drive can sustain. Because a hard drive reads and writes at different speeds, there are IOPS numbers for disk reads and writes. The average IOPS is the average between those two figures. See Disk subsystem in the Capacity Planning Manual for Splunk Enterprise for more about IOPS.

Network interface requirements

Splunk UBA requires at least one 1Gb ethernet interface on each node.

Configure each Splunk UBA node with at least one control plane interface and one data plane interface. Configure the control plane interfaces on one subnet, and the data plane interfaces on a separate subnet.

Connect all interfaces on the data plane network with at least one 10GbE or better ethernet interface. For larger clusters, use 25GbE, 40GbE or 50GbE network interfaces.

When installing Splunk UBA, ensure that the interface on which you are setting up Splunk UBA is associated with the "public" zone of the firewall and has the default route set for the interface.

Use the following command to check the firewall zone of the specific interface:

sudo firewall-cmd --get-zone-of-interface=<interface name>

Ensure that the default route is present for the specific interface:

sudo ip route | grep "default"

Example:

[caspida@uba-host-01 ~]$ sudo ip route | grep "default"
default via 10.159.248.1 dev ens34 proto dhcp src 10.159.249.161 metric 101

Install third-party agents after you install Splunk UBA

Third-party agents, such as anti-virus software, can block a successful installation of Splunk UBA.

Installing security agents similar to b9daemon (Carbon Black) can cause high CPU utilizations and detrimentally impact UBA functionality.

Install any desired third-party agents only after Splunk UBA is installed, and monitor Splunk UBA for any possible interference after the agents are installed and running.

If CPU utilization spikes when third-party agents are running on your UBA nodes, exclude the /var/vcap/store/docker/overlay2/ directory in the third-party agents setting.

Directories created or modified on the disk

Splunk UBA creates or modifies the following directories on the disk during installation.

Directory Disk Description of Contents Updated During Upgrade? Recommended Space
/home/caspida Disk 1 Contains the Splunk UBA installation and upgrade .tgz files. Yes 20 GB
/opt/caspida Disk 1 Contains the Splunk UBA software. Yes 10 GB
/opt/splunk Disk 1 Contains the Splunk forwarder to send data to the Splunk platform. Yes 10 GB
/etc/caspida/local/conf Disk 1 Contains custom configuration files affecting your local environment. No 1 GB
/var Disk 1 Contains various support files required by Splunk UBA. The /var/lib directory must have a minimum of 20 GB. Yes 50 GB
/usr Disk 1 Installs libraries UBA needs including Hadoop, Kafka, and Postgres. Yes 2.5 GB
/var/vcap Disk 2 Contains the following notable sub-directories:
  • /var/vcap/packages - Contains Spark and Spark configuration files.
  • /var/vcap/sys/run - Contains files used by some Splunk UBA processes during initial startup.
  • /var/vcap/sys/log - Contains Splunk UBA log files.
  • /var/vcap/sys/tmp - Contains temporary files for some Splunk UBA processes.
  • /var/vcap/store - Contains the metadata storage for Splunk UBA services such as Kafka, PostgreSQL, Hadoop, Influxdb, and Redis
Yes 1 TB
/var/vcap2 Disk 3 Contains the runtime data for Spark services. Yes 1 TB

Operating system requirements

Installing Splunk UBA on hardened operating systems is not supported.

You must install Splunk UBA on a server that uses one of the following operating systems:

Kernel version tested details are provided for informational purposes only. Installing Splunk UBA is not kernel version dependent.

Operating System Can I upgrade Splunk UBA on this OS version? Can I install Splunk UBA on this OS version? Kernel version tested Installation type and description
Red Hat Enterprise Linux (RHEL) 8.10 Yes Yes 4.18.0-553.16.1.el8_10.x86_64 Install on supported hardware. Obtain the software from Splunk UBA Software Installation Package on Splunkbase.
Red Hat Enterprise Linux (RHEL) 8.8 Yes Yes 4.18.0-477.27.1.el8_8.x86_64 Install on supported hardware. Obtain the software from Splunk UBA Software Installation Package on Splunkbase.
Red Hat Enterprise Linux (RHEL) 8.6 Yes No 4.18.0-372.32.1.el8_6.x86_64 Install on supported hardware. Obtain the software from Splunk UBA Software Installation Package on Splunkbase.
Oracle Enterprise Linux (OEL) 8.10 Yes Yes 5.4.17-2136.333.5.1.el8uek.x86_64 Install on supported hardware. Obtain the software from Splunk UBA Software Installation Package on Splunkbase.
Oracle Enterprise Linux (OEL) 8.9 Yes Yes 5.15.0-204.147.6.3.el8uek.x86_64 Install on supported hardware. Obtain the software from Splunk UBA Software Installation Package on Splunkbase.
Oracle Enterprise Linux (OEL) 8.8 Yes No 5.15.0-101.103.2.1.el8uek.x86_64 Install on supported hardware. Obtain the software from Splunk UBA Software Installation Package on Splunkbase.
Ubuntu 20.04 LTS Yes No 5.4.0-174-generic

5.15.0-1030-aws

This operating system is bundled with the AMI installation package.


If you want to install UBA version 5.4.1 on a single Amazon Web Services instance or on several Amazon Web Services instances, follow the instructions to install version 5.4.0 and then follow the instructions to Upgrade a single node AMI or OVA installation of Splunk UBA or Upgrade a distributed AMI or OVA installation of Splunk UBA.

Splunk UBA requires that the operating system and underlying component versions match exactly on all nodes in your deployment.

Updating the operating system or any components in your deployment can break dependencies that will cause Splunk UBA to stop working and is not recommended.

If you must update the operating system before the next release of Splunk UBA, do so in a test environment and verify that everything is working properly before applying the upgrade to your production environment.

Additional RHEL requirements

Make sure your RHEL server has access to the RHEL repositories, and the license includes the following subscription names:

  • Red Hat Enterprise Linux Server
  • Red Hat Enterprise Linux Server - Extended Update Support (EUS)

The RHEL EUS subscription enables you to remain with previous stable releases of RHEL for up to approximately 24 months.

On December 31, 2021, Red Hat's CentOS Linux reached End Of Life (EOL). Per Red Hat, Inc, CentOS Linux users must migrate to a new operating system to continue receiving updates, patches, and new features. Red Hat also encourages customers to migrate to RHEL. Additionally, Red Hat made the new "CentOS Stream" operating system a non-production, pre-build version of RHEL, with no long-term support model. Splunk UBA does not include CentOS Stream as a supported operating system. Customers must migrate to and adopt a supported production Linux distro of RHEL, Ubuntu, or OEL as a base OS for UBA version 5.3.0.

User access requirements

If you are installing Splunk UBA using an AMI image, perform all tasks as the caspida user and use sudo for tasks requiring root-level privileges.

If you are installing Splunk UBA on a supported Linux platform, you must be able to do the following:

  • Be able to log in as root, or log in as a different user and use su or sudo to have root privileges. This is required for preparing the servers prior to installing the Splunk UBA software.
  • Create the caspida user with the appropriate privileges. The caspida user is required to install the Splunk UBA software.
  • All user and group authentication must be performed locally on each Splunk UBA host. Authenticating users and groups using a centralized controller or user and group management system is not supported.


Validate the UMASK value

Ensure the UMASK value of the root user is set to 0002 or 0022, or grant read permissions for newly created files and directories to the caspida user.

Complete the following steps to validate the UMASK value:

  1. Check the UMASK value of the root user by running the following command. The value must be 0002 or 0022:
    umask
  2. Verify the UMASK value in the /etc/login.defs file:
    grep -i "^UMASK" /etc/login.defs

    The umask value specified in /etc/login.defs applies as the default for all users.

  3. Validate the permissions for new files and directories:
    1. As the caspida user, create a new file or directory using sudo to observe the permissions:
      sudo touch testfile.txt 
      sudo mkdir testdirectory
      
    2. Next, check the permissions of the created files and directories. The read permission for caspida(other) users is required:
      ls -l testfile.txt 
      ls -ld testdirectory
      

To set the required umask value, edit the /etc/login.defs file and set the UMASK value to 022.

If the caspida user does not have read permissions, update the UMASK value accordingly. Failure to provide the required permission to the caspida user will result in a UBA installation or upgrade failure.

Networking requirements

Perform the following tasks or verify specific information to meet the networking requirements for installing Splunk UBA:

Assign static IP addresses to Splunk UBA servers

Assign static IP addresses to Splunk UBA servers.

Inbound networking port requirements

Splunk UBA requires the following ports to be open to the outside world for other services to interact with Splunk UBA.

Service Port
SSH 22
HTTPS 443

Splunk UBA requires the following ports to be open internally among the nodes in a distributed Splunk UBA cluster to allow specific services to interact with each other.

Service Port
SSH 22
Redis 6379, 16379
PostgreSQL 5432
Zookeeper 2181, 2888, 3888
Apache Kafka 9092, 9901, 9093 (for Kafka ingestion), 32768 - 65535 (for JMX)
Job Manager 9002
Time Series Database 8086
Apache Impala 21000, 21050, 25000, 25010, 25020
Apache Spark 7077, 8080, 8081
Hadoop Namenode 8020
Hadoop Namenode WebUI 9870
Hadoop Yarn ResourceManager 8090
Hadoop Data Transfer Port 9866
Hadoop Datanodes 9867, 9864
Hadoop Secondary namenode 9868
Hive Metastore 9090, 9095
Kubernetes/etcd 2379, 2380, 5000, 6443, 10250, 10251, 10252, 10255, 30000 - 32767

For more details on services in Splunk UBA, see Monitoring the health of your Splunk UBA deployment in Administer Splunk User Behavior Analytics.

Splunk platform port requirements

The following ports must be open on the Splunk platform to interact with Splunk UBA:

Service Port
HTTPS authentication 443
HTTP authentication 80
HTTP Event Collector to send events from Splunk UBA to the Splunk Platform Default is 8088, unless otherwise configured.
Splunk Cloud uses 443.
REST services that enable Splunk UBA to communicate with Splunk search heads 8089
Port used to send alerts to Splunk Enterprise Security (ES) User-defined.
Port used by Splunk UF to send UBA host logs to Splunk Indexers 9997 or other custom ingest port.

Modify firewalls and proxies

Modify firewalls and proxies to support the inbound and outbound port requirements defined in this document so that requests to internal services do not attempt to travel externally.

  • Set the no_proxy environment variable for general HTTP communication between nodes.
  • Set the NO_PROXY environment variable for Splunk UBA's time series database (influxdb). Set NO_PROXY to the same values as no_proxy.

Perform the following tasks to configure your firewall and proxy settings:

  1. If you use an HTTP or HTTPS proxy, exclude localhost and the IP addresses and names of the Splunk UBA servers from the proxy. For example, in a 3-node cluster, add the following configuration to the /etc/environment file:
    # Proxy host/port for reference. These variables are not used below. 
    PROXY_IP="1.2.3.4"
    PROXY_PORT="3128"
    
    # Set the proxy variables based on the values above. Both upper and lower case, different services look for different casing.
    HTTP_PROXY="http://1.2.3.4:3128"
    http_proxy="http://1.2.3.4:3128"
    
    HTTPS_PROXY="https://1.2.3.4:3128"
    https_proxy="https://1.2.3.4:3128"
    
    # Exclude loopback addresses from the proxy
    # Note: CIDR ranges aren't supported by older tools so specify both IP and CIDR
    #
    # Proxy values to be set:
    # localhost: "localhost,127.0.0.1,127.0.1.1,127.0.0.0/8"
    # UBA Containers: 10.96.0.0/12,10.244.0.0/16,172.17.0.1,172.17.0.2,172.17.0.0/16"
    # Site Specific hosts by shortname, fqdn, ip: "ubanode01,ubanode01.mydomain.local,10.10.10.1",ubanode02,ubanode02.mydomain.local,10.10.10.2""
    
    # Set NO_PROXY and no_proxy
    NO_PROXY="localhost,127.0.0.1,127.0.1.1,127.0.0.0/8,10.96.0.0/12,10.244.0.0/16,172.17.0.1,172.17.0.2,172.17.0.0/16,ubanode01,ubanode01.mydomain.local,10.10.10.1,ubanode02,ubanode02.mydomain.local,10.10.10.2,ubanode03,ubanode03.mydomain.local,10.10.10.3"
    
    no_proxy="localhost,127.0.0.1,127.0.1.1,127.0.0.0/8,10.96.0.0/12,10.244.0.0/16,172.17.0.1,172.17.0.2,172.17.0.0/16,ubanode01,ubanode01.mydomain.local,10.10.10.1,ubanode02,ubanode02.mydomain.local,10.10.10.2,ubanode03,ubanode03.mydomain.local,10.10.10.3"
    
  2. Use the following commands to stop and restart all Splunk UBA services for the changes in /etc/environment to take effect:
    /opt/caspida/bin/Caspida stop-all
    /opt/caspida/bin/Caspida start-all
    
  3. Verify that the nslookup localhost command returns a 127.x.x.x IP address. For example:
    $ nslookup localhost
    Server:		10.160.20.4
    Address:	10.160.20.4#53
    
    Name:	localhost.sv.splunk.com
    Address: 127.0.0.1
    

Configure host name lookups and DNS

Configure your environment so that Splunk UBA can resolve host names properly.

Configure the name switching service

The name switching service in Linux environments determines the order in which services are queried for host name lookups. Use cat /etc/nsswitch.conf to verify that your name switching service is using files before DNS. Check the hosts line in the output:

  • If you see files dns it means that /etc/hosts will be queried before checking DNS.
  • If you see dns files it means that DNS will be queried before the /etc/hosts file.

Also make sure myhostname is the last item on the hosts line so that the system can determine its own host name from the local config files.

$ cat /etc/nsswitch.conf
# /etc/nsswitch.conf
#
# Example configuration of GNU Name Service Switch functionality.
# If you have the `glibc-doc-reference' and `info' packages installed, try:
# `info libc "Name Service Switch"' for information about this file.

passwd:         compat
group:          compat
shadow:         compat
gshadow:        files

hosts:          files dns myhostname
...

Configure the DNS resolver

Some Splunk UBA services use DNS during installation and while the product is running. All nodes in your Splunk UBA deployment must point to the same DNS server. Verify this is the case in the /etc/resolv.conf file on each node. Use the following command to check if /etc/resolv.conf exists on your system:

ls -lH /etc/resolv.conf

If the file does not exist, create the file by performing the following tasks:

  1. Run the following command:
    sudo systemctl enable resolvconf
  2. Restart the server.
  3. Run the ls -lH /etc/resolv.conf command again to verify that the /etc/resolv.conf exists.

Verify the network interface configuration

Verify that the network interface configuration has a dns-search value configured to match your domain, such as mgmt.corp.local. Check the /etc/resolv.conf file to see if search mgmt.corp.local is present so that any shortname lookups for other local nodes are resolved correctly.

  • On Ubuntu systems, the configuration is located in /etc/network/interfaces as:
    dns-search mgmt.corp.local
  • On RHEL and OEL systems, the configuration may be located in /etc/sysconfig/network-scripts/ifcfg-eth0 as:
    DOMAIN=mgmt.corp.local
    More recent RHEL, and OEL systems may use a different slot-based naming scheme. The exact name may vary depending on your specific environment.

Be consistent with your naming conventions and use either all fully qualified domain names (FQDN) such has host.example.com or all short names such as host. Do not use FQDNs in some places and short names in others.

Configure local DNS using the /etc/hosts file

Verify that the /etc/hosts file identifies each node in your Splunk UBA cluster using the following format:

<IP address> <FQDN> <short name> <alias>

For example:

192.168.10.1    spluba01.mgmt.corp.local    spluba01    ubanode01
192.168.10.2    spluba02.mgmt.corp.local    spluba02    ubanode02
192.168.10.3    spluba03.mgmt.corp.local    spluba03    ubanode03
192.168.10.4    spluba04.mgmt.corp.local    spluba04    ubanode04
192.168.10.5    spluba05.mgmt.corp.local    spluba05    ubanode05

In this example, host spluba01 has an IP address of 192.168.10.1 and its FQDN is spluba01.mgmt.corp.local. Anything after the first three field is considered an alias, and is optional. In this example, we use ubanode1 is used to identify node number 1, ubanode2 is used to identify node number 2, and so on.

If your environment contains both internal and external IP addresses, be sure to use the internal IP address when configuring Splunk UBA. You can use the ip route command to help you determine this.

Formatting your /etc/hosts file this way in conjunction with using files before DNS in /etc/nsswitch.conf means that both short names and FQDNs can be obtained without any DNS lookups.

If you choose to not include the FQDN in the /etc/hosts file, you must add the domain name into the /etc/resolv.conf file in order for DNS to work properly in your environment.

Verifying your name lookup and DNS settings

Test your name lookup and DNS settings to make sure you get the expected output.

  • Use various hostname commands and verify the expected output. For example, from the spluba01.mgmt.corp.local node:
    $ hostname
    spluba01
    $ hostname -s
    spluba01
    $ hostname --fqdn
    spluba01.mgmt.corp.local
    
  • Use the ping <short name> command from each Splunk UBA node to all other Splunk UBA nodes and verify that all nodes can be reached.
  • Use the ping <FQDN> command from each Splunk UBA node to all other Splunk UBA nodes and verify that all nodes can be reached.

Supported web browsers

Open Splunk UBA in the latest versions of any of the following web browsers. Splunk UBA does not support other web browsers, such as Internet Explorer.

  • Mozilla Firefox
  • Google Chrome
  • Apple Safari

Supported single sign-on identity providers

Splunk UBA supports single sign-on integration with the following identity providers:

  • Ping Identity
  • Okta
  • Microsoft ADFS
  • OneLogin

See Configure authentication using single sign-on in Administer Splunk User Behavior Analytics.

Requirements for connecting to and getting data from the Splunk platform

To send data from Splunk platform to Splunk UBA, you must have specific Splunk platform versions and a properly configured user account. See Splunk UBA product compatibility matrix in the Plan and scale your Splunk UBA Deployment manual.

Red Hat Enterprise Linux 8.x cryptographic policies

When installed on RHEL 8.x operating systems, Splunk UBA uses a 2048 bit RSA encryption key. The Splunk platform that communicates with Splunk UBA must also use a 2048 bit encryption key. If the Splunk platform uses a 1024 bit encryption key, you will see the following error in the job executor logs:

Caused by: java.security.cert.CertPathValidatorException: Algorithm constraints check failed on keysize limits. RSA 1024bit key used with certificate

To resolve this issue, do any one of the following:

  • Use a stronger certificate of at least 2048 bits in length on the Splunk platform.
  • Set the java.security.disableSystemPropertiesFile property to true in the $JAVA_HOME/jre/lib/security/java.security file in Splunk UBA.
    java.security.disableSystemPropertiesFile property = true
  • Perform the following steps to reduce the strength of the RHEL 8 cryptographic policies in Splunk UBA.
    1. Run the following command on the Splunk UBA management node:
      update-crypto-policies --set LEGACY
    2. Run the following commands on the Splunk UBA management node to stop and restart Splunk UBA:
      /opt/caspida/bin/Caspida stop-all
      /opt/caspida/bin/Caspida start-all
      

Requirements for the Splunk Enterprise user account

If you create a custom role, and the user in that role handles information related to data sources, make sure that custom role has permissions to edit_roles or admin_all_objects. These permissions allow the custom role user to drill down into event queries.

You can add capabilities to a role in Splunk Web. See Add and edit roles with Splunk Web in Securing Splunk Enterprise.

Verify that you have a Splunk Enterprise user account with the following capabilities and configurations:

  1. Capabilities to perform real-time search, perform REST API calls, and access to the data. The admin role in Splunk Enterprise has the required capabilities by default. If you use a different role, you need the rt_search, edit_forwarders, list_forwarders, and edit_uba_settings capabilities.
  2. Configure the search job limits for the Splunk Enterprise user account and role so that they are twice the number of maximum allowed data sources for your deployment.
    Size of cluster Max number of data sources User-level concurrent search job limit User-level concurrent real-time search job limit Role-level concurrent search job limit Role-level concurrent real-time search job limit
    1 Node 6 12 12 12 12
    3 Nodes 10 20 20 20 20
    5 Nodes 12 24 24 24 24
    7 Nodes 24 48 48 48 48
    10 Nodes 32 64 64 64 64
    20 Nodes - Classic 64 128 128 128 128
    20 Nodes - XL 64 128 128 128 128
  3. Configure the Splunk Enterprise user account to have sufficient disk usage quota. For example, 40GB. .

Send data to and receive data from Splunk Enterprise Security

To send and receive data from Splunk Enterprise Security, you must have the Splunk add-on for Splunk UBA installed and enabled on your search head with the ueba index deployed to your indexers. See Splunk UBA product compatibility matrix in the Plan and Scale your Splunk UBA Deployment manual for information about version compatibility among products.

Splunk Cloud customers must contact Splunk Support to fully integrate with Splunk UBA. The Splunk Cloud admin role cannot perform Splunk UBA setup.

Send data from Splunk Enterprise directly to Kafka in Splunk UBA

Use the Splunk UBA Kafka Ingestion App to send data from large data sets in Splunk Enterprise directly to Kafka in Splunk UBA. Sending data directly to Kafka offloads the processing task from the search heads to the indexers. See Requirements for Kafka data ingestion in the Splunk UBA Kafka Ingestion App manual.

Monitor Splunk UBA directly from Splunk Enterprise

Use the Splunk UBA Monitoring App to monitor the health of Splunk UBA and investigate Splunk UBA issues directly from Splunk Enterprise. See Splunk UBA Monitoring app requirements in the Splunk UBA Monitoring App manual.

Installing Splunk UBA in environments with no Internet access

Some environments require Splunk UBA to be installed without access to the Internet. In such cases, the functionality of Splunk UBA will be limited in the following areas:

  • Splunk UBA pages that normally show visual geographical location information about a device will show warnings that the Google Maps API cannot be reached. Perform the following tasks to disable Splunk UBA from using geographical location and displaying the warning:
    • In Splunk UBA, select Manage > Settings.
    • Select Geo Location.
    • Deselect the checkbox in the Show Geo Maps field.
    • Click OK.
  • Clicking the Learn more link on any Splunk UBA page will open a new tab with a link to quickdraw.splunk.com. This is the URL used to generate the correct help link to the Splunk UBA documentation.
Last modified on 08 November, 2024
Splunk UBA installation checklist   Check system status before and after installation

This documentation applies to the following versions of Splunk® User Behavior Analytics: 5.4.1


Was this topic useful?







You must be logged into splunk.com in order to post comments. Log in now.

Please try to keep this discussion focused on the content covered in this documentation topic. If you have a more general question about Splunk functionality or are experiencing a difficulty with Splunk, consider posting a question to Splunkbase Answers.

0 out of 1000 Characters