
Step 2: Set up your data
Once you have your virtual machine configured, you need to install the tutorial sample data,which can be downloaded here.
Note: Before you set up a provider, make sure that Hunk has the following permissions:
- Read-only access to the HDFS directory where your virtual index data resides.
- Read-write access to the HDFS directory where your Splunk instance is installed. (This is usually your
splunkMR
directory, for example:User/hue/splunk_mr/dispatch
). - Read-write access to the Datanode where your
/tmp
directory resides. This is the temp directory that thevix.splunk.home.datanode
value points to in your Provider settings.
1. SSH to your virtual machine, and move Hunkdata.json.gz
and your Splunk downloadto the HDFS user’s home directory. (If you are using the Cloudera quickstart VM, the password for root user is “cloudera”.)
For example:
scp Hunkdata.json.gz root@172.16.220.166:~ scp splunk-6.0-<version number>-Linux-x86_64.tgz root@172.16.220.166:~ ssh root@172.16.220.166 mv Hunkdata.json.gz ~hdfs (this is moves the data to the hdfs user)
2. Put the data into HDFS as the hdfs user. For example:
su - hdfs -c "hadoop fs -mkdir hdfs://localhost:8020/data" su - hdfs -c "hadoop fs -put ~/Hunkdata.json.gz hdfs://localhost:8020/data" su - hdfs -c "hadoop fs -ls hdfs://localhost:8020/data"
PREVIOUS Step 1: Set up a Hadoop Virtual Machine instance |
NEXT Step 3: Set up an HDFS directory for Hunk access |
This documentation applies to the following versions of Hunk®(Legacy): 6.0, 6.0.1, 6.0.2, 6.0.3, 6.1, 6.1.1, 6.1.2, 6.1.3, 6.2, 6.2.1, 6.2.2, 6.2.3, 6.2.4, 6.2.5, 6.2.6, 6.2.7, 6.2.8, 6.2.9, 6.2.10, 6.2.11, 6.2.12, 6.2.13, 6.3.0, 6.3.1, 6.3.2, 6.3.3, 6.3.4, 6.3.5, 6.3.6, 6.4.0, 6.4.1, 6.4.2, 6.4.3, 6.4.4, 6.4.5, 6.4.6, 6.4.7, 6.4.8, 6.4.9, 6.4.10, 6.4.11
Feedback submitted, thanks!