Getting Data In

What are other options for getting data into Splunk without a forwarder?

wegscd
Contributor

I have some computing antiques running Unix; I need to monitor some files on them, and get them into Splunk.

I read http://answers.splunk.com/answers/8328/best-practice-for-getting-data-into-splunk-without-a-forwarde... the "scripted/scheduled-copy-files-to-a-machine-that-does-have-a-forwarder" seems reasonable/doable, and is probably where I'll end up unless I find something better. BUT....

One thing I've been mentally toying with is running a (perl?) script to tail the files and ship them via TCP to an indexer listening on a port dedicated to the purpose.

I have a hard time believing that we are the first people going down this road; has anyone else done this?

has anyone cooked up any solutions other than the ones in http://answers.splunk.com/answers/8328/best-practice-for-getting-data-into-splunk-without-a-forwarde... ?

Tags (3)
0 Karma

martin_mueller
SplunkTrust
SplunkTrust

Ah. What kind of antique are we talking about?

0 Karma

wegscd
Contributor

very old AIX....

0 Karma

musskopf
Builder

Antique Unix Admin

musskopf
Builder

Don't forget that Splunk can be used as a Syslog server: http://docs.splunk.com/Documentation/Splunk/6.2.0/Data/SyslogTCP
(if your Splunk is not running as Root, just use iptables to redirect the TCP/514 and UDP/514 portd to the Splunk listening port)

Anyway, if you dislike the option above, do you have netcat available on the servers? Even if not, you might be able to easily compile and drop the binary on the servers: http://docs.splunk.com/Documentation/Storm/Storm/User/Howtoforwarddatavianetcat
You could simply use with tail as the example above or create a smarter script that runs every X seconds and sends only the deltas.

Well, I still would prefer to spend the time deploying a proper syslog deamon and use it, nothing complicated. You might find out other things you could be sending to that syslog server in the future to justify the effort.

0 Karma

wegscd
Contributor

the receiving end isn't the issue. the sending end is the issue.

netcat or socat can take care of the transport, now the issue is just having a tail-like utility that can persist it's view of where it left off (so we don't double index after a restart)...

0 Karma

cpetterborg
SplunkTrust
SplunkTrust

The biggest reason I can see for deploying a syslog server that will forward to Splunk is that you will drop lots of syslog packets when you restart a splunk indexer because it takes so much time to restart one, and syslog usually is UDP, so there is no re-try if the send failed because the index server was down.

The syslog server running with a forwarder is a good option, if you don't want to use a forwarder. But a forwarder is always better. If your network goes down, a syslog server (or splunk relying on syslog) never gets the data, while the forwarders will always start from the point of disconnect to the present when the network comes back up. No lost data.

You can use the REST API to send data to Splunk, but it isn't much different from using a forwarder. We have Cloud Foundry servers that use the REST API because they have more control over the data sent.

0 Karma

martin_mueller
SplunkTrust
SplunkTrust

Do your antiques speak syslog or a related network-based protocol?

0 Karma

wegscd
Contributor

the antiques do speak syslog, but do not have a syslog daemon available that allows one to send files via the syslog protocol.

0 Karma
Get Updates on the Splunk Community!

Routing logs with Splunk OTel Collector for Kubernetes

The Splunk Distribution of the OpenTelemetry (OTel) Collector is a product that provides a way to ingest ...

Welcome to the Splunk Community!

(view in My Videos) We're so glad you're here! The Splunk Community is place to connect, learn, give back, and ...

Tech Talk | Elevating Digital Service Excellence: The Synergy of Splunk RUM & APM

Elevating Digital Service Excellence: The Synergy of Real User Monitoring and Application Performance ...