2 hours 29 minutes
in this video, we'll discuss the parts of Splunk.
First, we'll talk about the data pipeline,
then going to spawn components
and have a little bit of a discussion on distributed versus non distributed versus clustered Environments.
The data pipeline US Blanc See's. It is made up of input, parsing, indexing and searching
the input. Part of that is just what you might think. *** is getting data
at this stage. There's also metadata at it, like source host and source type.
But the main focus is getting input. Those data coming in
each of these stages corresponds to different actual splint components. We'll talk about each of these, but input goes along with four orders universal or heavy foreigners,
and it can be done at the index for level
at the next stage, we have parsing
data is getting turned into events. At this stage,
this could be line breaks happening or data being transformed based on certain rules.
This can occur on an index or or heavy foreigner.
Indexing is taking the parsed events and putting them into an index for later use.
When you get to the searching stage, there's some interaction between the search head and indexers. The search had as responsible for search management.
This is where you would go to run a search and that search request get sent to an indexer, and then the results get sent back to the search head for you to view and work with.
At the searching level, you can have scheduled searches, alerts and dashboards
along the side. Here as a common set up, you might see
I have you f for universal foreigner.
You can think of a universal foreigner as something like an agent
may be installed on a server and set up to collect Windows Event logs. It's getting them put taking those windows event logs and sending them on to the indexer
on the index. Sir, you have parsing and
indexing is taking the data,
breaking it into events and organizing it in a place the search head can easily friend. Request for
search heads are what users typically interact with.
They perform search management.
You can, for example,
go into this box
and run a basic search, which the search head then distributes other requests to the different indexers and then displays the results from
you can do things like have
custom dashboards, alerts and reports
indexers received in next in store data.
They can provide the search head with needed information.
There's a bit of complexity around the term index, so I wanted to break down the different definitions.
Index as a noun is a data repositories.
By breaking up data into different indexes, you can improve performance, apply different data retention policies and limit access to different sets of data.
For example, if you're collecting firewall logs, you may have an index title firewall logs where logs from multiple firewalls get stored.
This could make it easier to limit your searches to the type of information you're looking for.
And if you have another team, say, helped us
that these two look at authentication logs, but maybe not Web traffic. You could easily limit them from viewing this data.
If you need to keep far wall logs for a set amount of time for audits,
you could specify this on the date of retention policies.
Index as a verb, is the process.
The theme of raw data, as in taking the data and handling and organizing it
and indexer is a particular Splunk index.
It's a particular ***. Instance that indexes data. This sentence might help you remember the different meetings
and indexer indexes data and puts it in an index.
Foreigners like I mentioned,
you can kind of think of them like agents. You installing a host, they send data onward.
There are several different types of foreigners. A light for it, er is deprecate ID meeting. There are newer versions of it, but it does exist.
Your first soul. Foreigners are typically what you want to install. One possible. They have a pretty light footprint and mostly just worked to send data onward. You can do some filter. And with universal four order such a cz
bye blacklisting certain event types. But if you want to do any more complex filtering, you're probably going to need to set up a heavy foreigner.
There are also different server rolls. We're going to talk about them too much in this course. But I want you to know they exist.
For example, things like a deployment server can help you manage foreigners and send APS by groups.
Distributed environments are basically once where different components of Splunk are broken out.
The set up in this course will be a simple non distributed environment or search head indexer and license master are all combined
for larger companies. Or, if you're handling a lot of data, you'll probably need to separate these pieces out.
This is sometimes thought of as horizontal scaling as you grow. You can add different parts to scale the environment
with the idea of different deployment scales. If you have a very small office working with less than 20 gigs a day with fear than 100 foreigners, you could probably get away with a non distributed environment like we're doing for this course.
For a larger company, you're probably going to need a distributed environment.
Clustering is a more advanced topic, but she should know what it is. At a basic level, it replicates data between different components to create redundancy so that there is duplicate data across multiple instances.
This is good to look at if you can't have any downtime in your environment or if you're worried about disaster recovery or the potential of losing data.
Question time. Universal Florida. It deals with the blank part of the data pipeline.
The answer is input.
A universal four. ITER helps to bring data into the *** environment
as a review for it. Er send data
indexers, turn data into events and place them in indexes. Search head, send search requests and display data.
A larger company will likely need a distributed environment, but for this course we will set up a simple non distributed environment.
Clustering also won't be covered in this course as it is a more advanced topic.
But you should know that it provides redundancy and is a good option for high availability and disaster recovery options.
There. Next to video, we will install ***.
Splunk online test helps to assess knowledge of Splunk which is an advanced, scalable and ...
Event Log Collection
In this lab you will use Splunk Enterprise to ingest logs from a local host ...