Previewing Collection Data
Preview Jobs can be used to see the shape of the data being collected, just like a normal capture. Let's run a preview job, and take a look at the data.
-
If you're not already here: Click
Cribl
on the top row of tabs, then withManage
active in Stream's top nav, selectData
and clickSources
. -
Click into
S3
under theCollectors
section. -
Click the
Run
button next to thepan-logs
collector you configured earlier, and theRun configuration
modal will appear. -
Accept the default values (which include
Preview
mode), and clickRun
. ThePreview
modal should appear and start capturing data. After either the time expires, or the defined number of events are captured, you should see something similar to this:noteNow is a great time to take a closer look at the data that we just captured.
- At the
Previewing
modal's upper right, click on the...
, and enable theShow Internal Fields
toggle. - You should now see a number of internal fields with names that start with
__
. A few that are relevant to collected data:
__channel Contains a unique identifier for one of the tasks' data streams. The Event Breaker engine uses this to ensure that streams are not mixed. __collectible This shows you metadata about the collection job itself. __inputId If data comes from a collection, the value of this will be collection:<collection job ID>
.__isBroken Indicates whether this record has been "broken" by an Event Breaker ruleset (e.g., the Cribl ruleset we used to configure the collector). Of course, any of these fields' values are available to you for Route/Pipeline/Function filters, etc.
- At the
-
We want to save this previewed data for later use, so click
Save as Sample File
and in theFile Name
field, entercollected-events.log
then click Save.
When you are done exploring the data, click the X
at the top right corner of the preview modal to dismiss it.
Next, let's run a discovery job...