Skip to main content

Previewing Collection Data

Preview Jobs can be used to see the shape of the data being collected, just like a normal capture. Let's run a preview job, and take a look at the data.

important
  1. If you're not already here: Click Cribl on the top row of tabs, then with Manage active in Stream's top nav, select Data and click Sources.

  2. Click into S3 under the Collectors section.

  3. Click the Run button next to the pan-logs collector you configured earlier, and the Run configuration modal will appear.

  4. Accept the default values (which include Preview mode), and click Run. The Preview modal should appear and start capturing data. After either the time expires, or the defined number of events are captured, you should see something similar to this: Preview Results

    note

    Now is a great time to take a closer look at the data that we just captured.

    1. At the Previewing modal's upper right, click on the ..., and enable the Show Internal Fields toggle.
    2. You should now see a number of internal fields with names that start with __. A few that are relevant to collected data:
    __channelContains a unique identifier for one of the tasks' data streams. The Event Breaker engine uses this to ensure that streams are not mixed.
    __collectibleThis shows you metadata about the collection job itself.
    __inputIdIf data comes from a collection, the value of this will be collection:<collection job ID>.
    __isBrokenIndicates whether this record has been "broken" by an Event Breaker ruleset (e.g., the Cribl ruleset we used to configure the collector).

    Of course, any of these fields' values are available to you for Route/Pipeline/Function filters, etc.

  5. We want to save this previewed data for later use, so click Save as Sample File and in the File Name field, enter collected-events.log then click Save.

When you are done exploring the data, click the X at the top right corner of the preview modal to dismiss it.

Next, let's run a discovery job...