sentinel-kit

πŸ“Š Data Ingestion Setup

This guide covers configuring log sources, setting up FluentBit collection, and monitoring data ingestion in Sentinel Kit.


Understanding Data Flow

Sentinel Kit uses FluentBit as the primary log collection engine, feeding data into Elasticsearch for storage and analysis:

Log Sources β†’ FluentBit β†’ Elasticsearch β†’ Sentinel-Kit frontend and Kibana

Supported Data Sources


πŸ’» Log Ingestion into Sentinel-Kit

This document outlines the various methods available for ingesting logs into the Sentinel-Kit Elastic Stack.


πŸš€ Ingestion Methods

Logs can be funneled into the Elastic Stack using one of the following methods:

  1. Direct Indexing: Placing pre-extracted log files directly into a specific directory.
  2. SFTP Transfer: Routing logs via the built-in SFTP service.
  3. HTTP Forwarding: Sending logs using a forwarder (like Logstash, Fluent Bit, etc.) to Sentinel-Kit’s dedicated HTTP ingestion service.

1. Direct Indexing

This is the fastest way to index data. To use this method, simply place the data you wish to index into the following directory: ./data/log_ingest_data

Supported Formats (Out-of-the-Box)

By default, the stack is configured for rapid indexing of the following log types:

Accessing Indexed Data

All indexed data is placed into Elasticsearch indices following the format: sentinelkit-ingest-<TYPE>-<YY>-<MM>-<DD>.

You can access and visualize this data via Kibana:


2. SFTP Transfer

Sentinel-Kit exposes an SFTP service for data transfer.

⚠️ Note on Accessibility: To make the SFTP service accessible over the internet, you must configure your network equipment (e.g., firewall rules).

Credentials

The SFTP credentials are found in your local .env file:

SFTP_USER=sentinel-kit_sftp_user
SFTP_PASSWORD=sentinel-kit_sftp_passwd

Data Handling

Once uploaded, the files will be available in:

./data/ftp_data

Important: Data placed here is NOT automatically indexed. You must manually move or copy the desired logs from this location into the ./data/log_ingest_data directory to trigger direct indexing (Method 1).


You can send logs, primarily in JSON format, using a dedicated forwarder. This requires setting up a data source via the backend console.

Step 1: Access the Backend Console

Execute the following command to enter the backend console application:

./launcher console

Step 2: Create a New Data Source

Once inside the container, run the following command to create a new ingestion endpoint:

sentinel-kit> app:datasource:create <name> <index> [<validFrom> [<validTo>]]

This command takes 4 arguments:

Example:

sentinel-kit> app:datasource:create MyIngestName temp_index 2020-01-01 2030-01-01
MyIngestName - temp_index
[OK] Datasource "MyIngestName" created successfully 
Valid from 2020-01-01
Valid to 2030-01-01
Ingest key (header X-Ingest-Key): M2VmYjRiZTMtYThmNi00ZDhlLTliZTQtMGFjYWNhZDVjY2Mw
Forwarder URL: https://backend.sentinel-kit.local/ingest/json

Step 3: Send logs

Once the source is created, logs must be sent to the Forwarder URL displayed in the console output.

βž• Additional Console Commands

Use these commands to manage your data sources:


Next Steps

With data ingestion configured:

  1. Avanced ingestion - Extend core capabilities with advanced logs ingestion
  2. Create Detection Rules - Build Sigma rules for your data
  3. Investigate Alerts - Learn alert analysis workflows
  4. Monitor Platform Health - Set up monitoring alerts

Next: Sigma Rules Management β†’ Β«Back to index