Sponsored by Spartan conseil
# What is Feedo ?
Feedo is an ETL, for Extract, Transform and Load. Basically, it gets data from files or database, process it thanks to pipelines and store data to a file or a database. It is very versatile and processing brick can be added without pain.
The purpose of Feedo is generic :
- ETL to convert database to another one
- Alerting like elastalert
- Gather information from agent like Fluentbit
- SIEM with correlation rule
- Intrusion detection thanks AI
- ...
The Feedo's design is for Security Operational Center (SOC). But if you need to play with data, you need Feedo as friend :)
They are many reasons why I decided to build Feedo.
Firstly, I work with RethinkDB as main database. It is amazingly easy to use, with enough performance for my needs. But the main drawback is about community tools. Briefly, they are no connector to work with, especially with Fluentd, Fluentbit or a clone of Elastalert.
Here we are: the second point ! I really appreciate Fluent family, especially Fluentbit fully written in C. Nevertheless, a drawback arrive when we talk about plugins or modifications. I worked many years with Fluentd and it can become painful when you need something was not shipped with.
So a sort of Python version a Fluent with rules and easy extension seems to me a good idea !
The installation is very easy :
pip3 install feedo
You just install Feedo and you want to test ? Let's do a basic example !
Create a file at /etc/feedo/default.yaml
and copy-paste that :
pipelines:
"pipeline#1":
- name : input_dummy
tag : "my_pypeline"
data : {"log":"my log"}
- name : output_stdout
match : "*"
Now execute feedo :
you@computer:>feedo
my_pypeline[1607608082]: {'log': 'my log'}
It works ! 🎉 You ran your first pipeline.
Keep in mind previous example, I will reuse it now.
The heart of the processing is based on pipeline. It is similar to pipe operator in Unix system : every action do a basic operation and forward data to the following action :
you@computer:>cat /var/log/auth.log | head | grep "sudo"
Feedo do processing like this but add a tag to data. This way following action can decide to process the data (if it match) or just forward it to the next action. Tag is added by the data producer ("my_pipeline" in input_dummy) and other action will try to match (" * " in output_stdout). In the Feedo context, we call data Event. Indeed diffent : Event contains data, called record (dict), an unix timestamp and the tag.
Actions are categorized in four cases :
- input
- output
- filter
- parser
Input produces events in the pipeline, including tag. If an input receives a event it forwards it to next action.
It is useful for testing purpose and forward events base on dicts.
Parameters :
tag
: events' tagdata
: a dict or a list of dict with fact
Example:
- name : input_dummy
tag : "my_pypeline"
data : {"log":"my log"}
It watch a path and load file if :
- The file exists on the startup
- The file was create (written and closed)
Parameters :
tag
: events' tagwatch_path
: the path watched, typically a directorypath_pattern
: provide a pattern which must match when a file is found with watch_pathremove=False
: remove the file once read
Example:
- name : input_file
tag : "logs"
watch_path : /var/log
path_pattern : /var/log/stuff.*.log
Based on Fluentbit-server-py, it allows to received event from Fluentbit agent using the forward protocol. It support authentication using shared key and TLS. No tag parameter is available since it is provided by the agent.
Parameters :
host="localhost"
: Used to bind socket serverport=24224
: Port used to bind the servertls_enable=False
: Used to enable TLS supportkey_file=None
: if TLS enabled, path to the key filecrt_file=None
: if TLS enabled, path to the certificate fileshared_key=None
: defined a shared key between servers an nodes. If set, enable authenticationserver_hostname=""
: define the hostname, useful for shared_key authentication
Warning : if you exposed the port to internet, use authentication and TLS. If you can, add firewall rule to decrease surface attack. BE CAREFUL !
Example:
- name : input_forward
port: 24224
host : 0.0.0.0
tls_enable : true
key_file : /etc/tls/foo.key
crt_file : /etc/tls/foo.crt
shared_key : my_pr1vate_sh4red_K3y
server_hostname : foo.com
It exports events out of the pipeline. It can be file, database, etc.
It stores in buffer event and it writes buffer in file.
Parameters :
match
: pattern to match tagtime_key
: used to extract time to interpolate path_templatepath_template
: Template of the file pathbuffer_size=1000
: Number of event stored before flushtimeout_flush=60
: Flush buffer after timeout, in second
Example :
- name : output_archive
time_key : timestamp
path_template : "/archives/{source}/log-%Y%m%d.json"
match : mylog
Notes :
- if the example received the event contains record
{"timestamp":1607618046, "source":foo, "data":"test"}
, the path will be/archives/foo/log-20201210.json
Store events in RethinkDB as time serie.
Parameters
match
: pattern to match tagtime_key
: used to extract time to interpolatetable_template
table_template
: Template of the table namebuffer_size=1000
: Number of event stored before flushdatabase="test"
: database nameip="localhost"
: f set, change the database destination ipport=None
: if set, change the database destination portwait_connection=30
: used to wait the database warmuptimeout_flush=60
: Flush buffer after timeout, in second
Example:
- name : output_rethinkdb
time_key : timestamp
table_template : "log-%Y%m%d"
database : test
ip : rethink.com
port : 28015
match : my_log
Display event in stdout.
Parameters
match
: pattern to match tag
Example:
- name : output_stdout
match : my_log
A parser take an event et parse one field with a specific format : regex
, json
, etc.
Read a field and add new fields to existing event.
Parameters :
match
: pattern to match tagkey
: Key to be parsedmode="merge"
: A string that can be "merge", "tree" or "add"
Example of modes :
merge
: {"key":"Z", "value":'{"aaa": "bb"}'} -> {"key":"Z", "aaa":"bb"}add
: {"key":"Z", "value":'{"aaa": "bb"}'} -> {"key":"Z", "value":'{"aaa": "bb"}', "aaa":"bb"}`tree
: {"key":"Z", "value":'{"aaa": "bb"}'} -> {"key":"Z", "value":{"aaa":"bb"}}
Example :
- name : parser_json
match : my_log
key : json_log
mode : add
Read a field and add new fields to existing event.
Parameters :
match
: pattern to match tagkey
: Key to be parsedregex
: define the behaviour. Use name group to create fieldmode="merge"
: A string that can be "merge", "tree" or "add"
Example of modes :
merge
: {"key":"Z", "value":'{"aaa": "bb"}'} -> {"key":"Z", "aaa":"bb"}`add
: {"key":"Z", "value":'{"aaa": "bb"}'} -> {"key":"Z", "value":'{"aaa": "bb"}', "aaa":"bb"}tree
: {"key":"Z", "value":'{"aaa": "bb"}'} -> {"key":"Z", "value":{"aaa":"bb"}}
Example :
- name : parser_regex
match : my_log
key : line
mode : merge
regex : ".+?(?P<name>\\{.+\\})"
Process event, may create or delete events.
This action will monitor a certain field and match if that field changes. The field must change with respect to the last event with the same query_key.
Parameters :
- match : pattern to match tag
- tag : tag used to generate new event on change
- alert : dict used to generate new event on change
- compare_key : key monitored to find change
- query_key : key used to group type of event
- ignore_null=True : ignore if compare_key is missing. If ignore_null if false, missing compare_key is a valid state
- db_path=None : file path to store internal state. None means only RAM is used.
Example :
- name : filter_change
match : my_log
tag : my_alert
alert :
title : The hostname change of status
priority : 2
compare_key : status
query_key : hostname
if events are :
- {"hostname":"foo.bar", "status":"on"}
- {"hostname":"foo.bar", "status":"off"}
Then a new event will be created on second event
This action performs a time parsing and allows to change the time format. Under the hook, it use Chronyk library so feel free refere about time format.
Parameters :
- match : pattern to match tag
- key : define the key to be parsed
- format=None : define the output format of key's value. None means unix timestamps
Example :
- name : filter_date
match : date
Event likes {"date":"Fri, 11 Dec 2020 08:30:13 +0000"} become {"date":1607675413}
It takes a date and convert it to timestamp
This action remove one or more key in event.
Parameters :
- match : pattern to match tag
- keys : on string or a list of string to describe keys to be removed.
Example :
- name : filter_remove_keys
match : date
keys :
- A
- B
Event likes {"A":1, "B":2, "C":3} become {"C":3}
### filter_retag
This action change the event's tag with a value in event or with a constant value.
Parameters :
- match : pattern to match tag
- value : New tag if key doesn't exist or if key=None
- key=None: event value used to retag event. Use value parameter if missing
Example :
- name : filter_retag
match : my_log
value : generic_log
key : source
If event looks like {"source":"auth", "data":"xxx"}, the new tag will be "auth". If event looks like {"data":"xxx"}, the new tag will be "generic_log".
Thanks to flaticon.com !