If you have found this tool useful, give it a star  on Github

Test Tokenizer Patterns for the Dissect filter

This app tries to parse a set of logfile samples with a given dissect tokenization pattern and return the matched fields for each log line.

Syntax compatible with Filebeat, Elasticsearch and Logstash processors/filters.

This instance is using a backend running v7.17.6 of Elastic Beats.
One sample per line
a b
One result per block

New in beats v7.9

beats v7.9 introduced a new feature that allows converting a field from a string (only data type supported until now) into one of integer, long, float, double, boolean, or ip. This type conversion step was usually delegated to a later step in your processing. To take advantage of this new feature you need to provide a valid type after the key using | as a separator. For example:

            - dissect:
            tokenizer: "%{key1} %{key2|integer}"
            field: "message"
            target_prefix: "dissect"

In this case key2 will be casted to an integer value. More information can be found in the official documentation.