-
Notifications
You must be signed in to change notification settings - Fork 306
Description
Hello,
Currently the Elasticsearch output in Logstash does not allow the creation of a custom data stream type, if you use the data_stream_*
settings of the output it will validate the data_stream_type
and it will only allow the following values:
logs
metrics
synthetics
traces
All of those types are also used by Elastic Agent and have system managed templates and lifecycle policies, so to use data stream now in logstash you would need to create some template for the type you want but make sure that this template will not override the system templates, this makes things more complex and there is always the risk of human error that would override the templates used by Elastic Agent and break things.
To be able to use custom data streams in logstash you need a trick on the output like the example below:
output {
elasticsearch {
hosts => ["HOSTS"]
index => "data-stream-name"
action => "create"
http_compression => true
data_stream => false
manage_template => false
ilm_enabled => false
cacert => 'ca.crt'
user => 'USER'
password => 'PASSWORD'
}
}
While this works, Logstash should allow the creation of data streams of custom types, which is not possible now.
Activity
yaauie commentedon Oct 12, 2023
Thoughts on implementation options:
\A(?!\.{1,2}$)[[:lower:][:digit:]][[:lower:][:digit:]\._+]{0,252}\Z
the successfully rejects known-invalid index prefixes while letting likely-valid ones through (limitation: composed index name length cannot be validated solely from a single component).#initialize
or#register
if and only ifdata_stream
is effectivelytrue
robbavey commentedon Mar 13, 2025
@jsvd Is this an actual issue that we should prioritize/work on, or is there a workaround that we can use?
leandrojmp commentedon Mar 27, 2025
Hello @robbavey, just a feedback as a user.
Currently Logstash is only capable of create data streams that follows the elastic naming scheme,
<type>-<dataset>-<namespace>
, so withdata_stream
astrue
, the user is limited to create data stream using one of the available types and also needs to provide a dataset and a namespace.In my case my data streams follows another naming pattern, I do not use
type
ornamespace
, some data streams have a prefix in the name and others are just the dataset name.So the configuration I shared works because logstash just send the request and the index is created as data stream because it is defined in the index template.
Being honest, just having these steps in the documentation could be enough.