Reputation: 823
I am new to Logstash filter and going through different blogs and links to understand in detail. I have few questions which are still unanswered.
. If my log file has different log pattern e.g.
2017-01-30 14:30:58 INFO ThreadName:33 - {"t":1485786658088,"h":"abcd1234", "l":"INFO", "cN":"org.logstash.demo", "mN":"getNextvalue", "m":"fetching next value"} 2017-01-30 14:30:58 INFO AnotherThread:33 -my log pattern is different
I have below filter which is successfully filtering line 1 of the log
grok
{
match => [ "message", "%{TIMESTAMP_ISO8601:LogDate} %{LOGLEVEL:loglevel} %{WORD:threadName}:%{NUMBER:ThreadID} - %{GREEDYDATA:Line}" ]
}
json
{
source => "Line"
}
Upvotes: 0
Views: 242
Reputation: 16362
As @darth_vader points out, you'll get a "grok_parsefailure" tag on each document that doesn't match your pattern(s) in a grok{} filter. However, how you handle this failure is up to you.
By default, all the events will fall through to your output{} section, which presumably would send them to elasticsearch. You could also have a conditional output{} section, which sent parsed logs to one output and unparsed logs to another (a file{} output, or a different index, or...).
As for examples, the official doc tends to include incomplete fragments (at best), so you're probably going to find better examples in random internet blogs.
Upvotes: 0
Reputation: 9454
To answer your questions:
The lines which cannot be filtered using grok
would end up in a
grok_parsefailure
. Make sure you handle it by dropping the lines
which don't actually match the filter criteria.
As far as I know you can't capture them separately and push it to ES. Maybe for this, you can have multiple grok patterns so that you can filter it out and send it to different ES indices thereafter.
I've added the links in the comment above.
This SO could come in handy. Hope it helps!
Upvotes: 0