Logstash JSON Grok 过滤器问题
Logstash JSON Grok filter issue
我设置了 squid 代理以通过 Logstash 将 JSON 格式的日志发送到 Elastic。我正在尝试使用 GROK 过滤来解析日志。该过滤器在 Kiabana Grok 调试器中工作,但在我重新启动 Logstash 时出现以下错误
Failed to execute action {:action=>LogStash::PipelineAction::Create/pipeline_id:squid_logs,
:exception=>"LogStash::ConfigurationError", :message=>"Expected one of [ \t\r\n], \"#\", \"
{\", \",\", \"]\" at line 10, column 62 (byte 137) after filter {\n grok {\n match => {\n
\"message\" => [ \"%{IPV4:vendor_ip}\", \"%{WORD:message}\"", :backtrace=>["/usr/share/logstash/logstash-core/lib/logstash/compiler.rb:32:in `compile_imperative'",
"org/logstash/execution/AbstractPipelineExt.java:184:in `initialize'",
"org/logstash/execution/JavaBasePipelineExt.java:69:in `initialize'",
"/usr/share/logstash/logstash-core/lib/logstash/java_pipeline.rb:47:in `initialize'",
"/usr/share/logstash/logstash-core/lib/logstash/pipeline_action/create.rb:52:in `execute'",
"/usr/share/logstash/logstash-core/lib/logstash/agent.rb:389:in `block in converge_state'"]}
我有以下 GROK 过滤器
"%{IPV4:vendor_ip}", "%{WORD:message}": "%{IPV4:clientip}", "%{WORD:message}": "%
{DATA:timestamp}", "%{WORD:message}": "%{WORD:verb}", "%{WORD:message}": "%{DATA:request}", "%
{WORD:message}": "%{URIPATHPARAM:path}"
在 Kibana Grok 调试器中,过滤器可以正常处理如下消息:
{ "vendor_ip": "x.x.x.x", "clientip": "x.x.x.x", "timestamp": "2021-04-09T13:58:38+0000",
"verb": "GET", "request": "https://domain", "path": "/somepath", "httpversion": "HTTP/1.1",
"response": 200, "bytes": 2518042, "referer": "-", "useragent": "Microsoft BITS/7.8",
"request_status": "HIER_DIRECT", "hierarchy_status": "HIER_DIRECT" }
Logstash 配置如下:
input {
beats {
port => 5045
}
}
filter {
grok {
match => {
"message" => [ "%{IPV4:vendor_ip}", "%{WORD:message}": "%{IPV4:clientip}", "%{WORD:message}": "%{DATA:timestamp}", "%{WORD:message}": "%{WORD:verb}", "%{WORD:message}": "%{DATA:request}", "%{WORD:message}": "%{URIPATHPARAM:path}" ]
}
}
}
output {
elasticsearch {
hosts => ["x.x.x.x:9200"]
index => "squid_logs"
}
}
使用 grok 过滤器解析 json 消息是错误的方法,没有必要这样做,因为您需要转义消息中的所有双引号,因此需要做很多工作否则您将遇到配置错误,这是您的情况。
使用 json 过滤器解析 json 条消息
只需在您的管道中使用它:
filter {
json {
source => "message"
}
}
我设置了 squid 代理以通过 Logstash 将 JSON 格式的日志发送到 Elastic。我正在尝试使用 GROK 过滤来解析日志。该过滤器在 Kiabana Grok 调试器中工作,但在我重新启动 Logstash 时出现以下错误
Failed to execute action {:action=>LogStash::PipelineAction::Create/pipeline_id:squid_logs,
:exception=>"LogStash::ConfigurationError", :message=>"Expected one of [ \t\r\n], \"#\", \"
{\", \",\", \"]\" at line 10, column 62 (byte 137) after filter {\n grok {\n match => {\n
\"message\" => [ \"%{IPV4:vendor_ip}\", \"%{WORD:message}\"", :backtrace=>["/usr/share/logstash/logstash-core/lib/logstash/compiler.rb:32:in `compile_imperative'",
"org/logstash/execution/AbstractPipelineExt.java:184:in `initialize'",
"org/logstash/execution/JavaBasePipelineExt.java:69:in `initialize'",
"/usr/share/logstash/logstash-core/lib/logstash/java_pipeline.rb:47:in `initialize'",
"/usr/share/logstash/logstash-core/lib/logstash/pipeline_action/create.rb:52:in `execute'",
"/usr/share/logstash/logstash-core/lib/logstash/agent.rb:389:in `block in converge_state'"]}
我有以下 GROK 过滤器
"%{IPV4:vendor_ip}", "%{WORD:message}": "%{IPV4:clientip}", "%{WORD:message}": "%
{DATA:timestamp}", "%{WORD:message}": "%{WORD:verb}", "%{WORD:message}": "%{DATA:request}", "%
{WORD:message}": "%{URIPATHPARAM:path}"
在 Kibana Grok 调试器中,过滤器可以正常处理如下消息:
{ "vendor_ip": "x.x.x.x", "clientip": "x.x.x.x", "timestamp": "2021-04-09T13:58:38+0000",
"verb": "GET", "request": "https://domain", "path": "/somepath", "httpversion": "HTTP/1.1",
"response": 200, "bytes": 2518042, "referer": "-", "useragent": "Microsoft BITS/7.8",
"request_status": "HIER_DIRECT", "hierarchy_status": "HIER_DIRECT" }
Logstash 配置如下:
input {
beats {
port => 5045
}
}
filter {
grok {
match => {
"message" => [ "%{IPV4:vendor_ip}", "%{WORD:message}": "%{IPV4:clientip}", "%{WORD:message}": "%{DATA:timestamp}", "%{WORD:message}": "%{WORD:verb}", "%{WORD:message}": "%{DATA:request}", "%{WORD:message}": "%{URIPATHPARAM:path}" ]
}
}
}
output {
elasticsearch {
hosts => ["x.x.x.x:9200"]
index => "squid_logs"
}
}
使用 grok 过滤器解析 json 消息是错误的方法,没有必要这样做,因为您需要转义消息中的所有双引号,因此需要做很多工作否则您将遇到配置错误,这是您的情况。
使用 json 过滤器解析 json 条消息
只需在您的管道中使用它:
filter {
json {
source => "message"
}
}