Groups events and applies aggregate functions to each group.
summarize (group|aggregation)..., [options={...}]Description
Section titled “Description”The summarize operator groups events according to certain fields and applies
aggregation functions to each group. By default,
the operator consumes the entire input before producing any output, and may reorder
the event stream.
With the options argument, you can enable periodic emission of results at fixed
intervals for real-time streaming analytics.
The order of the output fields follows the sequence of the provided arguments. Unspecified fields are dropped.
To group by a certain field, use the syntax <field> or <field>=<field>. For
each unique combination of the group fields, a single output event will be
returned.
aggregation
Section titled “aggregation”The aggregation functions applied to each group
are specified with f(…) or <field>=f(…), where f is the name of an
aggregation function (see below) and <field> is an optional name for the
result. The aggregation function will produce a single result for each group.
If no name is specified, the aggregation function call will automatically
generate one. If processing continues after summarize, we strongly recommend
to specify a custom name.
options
Section titled “options”The optional options argument accepts a record with the following fields:
frequency: duration
Section titled “frequency: duration”Emit aggregation results at the specified interval instead of waiting for the input to end. When set, results are emitted periodically and also when the input stream ends.
mode: string
Section titled “mode: string”Controls how aggregations behave between emissions. Only applies when
frequency is set. Possible values:
"reset"(default): Clear aggregations after each emission, providing per-interval metrics."cumulative": Keep accumulating values across emissions, providing running totals."update": Only emit groups whose values changed since the last emission, reducing output noise in monitoring scenarios.
Examples
Section titled “Examples”Compute the sum of a field over all events
Section titled “Compute the sum of a field over all events”from {x: 1}, {x: 2}summarize x=sum(x){x: 3}Group over y and compute the sum of x for each group:
from {x: 0, y: 0, z: 1}, {x: 1, y: 1, z: 2}, {x: 1, y: 1, z: 3}summarize y, x=sum(x){y: 0, x: 0}{y: 1, x: 2}Gather unique values in a list
Section titled “Gather unique values in a list”Group the input by src_ip and aggregate all unique dest_port values into a
list:
summarize src_ip, distinct(dest_port)Same as above, but produce a count of the unique number of values instead of a list:
summarize src_ip, count_distinct(dest_port)Compute min and max of a group
Section titled “Compute min and max of a group”Compute minimum and maximum of the timestamp field per src_ip group:
summarize min(timestamp), max(timestamp), src_ipCompute minimum and maximum of the timestamp field over all events:
summarize min(timestamp), max(timestamp)Check if any value of a group is true
Section titled “Check if any value of a group is true”Create a boolean flag originator that is true if any value in the src_ip
group is true:
summarize src_ip, originator=any(is_orig)Create 1-hour time buckets
Section titled “Create 1-hour time buckets”Create 1-hour groups and produce a summary of network traffic between host pairs:
ts = round(ts, 1h)summarize ts, src_ip, dest_ip, sum(bytes_in), sum(bytes_out)Emit aggregations every 5 seconds
Section titled “Emit aggregations every 5 seconds”Count events per source IP and emit results every 5 seconds, resetting the count after each emission:
summarize count(this), src_ip, options={frequency: 5s}Emit cumulative totals
Section titled “Emit cumulative totals”Sum bytes and emit running totals every minute:
summarize sum(bytes), options={frequency: 1min, mode: "cumulative"}Emit only when values change
Section titled “Emit only when values change”Count events by severity and only emit when counts change, reducing output noise:
summarize count(this), severity, options={frequency: 10s, mode: "update"}