Skip to content

Writes events to one or multiple files on a filesystem.

to_file url:string, [max_size=int, timeout=duration,
partition_by=list<field>] {}

The to_file operator writes events to local filesystems, automatically opening new files when a rotation condition triggers. It supports hive-style partitioning through a ** placeholder in the URL and per-partition unique filenames through a {uuid} placeholder.

Path or URL identifying where files should be written.

When ~ is the first character, it is substituted with the value of the $HOME environment variable. Relative paths are resolved against the current working directory.

The path may contain two placeholders:

  • ** marks the location where hive partition segments are inserted. When present, partition_by must also be set, and vice versa.
  • {uuid} expands to a fresh UUIDv7 for every file. This is required when partitioning or when rotation can produce multiple files, so that rotated or per-partition files do not overwrite each other.

Rotates to a new file after the current file exceeds this size in bytes. Because rotation only fires after the threshold is crossed, individual files may be slightly larger than max_size.

Defaults to 100M.

Rotates to a new file after the current file has been open for this duration. Rotation is measured from the time the file was first opened.

Defaults to 5min.

A list of fields used to partition events into separate files. For every distinct combination of partition-field values, a separate file (or group of rotated files) is written. The URL must contain a ** placeholder, which is replaced by the hive-style path field1=value1/field2=value2/….

Unlike to_hive, the partitioning fields are not stripped from the written events — they remain in each record.

Pipeline that transforms the incoming events into the byte stream that is written to each file. The pipeline must return bytes, so it must end with a writer such as write_ndjson, write_parquet, or write_csv, optionally followed by a compressor such as compress_gzip.

The subpipeline runs once per output file. When rotation or partitioning produces a new file, a new instance of the subpipeline is spawned for it.

from {msg: "hello"}, {msg: "world"}
to_file "/tmp/out.json" {
write_ndjson
}

Partition events by a field into separate files

Section titled “Partition events by a field into separate files”
from {region: "us", n: 1},
{region: "eu", n: 2},
{region: "us", n: 3}
to_file "/tmp/events/**/data_{uuid}.json", partition_by=[region] {
write_ndjson
}
// Produces files under:
// /tmp/events/region=us/data_<uuid>.json
// /tmp/events/region=eu/data_<uuid>.json
to_file "/tmp/events/**/data_{uuid}.parquet",
partition_by=[year, month] {
write_parquet
}
// Produces files under:
// /tmp/events/year=<year>/month=<month>/data_<uuid>.parquet
to_file "/tmp/logs/events_{uuid}.json", max_size=10M {
write_ndjson
}
to_file "/tmp/logs/events_{uuid}.json", timeout=1min {
write_ndjson
}

Last updated: