About Miller

Miller is like sed, awk, cut, join, and sort for name-indexed data such as CSV.

With Miller you get to use named fields without needing to count positional indices. For example:

% mlr --csv cut -f hostname,uptime mydata.csv
% mlr --csv filter '$status != "down" && $upsec >= 10000' *.csv
% mlr --nidx put '$sum = $7 + 2.1*$8' *.dat
% grep -v '^#' /etc/group | mlr --ifs : --nidx --opprint label group,pass,gid,member then sort -f group
% mlr join -j account_id -f accounts.dat then group-by account_name balances.dat
% mlr put '$attr = sub($attr, "([0-9]+)_([0-9]+)_.*", "\1:\2")' data/*
% mlr stats1 -a min,mean,max,p10,p50,p90 -f flag,u,v data/*
% mlr stats2 -a linreg-pca -f u,v -g shape data/*

This is something the Unix toolkit always could have done, and arguably always should have done. It operates on key-value-pair data while the familiar Unix tools operate on integer-indexed fields: if the natural data structure for the latter is the array, then Miller’s natural data structure is the insertion-ordered hash map. This encompasses a variety of data formats, including but not limited to the familiar CSV. (Miller can handle positionally-indexed data as a special case.)

Features:

  • I/O formats including tabular pretty-printing and positionally indexed (Unix-toolkit style)
  • Conversion between formats
  • Format-aware processing: e.g. CSV sort and tac keep header lines first
  • High-throughput performance on par with the Unix toolkit
  • Miller is pipe-friendly and interoperates with Unix toolkit.
  • Miller is streaming: most operations need only a single record in memory at a time, rather than ingesting all input before producing any output. For those operations which require deeper retention (sort, tac, stats1), Miller retains only as much data as needed. This means that whenever functionally possible you can operate on files which are larger than your system’s available RAM, and you can use Miller in tail -f contexts.
  • It complements SQL databases: you can slice, dice, and reformat data on the client side on its way into or out of a database. You can also reap some of the benefits of databases for quick, setup-free one-off tasks when just need to query some data in disk files in a hurry.
  • Likewise, you can use Miller’s text-reformatting strengths to (among other examples) snarf and munge log-file data into CSV format and then load that into R for further statistical and/or graphical processing.
  • Miller also goes beyond classic Unix tools by stepping into our modern, no-SQL world: its essential record-heterogeneity property allows it to operate on data where records with different schema (field names) are interleaved.
  • Not unlike jq (for JSON), Miller is written in modern C, and it has zero runtime dependencies. You can download or compile a single binary, scp it to a faraway machine, and expect it to work.

Releases and release notes: https://github.com/johnkerl/miller/releases.