bash - shuffle a file that is too large to fit in memory
Using a form of decorate-sort-undecorate pattern and awk
you can do something like:
$ seq 10 | awk 'BEGIN{srand();} {printf "%06d %s\n", rand()*1000000, $0;}' | sort -n | cut -c8-
8
5
1
9
6
3
7
2
10
4
For a file, you would do:
$ awk 'BEGIN{srand();} {printf "%06d %s\n", rand()*1000000, $0;}' SORTED.TXT | sort -n | cut -c8- > SHUFFLED.TXT
or cat
the file at the start of the pipeline.
This works by generating a column of random numbers between 000000
and 999999
inclusive (decorate); sorting on that column (sort); then deleting the column (undecorate). That should work on platforms where sort does not understand numerics by generating a column with leading zeros for lexicographic sorting.
You can increase that randomization, if desired, in several ways:
If your platform's
sort
understands numerical values (POSIX, GNU and BSD do) you can doawk 'BEGIN{srand();} {printf "%0.15f\t%s\n", rand(), $0;}' FILE.TXT | sort -n | cut -f 2-
to use a near double float for random representation.If you are limited to a lexicographic sort, just combine two calls to
rand
into one column like so:awk 'BEGIN{srand();} {printf "%06d%06d\t%s\n", rand()*1000000,rand()*1000000, $0;}' FILE.TXT | sort -n | cut -f 2-
which gives a composite 12 digits of randomization.
Count lines (wc -l
) and generate a list of numbers corresponding to line numbers, in a random order - perhaps by generating a list of numbers in a temp file (use /tmp/
, which is in RAM typically, and thus relatively fast). Then copy the line corresponding to each number to the target file in the order of the shuffled numbers.
This would be time-inefficient, because of the amount of seeking for newlines in the file, but it would work on almost any size of file.
Have a look at https://github.com/alexandres/terashuf . From page:
terashuf implements a quasi-shuffle algorithm for shuffling multi-terabyte text files using limited memory