OK, I guess I should contribute the solution I finally used. The
problem at hand was a 34 MB mailing list archive containing lots of
news excerpts (about 10000 messages). For an article I'm going to
write, I wanted to scan that archive for contributions concerning
certain topics.
My solution was this: First, using mutt, I stored that archive to a
maildir folder. I then created a sub-directory hashed/, and used a
small shell script (horribly inefficient) to move the message files
into a certain number of subdirectories under hashed/. That way,
file system overhead was under control again. I then indexed the
hashed/ subdirectory using glimpseindex, and finally used the
attached script to create hard links to the cur sub-directory, which
could then be read using mutt.
Note that this script has a kludge to emulate glimpse -l - that's
because glimpse -l dumps core on my system.
--
http://www.guug.de/~roessler/
#!/bin/sh --
glimpse -H . -1 -i "$@" | cut -f1 -d: | while read filename ; do
ln $filename cur/
done