However, for research purposes I would first dump TimeMaps of various URIs in separate files and process them later. This will allow me to revisit those dumps to extract more information later without hitting the network again.
# Download TimeMaps
# Extract per-archive summary
$ grep -v "^@" /PATH/TO/TIMEMAPS/URI-1.cdxj | awk -F'[/ ]' '{if(!($5 in a)){a[$5]["first"]=$1};a[$5]["last"]=$1;a[$5]["count"]++} END {for (k in a){print a[k]["count"]" - "k" - "a[k]["first"]" / "a[k]["last"]}}'
# Yearly memento counts
$ grep -v "^@" /PATH/TO/TIMEMAPS/URI-1.cdxj | cut -c-4 | uniq -c
I used CDXJ format for TimeMaps because it is easier to parse, but you can do it using standard Link format as well. I would note here that MemGaor does not aggregate from as many sources by default as LANL's TimeTravel service, but you can customize the list of archives to aggregate from using "-A" flag. Also, MemGator provides the complete TimeMap in one go, sorted by Datetime, without any pagination or nested index of TimeMaps, which makes it easier to process the response, but may be slower.
Best,
--
Department of Computer Science
Old Dominion University
Norfolk VA 23529