Hi Theo,
First, memory and disk space are two different things. In terms of storage, we're talking about disk space - AtoM does allow you to see uploads associated with a particular repository (and set limits per repository, if desired), as well as overall. See:
Keep in mind that this does not currently track space used by downloads - such as generated or uploaded finding aids, cached XML files, reports, etc. For more information on the structure of both the uploads and downloads directories, see:
Additionally, this is on a per-repository basis. It will be trickier to figure out per descriptive hierarchy. For each individual description, AtoM does show you file sizes for the uploaded master, but it doesn't currently provide additional technical metadata for the derivatives - though this is coming in our 2.7 release. In the meantime however, I imagine it might be cumbersome to have go through all your descriptions individually to collect some numbers. It might be possible for someone to create a script that could derive this information from the database, but that would require a deeper look than I can currently provide here.
As for the metadata itself - it would be much harder to calculate the disk space used per repository, as this data is held in a relational database. I can tell you, by way of example, that the entire dataset in our
public demo site is about 23MB when dumped to a SQL file. An individual description is likely only a couple KB (it's really just a bunch of connected rows in a series of tables), though as part of a SQL dump you can't really separate them out like that.
Memory, on the other hand, is not something that would be very easy to determine on a per-institution basis. The memory on your server is a pool that is drawn upon as needed, and then released for other uses - so it's usage will go up or down depending on the activities being performed. Some operations (for example, XML imports) require reading a lot of data into memory, to be able to quickly refer to parts of it while parsing the import - but once the import completes, that memory should become available again. The same is true for most processes - in general, I'd suspect that writes require more memory use than reads, especially if you are using caching on your public website. I've no idea how you would go about determining memory usage per institution, let alone per description. I will mention that there are tools available that will let you monitor system resource usage - we describe one command-line tool (
htop), in our documentation here:
This will show you memory usage at the server level - I suppose you could do some tests by performing actions in the user interface while monitoring with htop or a similar tool, to establish a rough benchmark, but it would be rough at best, and likely to vary depending on a number of interrelated factors. There may also be other types of tools out there that could help, but that's beyond my personal experience.
Likely not as straightforward of a response as you were hoping for, but I hope it helps!
Cheers,