Speaking only from my own experience with https://dataverse.unc.edu
, I doubt you'll push your storage very hard. We run our Dataverse in a shared environment (VMware/Dell cluster over NFS mounts from our NetApp). The storage I/O will be bursty only during uploads and downloads, and unless you'll have a very active user base with a ton of huge files I don't see you bogging it down.
Be sure to give Glassfish plenty of CPUs (for us, 4, the maximum supported by VMware Fault Tolerance) and RAM (for us, 64GB, the maximum supported by Fault Tolerance). Our Glassfish installation typically holds 18G-21G of active memory; the OS keeps 1G-2G free and uses the rest for buffers/cache.
If you install the Rapache module, Apache's CPU usage will increase significantly, but the one time I've seen our installation start to approach a system load of 1.0 was during our webinar when our archivist had all the folks following along at home publish their test datasets simultaneously. Solr indexing kept the machine busy for about 30 seconds, then things quieted back down.