On a practical level, you are more likely to be performance limited by
disk IOPS and RAM. If you are primary doing archival, where data is
never updated, and your primary index is ordered by insert time
(ObjectID is designed around this property), and you mostly access the
recent end of your data, you should be able to pack several TB onto a
single high-end node. However, if you do frequent random access, you
will need a much higher ram-to-data ratio.
[1] https://github.com/mongodb/mongo/blob/master/db/diskloc.h#L49
> --
> You received this message because you are subscribed to the Google Groups
> "mongodb-user" group.
> To post to this group, send email to mongod...@googlegroups.com.
> To unsubscribe from this group, send email to
> mongodb-user...@googlegroups.com.
> For more options, visit this group at
> http://groups.google.com/group/mongodb-user?hl=en.
>
Yes this applies to 64-bit OSs as well. We use the same data-file
format on all platforms so you can freely move your files around.
The 32-bit offset is mostly a space optimization. It allows us to
point to anywhere within a single db using a 64-bit DiskLoc stuct with
a 32-bit file number and a 32-bit offset into that file. This is what
mongo uses internally when storing "pointers" to disk.
There is no reason you can't store huge data sets in MongoDB. We have
some users with multiple TB of data.