On 08/27/2012 04:09 PM, auada wrote:
> Also I was thinking and wondering if, building a filesystem and
> splitting it among different S3 buckets would make any difference in
> relation to performance and security.
No, as far as I can see that would be pointless.
Best,
-Nikolaus
--
�Time flies like an arrow, fruit flies like a Banana.�
On 08/27/2012 09:12 PM, auada wrote:
> 1- regarding security and durability, if we use s3ql filesystem as a
> production resource, should we keep backups outside the filesystem, or
> the snapshots would be enough ?
That's something only you can decide.
Since the release of S3QL 1.0, any reports of data loss where restricted
to data that was in the process of being copied while either computer or
network connection crashed. I have not had any reports of data being
lost after it was written and the metadata was uploaded. For my own
data, I use S3QL backups exclusively.
> 2- would it be possible to sinchronize sql lite (or other db) among
> different instances and try to make multiple mounts of same file system,
> or there is more to multiple mounts than only the db ? I ask, because it
> is like a dream to be able to have multiple mounts and leave nfs behind
> and get rid of a single point of failure.
This has been discussed a few times on the list, you may want to take a
look at the archives. The short answer is: no, it's not feasible.
> 3- in the manual you say that standard S3 at Amazon has an eventual
> consistency window and other regions have immediate. Where did you take
> that from: from AWS or from experience ? And do you include Oregon as
> standard as well ?
That's from the AWS documentation. I don't know about Oregon on top of
my head, if you want to be sure you should probably check the AWS
documentation.
Just checked. Oregon is not Standard, so it´s immediate consistent.
> 4- what is your feeling about using s3ql filesystem for maildirs,
> regarding performance, durability and security ? Do you think we could
> have a bottle neck ?
I think you're trying to make a very complex question a bit too simple
here. This all depends on your requirements. S3QL is slower than
in-kernel file systems like ext4 for sure. It's probably more durable
than running ext3 on a hard disk, because Amazon S3 has much more
redundancy, but you can of course compensate for that by setting up a
giant RAID system yourself. I am not sure what you mean with security.
Every system always has a bottle neck. Whether it will be S3QL depends
on your hardware, software, user base and requirements.
Best,
-Nikolaus
--
�Time flies like an arrow, fruit flies like a Banana.�
On 08/28/2012 10:33 AM, auada wrote:
>
>
> Em ter�a-feira, 28 de agosto de 2012 09h56min14s UTC-3, Nikolaus Rath
> escreveu:
>
> On 08/27/2012 09:12 PM, auada wrote:
> > 1- regarding security and durability, if we use s3ql filesystem as a
> > production resource, should we keep backups outside the
> filesystem, or
> > the snapshots would be enough ?
>
> That's something only you can decide.
>
> Since the release of S3QL 1.0, any reports of data loss where
> restricted
> to data that was in the process of being copied while either
> computer or
> network connection crashed. I have not had any reports of data being
> lost after it was written and the metadata was uploaded. For my own
> data, I use S3QL backups exclusively.
>
>
> I wonder if from a server crash the db can get corrupted and the data
> not recoverable, even with s3ql.fsck ... is that a situation that can
> happen ?
The local copy of the metadata may get corrupted beyond recovery, but
the last remote copy (which is uploaded periodically even if the file
system is never unmounted) is always available.
> > 2- would it be possible to sinchronize sql lite (or other db) among
> > different instances and try to make multiple mounts of same file
> system,
> > or there is more to multiple mounts than only the db ? I ask,
> because it
> > is like a dream to be able to have multiple mounts and leave nfs
> behind
> > and get rid of a single point of failure.
>
> This has been discussed a few times on the list, you may want to take a
> look at the archives. The short answer is: no, it's not feasible.
>
>
> OK, short answer is enoguh, thanks.
>
> Regarding the nfs flag, how does it help NFS server ?
NFS occasionally needs to look up the name of a directory entry from its
inode. The --nfs flag creates a special index for that. Without this
index, the time required to process such a request is linear in the
number of directory entries (read: terrible). The flag increases the
size of the cached metadata (but not the amount of metadata that will be
uploaded) but has no adverse effects. I don't know how often an NFS
server performs this sort of query.
> I have been researching the list and I found something saying that for
> huge filesystem size, the netada can get too big. Would that be a bottle
> neck ? Would it be difficult to migrate the filesystem to another server
> ? What would be the consequences of this metada being too large ?
The metadata will grow, and will take longer and longer to upload. "Too
big" thus depends on how often you want to upload metadata, and how much
time and bandwith you are willing to invest in metadata uploads. But as
far as S3QL is concerned, there is no hard limit on the metadata size.
"Migrating" the file system just involves unmounting it on one server
and mounting it on the other, i.e. it means uploading and downloading
the metadata once. An S3QL file server should be thought of as being
located in S3, not on the computer that has mounted it at a given point
in time.
Best,
-Nikolaus
--
�Time flies like an arrow, fruit flies like a Banana.�
On 08/30/2012 11:08 AM, auada wrote:
> > I wonder if from a server crash the db can get corrupted and the data
> > not recoverable, even with s3ql.fsck ... is that a situation that can
> > happen ?
>
> The local copy of the metadata may get corrupted beyond recovery, but
> the last remote copy (which is uploaded periodically even if the file
> system is never unmounted) is always available.
>
> And this last copy, if not updated 100 % with local metada, will be able
> to access all files ?
All files that were present at the time the metadata was uploaded and
that have not been deleted afterwards.
> > I have been researching the list and I found something saying that
> for
> > huge filesystem size, the netada can get too big. Would that be a
> bottle
> > neck ? Would it be difficult to migrate the filesystem to another
> server
> > ? What would be the consequences of this metada being too large ?
>
> The metadata will grow, and will take longer and longer to upload. "Too
> big" thus depends on how often you want to upload metadata, and how
> much
> time and bandwith you are willing to invest in metadata uploads. But as
> far as S3QL is concerned, there is no hard limit on the metadata size.
>
>
> And can we (user) decide the frequency of uploads ? I haven�t seen that
> in the manual, but I might have missed it.
Look for --metadata-upload-interval in
http://www.rath.org/s3ql-docs/mount.html
> And is there any relation like, a ratio between the size of the
> filesystem and the metada ? Any rule of thumb ?
I haven't looked into that. If you decide to investigate that, please
share your results :-).
Best,
-Nikolaus
--
�Time flies like an arrow, fruit flies like a Banana.�
On 08/30/2012 11:14 AM, auada wrote:
> - if we build a cache in a local driver, but configure it to grow huge,
> with many more file descriptors and cache size, like trying to reach 10
> GBytes, would that bring performance benefit ? How do you see that ?
That depends on your access pattern. If you previously accessed more
files simultaneously than you used descriptors, then increasing the
number of descriptors will give you a dramatic performance increase.
Same with cache size. If you simultaneously access more blocks than you
have cache, increasing cache size will give a dramatic performance
increase. Otherwise you'll not gain anything.
I think the thing that would help you most is to simply set up a server
and test.
Best,
-Nikolaus
--
�Time flies like an arrow, fruit flies like a Banana.�
--
»Time flies like an arrow, fruit flies like a Banana.«
PGP fingerprint: 5B93 61F8 4EA2 E279 ABF6 02CF A9AD B7F8 AE4E 425C
--
You received this message because you are subscribed to the Google Groups "s3ql" group.
Brad, thanks a lot for your great explanation. Does it apply for web hosting as well, not considering DB´s of course, only web codes ? Regards, -- Ricardo Buchalla Auada