On 18/01/2021 11:28, Brock Palen wrote:
> Disclaimer I have not used s3 with bareos but done many cloud
> calculations.
>
> Few things to think about using cloud.
> Are you running your SD in the cloud?
> Are your backup clients in the cloud?
> If not what’s your bandwidth? It will impact your backup and restore
> times significantly if you have modest WAN capacity for local clients
> servers.
No, no. I was thinking about keeping an extra copy "off-site". I'm
mostly cloud-free at the moment and I do not wish to change it
significantly. I was thinking whether S3 could be an option for
extending my home backup setup.
Of course I understand the impact of bandwidth on the backup/restore
times. :-)
> As for s3 pricing read this carefully
>
>
https://aws.amazon.com/s3/pricing/ <
https://aws.amazon.com/s3/pricing/>
>
> You have three components to pricing with s3 and I expect only two
> move the needle on cost.
>
> Data stored
> Bandwidth and retrieval
> Operations
>
> Opts rations are so cheap and guessing how bareos uses virtual tape
> volumes it’s prob not a big issue. Someone who has used it though can
> speak.
That's a good observation. Thanks!
> Data stored its straight $/gb/month. So you need to estimate your
> total data stored for all your fulls and incrementals. Your right
> these costs decline when you look at glacier but there is a trade off.
> The cheaper to store the more expensive to access.
>
> Retrieval fees come in two forms. The first is bandwidth. Which fit
> most people is .09$/gb (unless your clients and servers are in the
> same aws region) for my cloud activities this is 50% of my monthly
> bill. It’s the thing that messes most cloud calculators for budget.
> That said if your sever is on prem you likely will never pay this if
> you don’t use always incremental or do any restores. So if your ok
> paying for restores maybe it’s ok.
>
> The cold tiers like glacier charge to access data. Again maybe fine if
> you almost never read it. Glacier runs $10/tb or more for transfer vs
> nothing for regular s3. With bandwidth your at ~$100/tb This is
> something to avoid deep archive. Their sla is many hours to get data.
> I don’t think deep archive is a backup replacement but a compliance
> archive replacement
Well, that's what I'm counting on - it's better to have backup copy and
not need to use it than not having it ;-)
What I was also interested in was also how to approach the long SLA
regarding Bareos SD operation. Would I have to firstly request access to
the glacier data independently of the SD and after receiving
confirmation of data availability would have to run a restore job? Or
would I just run a restore job from storage using cold-tiered bucket and
the job would simply wait for data availability (similar to mounting tape)?
> Also be aware glacier and deep archive have minimum retention times of
> 90 and 180 days. So you will always pay that at a minimum. Ok if your
> keeping fulls for a long time. Look at the auto tier options to
> manage aging volumes.
Yes, I noticed that
>
> So YMMV. If you are 100% in the cloud or you don’t use always
> incremental or have small data volumes or just a dr copy it works great.
>
> Personally I run my servers in aws and my full bareos setup on prem
> with a $400 tape library from eBay. This gives me diversity and most
> of the data in the cloud is small (websites email text) while the on
> prem is video photos and road warriors using always incremental.
So it all comes to "try the free tier and see for yourself" :-) I'll
have to do it anyway when I get some spare time just to see how it works
and get some understanding about achievable througputs, needed space and
so on.
Thanks for valuable insight!