> On Feb 5, 2021, at 6:14 PM, Sean Chittenden <se...@freebsd.org> wrote:
>
> To be clear, this is a known issue that needs attention: it is not a
So silly question ; how does FreeBSD work on Google gcp, Microsoft Azure , Digital Ocean?
Do they all suffer the same issue or is this just an Amazon issue ?
Also just a bit of advice; Contrary to popular belief Amazon does not actually sell magic beans .
---
Mark Saad | none...@longcount.org
> Also just a bit of advice; Contrary to popular belief Amazon does not actually sell magic beans .
AWS has 32% market share. I don't know if they beans are magic or not,
but this is the biggest cloud today. If FreeBSD want to be on this
train, it has to perform at least as good as competitors. It's simply
matter of survival for this project.
In 2007 about 90% of my machines were running FreeBSD (30-40 servers).
Now I run FreeBSD on 9 boxes, and 130 is running Linux. And I'm not the
only one out there who was forced to do this transition.
--
Best regards,
Łukasz Wąsikowski
To be clear we should check if the issue something that aws is doing with their xen platform , kvm/qemu one or common to all ? Also does that same issue appear on google and Microsoft’s platforms? This will at least get some bounds on the problem and what if any fixes may exist .
There are some commercial FreeBSD products running on aws . Maybe the vendors know some stuff that can help ?
Thoughts ?
---
Mark Saad | none...@longcount.org
I have a slight sign of hope also: in my latest Amazon Linux db server deployment
the performance is only half of what I got in my previous db server deployment.
Go figure. But it's a fresh launch and I can't figure out what else I had optimized
with the previous Amazon Linux box. So, while not improving anything, it at least
closes the inequality between Linux and FreeBSD in the socialistic way ;)
Now I am trying the FreeBSD install again with the same disk setup. Because one
thing I clearly know is that it makes a huge difference of having one large EBS
device and a single partition / file system vs. the same large EBS device with
many partitions and file systems, separating tables, indexes, temporary sort space,
etc. so that there is less random access contention on a single file system.
Why that is important, I wonder, actually. And it's not the underlying EBS that
matters so much. Like in bare metal world, the approach would be separate disks
vs. striping (RAID-0) to get more spindles involved instead of having disk seek.
But none of that should mater that much any more with SSD drives (which EBS gp2 and
gp3 are!)
Fortunately now Amazon has gp3 volumes also which support 3000 io transactions
per second (IOps) and 250 or up to 1000 MB/s throughput despite being as small as
4 GB. So what I'm doing now is instead of a partitioned 1 TB gp2 volume with many
partitions, I make myself over 20 separate smaller gp3 volumes, the advantage of
this is that I can individually resize those without colliding with the neighboring
partition.
Tomorrow I should have better comparison numbers for my database on both Linux and
FreeBSD configured the exact same way, and it might be closer now. But sadly only
in the socialist way.
regards,
-Gunther
Cloud has been an increasingly important workload for FreeBSD users. Or at
least, our community is moving a fair number of workloads to virtualized
metal, and running on the cloud is in need of attention as or more than
some of our other subsystems. Many of the people who have the skills to
jump in and diagnose and fix this type of problem have been sucked up with
other professional commitments and aren't available. Which is to say, we
need to increase the depth of our bench for performance-related work.
But if someone's reading this thinking, "well I'll wait until Core...."
don't wait! Jump in.
If people are looking for a single starting point, I'd suggest trying to
conjure up an iflib-backed ethernet driver and reducing the administrative
friction of spinning up one of our images in the Cloud. "Secure by
default" or conservative image defaults aren't doing anyone any favors in
the cloud era.
Or begin porting the io_uring kernel interface and begin working down the
stack into the driver layer.
$0.02. -sc