On 09/05/15 15:03, Srikanth Vavilapalli wrote:
> Hi
>
> This is the second time I have noticed this behavior with my openstack
> cluster created with "johnsond-osp" profile at clemson site.
>
> Initially the disk allocation looks as shown below after all Open stack
> comes up on the controller node:
>
> svavilap@ctl:~$ df -h
> Filesystem Size Used Avail Use% Mounted on
> udev 126G 0 126G 0% /dev
> tmpfs 26G 42M 26G 1% /run
> /dev/sda1 16G 2.9G 13G 20% /
> tmpfs 126G 0 126G 0% /dev/shm
> tmpfs 5.0M 0 5.0M 0% /run/lock
> tmpfs 126G 0 126G 0% /sys/fs/cgroup
> ops.clemson.cloudlab.us:/proj/xos-PG0 100G 851M 100G 1% /proj/xos-PG0
> ops.clemson.cloudlab.us:/share 97G 0 90G 0% /share
> tmpfs 26G 0 26G 0% /run/user/0
> tmpfs 26G 0 26G 0% /run/user/20001
Did you wait to check space usage until you've gotten the second email
from the experiment about being finished setting up? I would have
thought usage after setup would be more like 4.5GB, maybe over 5GB. The
base image for the controller node probably uncompresses to at least
2-3GB when it's loaded on disk (the other images are slightly different
due to installed packages). Then we download cloud images for the guest
VMs to run (and a few other things -- consumes maybe half a GB or so)
into /root/setup, and then import them into glance. Then there's setup
stuff in /root/setup (where the setup scripts store their state and tmp
files), and openstack runtime stuff in /var/lib/glance .
Anyway, you should get up to 5.5GB pretty fast after instantiation, I
think. The setup scripts configure many of the openstack services with
verbose or debug mode logging, so if you're pounding on the services,
those log directories in /var/log are going to be filling up fast. One
thing I will do for you is make the verbose and debug logging modes
optional; this may help. But you'll have to start using the "OpenStack"
profile instead of "johnsond-osp" -- it's the same thing, but more
featureful, tested, and complete. "johnsond-osp" was really just a
temporary stand-in, and it won't work or will disappear sometime later
on. I'll let you know when I have this done. You could also get rid of
the large image files in /root/setup ; those aren't necessary anymore
after they've been imported into glance.
I can't really comment on your docker container's size or any of the XOS
software's size... you'll have to see what's eating up disk using the
'du' command yourself.
> However after few days of usage, the /dev/sda is completely filled up
> and any operation on this node throws up an error: "*write failed:
> /tmp/sortko4fc2: No space left on device*". I have not downloaded any
> large files onto this machine other launching XOS docker container in
> addition to open stack services.
> When I created a cluster using the same profile on "Clouldlab Utah", the
> disk allocation on the controller node is much bigger as shown below
> (The /dev/sda1 is getting assigned with 110G), However I could not use
> this Utah site because of not able to consistently bring up my
> experiment using this "johnsond-osp" profile there (I have seen many
> times my experiment booting process fails)
>
> svavilap@ctl:~$ df -h
> Filesystem Size Used Avail Use% Mounted on
> udev 32G 0 32G 0% /dev
> tmpfs 6.3G 89M 6.3G 2% /run
> /dev/sda1 110G 2.7G 102G 3% /
> tmpfs 32G 0 32G 0% /dev/shm
> tmpfs 5.0M 0 5.0M 0% /run/lock
> tmpfs 32G 0 32G 0% /sys/fs/cgroup
> ops.utah.cloudlab.us:/share 97G 1.3G 88G 2% /share
> ops.utah.cloudlab.us:/proj/xos-PG0 100G 624K 100G 1% /proj/xos-PG0
> tmpfs 6.3G 0 6.3G 0% /run/user/0
> tmpfs 6.3G 0 6.3G 0% /run/user/20001
>
>
> Until this problem is root caused, would it be possible to increase the
> primary disk space on the controller and compute nodes at "Cloudlab
> Clemson" while using this "johnsond-osp" profile?
Sorry, it's non-trivial to expand the root partitions on the existing
disk images that the profile uses, for a lot of reasons. The profile's
disk images are based on our standard disk image, which provides a 16GB
root partition, and allows you, the experimenter, to use the remainder
of the disk as you see fit. We *always* have users who want physical
disk space that is not already partitioned (especially not in the root),
to use for LVM or any kind of storage experiment. Also, these standard
images have to work on a wide variety of machines (small to large disks,
sometimes only a single physical disk), and 16GB basically represents a
(currently generous) guess of how much space a base Linux OS
installation, plus a kernel compilation, would consume in the root
partition. So, our standard partition layout is carefully designed to
satisfy lots of customers :).
My first advice to you is that you try to put all the stuff you install
on the node in an LVM logical volume. The openstack profile sets up an
LVM physical volume on the controller node's disk using unpartitioned
space... for instance, swift uses this space. You could certainly add a
logical volume for your XOS stuff.
If you need extra space in the root partition, and it can't be solved by
decreasing openstack service logging, or moving your own stuff into an
LVM, you'll have to get really creative or make your own custom images.
> Srikanth
David