The VM's (running in a VMware environment) serving as LXC hosts were setup in a way, that the volume group (VG) used for the containers, can be dynamically increased without downtime.
The classical and probably easiest way is to just add a new virtual disk to the VM, create a new PV of this disk and use vgextend to add the new PV to the existing VG.
A nicer method however (as least imho) is to increase the already used virtual disk/PV.
Before I touched anything, I check the current available space of the volume group:
# vgs
VG #PV #LV #SN Attr VSize VFree
vglxc 1 5 0 wz--n- 50.00g 6.00g
vgsystem 1 2 0 wz--n- 6.61g 1.96g
1. In VMware you can increase the disk's size given you still have space in your volume.
2. You need to tell your OS (in my case this is an Ubuntu 14.04) that it should rescan the scsi bus to detect changes. As I increased the second disk, which is seen as /dev/sdb, I launched the following command for this:
# echo 1 > /sys/class/scsi_device/2\:0\:1\:0/device/rescan
If you're not sure which device number your disk (again, in my case sdb) has, you can easily double-check the device path by checking /sys/block/sdb/device:
# ll /sys/block/sdb/device
lrwxrwxrwx 1 root root 0 Jun 28 09:08 /sys/block/sdb/device -> ../../../2:0:1:0
On newer Linux releases you can directly run the echo command using the drive (sdb for example) :
# echo 1 > /sys/class/scsi_device/2\:0\:1\:0/device/rescan
3. Now you can verify that your disk's size has changed. In dmesg you should see a message that the drive size has changed:
# dmesg | tail
[17006442.602408] sdb: detected capacity change from 53687091200 to 75161927680
You can also use fdisk to check the new capacity:
# fdisk -l /dev/sdb
Disk /dev/sdb: 75.2 GB, 75161927680 bytes
255 heads, 63 sectors/track, 9137 cylinders, total 146800640 sectors
Units = sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disk identifier: 0x00000000
(the previous size was 25GB smaller)
4. All you need to do now is to tell LVM that the physical volume (ergo /dev/sdb in my case) has changed it's size:
# pvresize /dev/sdb
Physical volume "/dev/sdb" changed
1 physical volume(s) resized / 0 physical volume(s) not resized
5. Check again the available space of the volume group:
# vgs
VG #PV #LV #SN Attr VSize VFree
vglxc 1 5 0 wz--n- 70.00g 26.00g
vgsystem 1 2 0 wz--n- 6.61g 1.96g
VoilĂ . 20GB more space for my containers without downtime.
Note: This only works "that easy" because I used the complete disk /dev/sdb as physical volume. There are no partitions on sdb. If there were, it would be mandatory to increase the partition as well.
No comments yet.
AWS Android Ansible Apache Apple Atlassian BSD Backup Bash Bluecoat CMS Chef Cloud Coding Consul Containers CouchDB DB DNS Database Databases Docker ELK Elasticsearch Filebeat FreeBSD Galera Git GlusterFS Grafana Graphics HAProxy HTML Hacks Hardware Icinga Influx Internet Java KVM Kibana Kodi Kubernetes LVM LXC Linux Logstash Mac Macintosh Mail MariaDB Minio MongoDB Monitoring Multimedia MySQL NFS Nagios Network Nginx OSSEC OTRS Office OpenSearch PGSQL PHP Perl Personal PostgreSQL Postgres PowerDNS Proxmox Proxy Python Rancher Rant Redis Roundcube SSL Samba Seafile Security Shell SmartOS Solaris Surveillance Systemd TLS Tomcat Ubuntu Unix VMWare VMware Varnish Virtualization Windows Wireless Wordpress Wyse ZFS Zoneminder