Pigz abort
-
Something odd looks to be happening. The Image File size says that it’s only 3.97 GB, but the data that’s being copied is 173.88 GB? Does that sound right to you? It looks like it’s copying the image file size (3.97GB) then aborting the upload because it’s reached the file size. What if you try creating a new image with Multiple Partition, All Disks Non-resizeable. Then assign that image to the machine, then try to upload?
-
Actually, the “Image file size” continues to grow in size. Just did the all disk option, same (similar) results.
-
Can you try a system with a smaller image/partition table? Maybe even a different system altogether? This will let you know if it’s the system, or your fog server that can’t handle the issue.
Also, are you sure you have 1.2 T available? This sounds like your storage location is full.
-
Yes, Tom, I think that is part of the problem. I’ve looked at the server settings, but I don’t see where it’s misconfigured. The strange thing is that previously I didn’t have any problems; it’s worked great for more than 18 months or so. I guess I got smacked by an update or two.
-Chuck -
From the sounds of it, your storage node is full which would explain why it stops imaging at, very nearly, the same point every time. Maybe try adding some storage space, if you can, or add another storage node that has the space available on it.
-
I’ve checked the disks, and they look good. /storage shows 1.2T Available (and / shows 120 G available)
Somehow, the system doesn’t register what’s available. Any ideas where to look on the server? I’ve already looked up the Storage Management -> All Storage Nodes and double checked those settings, along with the config in /opt/fog/service/etc/config.php and /var/www/fog/commons/config.php.
Would the memory_limit, post_max_size, and upload_max_filesize all set at 1900M in /etc/php5/apache2/php.ini cause a problem? (I tried to image a 325 G drive today and I still ran into a problem…)
Thanks,
Chuck -
Does your fog server actually use /storage as the location for the images? Or is it the typical /images directory setup. If it’s /images, it’s trying to use your root system (120G) to store more than that worth of data which would fail. A bypass to that would be to move /images to /storage then link /storage/images to /images which could be done with->
mv /images /storage; ln -s /storage/images /
You wouldn’t have to make any configuration changes then. The other thing to check would be the /etc/exports file to see what your NFS system is trying to mount to store the image. My guess is it’s actually mounting /images even if your configuration is pointing to /storage
To fix that, you’d simply change the file reference to /storage and /storage/dev (assuming that’s how your system is setup)
Also check your .mntcheck files to see that they exist. They don’t actually contain information, but they need to be present. Make sure they exist by typing:
touch /images/.mntcheck; touch /images/dev/.mntcheck; touch /storage/.mntcheck /storage/dev/.mntcheck
chmod -R +x /images; chmod -R +x /storage
Of course remove the items you don’t need as I don’t know what your particular setup requires.
Then try again.
-
One last note,
If you do end up changing the /etc/exports file, make sure to restart the NFS server.
I think a simple:
/etc/init.d/nfs restart
will do the trick. -
Hey Tom,
Yes, I do have everything pointing to /storage… actually /storage/images, and when I try to create new images, the new files are written there. For example, when creating the all disk image you suggested it did the following:
/storage/images/4300AllDisk contains: d1.mbr, d1p1.img, d1p2.img, d1p3.img, d1p4.imgThe only thing is that d1p2.img is only 2G for a 456.7 G partition.
(I’ve checked /images and the only files there are the /dev directory and the .mntcheck file)
.mntcheck is present in /storage/images and set to 777 root:root
Oh, and /etc/exports looks ok to me with:
/storage/images *(ro,sync,no_wdelay,insecure_locks,no_root_squash,insecure)
/storage/images/dev *(rw,sync,no_wdelay,no_root_squash,insecure)Thanks again,
Chuck -
Then the next question is … Is /storage a raid system with a possible bad drive? I don’t know what else to check then and I’m sorry.
-
Thanks, Tom! That may very well be the problem. It appears I’m getting conflicting data…
vgdisplay shows only 68.70 GB free, while df -h shows 1.2 T available.-Chuck
-
So is /storage a part of an LVM?
What is the output of vgdisplay?
-
Here’s what my output looks like!
— Volume group —
VG Name vg_mastavirtual
System ID
Format lvm2
Metadata Areas 5
Metadata Sequence No 4
VG Access read/write
VG Status resizable
MAX LV 0
Cur LV 3
Open LV 3
Max PV 0
Cur PV 5
Act PV 5
VG Size 4.16 TiB
PE Size 4.00 MiB
Total PE 1089666
Alloc PE / Size 1089666 / 4.16 TiB
Free PE / Size 0 / 0
VG UUID Y86n60-8kTN-Akpl-Ixkm-jcUK-iqNG-UIFXV6Then a pvdisplay shows:
— Volume group —
VG Name vg_mastavirtual
System ID
Format lvm2
Metadata Areas 5
Metadata Sequence No 4
VG Access read/write
VG Status resizable
MAX LV 0
Cur LV 3
Open LV 3
Max PV 0
Cur PV 5
Act PV 5
VG Size 4.16 TiB
PE Size 4.00 MiB
Total PE 1089666
Alloc PE / Size 1089666 / 4.16 TiB
Free PE / Size 0 / 0
VG UUID Y86n60-8kTN-Akpl-Ixkm-jcUK-iqNG-UIFXV6[root@mastavirtual ~]# pvdisplay
— Physical volume —
PV Name /dev/sda1
VG Name vg_mastavirtual
PV Size 465.76 GiB / not usable 3.00 MiB
Allocatable yes (but full)
PE Size 4.00 MiB
Total PE 119234
Free PE 0
Allocated PE 119234
PV UUID NQYnUI-WNc1-4usj-uMSO-ZLTc-4O7k-pWWx2M— Physical volume —
PV Name /dev/sdb1
VG Name vg_mastavirtual
PV Size 698.64 GiB / not usable 3.00 MiB
Allocatable yes (but full)
PE Size 4.00 MiB
Total PE 178850
Free PE 0
Allocated PE 178850
PV UUID aBQhCX-0A1r-EupN-bJsv-tAWf-HEdA-soXPfC— Physical volume —
PV Name /dev/sdc1
VG Name vg_mastavirtual
PV Size 298.09 GiB / not usable 4.00 MiB
Allocatable yes (but full)
PE Size 4.00 MiB
Total PE 76310
Free PE 0
Allocated PE 76310
PV UUID cE19Nu-51wP-UdvP-K8ak-BeFA-PLtd-cr4h2l— Physical volume —
PV Name /dev/sdd1
VG Name vg_mastavirtual
PV Size 931.51 GiB / not usable 4.00 MiB
Allocatable yes (but full)
PE Size 4.00 MiB
Total PE 238466
Free PE 0
Allocated PE 238466
PV UUID Th8qJh-j1Y8-g20V-xlL9-Jkqv-WsRl-3T5gKy— Physical volume —
PV Name /dev/sde2
VG Name vg_mastavirtual
PV Size 1.82 TiB / not usable 4.00 MiB
Allocatable yes (but full)
PE Size 4.00 MiB
Total PE 476806
Free PE 0
Allocated PE 476806
PV UUID fUVEMi-fdqY-1QgK-yRVY-oU75-5oBb-XtmCZD -
Yes (/storage/images) is part of the LVM.
Here’s the output:
root@kiri1:~# pvdisplay
— Physical volume —
PV Name /dev/sdb1
VG Name kiri2
PV Size 1.36 TiB / not usable 1.29 MiB
Allocatable yes
PE Size 4.00 MiB
Total PE 357587
Free PE 17587
Allocated PE 340000
PV UUID glQE6b-qwEH-LQo3-k2an-owfH-OPvy-PXnj1I— Physical volume —
PV Name /dev/sda5
VG Name kiri1
PV Size 136.46 GiB / not usable 2.00 MiB
Allocatable yes
PE Size 4.00 MiB
Total PE 34933
Free PE 6
Allocated PE 34927
PV UUID GUvasn-41Eh-ozeI-bZnZ-UaXV-HsYC-r1C9MIroot@kiri1:~# vgdisplay
— Volume group —
VG Name kiri2
System ID
Format lvm2
Metadata Areas 1
Metadata Sequence No 2
VG Access read/write
VG Status resizable
MAX LV 0
Cur LV 1
Open LV 1
Max PV 0
Cur PV 1
Act PV 1
VG Size 1.36 TiB
PE Size 4.00 MiB
Total PE 357587
[COLOR=#ff0000]Alloc PE / Size 340000 / 1.30 TiB[/COLOR]
[COLOR=#ff0000]Free PE / Size 17587 / 68.70 GiB[/COLOR]
VG UUID Agux9x-QG3h-O5Gg-WZrc-wuDh-4oGO-twHaDg— Volume group —
VG Name kiri1
System ID
Format lvm2
Metadata Areas 1
Metadata Sequence No 3
VG Access read/write
VG Status resizable
MAX LV 0
Cur LV 2
Open LV 2
Max PV 0
Cur PV 1
Act PV 1
VG Size 136.46 GiB
PE Size 4.00 MiB
Total PE 34933
Alloc PE / Size 34927 / 136.43 GiB
Free PE / Size 6 / 24.00 MiB
VG UUID NRCufd-StrG-CQb0-87bS-2ehH-7Yea-PSMhKMroot@kiri1:~# lvdisplay
— Logical volume —
LV Name /dev/kiri2/storage
VG Name kiri2
LV UUID 75YgXN-ucDL-pgvA-E6lm-h6LK-sSTz-urI94a
LV Write Access read/write
LV Status availableopen 1
LV Size 1.30 TiB
Current LE 340000
Segments 1
Allocation inherit
Read ahead sectors auto- currently set to 256
Block device 252:0
— Logical volume —
LV Name /dev/kiri1/root
VG Name kiri1
LV UUID 2Z2aY7-qkTk-gu1y-j8qt-x9gO-IObv-f7Ht3M
LV Write Access read/write
LV Status availableopen 1
LV Size 130.86 GiB
Current LE 33499
Segments 1
Allocation inherit
Read ahead sectors auto- currently set to 256
Block device 252:1
— Logical volume —
LV Name /dev/kiri1/swap_1
VG Name kiri1
LV UUID u4ZjMQ-j9IO-n8Pt-GP5p-uGzv-R76q-HQmt0f
LV Write Access read/write
LV Status availableopen 2
LV Size 5.58 GiB
Current LE 1428
Segments 1
Allocation inherit
Read ahead sectors auto- currently set to 256
Block device 252:2
root@kiri1:~# df -h
Filesystem Size Used Avail Use% Mounted on
/dev/mapper/kiri1-root 129G 3.0G 120G 3% /
udev 5.9G 4.0K 5.9G 1% /dev
tmpfs 2.4G 336K 2.4G 1% /run
none 5.0M 0 5.0M 0% /run/lock
none 5.9G 0 5.9G 0% /run/shm
/dev/mapper/kiri2-storage 1.3T 141G 1.2T 11% /storage/dev/sda1 228M 47M 170M 22% /boot
- currently set to 256
-
Do you know of any tools to test the disks within the LVM to find out if your disks are okay?
-
pvck & vgck ? Right now, I replaced one of the drives on sda5, and one of the drives on sdb1. They will need some time to rebuild.
-
Well, I hope this works for you. I’m sorry I wasn’t much help, but it sounds like it was probably just a bad drive causing all of this headache. To test, while this one is rebuilding, do you have another storage node that has enough space that you could test with?
-
Hey Tom, I appreciate you going through all the steps with me. I just wanted to rule out something silly that I may have missed. I’m currently waiting to see if I can establish another storage node.
-Chuck -
Have you had any luck, or is it still rebuilding?
-
I’m uploading an image right now. Built a new simple (no bonding, no LVM, simple partitioning with default /images) VM Ubuntu 12.04.2 LTS with Fog 0.32 on a 200 GB drive.
The really funny thing is, it still reads:
“Available space for image:…1.66 GiB = 1778888704 bytes”