all lists on lists.proxmox.com
 help / color / mirror / Atom feed
From: Gilberto Nunes <gilberto.nunes32@gmail.com>
To: Aaron Lauterer <a.lauterer@proxmox.com>
Cc: Proxmox VE development discussion <pve-devel@lists.proxmox.com>
Subject: Re: [pve-devel] Qm move_disk bug (?)
Date: Wed, 30 Sep 2020 11:21:47 -0300	[thread overview]
Message-ID: <CAOKSTBuwgrKE1VKSmngoPOM1wmwUUGyfh_XhRMFg-h60xnHiZQ@mail.gmail.com> (raw)
In-Reply-To: <CAOKSTBtyDJi8fQZ9qBFSuXvvg33Jqwj-JxWYxmX=hzfeKSH8xg@mail.gmail.com>

Ok! Just to be sure, I did it again...

In the LVM-Thin I have an 100.00g vm disk. Note that only about 6% are
filled up.

lvs
  LV            VG  Attr       LSize   Pool Origin Data%  Meta%  Move Log
Cpy%Sync Convert
  data          pve twi-aotz--  18.87g             31.33  1.86

  root          pve -wi-ao----   9.75g

  swap          pve -wi-ao----   4.00g

  vm-100-disk-0 pve Vwi-aotz-- 100.00g data        5.91


No tried to use move_disk

cmd: qm move_disk 100 scsi0 VMS --format qcow2
(VMS is the Directory Storage)

Using this command to check the qcow2 file

cmd: watch -n 1 qemu-img info vm-100-disk-0.qcow2

Every 1.0s: qemu-img info vm-100-disk-0.qcow2
                                      proxmox01: Wed Sep 30 11:02:02 2020

image: vm-100-disk-0.qcow2
file format: qcow2
virtual size: 100 GiB (107374182400 bytes)
disk size: 21.2 GiB
cluster_size: 65536
Format specific information:
    compat: 1.1
    compression type: zlib
    lazy refcounts: false
    refcount bits: 16
    corrupt: false


After a while, all space in /DATA, which is the Directory Storage are full.
df -h
Filesystem            Size  Used Avail Use% Mounted on
udev                  1.9G     0  1.9G   0% /dev
tmpfs                 394M  5.8M  388M   2% /run
/dev/mapper/pve-root  9.8G  2.5G  7.4G  25% /
tmpfs                 2.0G   52M  1.9G   3% /dev/shm
tmpfs                 5.0M     0  5.0M   0% /run/lock
tmpfs                 2.0G     0  2.0G   0% /sys/fs/cgroup
/dev/vdb1              40G   40G  316K 100% /DATA
/dev/fuse              30M   16K   30M   1% /etc/pve
tmpfs                 394M     0  394M   0% /run/user/0

and the image has almost 40G filled....

qemu-img info vm-100-disk-0.qcow2
image: vm-100-disk-0.qcow2
file format: qcow2
virtual size: 100 GiB (107374182400 bytes)
disk size: 39.9 GiB
cluster_size: 65536
Format specific information:
    compat: 1.1
    compression type: zlib
    lazy refcounts: false
    refcount bits: 16
    corrupt: false

And the command qm move_disk got error after a while:

qm move_disk 100 scsi0 VMS --format qcow2
create full clone of drive scsi0 (local-lvm:vm-100-disk-0)
Formatting '/DATA/images/100/vm-100-disk-0.qcow2', fmt=qcow2
cluster_size=65536 preallocation=metadata compression_type=zlib
size=107374182400 lazy_refcounts=off refcount_bits=16
drive mirror is starting for drive-scsi0
drive-scsi0: transferred: 384827392 bytes remaining: 106989355008 bytes
total: 107374182400 bytes progression: 0.36 % busy: 1 ready: 0
...
...
drive-scsi0: transferred: 42833281024 bytes remaining: 64541097984 bytes
total: 107374379008 bytes progression: 39.89 % busy: 1 ready: 0
drive-scsi0: transferred: 42833281024 bytes remaining: 64541097984 bytes
total: 107374379008 bytes progression: 39.89 % busy: 1 ready: 0
drive-scsi0: Cancelling block job
drive-scsi0: Done.
storage migration failed: mirroring error: drive-scsi0: mirroring has been
cancelled

Then I tried to use qemu-img convert and everything works fine

qemu-img convert -O qcow2 /dev/pve/vm-100-disk-0
/DATA/images/100/vm-100-disk-0.qcow2

qemu-img info vm-100-disk-0.qcow2
image: vm-100-disk-0.qcow2
file format: qcow2
virtual size: 100 GiB (107374182400 bytes)
disk size: 6.01 GiB
cluster_size: 65536
Format specific information:
    compat: 1.1
    compression type: zlib
    lazy refcounts: false
    refcount bits: 16
    corrupt: false
---
Gilberto Nunes Ferreira





Em qua., 30 de set. de 2020 às 10:59, Gilberto Nunes <
gilberto.nunes32@gmail.com> escreveu:

> UPDATE
> From CLI I have used
>
> qm move_disk 100 scsi0 VMS --format qcow2
>
> ---
> Gilberto Nunes Ferreira
>
> (47) 3025-5907
> (47) 99676-7530 - Whatsapp / Telegram
>
> Skype: gilberto.nunes36
>
>
>
>
>
> Em qua., 30 de set. de 2020 às 10:26, Gilberto Nunes <
> gilberto.nunes32@gmail.com> escreveu:
>
>> >> How did you move the disk? GUI or CLI?
>> Both.
>> From CLI qm move_disk 100 scsi0 VMS   (VMS is the Directory Storage)
>>
>> Proxmox all up to date...
>> pveversion -v
>> proxmox-ve: 6.2-2 (running kernel: 5.4.65-1-pve)
>> pve-manager: 6.2-12 (running version: 6.2-12/b287dd27)
>> pve-kernel-5.4: 6.2-7
>> pve-kernel-helper: 6.2-7
>> pve-kernel-5.4.65-1-pve: 5.4.65-1
>> pve-kernel-5.4.34-1-pve: 5.4.34-2
>> ceph-fuse: 12.2.11+dfsg1-2.1+b1
>> corosync: 3.0.4-pve1
>> criu: 3.11-3
>> glusterfs-client: 5.5-3
>> ifupdown: 0.8.35+pve1
>> ksm-control-daemon: 1.3-1
>> libjs-extjs: 6.0.1-10
>> libknet1: 1.16-pve1
>> libproxmox-acme-perl: 1.0.5
>> libpve-access-control: 6.1-2
>> libpve-apiclient-perl: 3.0-3
>> libpve-common-perl: 6.2-2
>> libpve-guest-common-perl: 3.1-3
>> libpve-http-server-perl: 3.0-6
>> libpve-storage-perl: 6.2-6
>> libqb0: 1.0.5-1
>> libspice-server1: 0.14.2-4~pve6+1
>> lvm2: 2.03.02-pve4
>> lxc-pve: 4.0.3-1
>> lxcfs: 4.0.3-pve3
>> novnc-pve: 1.1.0-1
>> proxmox-backup-client: 0.8.21-1
>> proxmox-mini-journalreader: 1.1-1
>> proxmox-widget-toolkit: 2.2-12
>> pve-cluster: 6.1-8
>> pve-container: 3.2-2
>> pve-docs: 6.2-6
>> pve-edk2-firmware: 2.20200531-1
>> pve-firewall: 4.1-3
>> pve-firmware: 3.1-3
>> pve-ha-manager: 3.1-1
>> pve-i18n: 2.2-1
>> pve-qemu-kvm: 5.1.0-2
>> pve-xtermjs: 4.7.0-2
>> qemu-server: 6.2-14
>> smartmontools: 7.1-pve2
>> spiceterm: 3.1-1
>> vncterm: 1.6-2
>> zfsutils-linux: 0.8.4-pve1
>>
>>
>> >>> The VM disk (100G) or the physical disk of of the storage?
>>
>> The VM disk has 100G in size, but the storage has 40G... It's just a
>> lab...
>>
>>
>>
>> ---
>> Gilberto Nunes Ferreira
>>
>>
>> Em qua., 30 de set. de 2020 às 10:22, Aaron Lauterer <
>> a.lauterer@proxmox.com> escreveu:
>>
>>> Hey,
>>>
>>> How did you move the disk? GUI or CLI?
>>>
>>> If via CLI, could you post the command?
>>>
>>> Additionally, which versions are installed? (pveversion -v)
>>>
>>> One more question inline.
>>>
>>> On 9/30/20 3:16 PM, Gilberto Nunes wrote:
>>> > Hi all
>>> >
>>> > I tried to move a vm disk from LVM-thin to a Directory Storage but
>>> when I
>>> > did this, the qm move_disk just filled up the entire disk.
>>>
>>> The VM disk (100G) or the physical disk of of the storage?
>>>
>>> > The disk inside LVM-thin has 100G in size but only about 5G is
>>> occupied by
>>> > the OS.
>>> > I have used the qcow2 format.
>>> > However, if I do it from CLI with the command:
>>> >
>>> > qemu-img convert -O qcow2 /dev/pve/vm-100-disk-0
>>> > /DATA/images/100/vm-100-disk-0.qcow2
>>> >
>>> > It works nicely and just copied what the OS occupied inside the VM, but
>>> > created a virtual disk with 100GB.
>>> >
>>> > It's some kind of bug with qm move_disk???
>>> >
>>> > Thanks a lot
>>> >
>>> > ---
>>> > Gilberto Nunes Ferreira
>>> > _______________________________________________
>>> > pve-devel mailing list
>>> > pve-devel@lists.proxmox.com
>>> > https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-devel
>>> >
>>> >
>>>
>>>


      reply	other threads:[~2020-09-30 14:22 UTC|newest]

Thread overview: 5+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2020-09-30 13:16 Gilberto Nunes
2020-09-30 13:22 ` Aaron Lauterer
2020-09-30 13:26   ` Gilberto Nunes
2020-09-30 13:59     ` Gilberto Nunes
2020-09-30 14:21       ` Gilberto Nunes [this message]

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=CAOKSTBuwgrKE1VKSmngoPOM1wmwUUGyfh_XhRMFg-h60xnHiZQ@mail.gmail.com \
    --to=gilberto.nunes32@gmail.com \
    --cc=a.lauterer@proxmox.com \
    --cc=pve-devel@lists.proxmox.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.
Service provided by Proxmox Server Solutions GmbH | Privacy | Legal