From: "Сергей Цаболов" <tsabolov@t8.ru>
To: uwe.sauter.de@gmail.com,
Proxmox VE user list <pve-user@lists.proxmox.com>
Subject: Re: [PVE-User] [ceph-users] Re: Ceph Usage web and terminal.
Date: Wed, 29 Dec 2021 17:06:16 +0300 [thread overview]
Message-ID: <131ea5ec-89df-4c90-5808-451c33abbb05@t8.ru> (raw)
In-Reply-To: <550c21eb-5371-6f3e-f1f4-bccbc6b5384b@gmail.com>
Ok, I understand the case.
29.12.2021 16:13, Uwe Sauter пишет:
> Am 29.12.21 um 13:51 schrieb Сергей Цаболов:
>> Hi, Uwe
>>
>> 29.12.2021 14:16, Uwe Sauter пишет:
>>> Just a feeling but I'd say that the imbalance in OSDs (one host having many more disks than the
>>> rest) is your problem.
>> Yes, last node in cluster have more disk then the rest, but
>>
>> one disk is 12TB and all others 9 HD is 1TB
>>
>>> Assuming that your configuration keeps 3 copies of each VM image then the imbalance probably means
>>> that 2 of these 3 copies reside on pve-3111 and if this host is unavailable, all VM images with 2
>>> copies on that host become unresponsive, too.
>> In Proxmox web ceph pool I set the Size: 2 , Min.Size: 2
>>
> So this means that you want to have 2 copies in the regular case (size) and also 2 copies in the
> failure case (min size) so that the VMs stay available.
Yes I think before like you answer, but is not so worked.
>
> So you might solve your problem by decreasing min size to 1 (dangerous!!) or by increasing size to
> 3, which means that in the regular case you will have 3 copies but if only 2 are available, it will
> still work and re-sync the 3rd copy once it comes online again.
I understand if decreasing min.size to 1 is very (dangerous!!!)
If I increasing to 3 min.size keep 2 is default .
But I'm afraid if set the 3/2 (good choice) MAX AVAIL in pool is will
decrease in two or more space, or am I wrong?
For now I have with all disk :
CLASS SIZE AVAIL USED RAW USED %RAW USED
hdd `106 TiB 99 TiB 7.7 TiB 7.7 TiB 7.26
TOTAL 106 TiB 99 TiB 7.7 TiB 7.7 TiB 7.26
--- POOLS ---
POOL ID PGS STORED OBJECTS
USED %USED MAX AVAIL
device_health_metrics 1 1 8.3 MiB 22 17 MiB
0 44 TiB
vm.pool 2 1024 3.0 TiB 864.55k
6.0 TiB 6.39 44 TiB ( terminal 44 TiB = 48.37 ) in web I
see 51.50 TB
cephfs_data 3 32 874 GiB 223.76k 1.7
TiB 1.91 44 TiB
cephfs_metadata 4 32 25 MiB 27 51 MiB
0 44 TiB
Am I right in my reasoning ?
Thank you!
>
>> With : ceph osd map vm.pool object-name (vm ID) I see some of vm object one copy is on osd.12,
>> example :
>>
>> osdmap e14321 pool 'vm.pool' (2) object '114' -> pg 2.10486407 (2.7) -> up ([12,8], p12) acting
>> ([12,8], p12)
>>
>> But this example :
>>
>> osdmap e14321 pool 'vm.pool' (2) object '113' -> pg 2.8bd09f6d (2.36d) -> up ([10,7], p10) acting
>> ([10,7], p10)
>>
>> osd.10 and osd.7
>>
>>> Check your failure domain for Ceph and possibly change it from OSD to host. This should prevent that
>>> one host holds multiple copies of a VM image.
>> I didn 't understand a little what to check ?
>>
>> Can you explain me with example?
>>
> I don't have an example but you can read about the concept at:
>
> https://docs.ceph.com/en/latest/rados/operations/crush-map/#crush-maps
>
>
> Regards,
>
> Uwe
>
>
>
>>>
>>> Regards,
>>>
>>> Uwe
>>>
>>> Am 29.12.21 um 09:36 schrieb Сергей Цаболов:
>>>> Hello to all.
>>>>
>>>> In my case I have the 7 node cluster Proxmox and working Ceph (ceph version 15.2.15 octopus
>>>> (stable)": 7)
>>>>
>>>> Ceph HEALTH_OK
>>>>
>>>> ceph -s
>>>> cluster:
>>>> id: 9662e3fa-4ce6-41df-8d74-5deaa41a8dde
>>>> health: HEALTH_OK
>>>>
>>>> services:
>>>> mon: 7 daemons, quorum pve-3105,pve-3107,pve-3108,pve-3103,pve-3101,pve-3111,pve-3109 (age 17h)
>>>> mgr: pve-3107(active, since 41h), standbys: pve-3109, pve-3103, pve-3105, pve-3101, pve-3111,
>>>> pve-3108
>>>> mds: cephfs:1 {0=pve-3105=up:active} 6 up:standby
>>>> osd: 22 osds: 22 up (since 17h), 22 in (since 17h)
>>>>
>>>> task status:
>>>>
>>>> data:
>>>> pools: 4 pools, 1089 pgs
>>>> objects: 1.09M objects, 4.1 TiB
>>>> usage: 7.7 TiB used, 99 TiB / 106 TiB avail
>>>> pgs: 1089 active+clean
>>>>
>>>> ---------------------------------------------------------------------------------------------------------------------
>>>>
>>>>
>>>>
>>>> ceph osd tree
>>>>
>>>> ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF
>>>> -1 106.43005 root default
>>>> -13 14.55478 host pve-3101
>>>> 10 hdd 7.27739 osd.10 up 1.00000 1.00000
>>>> 11 hdd 7.27739 osd.11 up 1.00000 1.00000
>>>> -11 14.55478 host pve-3103
>>>> 8 hdd 7.27739 osd.8 up 1.00000 1.00000
>>>> 9 hdd 7.27739 osd.9 up 1.00000 1.00000
>>>> -3 14.55478 host pve-3105
>>>> 0 hdd 7.27739 osd.0 up 1.00000 1.00000
>>>> 1 hdd 7.27739 osd.1 up 1.00000 1.00000
>>>> -5 14.55478 host pve-3107
>>>> 2 hdd 7.27739 osd.2 up 1.00000 1.00000
>>>> 3 hdd 7.27739 osd.3 up 1.00000 1.00000
>>>> -9 14.55478 host pve-3108
>>>> 6 hdd 7.27739 osd.6 up 1.00000 1.00000
>>>> 7 hdd 7.27739 osd.7 up 1.00000 1.00000
>>>> -7 14.55478 host pve-3109
>>>> 4 hdd 7.27739 osd.4 up 1.00000 1.00000
>>>> 5 hdd 7.27739 osd.5 up 1.00000 1.00000
>>>> -15 19.10138 host pve-3111
>>>> 12 hdd 10.91409 osd.12 up 1.00000 1.00000
>>>> 13 hdd 0.90970 osd.13 up 1.00000 1.00000
>>>> 14 hdd 0.90970 osd.14 up 1.00000 1.00000
>>>> 15 hdd 0.90970 osd.15 up 1.00000 1.00000
>>>> 16 hdd 0.90970 osd.16 up 1.00000 1.00000
>>>> 17 hdd 0.90970 osd.17 up 1.00000 1.00000
>>>> 18 hdd 0.90970 osd.18 up 1.00000 1.00000
>>>> 19 hdd 0.90970 osd.19 up 1.00000 1.00000
>>>> 20 hdd 0.90970 osd.20 up 1.00000 1.00000
>>>> 21 hdd 0.90970 osd.21 up 1.00000 1.00000
>>>>
>>>> ---------------------------------------------------------------------------------------------------------------
>>>>
>>>>
>>>>
>>>> POOL ID PGS STORED OBJECTS USED %USED MAX AVAIL
>>>> vm.pool 2 1024 3.0 TiB 863.31k 6.0 TiB 6.38 44 TiB (this pool
>>>> have the all VM disk)
>>>>
>>>> ---------------------------------------------------------------------------------------------------------------
>>>>
>>>>
>>>>
>>>> ceph osd map vm.pool vm.pool.object
>>>> osdmap e14319 pool 'vm.pool' (2) object 'vm.pool.object' -> pg 2.196f68d5 (2.d5) -> up ([2,4], p2)
>>>> acting ([2,4], p2)
>>>>
>>>> -----------------------------------------------------------------------------------------------------------------------------------------------------------------------
>>>>
>>>>
>>>> pveversion -v
>>>> proxmox-ve: 6.4-1 (running kernel: 5.4.143-1-pve)
>>>> pve-manager: 6.4-13 (running version: 6.4-13/9f411e79)
>>>> pve-kernel-helper: 6.4-8
>>>> pve-kernel-5.4: 6.4-7
>>>> pve-kernel-5.4.143-1-pve: 5.4.143-1
>>>> pve-kernel-5.4.106-1-pve: 5.4.106-1
>>>> ceph: 15.2.15-pve1~bpo10
>>>> ceph-fuse: 15.2.15-pve1~bpo10
>>>> corosync: 3.1.2-pve1
>>>> criu: 3.11-3
>>>> glusterfs-client: 5.5-3
>>>> ifupdown: residual config
>>>> ifupdown2: 3.0.0-1+pve4~bpo10
>>>> ksm-control-daemon: 1.3-1
>>>> libjs-extjs: 6.0.1-10
>>>> libknet1: 1.22-pve1~bpo10+1
>>>> libproxmox-acme-perl: 1.1.0
>>>> libproxmox-backup-qemu0: 1.1.0-1
>>>> libpve-access-control: 6.4-3
>>>> libpve-apiclient-perl: 3.1-3
>>>> libpve-common-perl: 6.4-4
>>>> libpve-guest-common-perl: 3.1-5
>>>> libpve-http-server-perl: 3.2-3
>>>> libpve-storage-perl: 6.4-1
>>>> libqb0: 1.0.5-1
>>>> libspice-server1: 0.14.2-4~pve6+1
>>>> lvm2: 2.03.02-pve4
>>>> lxc-pve: 4.0.6-2
>>>> lxcfs: 4.0.6-pve1
>>>> novnc-pve: 1.1.0-1
>>>> proxmox-backup-client: 1.1.13-2
>>>> proxmox-mini-journalreader: 1.1-1
>>>> proxmox-widget-toolkit: 2.6-1
>>>> pve-cluster: 6.4-1
>>>> pve-container: 3.3-6
>>>> pve-docs: 6.4-2
>>>> pve-edk2-firmware: 2.20200531-1
>>>> pve-firewall: 4.1-4
>>>> pve-firmware: 3.3-2
>>>> pve-ha-manager: 3.1-1
>>>> pve-i18n: 2.3-1
>>>> pve-qemu-kvm: 5.2.0-6
>>>> pve-xtermjs: 4.7.0-3
>>>> qemu-server: 6.4-2
>>>> smartmontools: 7.2-pve2
>>>> spiceterm: 3.1-1
>>>> vncterm: 1.6-2
>>>> zfsutils-linux: 2.0.6-pve1~bpo10+1
>>>>
>>>> -----------------------------------------------------------------------------------------------------------------------------------------------------------------------
>>>>
>>>>
>>>>
>>>> And now my problem:
>>>>
>>>> For all VM I have one pool for VM disks
>>>>
>>>> When node/host pve-3111 is shutdown in many of other nodes/hosts pve-3107, pve-3105 VM not
>>>> shutdown but not available in network.
>>>>
>>>> After the node/host is up Ceph back to HEALTH_OK and the all VM back to access in Network (without
>>>> reboot).
>>>>
>>>> Can some one to suggest me what I can to check in Ceph ?
>>>>
>>>> Thanks.
>>>>
>
--
-------------------------
С уважением
Сергей Цаболов,
Системный администратор
ООО "Т8"
Тел.: +74992716161,
Моб: +79850334875
tsabolov@t8.ru
ООО «Т8», 107076, г. Москва, Краснобогатырская ул., д. 44, стр.1
www.t8.ru
next prev parent reply other threads:[~2021-12-29 14:06 UTC|newest]
Thread overview: 6+ messages / expand[flat|nested] mbox.gz Atom feed top
[not found] <e6b7d3f3-d6ce-ef41-cfa8-36b011243ebc@t8.ru>
[not found] ` <6f23d719-1931-cc81-899d-3202047c4a56@binovo.es>
[not found] ` <101971ad-519a-9af2-249e-433df28b1f1a@t8.ru>
2021-12-29 8:36 ` Сергей Цаболов
2021-12-29 11:16 ` Uwe Sauter
2021-12-29 12:51 ` Сергей Цаболов
2021-12-29 13:13 ` Uwe Sauter
2021-12-29 14:06 ` Сергей Цаболов [this message]
2021-12-29 14:13 ` Uwe Sauter
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=131ea5ec-89df-4c90-5808-451c33abbb05@t8.ru \
--to=tsabolov@t8.ru \
--cc=pve-user@lists.proxmox.com \
--cc=uwe.sauter.de@gmail.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.