From: Iztok Gregori <iztok.gregori@elettra.eu>
To: pve-user@lists.proxmox.com
Subject: Re: [PVE-User] Cannot list disks from an external CEPH pool
Date: Wed, 1 Jun 2022 12:30:29 +0200 [thread overview]
Message-ID: <28938ffa-9e73-1076-d1c3-10157d1d6618@elettra.eu> (raw)
In-Reply-To: <2757d9cfb5e8f35341b599493fb12a81@verdnatura.es>
On 01/06/22 12:11, nada wrote:
> hello, just correct IP address at pool1
> 1172.16.1.2 probably 172.16.1.2
1172.16.1 it was just a typo when i edited out the real addresses, the
monitor IPs are the same on both pools.
> and you may enforce access by krbd 1
What improvement should I see with enabling krbd in regards to my
original question?
> and simplify your list command by symlinks
CEPH CLI administration is done on a different, not proxmox, node. It
was my understanding that proxmox doesn't need ceph.conf file to access
an external CEPH cluster, so I never created any configuration/symlinks.
Am I missing something?
Cheers
Iztok
> example
>
> rbd: pool1
> content images
> krbd 1
> monhost 172.16.1.1,172.16.1.2,172.16.1.3
> pool pool1
> username admin
>
> # la /etc/ceph/
> total 20
> drwxr-xr-x 2 root root 7 Mar 6 13:03 .
> drwxr-xr-x 97 root root 193 May 19 03:10 ..
> lrwxrwxrwx 1 root root 27 Aug 4 2021 rbd.conf ->
> /etc/pve/priv/ceph/rbd.conf
> lrwxrwxrwx 1 root root 30 Aug 4 2021 rbd.keyring ->
> /etc/pve/priv/ceph/rbd.keyring
> -rw-r--r-- 1 root root 92 Aug 28 2019 rbdmap
> lrwxrwxrwx 1 root root 31 Feb 2 12:37 rbd_ssd.conf ->
> /etc/pve/priv/ceph/rbd_ssd.conf
> lrwxrwxrwx 1 root root 34 Feb 2 12:37 rbd_ssd.keyring ->
> /etc/pve/priv/ceph/rbd_ssd.keyring
>
> # pvesm list rbd
> Volid Format Type Size VMID
> rbd:vm-105-disk-0 raw images 42949672960 105
> rbd:vm-111-disk-0 raw images 42949672960 111
>
> # pvesm list rbd_ssd
> Volid Format Type Size VMID
> rbd_ssd:vm-102-disk-0 raw images 6442450944 102
> rbd_ssd:vm-103-disk-0 raw images 4294967296 103
>
> good luck
> Nada
>
>
> On 2022-06-01 11:51, Iztok Gregori wrote:
>> On 01/06/22 11:29, Aaron Lauterer wrote:
>>> Do you get additional errors if you run the following command?
>>> Assuming that the storage is also called pool1.
>>>
>>> pvesm list pool1
>>
>> No additional errors:
>>
>> root@pmx-14:~# pvesm list pool1
>> rbd error: rbd: listing images failed: (2) No such file or directory
>>
>>
>>
>>> Do you have VMs with disk images on that storage? If so, do they
>>> start normally?
>>
>> Yes, we have a lot of VMs with disk on that storage and yes they seems
>> to start normally (last start yesterday when we first notice the GUI
>> behaviour)
>>
>>>
>>> Can you show the configuration of that storage and the one of the
>>> working pool? (/etc/pve/storage.cfg)
>>
>> Sure (edited the IP addresses and pool names):
>>
>> [cit /etc/pve/storage.cfg]
>> ...
>> rbd: pool1
>> content images
>> monhost 172.16.1.1;1172.16.1.2;172.16.1.3
>> pool pool1
>> username admin
>>
>> rbd: pool2
>> content images
>> monhost 172.16.1.1;172.16.1.2;172.16.1.3
>> pool pool2
>> username admin
>> ...
>> [/cit]
>>
>> Thanks!
>>
>> Iztok
>>
>>>
>>> On 6/1/22 11:13, Iztok Gregori wrote:
>>>> Hi to all!
>>>>
>>>> I have a Proxmox cluster (7.1) connected to an external CEPH cluster
>>>> (octopus). From the GUI I cannot list the content (disks) of one
>>>> pool (but I'm able to list all the other pools):
>>>>
>>>> rbd error: rbd: listing images failed: (2) No such file or directory
>>>> (500)
>>>>
>>>> The pveproxy/access.log shows the error for "pool1":
>>>>
>>>> "GET /api2/json/nodes/pmx-14/storage/pool1/content?content=images
>>>> HTTP/1.1" 500 13
>>>>
>>>> but when I try another pool ("pool2") it works:
>>>>
>>>> "GET /api2/json/nodes/pmx-14/storage/pool2/content?content=images
>>>> HTTP/1.1" 200 841
>>>>
>>>> From the command line "rbd ls pool1" is working fine (because I
>>>> don't have a ceph.conf I ran it with "rbd -m 172.16.1.1 --keyring
>>>> /etc/pve/priv/ceph/pool1.keyring ls pool1") and I see the pool
>>>> contents.
>>>>
>>>> The cluster is running fine, the VMs access the disks on that pool
>>>> without a problem
>>>>
>>>> What can it be?
>>>>
>>>> The cluster is a mix of freshly installed nodes and upgraded ones,
>>>> all the 17 nodes (but one which is 6.4 but without any running VMs)
>>>> are running:
>>>>
>>>> root@pmx-14:~# pveversion -v
>>>> proxmox-ve: 7.1-1 (running kernel: 5.13.19-6-pve)
>>>> pve-manager: 7.1-12 (running version: 7.1-12/b3c09de3)
>>>> pve-kernel-helper: 7.1-14
>>>> pve-kernel-5.13: 7.1-9
>>>> pve-kernel-5.13.19-6-pve: 5.13.19-15
>>>> pve-kernel-5.13.19-2-pve: 5.13.19-4
>>>> ceph-fuse: 15.2.15-pve1
>>>> corosync: 3.1.5-pve2
>>>> criu: 3.15-1+pve-1
>>>> glusterfs-client: 9.2-1
>>>> ifupdown2: 3.1.0-1+pmx3
>>>> ksm-control-daemon: 1.4-1
>>>> libjs-extjs: 7.0.0-1
>>>> libknet1: 1.22-pve2
>>>> libproxmox-acme-perl: 1.4.1
>>>> libproxmox-backup-qemu0: 1.2.0-1
>>>> libpve-access-control: 7.1-7
>>>> libpve-apiclient-perl: 3.2-1
>>>> libpve-common-perl: 7.1-5
>>>> libpve-guest-common-perl: 4.1-1
>>>> libpve-http-server-perl: 4.1-1
>>>> libpve-storage-perl: 7.1-1
>>>> libspice-server1: 0.14.3-2.1
>>>> lvm2: 2.03.11-2.1
>>>> lxc-pve: 4.0.11-1
>>>> lxcfs: 4.0.11-pve1
>>>> novnc-pve: 1.3.0-2
>>>> proxmox-backup-client: 2.1.5-1
>>>> proxmox-backup-file-restore: 2.1.5-1
>>>> proxmox-mini-journalreader: 1.3-1
>>>> proxmox-widget-toolkit: 3.4-7
>>>> pve-cluster: 7.1-3
>>>> pve-container: 4.1-4
>>>> pve-docs: 7.1-2
>>>> pve-edk2-firmware: 3.20210831-2
>>>> pve-firewall: 4.2-5
>>>> pve-firmware: 3.3-6
>>>> pve-ha-manager: 3.3-3
>>>> pve-i18n: 2.6-2
>>>> pve-qemu-kvm: 6.1.1-2
>>>> pve-xtermjs: 4.16.0-1
>>>> qemu-server: 7.1-4
>>>> smartmontools: 7.2-1
>>>> spiceterm: 3.2-2
>>>> swtpm: 0.7.1~bpo11+1
>>>> vncterm: 1.7-1
>>>> zfsutils-linux: 2.1.4-pve1
>>>>
>>>> I can provide other information if it's needed.
>>>>
>>>> Cheers
>>>> Iztok Gregori
>>>>
>>>>
>>>
>>
>>
>> _______________________________________________
>> pve-user mailing list
>> pve-user@lists.proxmox.com
>> https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-user
>
> _______________________________________________
> pve-user mailing list
> pve-user@lists.proxmox.com
> https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-user
prev parent reply other threads:[~2022-06-01 10:30 UTC|newest]
Thread overview: 5+ messages / expand[flat|nested] mbox.gz Atom feed top
2022-06-01 9:13 Iztok Gregori
2022-06-01 9:29 ` Aaron Lauterer
2022-06-01 9:51 ` Iztok Gregori
2022-06-01 10:11 ` nada
2022-06-01 10:30 ` Iztok Gregori [this message]
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=28938ffa-9e73-1076-d1c3-10157d1d6618@elettra.eu \
--to=iztok.gregori@elettra.eu \
--cc=pve-user@lists.proxmox.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox