* [PVE-User] Python problem with upgrade to proxmoxVE6.3/ CEPH Nautilus 14.2.15 @ 2020-11-27 11:47 Jean-Luc Oms 2020-11-27 13:07 ` Lindsay Mathieson ` (2 more replies) 0 siblings, 3 replies; 11+ messages in thread From: Jean-Luc Oms @ 2020-11-27 11:47 UTC (permalink / raw) To: Proxmox VE user list Bonjour, Upgrading to last Proxmox VE / Ceph nautilus from the last 6.2 proxmox VE seems to introduce a python 2/3 version problem, dashboard healt stops working. root@ceph1:/usr/bin# ceph health HEALTH_ERR Module 'dashboard' has failed: ('invalid syntax', ('/usr/share/ceph/mgr/dashboard/controllers/orchestrator.py', 34, 11, ' result: dict = {}\n')) This syntax was introduced in 3.6, and using strace it seems python 2.7 is used. Any option to resolve this ? everything was ok in 6.2-15. Thanks -- Jean-Luc Oms /STI-ReseauX <https://rx.lirmm.fr>- LIRMM - CNRS/UM/ +33 4 67 41 85 93 <tel:+33-467-41-85-93> / +33 6 32 01 04 17 <tel:+33-632-01-04-17> ^ permalink raw reply [flat|nested] 11+ messages in thread
* Re: [PVE-User] Python problem with upgrade to proxmoxVE6.3/ CEPH Nautilus 14.2.15 2020-11-27 11:47 [PVE-User] Python problem with upgrade to proxmoxVE6.3/ CEPH Nautilus 14.2.15 Jean-Luc Oms @ 2020-11-27 13:07 ` Lindsay Mathieson [not found] ` <47b5a337-b2ca-ce6d-37c5-e904db8d6e03@univ-fcomte.fr> 2020-11-27 15:15 ` Jean-Luc Oms 2020-11-27 16:45 ` Marco M. Gabriel 2 siblings, 1 reply; 11+ messages in thread From: Lindsay Mathieson @ 2020-11-27 13:07 UTC (permalink / raw) To: pve-user On 27/11/2020 9:47 pm, Jean-Luc Oms wrote: > Upgrading to last Proxmox VE / Ceph nautilus from the last 6.2 proxmox > VE seems to introduce a python 2/3 version problem, dashboard healt > stops working. Was just about to report that :) Same here. -- Lindsay ^ permalink raw reply [flat|nested] 11+ messages in thread
[parent not found: <47b5a337-b2ca-ce6d-37c5-e904db8d6e03@univ-fcomte.fr>]
* Re: [PVE-User] Python problem with upgrade to proxmoxVE6.3/ CEPH Nautilus 14.2.15 [not found] ` <47b5a337-b2ca-ce6d-37c5-e904db8d6e03@univ-fcomte.fr> @ 2020-11-27 16:59 ` alexandre derumier 2020-11-27 17:12 ` [PVE-User] ProxmoxVE6.3/CEPH Octopus 15.2.6 was " Jean-Daniel TISSOT 2020-11-27 21:17 ` [PVE-User] " Lindsay Mathieson 2020-11-27 17:06 ` Marco M. Gabriel 1 sibling, 2 replies; 11+ messages in thread From: alexandre derumier @ 2020-11-27 16:59 UTC (permalink / raw) To: pve-user >>1 pools have too many placement groups Pool rbd has 128 placement groups, should have 32 >> >>I don't find any way to reduce placement groups tu 32 >> >>Any help welcome. you can't reduce PG on nautilus, only since octopus. (and ceph is able to do it automataaly with new pg autoscaler) I think it's a warning introduced in last nautilus update. If I remember, they are an option to disable this warning, (but I don't remember it) On 27/11/2020 17:29, Jean-Daniel TISSOT wrote: > Hi, > > I have another problem > > root@dmz-pve1:~ # ceph health HEALTH_WARN 1 pools have too many > placement groups root@dmz-pve1:~ # pveceph pool ls > ┌───────────────────────┬──────┬──────────┬────────┬───────────────────┬─────────────────┬──────────────────────┬──────────────┐ > │ Name │ Size │ Min Size │ PG Num │ PG Autoscale Mode > │ Crush Rule Name │ %-Used │ Used │ > ╞═══════════════════════╪══════╪══════════╪════════╪═══════════════════╪═════════════════╪══════════════════════╪══════════════╡ > │ device_health_metrics │ 3 │ 2 │ 1 │ on > │ replicated_rule │ 4.19273845864154e-07 │ 4534827 │ > ├───────────────────────┼──────┼──────────┼────────┼───────────────────┼─────────────────┼──────────────────────┼──────────────┤ > │ rbd │ 3 │ 2 │ 128 │ warn > │ replicated_rule │ 0.0116069903597236 │ 127014329075 │ > └───────────────────────┴──────┴──────────┴────────┴───────────────────┴─────────────────┴──────────────────────┴──────────────┘ > > > In the GUI : > > 1 pools have too many placement groups Pool rbd has 128 placement > groups, should have 32 > > I don't find any way to reduce placement groups tu 32 > > Any help welcome. > > Best regards, > > Le 27/11/2020 à 14:07, Lindsay Mathieson a écrit : >> On 27/11/2020 9:47 pm, Jean-Luc Oms wrote: >>> Upgrading to last Proxmox VE / Ceph nautilus from the last 6.2 proxmox >>> VE seems to introduce a python 2/3 version problem, dashboard healt >>> stops working. >> >> Was just about to report that :) >> >> >> Same here. >> ^ permalink raw reply [flat|nested] 11+ messages in thread
* [PVE-User] ProxmoxVE6.3/CEPH Octopus 15.2.6 was Python problem with upgrade to proxmoxVE6.3/ CEPH Nautilus 14.2.15 2020-11-27 16:59 ` alexandre derumier @ 2020-11-27 17:12 ` Jean-Daniel TISSOT 2020-11-27 21:17 ` [PVE-User] " Lindsay Mathieson 1 sibling, 0 replies; 11+ messages in thread From: Jean-Daniel TISSOT @ 2020-11-27 17:12 UTC (permalink / raw) To: pve-user Sorry to steal the thread. In fact before upgrading to Octopus, I don't remember any warning. I upgrade Proxmox an follow the wiki to upgrade CEPH. All things seam working (I don't test HA by migrate a VM work perfectly. I have just the warning on pool rdb (1 pools have too many placement groups Pool rbd has 128 placement groups, should have 32) Le 27/11/2020 à 17:59, alexandre derumier a écrit : > >>1 pools have too many placement groups Pool rbd has 128 placement > groups, should have 32 > >> > >>I don't find any way to reduce placement groups tu 32 > >> > >>Any help welcome. > > you can't reduce PG on nautilus, only since octopus. (and ceph is > able to do it automataaly with new pg autoscaler) > > I think it's a warning introduced in last nautilus update. > > If I remember, they are an option to disable this warning, (but I > don't remember it) > > > On 27/11/2020 17:29, Jean-Daniel TISSOT wrote: >> Hi, >> >> I have another problem >> >> root@dmz-pve1:~ # ceph health HEALTH_WARN 1 pools have too many >> placement groups root@dmz-pve1:~ # pveceph pool ls >> ┌───────────────────────┬──────┬──────────┬────────┬───────────────────┬─────────────────┬──────────────────────┬──────────────┐ >> │ Name │ Size │ Min Size │ PG Num │ PG Autoscale >> Mode │ Crush Rule Name │ %-Used │ Used │ >> ╞═══════════════════════╪══════╪══════════╪════════╪═══════════════════╪═════════════════╪══════════════════════╪══════════════╡ >> │ device_health_metrics │ 3 │ 2 │ 1 │ >> on │ replicated_rule │ 4.19273845864154e-07 │ 4534827 >> │ >> ├───────────────────────┼──────┼──────────┼────────┼───────────────────┼─────────────────┼──────────────────────┼──────────────┤ >> │ rbd │ 3 │ 2 │ 128 │ >> warn │ replicated_rule │ 0.0116069903597236 │ >> 127014329075 │ >> └───────────────────────┴──────┴──────────┴────────┴───────────────────┴─────────────────┴──────────────────────┴──────────────┘ >> >> >> In the GUI : >> >> 1 pools have too many placement groups Pool rbd has 128 placement >> groups, should have 32 >> >> I don't find any way to reduce placement groups tu 32 >> >> Any help welcome. >> >> Best regards, >> >> Le 27/11/2020 à 14:07, Lindsay Mathieson a écrit : >>> On 27/11/2020 9:47 pm, Jean-Luc Oms wrote: >>>> Upgrading to last Proxmox VE / Ceph nautilus from the last 6.2 proxmox >>>> VE seems to introduce a python 2/3 version problem, dashboard healt >>>> stops working. >>> >>> Was just about to report that :) >>> >>> >>> Same here. >>> > > > _______________________________________________ > pve-user mailing list > pve-user@lists.proxmox.com > https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-user -- Bien cordialement, Jean-Daniel Tissot - IE CNRS http://chrono-environnement.univ-fcomte.fr UMR 6249 - Laboratoire Chrono-environnement UMR CNRS-UFC Université de Franche-Comté, 16 route de Gray, 25030 Besançon Cedex, FRANCE Jean-Daniel.Tissot@univ-fcomte.fr tel:+33 3 81 666 440 Alabama, Mississippi, Minnesota, South Carolina, Oregon... not so sweet home Black Panther Party, renaît de tes cendres et reviens les aider https://www.youtube.com/watch?v=ZvilFSMVHTs ^ permalink raw reply [flat|nested] 11+ messages in thread
* Re: [PVE-User] Python problem with upgrade to proxmoxVE6.3/ CEPH Nautilus 14.2.15 2020-11-27 16:59 ` alexandre derumier 2020-11-27 17:12 ` [PVE-User] ProxmoxVE6.3/CEPH Octopus 15.2.6 was " Jean-Daniel TISSOT @ 2020-11-27 21:17 ` Lindsay Mathieson 1 sibling, 0 replies; 11+ messages in thread From: Lindsay Mathieson @ 2020-11-27 21:17 UTC (permalink / raw) To: pve-user On 28/11/2020 2:59 am, alexandre derumier wrote: > you can't reduce PG on nautilus, only since octopus. (and ceph is > able to do it automataaly with new pg autoscaler) Actually you can, has been the case since Nautilus. https://ceph.io/rados/new-in-nautilus-pg-merging-and-autotuning/ -- Lindsay ^ permalink raw reply [flat|nested] 11+ messages in thread
* Re: [PVE-User] Python problem with upgrade to proxmoxVE6.3/ CEPH Nautilus 14.2.15 [not found] ` <47b5a337-b2ca-ce6d-37c5-e904db8d6e03@univ-fcomte.fr> 2020-11-27 16:59 ` alexandre derumier @ 2020-11-27 17:06 ` Marco M. Gabriel 2020-11-27 17:08 ` alexandre derumier 2020-11-27 17:18 ` Jean-Daniel TISSOT 1 sibling, 2 replies; 11+ messages in thread From: Marco M. Gabriel @ 2020-11-27 17:06 UTC (permalink / raw) To: Proxmox VE user list You can enable the pg-autoscaler. It does the work for you and optimizes the number of placement groups on a given pool. The pg-autoscaler was introduced with nautilus and we ran it for a while without any problems. Here is an explanation of how to enable and how to use the autoscaler: https://ceph.io/rados/new-in-nautilus-pg-merging-and-autotuning/ Best regards, Marco Am Fr., 27. Nov. 2020 um 17:37 Uhr schrieb Jean-Daniel TISSOT <Jean-Daniel.Tissot@univ-fcomte.fr>: > > Hi, > > I have another problem > > root@dmz-pve1:~ # ceph health HEALTH_WARN 1 pools have too many > placement groups root@dmz-pve1:~ # pveceph pool ls > ┌───────────────────────┬──────┬──────────┬────────┬───────────────────┬─────────────────┬──────────────────────┬──────────────┐ > │ Name │ Size │ Min Size │ PG Num │ PG Autoscale Mode │ > Crush Rule Name │ %-Used │ Used │ > ╞═══════════════════════╪══════╪══════════╪════════╪═══════════════════╪═════════════════╪══════════════════════╪══════════════╡ > │ device_health_metrics │ 3 │ 2 │ 1 │ on │ > replicated_rule │ 4.19273845864154e-07 │ 4534827 │ > ├───────────────────────┼──────┼──────────┼────────┼───────────────────┼─────────────────┼──────────────────────┼──────────────┤ > │ rbd │ 3 │ 2 │ 128 │ warn │ > replicated_rule │ 0.0116069903597236 │ 127014329075 │ > └───────────────────────┴──────┴──────────┴────────┴───────────────────┴─────────────────┴──────────────────────┴──────────────┘ > > > In the GUI : > > 1 pools have too many placement groups Pool rbd has 128 placement > groups, should have 32 > > I don't find any way to reduce placement groups tu 32 > > Any help welcome. > > Best regards, > > Le 27/11/2020 à 14:07, Lindsay Mathieson a écrit : > > On 27/11/2020 9:47 pm, Jean-Luc Oms wrote: > >> Upgrading to last Proxmox VE / Ceph nautilus from the last 6.2 proxmox > >> VE seems to introduce a python 2/3 version problem, dashboard healt > >> stops working. > > > > Was just about to report that :) > > > > > > Same here. > > > -- > Bien cordialement, > Jean-Daniel Tissot - IE CNRS http://chrono-environnement.univ-fcomte.fr > UMR 6249 - Laboratoire Chrono-environnement UMR CNRS-UFC > Université de Franche-Comté, 16 route de Gray, 25030 Besançon Cedex, FRANCE > Jean-Daniel.Tissot@univ-fcomte.fr tel:+33 3 81 666 440 > > _______________________________________________ > pve-user mailing list > pve-user@lists.proxmox.com > https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-user ^ permalink raw reply [flat|nested] 11+ messages in thread
* Re: [PVE-User] Python problem with upgrade to proxmoxVE6.3/ CEPH Nautilus 14.2.15 2020-11-27 17:06 ` Marco M. Gabriel @ 2020-11-27 17:08 ` alexandre derumier 2020-11-27 17:18 ` Jean-Daniel TISSOT 1 sibling, 0 replies; 11+ messages in thread From: alexandre derumier @ 2020-11-27 17:08 UTC (permalink / raw) To: pve-user On 27/11/2020 18:06, Marco M. Gabriel wrote: > You can enable the pg-autoscaler. It does the work for you and > optimizes the number of placement groups on a given pool. Oh, yes indeed, I thinked it was introduced in Octopus, but it's indeed already avaible in Nautilus :) ^ permalink raw reply [flat|nested] 11+ messages in thread
* Re: [PVE-User] Python problem with upgrade to proxmoxVE6.3/ CEPH Nautilus 14.2.15 2020-11-27 17:06 ` Marco M. Gabriel 2020-11-27 17:08 ` alexandre derumier @ 2020-11-27 17:18 ` Jean-Daniel TISSOT 1 sibling, 0 replies; 11+ messages in thread From: Jean-Daniel TISSOT @ 2020-11-27 17:18 UTC (permalink / raw) To: pve-user Many thanks. Work like a charm. No more warning. Again many thanks Marco Best regards, Jean-Daniel Le 27/11/2020 à 18:06, Marco M. Gabriel a écrit : > You can enable the pg-autoscaler. It does the work for you and > optimizes the number of placement groups on a given pool. > > The pg-autoscaler was introduced with nautilus and we ran it for a > while without any problems. > > Here is an explanation of how to enable and how to use the autoscaler: > https://ceph.io/rados/new-in-nautilus-pg-merging-and-autotuning/ > > Best regards, > Marco > > > Am Fr., 27. Nov. 2020 um 17:37 Uhr schrieb Jean-Daniel TISSOT > <Jean-Daniel.Tissot@univ-fcomte.fr>: >> Hi, >> >> I have another problem >> >> root@dmz-pve1:~ # ceph health HEALTH_WARN 1 pools have too many >> placement groups root@dmz-pve1:~ # pveceph pool ls >> ┌───────────────────────┬──────┬──────────┬────────┬───────────────────┬─────────────────┬──────────────────────┬──────────────┐ >> │ Name │ Size │ Min Size │ PG Num │ PG Autoscale Mode │ >> Crush Rule Name │ %-Used │ Used │ >> ╞═══════════════════════╪══════╪══════════╪════════╪═══════════════════╪═════════════════╪══════════════════════╪══════════════╡ >> │ device_health_metrics │ 3 │ 2 │ 1 │ on │ >> replicated_rule │ 4.19273845864154e-07 │ 4534827 │ >> ├───────────────────────┼──────┼──────────┼────────┼───────────────────┼─────────────────┼──────────────────────┼──────────────┤ >> │ rbd │ 3 │ 2 │ 128 │ warn │ >> replicated_rule │ 0.0116069903597236 │ 127014329075 │ >> └───────────────────────┴──────┴──────────┴────────┴───────────────────┴─────────────────┴──────────────────────┴──────────────┘ >> >> >> In the GUI : >> >> 1 pools have too many placement groups Pool rbd has 128 placement >> groups, should have 32 >> >> I don't find any way to reduce placement groups tu 32 >> >> Any help welcome. >> >> Best regards, >> >> Le 27/11/2020 à 14:07, Lindsay Mathieson a écrit : >>> On 27/11/2020 9:47 pm, Jean-Luc Oms wrote: >>>> Upgrading to last Proxmox VE / Ceph nautilus from the last 6.2 proxmox >>>> VE seems to introduce a python 2/3 version problem, dashboard healt >>>> stops working. >>> Was just about to report that :) >>> >>> >>> Same here. >>> >> -- >> Bien cordialement, >> Jean-Daniel Tissot - IE CNRS http://chrono-environnement.univ-fcomte.fr >> UMR 6249 - Laboratoire Chrono-environnement UMR CNRS-UFC >> Université de Franche-Comté, 16 route de Gray, 25030 Besançon Cedex, FRANCE >> Jean-Daniel.Tissot@univ-fcomte.fr tel:+33 3 81 666 440 >> >> _______________________________________________ >> pve-user mailing list >> pve-user@lists.proxmox.com >> https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-user > _______________________________________________ > pve-user mailing list > pve-user@lists.proxmox.com > https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-user -- Bien cordialement, Jean-Daniel Tissot - IE CNRS http://chrono-environnement.univ-fcomte.fr UMR 6249 - Laboratoire Chrono-environnement UMR CNRS-UFC Université de Franche-Comté, 16 route de Gray, 25030 Besançon Cedex, FRANCE Jean-Daniel.Tissot@univ-fcomte.fr tel:+33 3 81 666 440 ^ permalink raw reply [flat|nested] 11+ messages in thread
* Re: [PVE-User] Python problem with upgrade to proxmoxVE6.3/ CEPH Nautilus 14.2.15 2020-11-27 11:47 [PVE-User] Python problem with upgrade to proxmoxVE6.3/ CEPH Nautilus 14.2.15 Jean-Luc Oms 2020-11-27 13:07 ` Lindsay Mathieson @ 2020-11-27 15:15 ` Jean-Luc Oms 2020-11-27 16:45 ` Marco M. Gabriel 2 siblings, 0 replies; 11+ messages in thread From: Jean-Luc Oms @ 2020-11-27 15:15 UTC (permalink / raw) To: pve-user Next step .... I've a small 'preprod' cluster for testing ... but without ceph. If I install ceph on oe node of this cluster this package is not installed: ceph-mgr-dashboard If I remove this package from my prod cluster, tested on the node running active manager, no dependencies and after manager restart, health is Ok ... Now i have installed : root@ceph1:~# dpkg -l | grep ceph ii ceph 14.2.15-pve2 amd64 distributed storage and file system ii ceph-base 14.2.15-pve2 amd64 common ceph daemon libraries and management tools ii ceph-common 14.2.15-pve2 amd64 common utilities to mount and interact with a ceph storage cluster ii ceph-fuse 14.2.15-pve2 amd64 FUSE-based client for the Ceph distributed file system ii ceph-mds 14.2.15-pve2 amd64 metadata server for the ceph distributed file system ii ceph-mgr 14.2.15-pve2 amd64 manager for the ceph distributed storage system ii ceph-mon 14.2.15-pve2 amd64 monitor server for the ceph storage system ii ceph-osd 14.2.15-pve2 amd64 OSD server for the ceph storage system ii libcephfs2 14.2.15-pve2 amd64 Ceph distributed file system client library ii python-ceph-argparse 14.2.15-pve2 all Python 2 utility libraries for Ceph CLI ii python-cephfs 14.2.15-pve2 amd64 Python 2 libraries for the Ceph libcephfs library Is this ok ? Is ceph-mgr-dashboard needed ? Thanks Le 27/11/2020 à 12:47, Jean-Luc Oms a écrit : > Bonjour, > > Upgrading to last Proxmox VE / Ceph nautilus from the last 6.2 proxmox > VE seems to introduce a python 2/3 version problem, dashboard healt > stops working. > > root@ceph1:/usr/bin# ceph health > HEALTH_ERR Module 'dashboard' has failed: ('invalid syntax', > ('/usr/share/ceph/mgr/dashboard/controllers/orchestrator.py', 34, 11, > ' result: dict = {}\n')) > > This syntax was introduced in 3.6, and using strace it seems python 2.7 > is used. > > Any option to resolve this ? everything was ok in 6.2-15. > > Thanks > > -- Jean-Luc Oms /STI-ReseauX <https://rx.lirmm.fr>- LIRMM - CNRS/UM/ +33 4 67 41 85 93 <tel:+33-467-41-85-93> / +33 6 32 01 04 17 <tel:+33-632-01-04-17> ^ permalink raw reply [flat|nested] 11+ messages in thread
* Re: [PVE-User] Python problem with upgrade to proxmoxVE6.3/ CEPH Nautilus 14.2.15 2020-11-27 11:47 [PVE-User] Python problem with upgrade to proxmoxVE6.3/ CEPH Nautilus 14.2.15 Jean-Luc Oms 2020-11-27 13:07 ` Lindsay Mathieson 2020-11-27 15:15 ` Jean-Luc Oms @ 2020-11-27 16:45 ` Marco M. Gabriel 2020-11-28 3:23 ` Lindsay Mathieson 2 siblings, 1 reply; 11+ messages in thread From: Marco M. Gabriel @ 2020-11-27 16:45 UTC (permalink / raw) To: Proxmox VE user list Same problem here after upgrading from 6.2.15 to 6.3 on a test cluster. But the problem disappeared suddenly when I upgraded Ceph from Nautilus to Octopus as well. Not sure if this is the reason why it disappeared and I wouldn't recommend doing this on a production cluster while ceph is in HEALTH_ERR. It would be fine if anyone could test and confirm that an upgrade to Ceph Octopus resolves the issue. Kind regards, Marco Am Fr., 27. Nov. 2020 um 12:54 Uhr schrieb Jean-Luc Oms <jean-luc.oms@lirmm.fr>: > > Bonjour, > > Upgrading to last Proxmox VE / Ceph nautilus from the last 6.2 proxmox > VE seems to introduce a python 2/3 version problem, dashboard healt > stops working. > > root@ceph1:/usr/bin# ceph health > HEALTH_ERR Module 'dashboard' has failed: ('invalid syntax', > ('/usr/share/ceph/mgr/dashboard/controllers/orchestrator.py', 34, 11, > ' result: dict = {}\n')) > > This syntax was introduced in 3.6, and using strace it seems python 2.7 > is used. > > Any option to resolve this ? everything was ok in 6.2-15. > > Thanks > > > -- > Jean-Luc Oms > /STI-ReseauX <https://rx.lirmm.fr>- LIRMM - CNRS/UM/ > +33 4 67 41 85 93 <tel:+33-467-41-85-93> / +33 6 32 01 04 17 > <tel:+33-632-01-04-17> > _______________________________________________ > pve-user mailing list > pve-user@lists.proxmox.com > https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-user ^ permalink raw reply [flat|nested] 11+ messages in thread
* Re: [PVE-User] Python problem with upgrade to proxmoxVE6.3/ CEPH Nautilus 14.2.15 2020-11-27 16:45 ` Marco M. Gabriel @ 2020-11-28 3:23 ` Lindsay Mathieson 0 siblings, 0 replies; 11+ messages in thread From: Lindsay Mathieson @ 2020-11-28 3:23 UTC (permalink / raw) To: pve-user On 28/11/2020 2:45 am, Marco M. Gabriel wrote: > It would be fine if anyone could test and confirm that an upgrade to > Ceph Octopus resolves the issue. Regards the dashboard not working on Proxmox 6.3 - upgrading to ceph Octopus fixed it for me. Also the dashboard looks a lot more swish :) -- Lindsay ^ permalink raw reply [flat|nested] 11+ messages in thread
end of thread, other threads:[~2020-11-28 3:24 UTC | newest] Thread overview: 11+ messages (download: mbox.gz / follow: Atom feed) -- links below jump to the message on this page -- 2020-11-27 11:47 [PVE-User] Python problem with upgrade to proxmoxVE6.3/ CEPH Nautilus 14.2.15 Jean-Luc Oms 2020-11-27 13:07 ` Lindsay Mathieson [not found] ` <47b5a337-b2ca-ce6d-37c5-e904db8d6e03@univ-fcomte.fr> 2020-11-27 16:59 ` alexandre derumier 2020-11-27 17:12 ` [PVE-User] ProxmoxVE6.3/CEPH Octopus 15.2.6 was " Jean-Daniel TISSOT 2020-11-27 21:17 ` [PVE-User] " Lindsay Mathieson 2020-11-27 17:06 ` Marco M. Gabriel 2020-11-27 17:08 ` alexandre derumier 2020-11-27 17:18 ` Jean-Daniel TISSOT 2020-11-27 15:15 ` Jean-Luc Oms 2020-11-27 16:45 ` Marco M. Gabriel 2020-11-28 3:23 ` Lindsay Mathieson
This is a public inbox, see mirroring instructions for how to clone and mirror all data and code used for this inbox