* [PVE-User] confirmation on osd replacement @ 2020-11-25 8:18 mj [not found] ` <E9407153-C09E-4098-BC0B-F0605CC43E26@suse.com> 0 siblings, 1 reply; 3+ messages in thread From: mj @ 2020-11-25 8:18 UTC (permalink / raw) To: pve-user Hi, I would just like to verify/confirm something here, as we are going to replace our spinning OSDs with SSDs. We have 8 OSDs per server, and two empty front driveslots available. The proxmox boot disk is internal, and currently known as /dev/sdk Suppose I insert two new SSDs in the (two empty) front drive bays, I expect the internal boot disk to shift from /dev/sdk to /dev/sdm The questions: - should we expect boot problems or other side-effects of doing that? (of course I will test on the first server, I'd just like to know what to expect) And then I am going to first add per server two new bluestore SSDs, making a (temporarily) total of 10 OSDs per server. And then I want to replace the 8 remaining filestore spinning OSDs with 6 bluestore SSDs. Making again a total of 8 OSDs per server. The idea is: first add two SSDs to increase IO capacity for the rest of the procedure, while at the same time reducing stress on our filestore journal SSD (wear level=75%) Any comments? MJ ^ permalink raw reply [flat|nested] 3+ messages in thread
[parent not found: <E9407153-C09E-4098-BC0B-F0605CC43E26@suse.com>]
* Re: [PVE-User] confirmation on osd replacement [not found] ` <E9407153-C09E-4098-BC0B-F0605CC43E26@suse.com> @ 2020-11-26 7:54 ` mj [not found] ` <7B860047-76E2-44B5-8F66-D04FFA216C07@suse.com> 0 siblings, 1 reply; 3+ messages in thread From: mj @ 2020-11-26 7:54 UTC (permalink / raw) To: Alejandro Bonilla, Proxmox VE user list Hi, Yes, perhaps I should have given more details :-) On 11/25/20 3:03 PM, Alejandro Bonilla wrote: > Have a look at /etc/fstab for any disk path mounts - since I think Proxmox uses lvm mostly, you shouldn’t see a problem. I will, thanks! > What is the pool replication configuration or ec-profile? How many nodes in the cluster? We're 3/2 replication, no ec. It's a three-node (small) cluster, 8 filestore OSDs per node, with an SSD journal (wear evel 75%) We will be using samsung PM833 bluestore of the same size (4TB spinners to 3.83GB PM833 SSDs) > Are you planning to remove all disks per server at once or disk by disk? I was planning to: - first add two SSDs to each server, and gradually increase their weight - then, disk-by-disk, replace the 8 (old) spinners with the 6 remaining SSDs > Will all new drives equal or increase the disk capacity of the cluster? Approx equal yes. The aim in not to increase space. MJ ^ permalink raw reply [flat|nested] 3+ messages in thread
[parent not found: <7B860047-76E2-44B5-8F66-D04FFA216C07@suse.com>]
* Re: [PVE-User] confirmation on osd replacement [not found] ` <7B860047-76E2-44B5-8F66-D04FFA216C07@suse.com> @ 2020-11-26 20:31 ` mj 0 siblings, 0 replies; 3+ messages in thread From: mj @ 2020-11-26 20:31 UTC (permalink / raw) To: Alejandro Bonilla; +Cc: Proxmox VE user list Hi Alejandro, Thanks for your feedback, much appreciated! Enjoy your weekend! MJ On 11/26/20 4:39 PM, Alejandro Bonilla wrote: > > >> On Nov 26, 2020, at 2:54 AM, mj <lists@merit.unu.edu> wrote: >> >> Hi, >> >> Yes, perhaps I should have given more details :-) >> >> On 11/25/20 3:03 PM, Alejandro Bonilla wrote: >> >>> Have a look at /etc/fstab for any disk path mounts - since I think Proxmox uses lvm mostly, you shouldn’t see a problem. >> I will, thanks! >> >>> What is the pool replication configuration or ec-profile? How many nodes in the cluster? >> We're 3/2 replication, no ec. It's a three-node (small) cluster, 8 filestore OSDs per node, with an SSD journal (wear evel 75%) > > If it’s 3 replicas, min 2, then you should be able to clear all drives from a system at once and replace them all to minimize the amount of times the cluster will end up rebalancing. > >> >> We will be using samsung PM833 bluestore of the same size >> (4TB spinners to 3.83GB PM833 SSDs) >> >>> Are you planning to remove all disks per server at once or disk by disk? >> I was planning to: >> - first add two SSDs to each server, and gradually increase their weight > > Two per server to ensure the disk replacement will work as expected is a good idea - I don’t think you’ll gain anything with a gradual re-weight. > >> - then, disk-by-disk, replace the 8 (old) spinners with the 6 remaining SSDs > > IF you have two other replicas, then a full system disk replacement should be no trouble - especially after two other SSDs were added and most data was shuffled around. > >> >>> Will all new drives equal or increase the disk capacity of the cluster? >> Approx equal yes. >> The aim in not to increase space. > > There are other reasons why I ask, specifically based on PG count and balancing of the cluster. > >> >> MJ >> > ^ permalink raw reply [flat|nested] 3+ messages in thread
end of thread, other threads:[~2020-11-26 20:32 UTC | newest] Thread overview: 3+ messages (download: mbox.gz / follow: Atom feed) -- links below jump to the message on this page -- 2020-11-25 8:18 [PVE-User] confirmation on osd replacement mj [not found] ` <E9407153-C09E-4098-BC0B-F0605CC43E26@suse.com> 2020-11-26 7:54 ` mj [not found] ` <7B860047-76E2-44B5-8F66-D04FFA216C07@suse.com> 2020-11-26 20:31 ` mj
This is a public inbox, see mirroring instructions for how to clone and mirror all data and code used for this inbox