From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: <pve-devel-bounces@lists.proxmox.com> Received: from firstgate.proxmox.com (firstgate.proxmox.com [212.224.123.68]) by lore.proxmox.com (Postfix) with ESMTPS id CD9801FF15C for <inbox@lore.proxmox.com>; Wed, 5 Feb 2025 11:09:52 +0100 (CET) Received: from firstgate.proxmox.com (localhost [127.0.0.1]) by firstgate.proxmox.com (Proxmox) with ESMTP id C54D17561; Wed, 5 Feb 2025 11:09:40 +0100 (CET) From: Alexander Zeidler <a.zeidler@proxmox.com> To: pve-devel@lists.proxmox.com Date: Wed, 5 Feb 2025 11:08:48 +0100 Message-Id: <20250205100850.3-4-a.zeidler@proxmox.com> X-Mailer: git-send-email 2.39.5 In-Reply-To: <20250205100850.3-1-a.zeidler@proxmox.com> References: <20250205100850.3-1-a.zeidler@proxmox.com> MIME-Version: 1.0 X-SPAM-LEVEL: Spam detection results: 0 AWL 0.086 Adjusted score from AWL reputation of From: address BAYES_00 -1.9 Bayes spam probability is 0 to 1% DMARC_MISSING 0.1 Missing DMARC policy KAM_DMARC_STATUS 0.01 Test Rule for DKIM or SPF Failure with Strict Alignment RCVD_IN_VALIDITY_CERTIFIED_BLOCKED 0.001 ADMINISTRATOR NOTICE: The query to Validity was blocked. See https://knowledge.validity.com/hc/en-us/articles/20961730681243 for more information. RCVD_IN_VALIDITY_RPBL_BLOCKED 0.001 ADMINISTRATOR NOTICE: The query to Validity was blocked. See https://knowledge.validity.com/hc/en-us/articles/20961730681243 for more information. RCVD_IN_VALIDITY_SAFE_BLOCKED 0.001 ADMINISTRATOR NOTICE: The query to Validity was blocked. See https://knowledge.validity.com/hc/en-us/articles/20961730681243 for more information. SPF_HELO_NONE 0.001 SPF: HELO does not publish an SPF Record SPF_PASS -0.001 SPF: sender matches SPF record Subject: [pve-devel] [PATCH docs v2 4/6] ceph: osd: revise and expand the section "Destroy OSDs" X-BeenThere: pve-devel@lists.proxmox.com X-Mailman-Version: 2.1.29 Precedence: list List-Id: Proxmox VE development discussion <pve-devel.lists.proxmox.com> List-Unsubscribe: <https://lists.proxmox.com/cgi-bin/mailman/options/pve-devel>, <mailto:pve-devel-request@lists.proxmox.com?subject=unsubscribe> List-Archive: <http://lists.proxmox.com/pipermail/pve-devel/> List-Post: <mailto:pve-devel@lists.proxmox.com> List-Help: <mailto:pve-devel-request@lists.proxmox.com?subject=help> List-Subscribe: <https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-devel>, <mailto:pve-devel-request@lists.proxmox.com?subject=subscribe> Reply-To: Proxmox VE development discussion <pve-devel@lists.proxmox.com> Content-Type: text/plain; charset="us-ascii" Content-Transfer-Encoding: 7bit Errors-To: pve-devel-bounces@lists.proxmox.com Sender: "pve-devel" <pve-devel-bounces@lists.proxmox.com> Existing information is slightly modified and retained. Add information: * Mention and link to the sections "Troubleshooting" and "Replace OSDs" * CLI commands (pveceph) must be executed on the affected node * Check in advance the "Used (%)" of OSDs to avoid blocked I/O * Check and wait until the OSD can be stopped safely * Use `pveceph stop` instead of `systemctl stop ceph-osd@<ID>.service` * Explain cleanup option a bit more Signed-off-by: Alexander Zeidler <a.zeidler@proxmox.com> --- v2: * implement both suggestions from Max Carrara ** mention what the warning is about (unsafe to stop OSD yet) ** use WARNING admonition and adapt the point accordingly pveceph.adoc | 61 +++++++++++++++++++++++++++++----------------------- 1 file changed, 34 insertions(+), 27 deletions(-) diff --git a/pveceph.adoc b/pveceph.adoc index 7401d2b..81a6cc7 100644 --- a/pveceph.adoc +++ b/pveceph.adoc @@ -502,33 +502,40 @@ ceph-volume lvm create --filestore --data /dev/sd[X] --journal /dev/sd[Y] Destroy OSDs ~~~~~~~~~~~~ -To remove an OSD via the GUI, first select a {PVE} node in the tree view and go -to the **Ceph -> OSD** panel. Then select the OSD to destroy and click the **OUT** -button. Once the OSD status has changed from `in` to `out`, click the **STOP** -button. Finally, after the status has changed from `up` to `down`, select -**Destroy** from the `More` drop-down menu. - -To remove an OSD via the CLI run the following commands. - -[source,bash] ----- -ceph osd out <ID> -systemctl stop ceph-osd@<ID>.service ----- - -NOTE: The first command instructs Ceph not to include the OSD in the data -distribution. The second command stops the OSD service. Until this time, no -data is lost. - -The following command destroys the OSD. Specify the '-cleanup' option to -additionally destroy the partition table. - -[source,bash] ----- -pveceph osd destroy <ID> ----- - -WARNING: The above command will destroy all data on the disk! +If you experience problems with an OSD or its disk, try to +xref:pve_ceph_mon_and_ts[troubleshoot] them first to decide if a +xref:pve_ceph_osd_replace[replacement] is needed. + +To destroy an OSD: + +. Either open the web interface and select any {pve} node in the tree +view, or open a shell on the node where the OSD to be deleted is +located. + +. Go to the __Ceph -> OSD__ panel (`ceph osd df tree`). If the OSD to +be deleted is still `up` and `in` (non-zero value at `AVAIL`), make +sure that all OSDs have their `Used (%)` value well below the +`nearfull_ratio` of default `85%`. In this way you can reduce the risk +from the upcoming rebalancing, which may cause OSDs to run full and +thereby blocking I/O on Ceph pools. + +. If the deletable OSD is not `out` yet, select the OSD and click on +**Out** (`ceph osd out <id>`). This will exclude it from data +distribution and starts a rebalance. + +. Click on **Stop**, if stopping is not safe yet, a warning will +appear and you should click on **Cancel**, please try again shortly +afterwards. When using the shell, check if it is safe to stop by +reading the output from `ceph osd ok-to-stop <id>`, once true, run +`pveceph stop --service osd.<id>` . + +. Finally: ++ +[WARNING] +To remove the OSD from Ceph and delete all disk data, first click on +**More -> Destroy**. Use the cleanup option to clean up the partition +table and similar, enabling an immediate reuse of the disk in {pve}. +Finally, click on **Remove** (`pveceph osd destroy <id> [--cleanup]`). [[pve_ceph_pools]] -- 2.39.5 _______________________________________________ pve-devel mailing list pve-devel@lists.proxmox.com https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-devel