public inbox for pve-devel@lists.proxmox.com
 help / color / mirror / Atom feed
From: Matthias Heiserer <m.heiserer@proxmox.com>
To: pve-devel@lists.proxmox.com
Subject: [pve-devel] [PATCH docs 03/10] consistently capitalize Ceph
Date: Wed,  9 Nov 2022 12:58:21 +0100	[thread overview]
Message-ID: <20221109115828.137770-4-m.heiserer@proxmox.com> (raw)
In-Reply-To: <20221109115828.137770-1-m.heiserer@proxmox.com>

Signed-off-by: Matthias Heiserer <m.heiserer@proxmox.com>
---
 hyper-converged-infrastructure.adoc | 4 ++--
 pve-storage-rbd.adoc                | 4 ++--
 pveceph.adoc                        | 6 +++---
 3 files changed, 7 insertions(+), 7 deletions(-)

diff --git a/hyper-converged-infrastructure.adoc b/hyper-converged-infrastructure.adoc
index ee9f185..4616392 100644
--- a/hyper-converged-infrastructure.adoc
+++ b/hyper-converged-infrastructure.adoc
@@ -48,9 +48,9 @@ Hyper-Converged Infrastructure: Storage
 infrastructure. You can, for example, deploy and manage the following two
 storage technologies by using the web interface only:
 
-- *ceph*: a both self-healing and self-managing shared, reliable and highly
+- *Ceph*: a both self-healing and self-managing shared, reliable and highly
   scalable storage system. Checkout
-  xref:chapter_pveceph[how to manage ceph services on {pve} nodes]
+  xref:chapter_pveceph[how to manage Ceph services on {pve} nodes]
 
 - *ZFS*: a combined file system and logical volume manager with extensive
   protection against data corruption, various RAID modes, fast and cheap
diff --git a/pve-storage-rbd.adoc b/pve-storage-rbd.adoc
index 5f8619a..5fe558a 100644
--- a/pve-storage-rbd.adoc
+++ b/pve-storage-rbd.adoc
@@ -109,9 +109,9 @@ management, see the Ceph docs.footnoteref:[cephusermgmt,{cephdocs-url}/rados/ope
 Ceph client configuration (optional)
 ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
 
-Connecting to an external ceph storage doesn't always allow setting
+Connecting to an external Ceph storage doesn't always allow setting
 client-specific options in the config DB on the external cluster. You can add a
-`ceph.conf` beside the ceph keyring to change the ceph client configuration for
+`ceph.conf` beside the Ceph keyring to change the Ceph client configuration for
 the storage.
 
 The ceph.conf needs to have the same name as the storage.
diff --git a/pveceph.adoc b/pveceph.adoc
index 54fb214..fdd4cf6 100644
--- a/pveceph.adoc
+++ b/pveceph.adoc
@@ -636,7 +636,7 @@ pvesm add rbd <storage-name> --pool <replicated-pool> --data-pool <ec-pool>
 ----
 
 TIP: Do not forget to add the `keyring` and `monhost` option for any external
-ceph clusters, not managed by the local {pve} cluster.
+Ceph clusters, not managed by the local {pve} cluster.
 
 Destroy Pools
 ~~~~~~~~~~~~~
@@ -761,7 +761,7 @@ ceph osd crush rule create-replicated <rule-name> <root> <failure-domain> <class
 [frame="none",grid="none", align="left", cols="30%,70%"]
 |===
 |<rule-name>|name of the rule, to connect with a pool (seen in GUI & CLI)
-|<root>|which crush root it should belong to (default ceph root "default")
+|<root>|which crush root it should belong to (default Ceph root "default")
 |<failure-domain>|at which failure-domain the objects should be distributed (usually host)
 |<class>|what type of OSD backing store to use (e.g., nvme, ssd, hdd)
 |===
@@ -943,7 +943,7 @@ servers.
 pveceph fs destroy NAME --remove-storages --remove-pools
 ----
 +
-This will automatically destroy the underlying ceph pools as well as remove
+This will automatically destroy the underlying Ceph pools as well as remove
 the storages from pve config.
 
 After these steps, the CephFS should be completely removed and if you have
-- 
2.30.2





  parent reply	other threads:[~2022-11-09 11:59 UTC|newest]

Thread overview: 13+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2022-11-09 11:58 [pve-devel] [PATCH docs 00/10] docs (mostly) spelling update Matthias Heiserer
2022-11-09 11:58 ` [pve-devel] [PATCH docs 01/10] consistently use 'web interface' (two words) Matthias Heiserer
2022-11-09 11:58 ` [pve-devel] [PATCH docs 02/10] remove misplaced comma Matthias Heiserer
2022-11-09 11:58 ` Matthias Heiserer [this message]
2022-11-09 11:58 ` [pve-devel] [PATCH docs 04/10] add note that secure boot needs to be disabled for host installation Matthias Heiserer
2022-11-09 11:58 ` [pve-devel] [PATCH docs 05/10] move installation options to advanced folder Matthias Heiserer
2022-11-09 14:46   ` Thomas Lamprecht
2022-11-09 11:58 ` [pve-devel] [PATCH docs 06/10] add missing comma in if sentences Matthias Heiserer
2022-11-09 11:58 ` [pve-devel] [PATCH docs 07/10] rephrase when root password is set to make it clear that when logging it, it is already set Matthias Heiserer
2022-11-09 11:58 ` [pve-devel] [PATCH docs 08/10] add missing letter Matthias Heiserer
2022-11-09 11:58 ` [pve-devel] [PATCH docs 09/10] it's -> its because it isn't 'it is' Matthias Heiserer
2022-11-09 11:58 ` [pve-devel] [PATCH docs 10/10] update link to debian handbook. current one is 404 Matthias Heiserer
2022-11-09 14:51 ` [pve-devel] applied-series: [PATCH docs 00/10] docs (mostly) spelling update Thomas Lamprecht

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20221109115828.137770-4-m.heiserer@proxmox.com \
    --to=m.heiserer@proxmox.com \
    --cc=pve-devel@lists.proxmox.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox
Service provided by Proxmox Server Solutions GmbH | Privacy | Legal