public inbox for pbs-devel@lists.proxmox.com
 help / color / mirror / Atom feed
From: Lukas Wagner <l.wagner@proxmox.com>
To: Proxmox Backup Server development discussion
	<pbs-devel@lists.proxmox.com>, Stefan Lendl <s.lendl@proxmox.com>
Subject: Re: [pbs-devel] [PATCH v3 proxmox-backup 1/8] api: garbage collect job status
Date: Thu, 21 Mar 2024 11:23:10 +0100	[thread overview]
Message-ID: <39516c32-0e25-4db1-a4ce-bf6d39e7a9a8@proxmox.com> (raw)
In-Reply-To: <20240208135953.2078165-2-s.lendl@proxmox.com>



On  2024-02-08 14:59, Stefan Lendl wrote:
> Adds an api endpoint on the datastore that reports the gc job status
> such as:
>  - Schedule
>  - State (of last run)
>  - Duration (of last run)
>  - Last Run
>  - Next Run (if scheduled)
>  - Pending Chunks (of last run)
>  - Removed Chunks (of last run)

From a user's perspective I think it would make more sense to report
bytes, not chunks, especially since we deal with non-constant chunk
sizes.

> 
> Adds a dedicated endpoint admin/gc that reports gc job status for all
> datastores including the onces without a gc-schedule.
> 
> Originally-by: Gabriel Goller <g.goller@proxmox.com>
> Signed-off-by: Stefan Lendl <s.lendl@proxmox.com>
> Tested-by: Gabriel Goller <g.goller@proxmox.com>
> Reviewd-by: Gabriel Goller <g.goller@proxmox.com>
        ^ typo, also applies to the other patches

> ---
>  pbs-api-types/src/datastore.rs |  40 ++++++++++
>  src/api2/admin/datastore.rs    | 129 ++++++++++++++++++++++++++++++++-
>  src/api2/admin/gc.rs           |  57 +++++++++++++++
>  src/api2/admin/mod.rs          |   2 +
>  4 files changed, 225 insertions(+), 3 deletions(-)
>  create mode 100644 src/api2/admin/gc.rs
> 
> diff --git a/pbs-api-types/src/datastore.rs b/pbs-api-types/src/datastore.rs
> index cce9888b..ba3879c9 100644
> --- a/pbs-api-types/src/datastore.rs
> +++ b/pbs-api-types/src/datastore.rs
> @@ -1270,6 +1270,46 @@ pub struct GarbageCollectionStatus {
>      pub still_bad: usize,
>  }
>  
> +#[api(
> +    properties: {
> +        "last-run-upid": {
> +            optional: true,
> +            type: UPID,
> +        },
> +    },
> +)]
> +#[derive(Clone, Debug, Default, Serialize, Deserialize, PartialEq)]
> +#[serde(rename_all = "kebab-case")]
> +/// Garbage Collection general info
> +pub struct GarbageCollectionJobStatus {
> +    /// Datastore
> +    pub store: String,
> +    /// upid of the last run gc job
> +    #[serde(skip_serializing_if = "Option::is_none")]
> +    pub last_run_upid: Option<String>,
> +    /// Number of removed chunks
> +    #[serde(skip_serializing_if = "Option::is_none")]
> +    pub removed_chunks: Option<usize>,
> +    /// Number of pending chunks
> +    #[serde(skip_serializing_if = "Option::is_none")]
> +    pub pending_chunks: Option<usize>,
> +    /// Schedule of the gc job
> +    #[serde(skip_serializing_if = "Option::is_none")]
> +    pub schedule: Option<String>,
> +    /// Time of the next gc run
> +    #[serde(skip_serializing_if = "Option::is_none")]
> +    pub next_run: Option<i64>,
> +    /// Endtime of the last gc run
> +    #[serde(skip_serializing_if = "Option::is_none")]
> +    pub last_run_endtime: Option<i64>,
> +    /// State of the last gc run
> +    #[serde(skip_serializing_if = "Option::is_none")]
> +    pub last_run_state: Option<String>,
> +    /// Duration of last gc run
> +    #[serde(skip_serializing_if = "Option::is_none")]
> +    pub duration: Option<i64>,
> +}
> +
>  #[api(
>      properties: {
>          "gc-status": {
> diff --git a/src/api2/admin/datastore.rs b/src/api2/admin/datastore.rs
> index a95031e7..357cae0a 100644
> --- a/src/api2/admin/datastore.rs
> +++ b/src/api2/admin/datastore.rs
> @@ -27,18 +27,20 @@ use proxmox_sys::fs::{
>      file_read_firstline, file_read_optional_string, replace_file, CreateOptions,
>  };
>  use proxmox_sys::{task_log, task_warn};
> +use proxmox_time::CalendarEvent;
>  
>  use pxar::accessor::aio::Accessor;
>  use pxar::EntryKind;
>  
>  use pbs_api_types::{
>      print_ns_and_snapshot, print_store_and_ns, Authid, BackupContent, BackupNamespace, BackupType,
> -    Counts, CryptMode, DataStoreListItem, DataStoreStatus, GarbageCollectionStatus, GroupListItem,
> +    Counts, CryptMode, DataStoreConfig, DataStoreListItem, DataStoreStatus,
> +    GarbageCollectionJobStatus, GarbageCollectionStatus, GroupListItem, JobScheduleStatus,
>      KeepOptions, Operation, PruneJobOptions, RRDMode, RRDTimeFrame, SnapshotListItem,
>      SnapshotVerifyState, BACKUP_ARCHIVE_NAME_SCHEMA, BACKUP_ID_SCHEMA, BACKUP_NAMESPACE_SCHEMA,
>      BACKUP_TIME_SCHEMA, BACKUP_TYPE_SCHEMA, DATASTORE_SCHEMA, IGNORE_VERIFIED_BACKUPS_SCHEMA,
>      MAX_NAMESPACE_DEPTH, NS_MAX_DEPTH_SCHEMA, PRIV_DATASTORE_AUDIT, PRIV_DATASTORE_BACKUP,
> -    PRIV_DATASTORE_MODIFY, PRIV_DATASTORE_PRUNE, PRIV_DATASTORE_READ, PRIV_DATASTORE_VERIFY,
> +    PRIV_DATASTORE_MODIFY, PRIV_DATASTORE_PRUNE, PRIV_DATASTORE_READ, PRIV_DATASTORE_VERIFY, UPID,
>      UPID_SCHEMA, VERIFICATION_OUTDATED_AFTER_SCHEMA,
>  };
>  use pbs_client::pxar::{create_tar, create_zip};
> @@ -67,7 +69,7 @@ use crate::backup::{
>      ListAccessibleBackupGroups, NS_PRIVS_OK,
>  };
>  
> -use crate::server::jobstate::Job;
> +use crate::server::jobstate::{compute_schedule_status, Job, JobState};
>  
>  const GROUP_NOTES_FILE_NAME: &str = "notes";
>  
> @@ -1199,6 +1201,123 @@ pub fn garbage_collection_status(
>      Ok(status)
>  }
>  
> +#[api(
> +    input: {
> +        properties: {
> +            store: {
> +                schema: DATASTORE_SCHEMA,
> +            },
> +        },
> +    },
> +    returns: {
> +        type: GarbageCollectionJobStatus,
> +    },
> +    access: {
> +        permission: &Permission::Privilege(&["datastore", "{store}"], PRIV_DATASTORE_AUDIT, false),
> +    },
> +)]
> +/// Garbage collection status.
> +pub fn garbage_collection_job_status(
> +    store: String,
> +    _info: &ApiMethod,
> +    _rpcenv: &mut dyn RpcEnvironment,
> +) -> Result<GarbageCollectionJobStatus, Error> {
> +    let (config, _) = pbs_config::datastore::config()?;
> +    let store_config: DataStoreConfig = config.lookup("datastore", &store)?;
> +
> +    let mut info = GarbageCollectionJobStatus {
> +        store: store.clone(),
> +        schedule: store_config.gc_schedule,
> +        ..Default::default()
> +    };
> +
> +    let datastore = DataStore::lookup_datastore(&store, Some(Operation::Read))?;
> +    let status_in_memory = datastore.last_gc_status();
> +    let state_file = JobState::load("garbage_collection", &store)
> +        .map_err(|err| {
> +            log::error!(
> +                "could not open statefile for {:?}: {}",
> +                info.last_run_upid,
> +                err
> +            )
> +        })
> +        .ok();
> +
> +    let mut selected_upid = None;
> +    if status_in_memory.upid.is_some() {
> +        selected_upid = status_in_memory.upid;
> +    } else if let Some(JobState::Finished { upid, .. }) = &state_file {
> +        selected_upid = Some(upid.to_owned());
> +    }
> +
> +    info.last_run_upid = selected_upid.clone();
> +
> +    match selected_upid {
> +        Some(upid) => {
> +            info.removed_chunks = Some(status_in_memory.removed_chunks);
> +            info.pending_chunks = Some(status_in_memory.pending_chunks);
> +

... seems like the gc-stats also contain the removed/pending bytes, so this should be
an easy change. :)

-- 
- Lukas




  reply	other threads:[~2024-03-21 10:23 UTC|newest]

Thread overview: 15+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2024-02-08 13:59 [pbs-devel] [PATCH v3 proxmox-backup 0/8] *** SUBJECT HERE *** Stefan Lendl
2024-02-08 13:59 ` [pbs-devel] [PATCH v3 proxmox-backup 1/8] api: garbage collect job status Stefan Lendl
2024-03-21 10:23   ` Lukas Wagner [this message]
2024-02-08 13:59 ` [pbs-devel] [PATCH v3 proxmox-backup 2/8] gc: global prune and gc job view Stefan Lendl
2024-02-08 13:59 ` [pbs-devel] [PATCH v3 proxmox-backup 3/8] gc: move datastore/PruneAndGC to config/PruneAndGC Stefan Lendl
2024-02-08 13:59 ` [pbs-devel] [PATCH v3 proxmox-backup 4/8] gc: hide datastore column in local gc view Stefan Lendl
2024-02-08 13:59 ` [pbs-devel] [PATCH v3 proxmox-backup 5/8] ui: order Prune&GC before SyncJobs Stefan Lendl
2024-02-08 13:59 ` [pbs-devel] [PATCH v3 proxmox-backup 6/8] cli: list gc jobs with proxmox-backup-manager Stefan Lendl
2024-02-08 13:59 ` [pbs-devel] [PATCH v3 proxmox-backup 7/8] gc: show removed and pending chunks of last run in ui Stefan Lendl
2024-02-08 13:59 ` [pbs-devel] [PATCH v3 proxmox-backup 8/8] gc: configure width and flex on GC Jobs columns Stefan Lendl
2024-02-21  9:56 ` [pbs-devel] [PATCH v3 proxmox-backup 0/8] Add GC job status to datastore and global prune job view Stefan Lendl
2024-02-21  9:58 ` Stefan Lendl
2024-03-07 13:16   ` Stefan Lendl
2024-03-21 10:24 ` [pbs-devel] [PATCH v3 proxmox-backup 0/8] *** SUBJECT HERE *** Lukas Wagner
2024-04-04 13:53 ` Stefan Lendl

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=39516c32-0e25-4db1-a4ce-bf6d39e7a9a8@proxmox.com \
    --to=l.wagner@proxmox.com \
    --cc=pbs-devel@lists.proxmox.com \
    --cc=s.lendl@proxmox.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox
Service provided by Proxmox Server Solutions GmbH | Privacy | Legal