From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from firstgate.proxmox.com (firstgate.proxmox.com [212.224.123.68]) by lore.proxmox.com (Postfix) with ESMTPS id E9B2D1FF13A for ; Wed, 01 Apr 2026 15:55:24 +0200 (CEST) Received: from firstgate.proxmox.com (localhost [127.0.0.1]) by firstgate.proxmox.com (Proxmox) with ESMTP id 7C46E1E53D; Wed, 1 Apr 2026 15:55:53 +0200 (CEST) From: Christian Ebner To: pbs-devel@lists.proxmox.com Subject: [PATCH proxmox-backup v7 33/34] bin: proxy: periodically schedule counter reset task Date: Wed, 1 Apr 2026 15:48:16 +0200 Message-ID: <20260401134817.926499-34-c.ebner@proxmox.com> X-Mailer: git-send-email 2.47.3 In-Reply-To: <20260401134817.926499-1-c.ebner@proxmox.com> References: <20260401134817.926499-1-c.ebner@proxmox.com> MIME-Version: 1.0 Content-Transfer-Encoding: 8bit X-Bm-Milter-Handled: 55990f41-d878-4baa-be0a-ee34c49e34d2 X-Bm-Transport-Timestamp: 1775051267948 X-SPAM-LEVEL: Spam detection results: 0 AWL -1.435 Adjusted score from AWL reputation of From: address BAYES_00 -1.9 Bayes spam probability is 0 to 1% DMARC_MISSING 0.1 Missing DMARC policy KAM_DMARC_STATUS 0.01 Test Rule for DKIM or SPF Failure with Strict Alignment RCVD_IN_VALIDITY_CERTIFIED_BLOCKED 1 ADMINISTRATOR NOTICE: The query to Validity was blocked. See https://knowledge.validity.com/hc/en-us/articles/20961730681243 for more information. RCVD_IN_VALIDITY_RPBL_BLOCKED 1 ADMINISTRATOR NOTICE: The query to Validity was blocked. See https://knowledge.validity.com/hc/en-us/articles/20961730681243 for more information. RCVD_IN_VALIDITY_SAFE_BLOCKED 1 ADMINISTRATOR NOTICE: The query to Validity was blocked. See https://knowledge.validity.com/hc/en-us/articles/20961730681243 for more information. SPF_HELO_NONE 0.001 SPF: HELO does not publish an SPF Record SPF_PASS -0.001 SPF: sender matches SPF record Message-ID-Hash: OPETETD5POFQ4LGSNKQOBRXBUUP7SFA5 X-Message-ID-Hash: OPETETD5POFQ4LGSNKQOBRXBUUP7SFA5 X-MailFrom: c.ebner@proxmox.com X-Mailman-Rule-Misses: dmarc-mitigation; no-senders; approved; loop; banned-address; emergency; member-moderation; nonmember-moderation; administrivia; implicit-dest; max-recipients; max-size; news-moderation; no-subject; digests; suspicious-header X-Mailman-Version: 3.3.10 Precedence: list List-Id: Proxmox Backup Server development discussion List-Help: List-Owner: List-Post: List-Subscribe: List-Unsubscribe: Analogous to other recurring scheduled tasks, check the configured counter reset schedule for each datastore and periodically execute the reset task if set. By performing this as a dedicated job, it is assured to keep track of the scheduled executions. Signed-off-by: Christian Ebner --- src/bin/proxmox-backup-proxy.rs | 81 ++++++++++++++++++++++++++++++++- 1 file changed, 79 insertions(+), 2 deletions(-) diff --git a/src/bin/proxmox-backup-proxy.rs b/src/bin/proxmox-backup-proxy.rs index b0efa78ae..238637989 100644 --- a/src/bin/proxmox-backup-proxy.rs +++ b/src/bin/proxmox-backup-proxy.rs @@ -1,5 +1,6 @@ use std::path::{Path, PathBuf}; use std::pin::pin; +use std::sync::atomic::Ordering; use std::sync::{Arc, Mutex}; use anyhow::{bail, format_err, Context, Error}; @@ -17,6 +18,7 @@ use url::form_urlencoded; use proxmox_http::Body; use proxmox_http::RateLimiterTag; +use proxmox_human_byte::HumanByte; use proxmox_lang::try_block; use proxmox_rest_server::{ cleanup_old_tasks, cookie_from_header, rotate_task_log_archive, ApiConfig, Redirector, @@ -40,8 +42,8 @@ use pbs_buildcfg::configdir; use proxmox_time::CalendarEvent; use pbs_api_types::{ - Authid, DataStoreConfig, Operation, PruneJobConfig, SyncJobConfig, TapeBackupJobConfig, - VerificationJobConfig, + Authid, DataStoreConfig, DatastoreBackendConfig, Operation, PruneJobConfig, SyncJobConfig, + TapeBackupJobConfig, VerificationJobConfig, }; use proxmox_backup::auth_helpers::*; @@ -508,6 +510,7 @@ async fn schedule_tasks() -> Result<(), Error> { schedule_datastore_verify_jobs().await; schedule_tape_backup_jobs().await; schedule_task_log_rotate().await; + schedule_notification_threshold_counter_reset().await; Ok(()) } @@ -881,6 +884,80 @@ async fn schedule_task_log_rotate() { } } +async fn schedule_notification_threshold_counter_reset() { + let config = match pbs_config::datastore::config() { + Err(err) => { + eprintln!("unable to read datastore config - {err}"); + return; + } + Ok((config, _digest)) => config, + }; + + for (store, (_, store_config)) in config.sections { + let store_config: DataStoreConfig = match serde_json::from_value(store_config) { + Ok(c) => c, + Err(err) => { + eprintln!("datastore config from_value failed - {err}"); + continue; + } + }; + + let event_str = match &store_config.counter_reset_schedule { + Some(event_str) => event_str, + None => continue, + }; + + let worker_type = "notification-threshold-counter-reset"; + if check_schedule(worker_type, event_str, &store) { + let mut job = match Job::new(worker_type, &store) { + Ok(job) => job, + Err(_) => continue, // could not get lock + }; + + if let Err(err) = WorkerTask::new_thread( + worker_type, + None, + Authid::root_auth_id().to_string(), + false, + move |worker| { + job.start(&worker.upid().to_string())?; + info!("executing counter reset for {store}"); + + let result = try_block!({ + let backend_config: DatastoreBackendConfig = + store_config.backend.as_deref().unwrap_or("").parse()?; + let request_counters = + DataStore::request_counters(&store_config, &backend_config)?; + let last_values = request_counters.reset(Ordering::Release); + info!("Last counter values before reset:"); + info!("Request traffic volume:"); + info!("Uploaded: {}", HumanByte::from(last_values.upload)); + info!("Downloaded: {}", HumanByte::from(last_values.download)); + info!("Request count by method:"); + info!("GET: {}", last_values.get); + info!("PUT: {}", last_values.put); + info!("POST: {}", last_values.post); + info!("HEAD: {}", last_values.head); + info!("DELETE: {}", last_values.delete); + + Ok(()) + }); + + let status = worker.create_state(&result); + + if let Err(err) = job.finish(status) { + eprintln!("could not finish job state for {worker_type}: {err}"); + } + + result + }, + ) { + eprintln!("unable to start counter reset task: {err}"); + } + } + } +} + async fn command_reopen_access_logfiles() -> Result<(), Error> { // only care about the most recent daemon instance for each, proxy & api, as other older ones // should not respond to new requests anyway, but only finish their current one and then exit. -- 2.47.3