1 use std
::collections
::{HashMap, VecDeque}
;
3 use std
::path
::PathBuf
;
4 use std
::io
::{Read, Write, BufRead, BufReader}
;
5 use std
::panic
::UnwindSafe
;
6 use std
::sync
::atomic
::{AtomicBool, Ordering}
;
7 use std
::sync
::{Arc, Mutex}
;
8 use std
::time
::{SystemTime, Duration}
;
10 use anyhow
::{bail, format_err, Error}
;
12 use lazy_static
::lazy_static
;
13 use serde_json
::{json, Value}
;
14 use serde
::{Serialize, Deserialize}
;
15 use tokio
::sync
::oneshot
;
16 use nix
::fcntl
::OFlag
;
17 use once_cell
::sync
::OnceCell
;
19 use proxmox
::sys
::linux
::procfs
;
20 use proxmox
::tools
::fs
::{create_path, replace_file, atomic_open_or_create_file, CreateOptions}
;
21 use proxmox_lang
::try_block
;
22 use proxmox_schema
::upid
::UPID
;
24 use proxmox_sys
::worker_task_context
::{WorkerTaskContext}
;
25 use proxmox_sys
::logrotate
::{LogRotate, LogRotateFiles}
;
27 use crate::{CommandSocket, FileLogger, FileLogOptions}
;
29 struct TaskListLockGuard(File
);
31 struct WorkerTaskSetup
{
32 file_opts
: CreateOptions
,
34 task_lock_fn
: PathBuf
,
35 active_tasks_fn
: PathBuf
,
36 task_index_fn
: PathBuf
,
37 task_archive_fn
: PathBuf
,
40 static WORKER_TASK_SETUP
: OnceCell
<WorkerTaskSetup
> = OnceCell
::new();
42 fn worker_task_setup() -> Result
<&'
static WorkerTaskSetup
, Error
> {
43 WORKER_TASK_SETUP
.get()
44 .ok_or_else(|| format_err
!("WorkerTask library is not initialized"))
47 impl WorkerTaskSetup
{
49 fn new(basedir
: PathBuf
, file_opts
: CreateOptions
) -> Self {
51 let mut taskdir
= basedir
.clone();
52 taskdir
.push("tasks");
54 let mut task_lock_fn
= taskdir
.clone();
55 task_lock_fn
.push(".active.lock");
57 let mut active_tasks_fn
= taskdir
.clone();
58 active_tasks_fn
.push("active");
60 let mut task_index_fn
= taskdir
.clone();
61 task_index_fn
.push("index");
63 let mut task_archive_fn
= taskdir
.clone();
64 task_archive_fn
.push("archive");
76 fn lock_task_list_files(&self, exclusive
: bool
) -> Result
<TaskListLockGuard
, Error
> {
77 let options
= self.file_opts
.clone()
78 .perm(nix
::sys
::stat
::Mode
::from_bits_truncate(0o660));
80 let timeout
= std
::time
::Duration
::new(10, 0);
82 let file
= proxmox
::tools
::fs
::open_file_locked(
89 Ok(TaskListLockGuard(file
))
92 fn log_path(&self, upid
: &UPID
) -> std
::path
::PathBuf
{
93 let mut path
= self.taskdir
.clone();
94 path
.push(format
!("{:02X}", upid
.pstart
% 256));
95 path
.push(upid
.to_string());
99 // atomically read/update the task list, update status of finished tasks
100 // new_upid is added to the list when specified.
101 fn update_active_workers(&self, new_upid
: Option
<&UPID
>) -> Result
<(), Error
> {
103 let lock
= self.lock_task_list_files(true)?
;
105 // TODO remove with 1.x
106 let mut finish_list
: Vec
<TaskListInfo
> = read_task_file_from_path(&self.task_index_fn
)?
;
107 let had_index_file
= !finish_list
.is_empty();
109 // We use filter_map because one negative case wants to *move* the data into `finish_list`,
110 // clippy doesn't quite catch this!
111 #[allow(clippy::unnecessary_filter_map)]
112 let mut active_list
: Vec
<TaskListInfo
> = read_task_file_from_path(&self.active_tasks_fn
)?
115 if info
.state
.is_some() {
116 // this can happen when the active file still includes finished tasks
117 finish_list
.push(info
);
121 if !worker_is_active_local(&info
.upid
) {
122 // println!("Detected stopped task '{}'", &info.upid_str);
123 let now
= proxmox_time
::epoch_i64();
124 let status
= upid_read_status(&info
.upid
).unwrap_or(TaskState
::Unknown { endtime: now }
);
125 finish_list
.push(TaskListInfo
{
127 upid_str
: info
.upid_str
,
136 if let Some(upid
) = new_upid
{
137 active_list
.push(TaskListInfo { upid: upid.clone(), upid_str: upid.to_string(), state: None }
);
140 let active_raw
= render_task_list(&active_list
);
142 let options
= self.file_opts
.clone()
143 .perm(nix
::sys
::stat
::Mode
::from_bits_truncate(0o660));
146 &self.active_tasks_fn
,
147 active_raw
.as_bytes(),
152 finish_list
.sort_unstable_by(|a
, b
| {
153 match (&a
.state
, &b
.state
) {
154 (Some(s1
), Some(s2
)) => s1
.cmp(&s2
),
155 (Some(_
), None
) => std
::cmp
::Ordering
::Less
,
156 (None
, Some(_
)) => std
::cmp
::Ordering
::Greater
,
157 _
=> a
.upid
.starttime
.cmp(&b
.upid
.starttime
),
161 if !finish_list
.is_empty() {
162 let options
= self.file_opts
.clone()
163 .perm(nix
::sys
::stat
::Mode
::from_bits_truncate(0o660));
165 let mut writer
= atomic_open_or_create_file(
166 &self.task_archive_fn
,
167 OFlag
::O_APPEND
| OFlag
::O_RDWR
,
172 for info
in &finish_list
{
173 writer
.write_all(render_task_line(&info
).as_bytes())?
;
177 // TODO Remove with 1.x
178 // for compatibility, if we had an INDEX file, we do not need it anymore
180 let _
= nix
::unistd
::unlink(&self.task_index_fn
);
188 // Create task log directory with correct permissions
189 fn create_task_log_dirs(&self) -> Result
<(), Error
> {
192 let dir_opts
= self.file_opts
.clone()
193 .perm(nix
::sys
::stat
::Mode
::from_bits_truncate(0o755));
195 create_path(&self.taskdir
, Some(dir_opts
.clone()), Some(dir_opts
.clone()))?
;
196 // fixme:??? create_path(pbs_buildcfg::PROXMOX_BACKUP_RUN_DIR, None, Some(opts))?;
198 }).map_err(|err
: Error
| format_err
!("unable to create task log dir - {}", err
))
202 /// Initialize the WorkerTask library
203 pub fn init_worker_tasks(basedir
: PathBuf
, file_opts
: CreateOptions
) -> Result
<(), Error
> {
204 let setup
= WorkerTaskSetup
::new(basedir
, file_opts
);
205 setup
.create_task_log_dirs()?
;
206 WORKER_TASK_SETUP
.set(setup
)
207 .map_err(|_
| format_err
!("init_worker_tasks failed - already initialized"))
210 /// checks if the Task Archive is bigger that 'size_threshold' bytes, and
211 /// rotates it if it is
212 pub fn rotate_task_log_archive(
215 max_files
: Option
<usize>,
216 options
: Option
<CreateOptions
>,
217 ) -> Result
<bool
, Error
> {
219 let setup
= worker_task_setup()?
;
221 let _lock
= setup
.lock_task_list_files(true)?
;
223 let mut logrotate
= LogRotate
::new(
224 &setup
.task_archive_fn
,
230 logrotate
.rotate(size_threshold
)
233 /// removes all task logs that are older than the oldest task entry in the
235 pub fn cleanup_old_tasks(compressed
: bool
) -> Result
<(), Error
> {
236 let setup
= worker_task_setup()?
;
238 let _lock
= setup
.lock_task_list_files(true)?
;
240 let logrotate
= LogRotate
::new(
241 &setup
.task_archive_fn
,
247 let mut timestamp
= None
;
248 if let Some(last_file
) = logrotate
.files().last() {
249 let reader
= BufReader
::new(last_file
);
250 for line
in reader
.lines() {
252 if let Ok((_
, _
, Some(state
))) = parse_worker_status_line(&line
) {
253 timestamp
= Some(state
.endtime());
259 fn get_modified(entry
: std
::fs
::DirEntry
) -> Result
<SystemTime
, std
::io
::Error
> {
260 entry
.metadata()?
.modified()
263 if let Some(timestamp
) = timestamp
{
264 let cutoff_time
= if timestamp
> 0 {
265 SystemTime
::UNIX_EPOCH
.checked_add(Duration
::from_secs(timestamp
as u64))
267 SystemTime
::UNIX_EPOCH
.checked_sub(Duration
::from_secs(-timestamp
as u64))
268 }.ok_or_else(|| format_err
!("could not calculate cutoff time"))?
;
271 let mut path
= setup
.taskdir
.clone();
272 path
.push(format
!("{:02X}", i
));
273 for file
in std
::fs
::read_dir(path
)?
{
275 let path
= file
.path();
277 let modified
= get_modified(file
)
278 .map_err(|err
| format_err
!("error getting mtime for {:?}: {}", path
, err
))?
;
280 if modified
< cutoff_time
{
281 match std
::fs
::remove_file(path
) {
283 Err(err
) if err
.kind() == std
::io
::ErrorKind
::NotFound
=> {}
,
284 Err(err
) => bail
!("could not remove file: {}", err
),
295 /// Path to the worker log file
296 pub fn upid_log_path(upid
: &UPID
) -> Result
<std
::path
::PathBuf
, Error
> {
297 let setup
= worker_task_setup()?
;
298 Ok(setup
.log_path(upid
))
301 /// Read endtime (time of last log line) and exitstatus from task log file
302 /// If there is not a single line with at valid datetime, we assume the
303 /// starttime to be the endtime
304 pub fn upid_read_status(upid
: &UPID
) -> Result
<TaskState
, Error
> {
306 let setup
= worker_task_setup()?
;
308 let mut status
= TaskState
::Unknown { endtime: upid.starttime }
;
310 let path
= setup
.log_path(upid
);
312 let mut file
= File
::open(path
)?
;
314 /// speedup - only read tail
316 use std
::io
::SeekFrom
;
317 let _
= file
.seek(SeekFrom
::End(-8192)); // ignore errors
319 let mut data
= Vec
::with_capacity(8192);
320 file
.read_to_end(&mut data
)?
;
322 // strip newlines at the end of the task logs
323 while data
.last() == Some(&b'
\n'
) {
327 let last_line
= match data
.iter().rposition(|c
| *c
== b'
\n'
) {
328 Some(start
) if data
.len() > (start
+1) => &data
[start
+1..],
329 Some(_
) => &data
, // should not happen, since we removed all trailing newlines
333 let last_line
= std
::str::from_utf8(last_line
)
334 .map_err(|err
| format_err
!("upid_read_status: utf8 parse failed: {}", err
))?
;
336 let mut iter
= last_line
.splitn(2, ": ");
337 if let Some(time_str
) = iter
.next() {
338 if let Ok(endtime
) = proxmox_time
::parse_rfc3339(time_str
) {
339 // set the endtime even if we cannot parse the state
340 status
= TaskState
::Unknown { endtime }
;
341 if let Some(rest
) = iter
.next().and_then(|rest
| rest
.strip_prefix("TASK ")) {
342 if let Ok(state
) = TaskState
::from_endtime_and_message(endtime
, rest
) {
353 static ref WORKER_TASK_LIST
: Mutex
<HashMap
<usize, Arc
<WorkerTask
>>> = Mutex
::new(HashMap
::new());
356 /// checks if the task UPID refers to a worker from this process
357 fn is_local_worker(upid
: &UPID
) -> bool
{
358 upid
.pid
== crate::pid() && upid
.pstart
== crate::pstart()
361 /// Test if the task is still running
362 pub async
fn worker_is_active(upid
: &UPID
) -> Result
<bool
, Error
> {
363 if is_local_worker(upid
) {
364 return Ok(WORKER_TASK_LIST
.lock().unwrap().contains_key(&upid
.task_id
));
367 if procfs
::check_process_running_pstart(upid
.pid
, upid
.pstart
).is_none() {
371 let sock
= crate::ctrl_sock_from_pid(upid
.pid
);
373 "command": "worker-task-status",
375 "upid": upid
.to_string(),
378 let status
= crate::send_command(sock
, &cmd
).await?
;
380 if let Some(active
) = status
.as_bool() {
383 bail
!("got unexpected result {:?} (expected bool)", status
);
387 /// Test if the task is still running (fast but inaccurate implementation)
389 /// If the task is spawned from a different process, we simply return if
390 /// that process is still running. This information is good enough to detect
392 pub fn worker_is_active_local(upid
: &UPID
) -> bool
{
393 if is_local_worker(upid
) {
394 WORKER_TASK_LIST
.lock().unwrap().contains_key(&upid
.task_id
)
396 procfs
::check_process_running_pstart(upid
.pid
, upid
.pstart
).is_some()
400 /// Register task control command on a [CommandSocket].
402 /// This create two commands:
404 /// * ``worker-task-abort <UPID>``: calls [abort_local_worker]
406 /// * ``worker-task-status <UPID>``: return true of false, depending on
407 /// whether the worker is running or stopped.
408 pub fn register_task_control_commands(
409 commando_sock
: &mut CommandSocket
,
410 ) -> Result
<(), Error
> {
411 fn get_upid(args
: Option
<&Value
>) -> Result
<UPID
, Error
> {
412 let args
= if let Some(args
) = args { args }
else { bail!("missing args") }
;
413 let upid
= match args
.get("upid") {
414 Some(Value
::String(upid
)) => upid
.parse
::<UPID
>()?
,
415 None
=> bail
!("no upid in args"),
416 _
=> bail
!("unable to parse upid"),
418 if !is_local_worker(&upid
) {
419 bail
!("upid does not belong to this process");
424 commando_sock
.register_command("worker-task-abort".into(), move |args
| {
425 let upid
= get_upid(args
)?
;
427 abort_local_worker(upid
);
431 commando_sock
.register_command("worker-task-status".into(), move |args
| {
432 let upid
= get_upid(args
)?
;
434 let active
= WORKER_TASK_LIST
.lock().unwrap().contains_key(&upid
.task_id
);
442 /// Try to abort a worker task, but do no wait
444 /// Errors (if any) are simply logged.
445 pub fn abort_worker_nowait(upid
: UPID
) {
446 tokio
::spawn(async
move {
447 if let Err(err
) = abort_worker(upid
).await
{
448 log
::error
!("abort worker task failed - {}", err
);
453 /// Abort a worker task
455 /// By sending ``worker-task-abort`` to the control socket.
456 pub async
fn abort_worker(upid
: UPID
) -> Result
<(), Error
> {
458 let sock
= crate::ctrl_sock_from_pid(upid
.pid
);
460 "command": "worker-task-abort",
462 "upid": upid
.to_string(),
465 crate::send_command(sock
, &cmd
).map_ok(|_
| ()).await
468 fn parse_worker_status_line(line
: &str) -> Result
<(String
, UPID
, Option
<TaskState
>), Error
> {
470 let data
= line
.splitn(3, ' '
).collect
::<Vec
<&str>>();
472 let len
= data
.len();
475 1 => Ok((data
[0].to_owned(), data
[0].parse
::<UPID
>()?
, None
)),
477 let endtime
= i64::from_str_radix(data
[1], 16)?
;
478 let state
= TaskState
::from_endtime_and_message(endtime
, data
[2])?
;
479 Ok((data
[0].to_owned(), data
[0].parse
::<UPID
>()?
, Some(state
)))
481 _
=> bail
!("wrong number of components"),
486 #[derive(Debug, PartialEq, Eq, Serialize, Deserialize)]
488 /// The Task ended with an undefined state
489 Unknown { endtime: i64 }
,
490 /// The Task ended and there were no errors or warnings
492 /// The Task had 'count' amount of warnings and no errors
493 Warning { count: u64, endtime: i64 }
,
494 /// The Task ended with the error described in 'message'
495 Error { message: String, endtime: i64 }
,
499 pub fn endtime(&self) -> i64 {
501 TaskState
::Unknown { endtime }
=> endtime
,
502 TaskState
::OK { endtime }
=> endtime
,
503 TaskState
::Warning { endtime, .. }
=> endtime
,
504 TaskState
::Error { endtime, .. }
=> endtime
,
508 fn result_text(&self) -> String
{
510 TaskState
::Error { message, .. }
=> format
!("TASK ERROR: {}", message
),
511 other
=> format
!("TASK {}", other
),
515 fn from_endtime_and_message(endtime
: i64, s
: &str) -> Result
<Self, Error
> {
517 Ok(TaskState
::Unknown { endtime }
)
518 } else if s
== "OK" {
519 Ok(TaskState
::OK { endtime }
)
520 } else if let Some(warnings
) = s
.strip_prefix("WARNINGS: ") {
521 let count
: u64 = warnings
.parse()?
;
522 Ok(TaskState
::Warning{ count, endtime }
)
523 } else if !s
.is_empty() {
524 let message
= if let Some(err
) = s
.strip_prefix("ERROR: ") { err }
else { s }
.to_string();
525 Ok(TaskState
::Error{ message, endtime }
)
527 bail
!("unable to parse Task Status '{}'", s
);
532 impl std
::cmp
::PartialOrd
for TaskState
{
533 fn partial_cmp(&self, other
: &Self) -> Option
<std
::cmp
::Ordering
> {
534 Some(self.endtime().cmp(&other
.endtime()))
538 impl std
::cmp
::Ord
for TaskState
{
539 fn cmp(&self, other
: &Self) -> std
::cmp
::Ordering
{
540 self.endtime().cmp(&other
.endtime())
544 impl std
::fmt
::Display
for TaskState
{
545 fn fmt(&self, f
: &mut std
::fmt
::Formatter
<'_
>) -> std
::fmt
::Result
{
547 TaskState
::Unknown { .. }
=> write
!(f
, "unknown"),
548 TaskState
::OK { .. }
=> write
!(f
, "OK"),
549 TaskState
::Warning { count, .. }
=> write
!(f
, "WARNINGS: {}", count
),
550 TaskState
::Error { message, .. }
=> write
!(f
, "{}", message
),
555 /// Task details including parsed UPID
557 /// If there is no `state`, the task is still running.
559 pub struct TaskListInfo
{
562 /// UPID string representation
563 pub upid_str
: String
,
564 /// Task `(endtime, status)` if already finished
565 pub state
: Option
<TaskState
>, // endtime, status
568 fn render_task_line(info
: &TaskListInfo
) -> String
{
569 let mut raw
= String
::new();
570 if let Some(status
) = &info
.state
{
571 raw
.push_str(&format
!("{} {:08X} {}\n", info
.upid_str
, status
.endtime(), status
));
573 raw
.push_str(&info
.upid_str
);
580 fn render_task_list(list
: &[TaskListInfo
]) -> String
{
581 let mut raw
= String
::new();
583 raw
.push_str(&render_task_line(&info
));
588 // note this is not locked, caller has to make sure it is
589 // this will skip (and log) lines that are not valid status lines
590 fn read_task_file
<R
: Read
>(reader
: R
) -> Result
<Vec
<TaskListInfo
>, Error
>
592 let reader
= BufReader
::new(reader
);
593 let mut list
= Vec
::new();
594 for line
in reader
.lines() {
596 match parse_worker_status_line(&line
) {
597 Ok((upid_str
, upid
, state
)) => list
.push(TaskListInfo
{
603 log
::warn
!("unable to parse worker status '{}' - {}", line
, err
);
612 // note this is not locked, caller has to make sure it is
613 fn read_task_file_from_path
<P
>(path
: P
) -> Result
<Vec
<TaskListInfo
>, Error
>
615 P
: AsRef
<std
::path
::Path
> + std
::fmt
::Debug
,
617 let file
= match File
::open(&path
) {
619 Err(err
) if err
.kind() == std
::io
::ErrorKind
::NotFound
=> return Ok(Vec
::new()),
620 Err(err
) => bail
!("unable to open task list {:?} - {}", path
, err
),
626 /// Iterate over existing/active worker tasks
627 pub struct TaskListInfoIterator
{
628 list
: VecDeque
<TaskListInfo
>,
630 archive
: Option
<LogRotateFiles
>,
631 lock
: Option
<TaskListLockGuard
>,
634 impl TaskListInfoIterator
{
635 /// Creates a new iterator instance.
636 pub fn new(active_only
: bool
) -> Result
<Self, Error
> {
638 let setup
= worker_task_setup()?
;
640 let (read_lock
, active_list
) = {
641 let lock
= setup
.lock_task_list_files(false)?
;
642 let active_list
= read_task_file_from_path(&setup
.active_tasks_fn
)?
;
644 let needs_update
= active_list
646 .any(|info
| info
.state
.is_some() || !worker_is_active_local(&info
.upid
));
648 // TODO remove with 1.x
649 let index_exists
= setup
.task_index_fn
.is_file();
651 if needs_update
|| index_exists
{
653 setup
.update_active_workers(None
)?
;
654 let lock
= setup
.lock_task_list_files(false)?
;
655 let active_list
= read_task_file_from_path(&setup
.active_tasks_fn
)?
;
662 let archive
= if active_only
{
665 let logrotate
= LogRotate
::new(&setup
.task_archive_fn
, true, None
, None
)?
;
666 Some(logrotate
.files())
669 let lock
= if active_only { None }
else { Some(read_lock) }
;
672 list
: active_list
.into(),
680 impl Iterator
for TaskListInfoIterator
{
681 type Item
= Result
<TaskListInfo
, Error
>;
683 fn next(&mut self) -> Option
<Self::Item
> {
685 if let Some(element
) = self.list
.pop_back() {
686 return Some(Ok(element
));
690 if let Some(mut archive
) = self.archive
.take() {
691 if let Some(file
) = archive
.next() {
692 let list
= match read_task_file(file
) {
694 Err(err
) => return Some(Err(err
)),
696 self.list
.append(&mut list
.into());
697 self.archive
= Some(archive
);
709 /// Launch long running worker tasks.
711 /// A worker task can either be a whole thread, or a simply tokio
712 /// task/future. Each task can `log()` messages, which are stored
713 /// persistently to files. Task should poll the `abort_requested`
714 /// flag, and stop execution when requested.
715 pub struct WorkerTask
{
716 setup
: &'
static WorkerTaskSetup
,
718 data
: Mutex
<WorkerTaskData
>,
719 abort_requested
: AtomicBool
,
722 impl std
::fmt
::Display
for WorkerTask
{
724 fn fmt(&self, f
: &mut std
::fmt
::Formatter
) -> std
::fmt
::Result
{
729 struct WorkerTaskData
{
731 progress
: f64, // 0..1
733 pub abort_listeners
: Vec
<oneshot
::Sender
<()>>,
740 worker_id
: Option
<String
>,
743 ) -> Result
<Arc
<Self>, Error
> {
745 let setup
= worker_task_setup()?
;
747 let upid
= UPID
::new(worker_type
, worker_id
, auth_id
)?
;
748 let task_id
= upid
.task_id
;
750 let mut path
= setup
.taskdir
.clone();
752 path
.push(format
!("{:02X}", upid
.pstart
& 255));
754 let dir_opts
= setup
.file_opts
.clone()
755 .perm(nix
::sys
::stat
::Mode
::from_bits_truncate(0o755));
757 create_path(&path
, None
, Some(dir_opts
))?
;
759 path
.push(upid
.to_string());
761 let logger_options
= FileLogOptions
{
766 file_opts
: setup
.file_opts
.clone(),
769 let logger
= FileLogger
::new(&path
, logger_options
)?
;
771 let worker
= Arc
::new(Self {
774 abort_requested
: AtomicBool
::new(false),
775 data
: Mutex
::new(WorkerTaskData
{
779 abort_listeners
: vec
![],
783 // scope to drop the lock again after inserting
785 let mut hash
= WORKER_TASK_LIST
.lock().unwrap();
786 hash
.insert(task_id
, worker
.clone());
787 crate::set_worker_count(hash
.len());
790 setup
.update_active_workers(Some(&upid
))?
;
795 /// Spawn a new tokio task/future.
798 worker_id
: Option
<String
>,
802 ) -> Result
<String
, Error
>
803 where F
: Send
+ '
static + FnOnce(Arc
<WorkerTask
>) -> T
,
804 T
: Send
+ '
static + Future
<Output
= Result
<(), Error
>>,
806 let worker
= WorkerTask
::new(worker_type
, worker_id
, auth_id
, to_stdout
)?
;
807 let upid_str
= worker
.upid
.to_string();
808 let f
= f(worker
.clone());
809 tokio
::spawn(async
move {
810 let result
= f
.await
;
811 worker
.log_result(&result
);
817 /// Create a new worker thread.
818 pub fn new_thread
<F
>(
820 worker_id
: Option
<String
>,
824 ) -> Result
<String
, Error
>
825 where F
: Send
+ UnwindSafe
+ '
static + FnOnce(Arc
<WorkerTask
>) -> Result
<(), Error
>
827 let worker
= WorkerTask
::new(worker_type
, worker_id
, auth_id
, to_stdout
)?
;
828 let upid_str
= worker
.upid
.to_string();
830 let _child
= std
::thread
::Builder
::new().name(upid_str
.clone()).spawn(move || {
831 let worker1
= worker
.clone();
832 let result
= match std
::panic
::catch_unwind(move || f(worker1
)) {
835 match panic
.downcast
::<&str>() {
837 Err(format_err
!("worker panicked: {}", panic_msg
))
840 Err(format_err
!("worker panicked: unknown type."))
846 worker
.log_result(&result
);
852 /// create state from self and a result
853 pub fn create_state(&self, result
: &Result
<(), Error
>) -> TaskState
{
854 let warn_count
= self.data
.lock().unwrap().warn_count
;
856 let endtime
= proxmox_time
::epoch_i64();
858 if let Err(err
) = result
{
859 TaskState
::Error { message: err.to_string(), endtime }
860 } else if warn_count
> 0 {
861 TaskState
::Warning { count: warn_count, endtime }
863 TaskState
::OK { endtime }
867 /// Log task result, remove task from running list
868 pub fn log_result(&self, result
: &Result
<(), Error
>) {
869 let state
= self.create_state(result
);
870 self.log_message(state
.result_text());
872 WORKER_TASK_LIST
.lock().unwrap().remove(&self.upid
.task_id
);
873 let _
= self.setup
.update_active_workers(None
);
874 crate::set_worker_count(WORKER_TASK_LIST
.lock().unwrap().len());
878 pub fn log_message
<S
: AsRef
<str>>(&self, msg
: S
) {
879 let mut data
= self.data
.lock().unwrap();
880 data
.logger
.log(msg
);
883 /// Log a message as warning.
884 pub fn log_warning
<S
: AsRef
<str>>(&self, msg
: S
) {
885 let mut data
= self.data
.lock().unwrap();
886 data
.logger
.log(format
!("WARN: {}", msg
.as_ref()));
887 data
.warn_count
+= 1;
890 /// Set progress indicator
891 pub fn progress(&self, progress
: f64) {
892 if progress
>= 0.0 && progress
<= 1.0 {
893 let mut data
= self.data
.lock().unwrap();
894 data
.progress
= progress
;
896 // fixme: log!("task '{}': ignoring strange value for progress '{}'", self.upid, progress);
901 pub fn request_abort(&self) {
902 let prev_abort
= self.abort_requested
.swap(true, Ordering
::SeqCst
);
903 if !prev_abort
{ // log abort one time
904 self.log_message(format
!("received abort request ..."));
907 let mut data
= self.data
.lock().unwrap();
909 match data
.abort_listeners
.pop() {
912 let _
= ch
.send(()); // ignore errors here
918 /// Get a future which resolves on task abort
919 pub fn abort_future(&self) -> oneshot
::Receiver
<()> {
920 let (tx
, rx
) = oneshot
::channel
::<()>();
922 let mut data
= self.data
.lock().unwrap();
923 if self.abort_requested() {
926 data
.abort_listeners
.push(tx
);
931 pub fn upid(&self) -> &UPID
{
936 impl WorkerTaskContext
for WorkerTask
{
938 fn abort_requested(&self) -> bool
{
939 self.abort_requested
.load(Ordering
::SeqCst
)
942 fn shutdown_requested(&self) -> bool
{
943 crate::shutdown_requested()
946 fn fail_on_shutdown(&self) -> Result
<(), Error
> {
947 crate::fail_on_shutdown()
950 fn log(&self, level
: log
::Level
, message
: &std
::fmt
::Arguments
) {
952 log
::Level
::Error
=> self.log_warning(&message
.to_string()),
953 log
::Level
::Warn
=> self.log_warning(&message
.to_string()),
954 log
::Level
::Info
=> self.log_message(&message
.to_string()),
955 log
::Level
::Debug
=> self.log_message(&format
!("DEBUG: {}", message
)),
956 log
::Level
::Trace
=> self.log_message(&format
!("TRACE: {}", message
)),
961 /// Wait for a locally spanned worker task
963 /// Note: local workers should print logs to stdout, so there is no
964 /// need to fetch/display logs. We just wait for the worker to finish.
965 pub async
fn wait_for_local_worker(upid_str
: &str) -> Result
<(), Error
> {
967 let upid
: UPID
= upid_str
.parse()?
;
969 let sleep_duration
= core
::time
::Duration
::new(0, 100_000_000);
972 if worker_is_active_local(&upid
) {
973 tokio
::time
::sleep(sleep_duration
).await
;
981 /// Request abort of a local worker (if existing and running)
982 pub fn abort_local_worker(upid
: UPID
) {
983 if let Some(ref worker
) = WORKER_TASK_LIST
.lock().unwrap().get(&upid
.task_id
) {
984 worker
.request_abort();