&& owner.user() == auth_id.user())
},
// default sync owner
- None => auth_id == Authid::backup_auth_id(),
+ None => auth_id == Authid::root_auth_id(),
};
// same permission as changing ownership after syncing
job.owner = Some(read_auth_id.clone());
assert_eq!(check_sync_job_modify_access(&user_info, &write_auth_id, &job), false);
- // also not to the default 'backup@pam'
+ // also not to the default 'root@pam'
job.owner = None;
assert_eq!(check_sync_job_modify_access(&user_info, &write_auth_id, &job), false);
let worker_future = async move {
let delete = sync_job.remove_vanished.unwrap_or(true);
- let sync_owner = sync_job.owner.unwrap_or(Authid::backup_auth_id().clone());
+ let sync_owner = sync_job.owner.unwrap_or(Authid::root_auth_id().clone());
let (client, src_repo, tgt_store) = get_pull_parameters(&sync_job.store, &sync_job.remote, &sync_job.remote_store).await?;
worker.log(format!("Starting datastore sync job '{}'", job_id));
&self.data
}
- /// Get the "backup@pam" user id.
- pub fn backup_userid() -> &'static Self {
- &*BACKUP_USERID
- }
-
/// Get the "root@pam" user id.
pub fn root_userid() -> &'static Self {
&*ROOT_USERID
}
lazy_static! {
- pub static ref BACKUP_USERID: Userid = Userid::new("backup@pam".to_string(), 6);
pub static ref ROOT_USERID: Userid = Userid::new("root@pam".to_string(), 4);
}
}
}
- /// Get the "backup@pam" auth id.
- pub fn backup_auth_id() -> &'static Self {
- &*BACKUP_AUTHID
- }
-
/// Get the "root@pam" auth id.
pub fn root_auth_id() -> &'static Self {
&*ROOT_AUTHID
}
lazy_static! {
- pub static ref BACKUP_AUTHID: Authid = Authid::from(Userid::new("backup@pam".to_string(), 6));
pub static ref ROOT_AUTHID: Authid = Authid::from(Userid::new("root@pam".to_string(), 4));
}
Err(_) => continue, // could not get lock
};
- let auth_id = Authid::backup_auth_id();
+ let auth_id = Authid::root_auth_id();
if let Err(err) = crate::server::do_garbage_collection_job(job, datastore, auth_id, Some(event_str), false) {
eprintln!("unable to start garbage collection job on datastore {} - {}", store, err);
Err(_) => continue, // could not get lock
};
- let auth_id = Authid::backup_auth_id().clone();
+ let auth_id = Authid::root_auth_id().clone();
if let Err(err) = do_prune_job(job, prune_options, store.clone(), &auth_id, Some(event_str)) {
eprintln!("unable to start datastore prune job {} - {}", &store, err);
}
Err(_) => continue, // could not get lock
};
- let auth_id = Authid::backup_auth_id().clone();
+ let auth_id = Authid::root_auth_id().clone();
if let Err(err) = do_sync_job(job, job_config, &auth_id, Some(event_str)) {
eprintln!("unable to start datastore sync job {} - {}", &job_id, err);
}
};
let worker_type = "verificationjob";
- let auth_id = Authid::backup_auth_id().clone();
+ let auth_id = Authid::root_auth_id().clone();
if check_schedule(worker_type, &event_str, &job_id) {
let job = match Job::new(&worker_type, &job_id) {
Ok(job) => job,
if let Err(err) = WorkerTask::new_thread(
worker_type,
None,
- Authid::backup_auth_id().clone(),
+ Authid::root_auth_id().clone(),
false,
move |worker| {
job.start(&worker.upid().to_string())?;
}
/// Lookup users email address
-///
-/// For "backup@pam", this returns the address from "root@pam".
fn lookup_user_email(userid: &Userid) -> Option<String> {
use crate::config::user::{self, User};
- if userid == Userid::backup_userid() {
- return lookup_user_email(Userid::root_userid());
- }
-
if let Ok(user_config) = user::cached_config() {
if let Ok(user) = user_config.lookup::<User>("user", userid.as_str()) {
return user.email.clone();
email = match config.notify_user {
Some(ref userid) => lookup_user_email(userid),
- None => lookup_user_email(Userid::backup_userid()),
+ None => lookup_user_email(Userid::root_userid()),
};
let notify_str = config.notify.unwrap_or(String::new());
name: 'owner',
allowBlank: true,
value: null,
- emptyText: 'backup@pam',
+ emptyText: 'root@pam',
skipEmptyText: true,
cbind: {
deleteEmpty: '{!isCreate}',