})
.collect()
} else {
- iter.map(|user: user::User| UserWithTokens::new(user))
+ iter.map(UserWithTokens::new)
.collect()
};
job.last_run_state = state;
job.last_run_endtime = endtime;
- let last = job.last_run_endtime.unwrap_or_else(|| starttime);
+ let last = job.last_run_endtime.unwrap_or(starttime);
job.next_run = (|| -> Option<i64> {
let schedule = job.schedule.as_ref()?;
let event = parse_calendar_event(&schedule).ok()?;
// ignore errors
- compute_next_event(&event, last, false).unwrap_or_else(|_| None)
+ compute_next_event(&event, last, false).unwrap_or(None)
})();
}
job.last_run_state = state;
job.last_run_endtime = endtime;
- let last = job.last_run_endtime.unwrap_or_else(|| starttime);
+ let last = job.last_run_endtime.unwrap_or(starttime);
job.next_run = (|| -> Option<i64> {
let schedule = job.schedule.as_ref()?;
let event = parse_calendar_event(&schedule).ok()?;
// ignore errors
- compute_next_event(&event, last, false).unwrap_or_else(|_| None)
+ compute_next_event(&event, last, false).unwrap_or(None)
})();
}
pub fn open(path: &Path) -> Result<Self, Error> {
File::open(path)
.map_err(Error::from)
- .and_then(|file| Self::new(file))
+ .and_then(Self::new)
.map_err(|err| format_err!("Unable to open fixed index {:?} - {}", path, err))
}
}
}
- let backup_time = backup_time_opt.unwrap_or_else(|| epoch_i64());
+ let backup_time = backup_time_opt.unwrap_or_else(epoch_i64);
let client = connect(&repo)?;
record_repository(&repo);
}
fn insert_group_role(&mut self, group: String, role: String, propagate: bool) {
- let map = self.groups.entry(group).or_insert_with(|| HashMap::new());
+ let map = self.groups.entry(group).or_insert_with(HashMap::new);
if role == ROLE_NAME_NO_ACCESS {
map.clear();
map.insert(role, propagate);
}
fn insert_user_role(&mut self, auth_id: Authid, role: String, propagate: bool) {
- let map = self.users.entry(auth_id).or_insert_with(|| HashMap::new());
+ let map = self.users.entry(auth_id).or_insert_with(HashMap::new);
if role == ROLE_NAME_NO_ACCESS {
map.clear();
map.insert(role, propagate);
node = node
.children
.entry(String::from(*comp))
- .or_insert_with(|| AclTreeNode::new());
+ .or_insert_with(AclTreeNode::new);
}
node
}
if *propagate {
role_ug_map1
.entry(role)
- .or_insert_with(|| BTreeSet::new())
+ .or_insert_with(BTreeSet::new)
.insert(auth_id);
} else {
role_ug_map0
.entry(role)
- .or_insert_with(|| BTreeSet::new())
+ .or_insert_with(BTreeSet::new)
.insert(auth_id);
}
}
if *propagate {
role_ug_map1
.entry(role)
- .or_insert_with(|| BTreeSet::new())
+ .or_insert_with(BTreeSet::new)
.insert(group);
} else {
role_ug_map0
.entry(role)
- .or_insert_with(|| BTreeSet::new())
+ .or_insert_with(BTreeSet::new)
.insert(group);
}
}
});
result_map
.entry(item_list)
- .or_insert_with(|| BTreeSet::new())
+ .or_insert_with(BTreeSet::new)
.insert(item.to_string());
}
result_map
allow_existing_dirs,
feature_flags,
current_path: Arc::new(Mutex::new(OsString::new())),
- on_error: Box::new(|err| Err(err)),
+ on_error: Box::new(Err),
}
}
if !worker_is_active_local(&info.upid) {
// println!("Detected stopped task '{}'", &info.upid_str);
let now = proxmox::tools::time::epoch_i64();
- let status = upid_read_status(&info.upid)
- .unwrap_or_else(|_| TaskState::Unknown { endtime: now });
+ let status = upid_read_status(&info.upid).unwrap_or(TaskState::Unknown { endtime: now });
finish_list.push(TaskListInfo {
upid: info.upid,
upid_str: info.upid_str,
let result: Result<u64, String> = serde_json::from_str(&output)?;
result
.map_err(|err| format_err!("{}", err))
- .map(|bits| TapeAlertFlags::from_bits_truncate(bits))
+ .map(TapeAlertFlags::from_bits_truncate)
}
/// Set or clear encryption key
};
serde_json::from_str(&serialized_state)
- .map(|s| Some(s))
+ .map(Some)
.map_err(|err| format_err!("could not parse cached package status - {}", err))
}