]>
Commit | Line | Data |
---|---|---|
1 | use std::path::Path; | |
2 | use std::sync::{Mutex, Arc}; | |
3 | ||
4 | use anyhow::{bail, format_err, Error}; | |
5 | use serde_json::Value; | |
6 | ||
7 | use proxmox_lang::try_block; | |
8 | use proxmox_router::{Permission, Router, RpcEnvironment, RpcEnvironmentType}; | |
9 | use proxmox_schema::api; | |
10 | ||
11 | use pbs_api_types::{ | |
12 | Authid, Userid, TapeBackupJobConfig, TapeBackupJobSetup, TapeBackupJobStatus, MediaPoolConfig, | |
13 | UPID_SCHEMA, JOB_ID_SCHEMA, PRIV_DATASTORE_READ, PRIV_TAPE_AUDIT, PRIV_TAPE_WRITE, | |
14 | }; | |
15 | ||
16 | use pbs_datastore::{DataStore, StoreProgress, SnapshotReader}; | |
17 | use pbs_datastore::backup_info::{BackupDir, BackupInfo}; | |
18 | use pbs_tools::{task_log, task_warn, task::WorkerTaskContext}; | |
19 | use pbs_config::CachedUserInfo; | |
20 | use proxmox_rest_server::WorkerTask; | |
21 | ||
22 | use crate::{ | |
23 | server::{ | |
24 | lookup_user_email, | |
25 | TapeBackupJobSummary, | |
26 | jobstate::{ | |
27 | Job, | |
28 | JobState, | |
29 | compute_schedule_status, | |
30 | }, | |
31 | }, | |
32 | tape::{ | |
33 | TAPE_STATUS_DIR, | |
34 | Inventory, | |
35 | PoolWriter, | |
36 | MediaPool, | |
37 | drive::{ | |
38 | media_changer, | |
39 | lock_tape_device, | |
40 | TapeLockError, | |
41 | set_tape_device_state, | |
42 | }, | |
43 | changer::update_changer_online_status, | |
44 | }, | |
45 | }; | |
46 | ||
47 | const TAPE_BACKUP_JOB_ROUTER: Router = Router::new() | |
48 | .post(&API_METHOD_RUN_TAPE_BACKUP_JOB); | |
49 | ||
50 | pub const ROUTER: Router = Router::new() | |
51 | .get(&API_METHOD_LIST_TAPE_BACKUP_JOBS) | |
52 | .post(&API_METHOD_BACKUP) | |
53 | .match_all("id", &TAPE_BACKUP_JOB_ROUTER); | |
54 | ||
55 | fn check_backup_permission( | |
56 | auth_id: &Authid, | |
57 | store: &str, | |
58 | pool: &str, | |
59 | drive: &str, | |
60 | ) -> Result<(), Error> { | |
61 | ||
62 | let user_info = CachedUserInfo::new()?; | |
63 | ||
64 | let privs = user_info.lookup_privs(auth_id, &["datastore", store]); | |
65 | if (privs & PRIV_DATASTORE_READ) == 0 { | |
66 | bail!("no permissions on /datastore/{}", store); | |
67 | } | |
68 | ||
69 | let privs = user_info.lookup_privs(auth_id, &["tape", "drive", drive]); | |
70 | if (privs & PRIV_TAPE_WRITE) == 0 { | |
71 | bail!("no permissions on /tape/drive/{}", drive); | |
72 | } | |
73 | ||
74 | let privs = user_info.lookup_privs(auth_id, &["tape", "pool", pool]); | |
75 | if (privs & PRIV_TAPE_WRITE) == 0 { | |
76 | bail!("no permissions on /tape/pool/{}", pool); | |
77 | } | |
78 | ||
79 | Ok(()) | |
80 | } | |
81 | ||
82 | #[api( | |
83 | returns: { | |
84 | description: "List configured thape backup jobs and their status", | |
85 | type: Array, | |
86 | items: { type: TapeBackupJobStatus }, | |
87 | }, | |
88 | access: { | |
89 | description: "List configured tape jobs filtered by Tape.Audit privileges", | |
90 | permission: &Permission::Anybody, | |
91 | }, | |
92 | )] | |
93 | /// List all tape backup jobs | |
94 | pub fn list_tape_backup_jobs( | |
95 | _param: Value, | |
96 | mut rpcenv: &mut dyn RpcEnvironment, | |
97 | ) -> Result<Vec<TapeBackupJobStatus>, Error> { | |
98 | let auth_id: Authid = rpcenv.get_auth_id().unwrap().parse()?; | |
99 | let user_info = CachedUserInfo::new()?; | |
100 | ||
101 | let (job_config, digest) = pbs_config::tape_job::config()?; | |
102 | let (pool_config, _pool_digest) = pbs_config::media_pool::config()?; | |
103 | let (drive_config, _digest) = pbs_config::drive::config()?; | |
104 | ||
105 | let job_list_iter = job_config | |
106 | .convert_to_typed_array("backup")? | |
107 | .into_iter() | |
108 | .filter(|_job: &TapeBackupJobConfig| { | |
109 | // fixme: check access permission | |
110 | true | |
111 | }); | |
112 | ||
113 | let mut list = Vec::new(); | |
114 | let status_path = Path::new(TAPE_STATUS_DIR); | |
115 | let current_time = proxmox_time::epoch_i64(); | |
116 | ||
117 | for job in job_list_iter { | |
118 | let privs = user_info.lookup_privs(&auth_id, &["tape", "job", &job.id]); | |
119 | if (privs & PRIV_TAPE_AUDIT) == 0 { | |
120 | continue; | |
121 | } | |
122 | ||
123 | let last_state = JobState::load("tape-backup-job", &job.id) | |
124 | .map_err(|err| format_err!("could not open statefile for {}: {}", &job.id, err))?; | |
125 | ||
126 | let status = compute_schedule_status(&last_state, job.schedule.as_deref())?; | |
127 | ||
128 | let next_run = status.next_run.unwrap_or(current_time); | |
129 | ||
130 | let mut next_media_label = None; | |
131 | ||
132 | if let Ok(pool) = pool_config.lookup::<MediaPoolConfig>("pool", &job.setup.pool) { | |
133 | let mut changer_name = None; | |
134 | if let Ok(Some((_, name))) = media_changer(&drive_config, &job.setup.drive) { | |
135 | changer_name = Some(name); | |
136 | } | |
137 | if let Ok(mut pool) = MediaPool::with_config(status_path, &pool, changer_name, true) { | |
138 | if pool.start_write_session(next_run, false).is_ok() { | |
139 | if let Ok(media_id) = pool.guess_next_writable_media(next_run) { | |
140 | next_media_label = Some(media_id.label.label_text); | |
141 | } | |
142 | } | |
143 | } | |
144 | } | |
145 | ||
146 | list.push(TapeBackupJobStatus { config: job, status, next_media_label }); | |
147 | } | |
148 | ||
149 | rpcenv["digest"] = proxmox::tools::digest_to_hex(&digest).into(); | |
150 | ||
151 | Ok(list) | |
152 | } | |
153 | ||
154 | pub fn do_tape_backup_job( | |
155 | mut job: Job, | |
156 | setup: TapeBackupJobSetup, | |
157 | auth_id: &Authid, | |
158 | schedule: Option<String>, | |
159 | to_stdout: bool, | |
160 | ) -> Result<String, Error> { | |
161 | ||
162 | let job_id = format!("{}:{}:{}:{}", | |
163 | setup.store, | |
164 | setup.pool, | |
165 | setup.drive, | |
166 | job.jobname()); | |
167 | ||
168 | let worker_type = job.jobtype().to_string(); | |
169 | ||
170 | let datastore = DataStore::lookup_datastore(&setup.store)?; | |
171 | ||
172 | let (config, _digest) = pbs_config::media_pool::config()?; | |
173 | let pool_config: MediaPoolConfig = config.lookup("pool", &setup.pool)?; | |
174 | ||
175 | let (drive_config, _digest) = pbs_config::drive::config()?; | |
176 | ||
177 | // for scheduled jobs we acquire the lock later in the worker | |
178 | let drive_lock = if schedule.is_some() { | |
179 | None | |
180 | } else { | |
181 | Some(lock_tape_device(&drive_config, &setup.drive)?) | |
182 | }; | |
183 | ||
184 | let notify_user = setup.notify_user.as_ref().unwrap_or_else(|| &Userid::root_userid()); | |
185 | let email = lookup_user_email(notify_user); | |
186 | ||
187 | let upid_str = WorkerTask::new_thread( | |
188 | &worker_type, | |
189 | Some(job_id.clone()), | |
190 | auth_id.to_string(), | |
191 | to_stdout, | |
192 | move |worker| { | |
193 | job.start(&worker.upid().to_string())?; | |
194 | let mut drive_lock = drive_lock; | |
195 | ||
196 | let mut summary = Default::default(); | |
197 | let job_result = try_block!({ | |
198 | if schedule.is_some() { | |
199 | // for scheduled tape backup jobs, we wait indefinitely for the lock | |
200 | task_log!(worker, "waiting for drive lock..."); | |
201 | loop { | |
202 | worker.check_abort()?; | |
203 | match lock_tape_device(&drive_config, &setup.drive) { | |
204 | Ok(lock) => { | |
205 | drive_lock = Some(lock); | |
206 | break; | |
207 | } | |
208 | Err(TapeLockError::TimeOut) => continue, | |
209 | Err(TapeLockError::Other(err)) => return Err(err), | |
210 | } | |
211 | } | |
212 | } | |
213 | set_tape_device_state(&setup.drive, &worker.upid().to_string())?; | |
214 | ||
215 | task_log!(worker,"Starting tape backup job '{}'", job_id); | |
216 | if let Some(event_str) = schedule { | |
217 | task_log!(worker,"task triggered by schedule '{}'", event_str); | |
218 | } | |
219 | ||
220 | ||
221 | backup_worker( | |
222 | &worker, | |
223 | datastore, | |
224 | &pool_config, | |
225 | &setup, | |
226 | email.clone(), | |
227 | &mut summary, | |
228 | false, | |
229 | ) | |
230 | }); | |
231 | ||
232 | let status = worker.create_state(&job_result); | |
233 | ||
234 | if let Some(email) = email { | |
235 | if let Err(err) = crate::server::send_tape_backup_status( | |
236 | &email, | |
237 | Some(job.jobname()), | |
238 | &setup, | |
239 | &job_result, | |
240 | summary, | |
241 | ) { | |
242 | eprintln!("send tape backup notification failed: {}", err); | |
243 | } | |
244 | } | |
245 | ||
246 | if let Err(err) = job.finish(status) { | |
247 | eprintln!( | |
248 | "could not finish job state for {}: {}", | |
249 | job.jobtype().to_string(), | |
250 | err | |
251 | ); | |
252 | } | |
253 | ||
254 | if let Err(err) = set_tape_device_state(&setup.drive, "") { | |
255 | eprintln!( | |
256 | "could not unset drive state for {}: {}", | |
257 | setup.drive, | |
258 | err | |
259 | ); | |
260 | } | |
261 | ||
262 | job_result | |
263 | } | |
264 | )?; | |
265 | ||
266 | Ok(upid_str) | |
267 | } | |
268 | ||
269 | #[api( | |
270 | input: { | |
271 | properties: { | |
272 | id: { | |
273 | schema: JOB_ID_SCHEMA, | |
274 | }, | |
275 | }, | |
276 | }, | |
277 | access: { | |
278 | // Note: parameters are from job config, so we need to test inside function body | |
279 | description: "The user needs Tape.Write privilege on /tape/pool/{pool} \ | |
280 | and /tape/drive/{drive}, Datastore.Read privilege on /datastore/{store}.", | |
281 | permission: &Permission::Anybody, | |
282 | }, | |
283 | )] | |
284 | /// Runs a tape backup job manually. | |
285 | pub fn run_tape_backup_job( | |
286 | id: String, | |
287 | rpcenv: &mut dyn RpcEnvironment, | |
288 | ) -> Result<String, Error> { | |
289 | let auth_id: Authid = rpcenv.get_auth_id().unwrap().parse()?; | |
290 | ||
291 | let (config, _digest) = pbs_config::tape_job::config()?; | |
292 | let backup_job: TapeBackupJobConfig = config.lookup("backup", &id)?; | |
293 | ||
294 | check_backup_permission( | |
295 | &auth_id, | |
296 | &backup_job.setup.store, | |
297 | &backup_job.setup.pool, | |
298 | &backup_job.setup.drive, | |
299 | )?; | |
300 | ||
301 | let job = Job::new("tape-backup-job", &id)?; | |
302 | ||
303 | let to_stdout = rpcenv.env_type() == RpcEnvironmentType::CLI; | |
304 | ||
305 | let upid_str = do_tape_backup_job(job, backup_job.setup, &auth_id, None, to_stdout)?; | |
306 | ||
307 | Ok(upid_str) | |
308 | } | |
309 | ||
310 | #[api( | |
311 | input: { | |
312 | properties: { | |
313 | setup: { | |
314 | type: TapeBackupJobSetup, | |
315 | flatten: true, | |
316 | }, | |
317 | "force-media-set": { | |
318 | description: "Ignore the allocation policy and start a new media-set.", | |
319 | optional: true, | |
320 | type: bool, | |
321 | default: false, | |
322 | }, | |
323 | }, | |
324 | }, | |
325 | returns: { | |
326 | schema: UPID_SCHEMA, | |
327 | }, | |
328 | access: { | |
329 | // Note: parameters are no uri parameter, so we need to test inside function body | |
330 | description: "The user needs Tape.Write privilege on /tape/pool/{pool} \ | |
331 | and /tape/drive/{drive}, Datastore.Read privilege on /datastore/{store}.", | |
332 | permission: &Permission::Anybody, | |
333 | }, | |
334 | )] | |
335 | /// Backup datastore to tape media pool | |
336 | pub fn backup( | |
337 | setup: TapeBackupJobSetup, | |
338 | force_media_set: bool, | |
339 | rpcenv: &mut dyn RpcEnvironment, | |
340 | ) -> Result<Value, Error> { | |
341 | ||
342 | let auth_id: Authid = rpcenv.get_auth_id().unwrap().parse()?; | |
343 | ||
344 | check_backup_permission( | |
345 | &auth_id, | |
346 | &setup.store, | |
347 | &setup.pool, | |
348 | &setup.drive, | |
349 | )?; | |
350 | ||
351 | let datastore = DataStore::lookup_datastore(&setup.store)?; | |
352 | ||
353 | let (config, _digest) = pbs_config::media_pool::config()?; | |
354 | let pool_config: MediaPoolConfig = config.lookup("pool", &setup.pool)?; | |
355 | ||
356 | let (drive_config, _digest) = pbs_config::drive::config()?; | |
357 | ||
358 | // early check/lock before starting worker | |
359 | let drive_lock = lock_tape_device(&drive_config, &setup.drive)?; | |
360 | ||
361 | let to_stdout = rpcenv.env_type() == RpcEnvironmentType::CLI; | |
362 | ||
363 | let job_id = format!("{}:{}:{}", setup.store, setup.pool, setup.drive); | |
364 | ||
365 | let notify_user = setup.notify_user.as_ref().unwrap_or_else(|| &Userid::root_userid()); | |
366 | let email = lookup_user_email(notify_user); | |
367 | ||
368 | let upid_str = WorkerTask::new_thread( | |
369 | "tape-backup", | |
370 | Some(job_id), | |
371 | auth_id.to_string(), | |
372 | to_stdout, | |
373 | move |worker| { | |
374 | let _drive_lock = drive_lock; // keep lock guard | |
375 | set_tape_device_state(&setup.drive, &worker.upid().to_string())?; | |
376 | ||
377 | let mut summary = Default::default(); | |
378 | let job_result = backup_worker( | |
379 | &worker, | |
380 | datastore, | |
381 | &pool_config, | |
382 | &setup, | |
383 | email.clone(), | |
384 | &mut summary, | |
385 | force_media_set, | |
386 | ); | |
387 | ||
388 | if let Some(email) = email { | |
389 | if let Err(err) = crate::server::send_tape_backup_status( | |
390 | &email, | |
391 | None, | |
392 | &setup, | |
393 | &job_result, | |
394 | summary, | |
395 | ) { | |
396 | eprintln!("send tape backup notification failed: {}", err); | |
397 | } | |
398 | } | |
399 | ||
400 | // ignore errors | |
401 | let _ = set_tape_device_state(&setup.drive, ""); | |
402 | job_result | |
403 | } | |
404 | )?; | |
405 | ||
406 | Ok(upid_str.into()) | |
407 | } | |
408 | ||
409 | fn backup_worker( | |
410 | worker: &WorkerTask, | |
411 | datastore: Arc<DataStore>, | |
412 | pool_config: &MediaPoolConfig, | |
413 | setup: &TapeBackupJobSetup, | |
414 | email: Option<String>, | |
415 | summary: &mut TapeBackupJobSummary, | |
416 | force_media_set: bool, | |
417 | ) -> Result<(), Error> { | |
418 | ||
419 | let status_path = Path::new(TAPE_STATUS_DIR); | |
420 | let start = std::time::Instant::now(); | |
421 | ||
422 | task_log!(worker, "update media online status"); | |
423 | let changer_name = update_media_online_status(&setup.drive)?; | |
424 | ||
425 | let pool = MediaPool::with_config(status_path, &pool_config, changer_name, false)?; | |
426 | ||
427 | let mut pool_writer = PoolWriter::new( | |
428 | pool, | |
429 | &setup.drive, | |
430 | worker, | |
431 | email, | |
432 | force_media_set | |
433 | )?; | |
434 | ||
435 | let mut group_list = BackupInfo::list_backup_groups(&datastore.base_path())?; | |
436 | ||
437 | group_list.sort_unstable(); | |
438 | ||
439 | let group_count = group_list.len(); | |
440 | task_log!(worker, "found {} groups", group_count); | |
441 | ||
442 | let mut progress = StoreProgress::new(group_count as u64); | |
443 | ||
444 | let latest_only = setup.latest_only.unwrap_or(false); | |
445 | ||
446 | if latest_only { | |
447 | task_log!(worker, "latest-only: true (only considering latest snapshots)"); | |
448 | } | |
449 | ||
450 | let datastore_name = datastore.name(); | |
451 | ||
452 | let mut errors = false; | |
453 | ||
454 | let mut need_catalog = false; // avoid writing catalog for empty jobs | |
455 | ||
456 | for (group_number, group) in group_list.into_iter().enumerate() { | |
457 | progress.done_groups = group_number as u64; | |
458 | progress.done_snapshots = 0; | |
459 | progress.group_snapshots = 0; | |
460 | ||
461 | let snapshot_list = group.list_backups(&datastore.base_path())?; | |
462 | ||
463 | // filter out unfinished backups | |
464 | let mut snapshot_list: Vec<_> = snapshot_list | |
465 | .into_iter() | |
466 | .filter(|item| item.is_finished()) | |
467 | .collect(); | |
468 | ||
469 | if snapshot_list.is_empty() { | |
470 | task_log!(worker, "group {} was empty", group); | |
471 | continue; | |
472 | } | |
473 | ||
474 | BackupInfo::sort_list(&mut snapshot_list, true); // oldest first | |
475 | ||
476 | if latest_only { | |
477 | progress.group_snapshots = 1; | |
478 | if let Some(info) = snapshot_list.pop() { | |
479 | if pool_writer.contains_snapshot(datastore_name, &info.backup_dir.to_string()) { | |
480 | task_log!(worker, "skip snapshot {}", info.backup_dir); | |
481 | continue; | |
482 | } | |
483 | ||
484 | need_catalog = true; | |
485 | ||
486 | let snapshot_name = info.backup_dir.to_string(); | |
487 | if !backup_snapshot(worker, &mut pool_writer, datastore.clone(), info.backup_dir)? { | |
488 | errors = true; | |
489 | } else { | |
490 | summary.snapshot_list.push(snapshot_name); | |
491 | } | |
492 | progress.done_snapshots = 1; | |
493 | task_log!( | |
494 | worker, | |
495 | "percentage done: {}", | |
496 | progress | |
497 | ); | |
498 | } | |
499 | } else { | |
500 | progress.group_snapshots = snapshot_list.len() as u64; | |
501 | for (snapshot_number, info) in snapshot_list.into_iter().enumerate() { | |
502 | if pool_writer.contains_snapshot(datastore_name, &info.backup_dir.to_string()) { | |
503 | task_log!(worker, "skip snapshot {}", info.backup_dir); | |
504 | continue; | |
505 | } | |
506 | ||
507 | need_catalog = true; | |
508 | ||
509 | let snapshot_name = info.backup_dir.to_string(); | |
510 | if !backup_snapshot(worker, &mut pool_writer, datastore.clone(), info.backup_dir)? { | |
511 | errors = true; | |
512 | } else { | |
513 | summary.snapshot_list.push(snapshot_name); | |
514 | } | |
515 | progress.done_snapshots = snapshot_number as u64 + 1; | |
516 | task_log!( | |
517 | worker, | |
518 | "percentage done: {}", | |
519 | progress | |
520 | ); | |
521 | } | |
522 | } | |
523 | } | |
524 | ||
525 | pool_writer.commit()?; | |
526 | ||
527 | if need_catalog { | |
528 | task_log!(worker, "append media catalog"); | |
529 | ||
530 | let uuid = pool_writer.load_writable_media(worker)?; | |
531 | let done = pool_writer.append_catalog_archive(worker)?; | |
532 | if !done { | |
533 | task_log!(worker, "catalog does not fit on tape, writing to next volume"); | |
534 | pool_writer.set_media_status_full(&uuid)?; | |
535 | pool_writer.load_writable_media(worker)?; | |
536 | let done = pool_writer.append_catalog_archive(worker)?; | |
537 | if !done { | |
538 | bail!("write_catalog_archive failed on second media"); | |
539 | } | |
540 | } | |
541 | } | |
542 | ||
543 | if setup.export_media_set.unwrap_or(false) { | |
544 | pool_writer.export_media_set(worker)?; | |
545 | } else if setup.eject_media.unwrap_or(false) { | |
546 | pool_writer.eject_media(worker)?; | |
547 | } | |
548 | ||
549 | if errors { | |
550 | bail!("Tape backup finished with some errors. Please check the task log."); | |
551 | } | |
552 | ||
553 | summary.duration = start.elapsed(); | |
554 | ||
555 | Ok(()) | |
556 | } | |
557 | ||
558 | // Try to update the the media online status | |
559 | fn update_media_online_status(drive: &str) -> Result<Option<String>, Error> { | |
560 | ||
561 | let (config, _digest) = pbs_config::drive::config()?; | |
562 | ||
563 | if let Ok(Some((mut changer, changer_name))) = media_changer(&config, drive) { | |
564 | ||
565 | let label_text_list = changer.online_media_label_texts()?; | |
566 | ||
567 | let status_path = Path::new(TAPE_STATUS_DIR); | |
568 | let mut inventory = Inventory::load(status_path)?; | |
569 | ||
570 | update_changer_online_status( | |
571 | &config, | |
572 | &mut inventory, | |
573 | &changer_name, | |
574 | &label_text_list, | |
575 | )?; | |
576 | ||
577 | Ok(Some(changer_name)) | |
578 | } else { | |
579 | Ok(None) | |
580 | } | |
581 | } | |
582 | ||
583 | pub fn backup_snapshot( | |
584 | worker: &WorkerTask, | |
585 | pool_writer: &mut PoolWriter, | |
586 | datastore: Arc<DataStore>, | |
587 | snapshot: BackupDir, | |
588 | ) -> Result<bool, Error> { | |
589 | ||
590 | task_log!(worker, "backup snapshot {}", snapshot); | |
591 | ||
592 | let snapshot_reader = match SnapshotReader::new(datastore.clone(), snapshot.clone()) { | |
593 | Ok(reader) => reader, | |
594 | Err(err) => { | |
595 | // ignore missing snapshots and continue | |
596 | task_warn!(worker, "failed opening snapshot '{}': {}", snapshot, err); | |
597 | return Ok(false); | |
598 | } | |
599 | }; | |
600 | ||
601 | let snapshot_reader = Arc::new(Mutex::new(snapshot_reader)); | |
602 | ||
603 | let (reader_thread, chunk_iter) = pool_writer.spawn_chunk_reader_thread( | |
604 | datastore.clone(), | |
605 | snapshot_reader.clone(), | |
606 | )?; | |
607 | ||
608 | let mut chunk_iter = chunk_iter.peekable(); | |
609 | ||
610 | loop { | |
611 | worker.check_abort()?; | |
612 | ||
613 | // test is we have remaining chunks | |
614 | match chunk_iter.peek() { | |
615 | None => break, | |
616 | Some(Ok(_)) => { /* Ok */ }, | |
617 | Some(Err(err)) => bail!("{}", err), | |
618 | } | |
619 | ||
620 | let uuid = pool_writer.load_writable_media(worker)?; | |
621 | ||
622 | worker.check_abort()?; | |
623 | ||
624 | let (leom, _bytes) = pool_writer.append_chunk_archive(worker, &mut chunk_iter, datastore.name())?; | |
625 | ||
626 | if leom { | |
627 | pool_writer.set_media_status_full(&uuid)?; | |
628 | } | |
629 | } | |
630 | ||
631 | if let Err(_) = reader_thread.join() { | |
632 | bail!("chunk reader thread failed"); | |
633 | } | |
634 | ||
635 | worker.check_abort()?; | |
636 | ||
637 | let uuid = pool_writer.load_writable_media(worker)?; | |
638 | ||
639 | worker.check_abort()?; | |
640 | ||
641 | let snapshot_reader = snapshot_reader.lock().unwrap(); | |
642 | ||
643 | let (done, _bytes) = pool_writer.append_snapshot_archive(worker, &snapshot_reader)?; | |
644 | ||
645 | if !done { | |
646 | // does not fit on tape, so we try on next volume | |
647 | pool_writer.set_media_status_full(&uuid)?; | |
648 | ||
649 | worker.check_abort()?; | |
650 | ||
651 | pool_writer.load_writable_media(worker)?; | |
652 | let (done, _bytes) = pool_writer.append_snapshot_archive(worker, &snapshot_reader)?; | |
653 | ||
654 | if !done { | |
655 | bail!("write_snapshot_archive failed on second media"); | |
656 | } | |
657 | } | |
658 | ||
659 | task_log!(worker, "end backup {}:{}", datastore.name(), snapshot); | |
660 | ||
661 | Ok(true) | |
662 | } |