]>
git.proxmox.com Git - pve-manager.git/blob - PVE/CLI/pveceph.pm
1 package PVE
::CLI
::pveceph
;
16 use PVE
::RPCEnvironment
;
18 use PVE
::Tools
qw(run_command);
19 use PVE
::JSONSchema
qw(get_standard_option);
21 use PVE
::Ceph
::Services
;
23 use PVE
::API2
::Ceph
::FS
;
24 use PVE
::API2
::Ceph
::MDS
;
25 use PVE
::API2
::Ceph
::MGR
;
26 use PVE
::API2
::Ceph
::MON
;
27 use PVE
::API2
::Ceph
::OSD
;
31 use base
qw(PVE::CLIHandler);
33 my $nodename = PVE
::INotify
::nodename
();
37 my $status = PVE
::Tools
::upid_read_status
($upid);
38 exit(PVE
::Tools
::upid_status_is_error
($status) ?
-1 : 0);
41 sub setup_environment
{
42 PVE
::RPCEnvironment-
>setup_default_cli_env();
45 __PACKAGE__-
>register_method ({
49 description
=> "Destroy ceph related data and configuration files.",
51 additionalProperties
=> 0,
54 description
=> 'Additionally purge Ceph logs, /var/log/ceph.',
59 description
=> 'Additionally purge Ceph crash logs, /var/lib/ceph/crash.',
65 returns
=> { type
=> 'null' },
76 my $rados = PVE
::RADOS-
>new();
77 $pools = PVE
::Ceph
::Tools
::ls_pools
(undef, $rados);
78 $monstat = PVE
::Ceph
::Services
::get_services_info
('mon', undef, $rados);
79 $mdsstat = PVE
::Ceph
::Services
::get_services_info
('mds', undef, $rados);
80 $osdstat = $rados->mon_command({ prefix
=> 'osd metadata' });
82 warn "Error gathering ceph info, already purged? Message: $@" if $@;
84 my $osd = grep { $_->{hostname
} eq $nodename } @$osdstat;
85 my $mds = grep { $mdsstat->{$_}->{host
} eq $nodename } keys %$mdsstat;
86 my $mon = grep { $monstat->{$_}->{host
} eq $nodename } keys %$monstat;
89 $message .= "- remove pools, this will !!DESTROY DATA!!\n" if @$pools;
90 $message .= "- remove active OSD on $nodename\n" if $osd;
91 $message .= "- remove active MDS on $nodename\n" if $mds;
92 $message .= "- remove other MONs, $nodename is not the last MON\n"
93 if scalar(keys %$monstat) > 1 && $mon;
95 # display all steps at once
96 die "Unable to purge Ceph!\n\nTo continue:\n$message" if $message;
98 my $services = PVE
::Ceph
::Services
::get_local_services
();
99 $services->{mon
} = $monstat if $mon;
100 $services->{crash
}->{$nodename} = { direxists
=> 1 } if $param->{crash
};
101 $services->{logs
}->{$nodename} = { direxists
=> 1 } if $param->{logs
};
103 PVE
::Ceph
::Tools
::purge_all_ceph_services
($services);
104 PVE
::Ceph
::Tools
::purge_all_ceph_files
($services);
109 my $supported_ceph_versions = ['quincy'];
110 my $default_ceph_version = 'quincy';
112 __PACKAGE__-
>register_method ({
116 description
=> "Install ceph related packages.",
118 additionalProperties
=> 0,
122 enum
=> $supported_ceph_versions,
123 default => $default_ceph_version,
124 description
=> "Ceph version to install.",
127 'allow-experimental' => {
131 description
=> "Allow experimental versions. Use with care!",
133 'test-repository' => {
137 description
=> "Use the test, not the main repository. Use with care!",
141 returns
=> { type
=> 'null' },
145 my $cephver = $param->{version
} || $default_ceph_version;
147 my $repo = $param->{'test-repository'} ?
'test' : 'main';
150 if ($cephver eq 'quincy') {
151 $repolist = "deb http://download.proxmox.com/debian/ceph-quincy bookworm $repo\n";
153 die "unsupported ceph version: $cephver";
155 PVE
::Tools
::file_set_contents
("/etc/apt/sources.list.d/ceph.list", $repolist);
157 my $supported_re = join('|', $supported_ceph_versions->@*);
158 warn "WARNING: installing non-default ceph release '$cephver'!\n" if $cephver !~ qr/^(?:$supported_re)$/;
160 local $ENV{DEBIAN_FRONTEND
} = 'noninteractive';
161 print "update available package list\n";
164 ['apt-get', '-q', 'update'],
166 errfunc
=> sub { print STDERR
"$_[0]\n" },
170 my @apt_install = qw(apt-get --no-install-recommends -o Dpkg::Options::=--force-confnew install --);
171 my @ceph_packages = qw(
180 # got split out with quincy and is required by PVE tooling, conditionally exclude it for older
181 # FIXME: remove condition with PVE 8.0, i.e., once we only support quincy+ new installations
182 if ($cephver ne 'octopus' and $cephver ne 'pacific') {
183 push @ceph_packages, 'ceph-volume';
186 print "start installation\n";
188 # this flag helps to determine when apt is actually done installing (vs. partial extracing)
189 my $install_flag_fn = PVE
::Ceph
::Tools
::ceph_install_flag_file
();
190 open(my $install_flag, '>', $install_flag_fn) or die "could not create install flag - $!\n";
193 if (system(@apt_install, @ceph_packages) != 0) {
194 unlink $install_flag_fn or warn "could not remove Ceph installation flag - $!";
195 die "apt failed during ceph installation ($?)\n";
198 print "\ninstalled ceph $cephver successfully!\n";
199 # done: drop flag file so that the PVE::Ceph::Tools check returns Ok now.
200 unlink $install_flag_fn or warn "could not remove Ceph installation flag - $!";
202 print "\nreloading API to load new Ceph RADOS library...\n";
204 'systemctl', 'try-reload-or-restart', 'pvedaemon.service', 'pveproxy.service'
210 __PACKAGE__-
>register_method ({
214 description
=> "Get Ceph Status.",
216 additionalProperties
=> 0,
218 returns
=> { type
=> 'null' },
220 PVE
::Ceph
::Tools
::check_ceph_inited
();
224 outfunc
=> sub { print "$_[0]\n" },
225 errfunc
=> sub { print STDERR
"$_[0]\n" },
231 my $get_storages = sub {
232 my ($fs, $is_default) = @_;
234 my $cfg = PVE
::Storage
::config
();
236 my $storages = $cfg->{ids
};
238 foreach my $storeid (keys %$storages) {
239 my $curr = $storages->{$storeid};
240 next if $curr->{type
} ne 'cephfs';
241 my $cur_fs = $curr->{'fs-name'};
242 $res->{$storeid} = $storages->{$storeid}
243 if (!defined($cur_fs) && $is_default) || (defined($cur_fs) && $fs eq $cur_fs);
249 __PACKAGE__-
>register_method ({
253 description
=> "Destroy a Ceph filesystem",
255 additionalProperties
=> 0,
257 node
=> get_standard_option
('pve-node'),
259 description
=> "The ceph filesystem name.",
262 'remove-storages' => {
263 description
=> "Remove all pveceph-managed storages configured for this fs.",
269 description
=> "Remove data and metadata pools configured for this fs.",
276 returns
=> { type
=> 'string' },
280 PVE
::Ceph
::Tools
::check_ceph_inited
();
282 my $rpcenv = PVE
::RPCEnvironment
::get
();
283 my $user = $rpcenv->get_user();
285 my $fs_name = $param->{name
};
288 my $fs_list = PVE
::Ceph
::Tools
::ls_fs
();
289 for my $entry (@$fs_list) {
290 next if $entry->{name
} ne $fs_name;
294 die "no such cephfs '$fs_name'\n" if !$fs;
297 my $rados = PVE
::RADOS-
>new();
299 if ($param->{'remove-storages'}) {
301 my $fs_dump = $rados->mon_command({ prefix
=> "fs dump" });
302 for my $fs ($fs_dump->{filesystems
}->@*) {
303 next if $fs->{id
} != $fs_dump->{default_fscid
};
304 $defaultfs = $fs->{mdsmap
}->{fs_name
};
306 warn "no default fs found, maybe not all relevant storages are removed\n"
307 if !defined($defaultfs);
309 my $storages = $get_storages->($fs_name, $fs_name eq ($defaultfs // ''));
310 for my $storeid (keys %$storages) {
311 my $store = $storages->{$storeid};
312 if (!$store->{disable
}) {
313 die "storage '$storeid' is not disabled, make sure to disable ".
314 "and unmount the storage first\n";
319 for my $storeid (keys %$storages) {
320 # skip external clusters, not managed by pveceph
321 next if $storages->{$storeid}->{monhost
};
322 eval { PVE
::API2
::Storage
::Config-
>delete({storage
=> $storeid}) };
324 warn "failed to remove storage '$storeid': $@\n";
328 die "failed to remove (some) storages - check log and remove manually!\n"
332 PVE
::Ceph
::Tools
::destroy_fs
($fs_name, $rados);
334 if ($param->{'remove-pools'}) {
335 warn "removing metadata pool '$fs->{metadata_pool}'\n";
336 eval { PVE
::Ceph
::Tools
::destroy_pool
($fs->{metadata_pool
}, $rados) };
339 foreach my $pool ($fs->{data_pools
}->@*) {
340 warn "removing data pool '$pool'\n";
341 eval { PVE
::Ceph
::Tools
::destroy_pool
($pool, $rados) };
347 return $rpcenv->fork_worker('cephdestroyfs', $fs_name, $user, $worker);
351 init
=> [ 'PVE::API2::Ceph', 'init', [], { node
=> $nodename } ],
353 ls
=> [ 'PVE::API2::Ceph::Pool', 'lspools', [], { node
=> $nodename }, sub {
354 my ($data, $schema, $options) = @_;
355 PVE
::CLIFormatter
::print_api_result
($data, $schema,
371 }, $PVE::RESTHandler
::standard_output_options
],
372 create
=> [ 'PVE::API2::Ceph::Pool', 'createpool', ['name'], { node
=> $nodename }],
373 destroy
=> [ 'PVE::API2::Ceph::Pool', 'destroypool', ['name'], { node
=> $nodename } ],
374 set
=> [ 'PVE::API2::Ceph::Pool', 'setpool', ['name'], { node
=> $nodename } ],
375 get
=> [ 'PVE::API2::Ceph::Pool', 'getpool', ['name'], { node
=> $nodename }, sub {
376 my ($data, $schema, $options) = @_;
377 PVE
::CLIFormatter
::print_api_result
($data, $schema, undef, $options);
378 }, $PVE::RESTHandler
::standard_output_options
],
380 lspools
=> { alias
=> 'pool ls' },
381 createpool
=> { alias
=> 'pool create' },
382 destroypool
=> { alias
=> 'pool destroy' },
384 create
=> [ 'PVE::API2::Ceph::FS', 'createfs', [], { node
=> $nodename }],
385 destroy
=> [ __PACKAGE__
, 'destroyfs', ['name'], { node
=> $nodename }],
388 create
=> [ 'PVE::API2::Ceph::OSD', 'createosd', ['dev'], { node
=> $nodename }, $upid_exit],
389 destroy
=> [ 'PVE::API2::Ceph::OSD', 'destroyosd', ['osdid'], { node
=> $nodename }, $upid_exit],
391 createosd
=> { alias
=> 'osd create' },
392 destroyosd
=> { alias
=> 'osd destroy' },
394 create
=> [ 'PVE::API2::Ceph::MON', 'createmon', [], { node
=> $nodename }, $upid_exit],
395 destroy
=> [ 'PVE::API2::Ceph::MON', 'destroymon', ['monid'], { node
=> $nodename }, $upid_exit],
397 createmon
=> { alias
=> 'mon create' },
398 destroymon
=> { alias
=> 'mon destroy' },
400 create
=> [ 'PVE::API2::Ceph::MGR', 'createmgr', [], { node
=> $nodename }, $upid_exit],
401 destroy
=> [ 'PVE::API2::Ceph::MGR', 'destroymgr', ['id'], { node
=> $nodename }, $upid_exit],
403 createmgr
=> { alias
=> 'mgr create' },
404 destroymgr
=> { alias
=> 'mgr destroy' },
406 create
=> [ 'PVE::API2::Ceph::MDS', 'createmds', [], { node
=> $nodename }, $upid_exit],
407 destroy
=> [ 'PVE::API2::Ceph::MDS', 'destroymds', ['name'], { node
=> $nodename }, $upid_exit],
409 start
=> [ 'PVE::API2::Ceph', 'start', [], { node
=> $nodename }, $upid_exit],
410 stop
=> [ 'PVE::API2::Ceph', 'stop', [], { node
=> $nodename }, $upid_exit],
411 install
=> [ __PACKAGE__
, 'install', [] ],
412 purge
=> [ __PACKAGE__
, 'purge', [] ],
413 status
=> [ __PACKAGE__
, 'status', []],