cmake_minimum_required(VERSION 3.16)
project(ceph
- VERSION 17.2.4
+ VERSION 17.2.5
LANGUAGES CXX C ASM)
cmake_policy(SET CMP0028 NEW)
# main package definition
#################################################################################
Name: ceph
-Version: 17.2.4
+Version: 17.2.5
Release: 0%{?dist}
%if 0%{?fedora} || 0%{?rhel}
Epoch: 2
Group: System/Filesystems
%endif
URL: http://ceph.com/
-Source0: %{?_remote_tarball_prefix}ceph-17.2.4.tar.bz2
+Source0: %{?_remote_tarball_prefix}ceph-17.2.5.tar.bz2
%if 0%{?suse_version}
# _insert_obs_source_lines_here
ExclusiveArch: x86_64 aarch64 ppc64le s390x
# common
#################################################################################
%prep
-%autosetup -p1 -n ceph-17.2.4
+%autosetup -p1 -n ceph-17.2.5
%build
# Disable lto on systems that do not support symver attribute
+ceph (17.2.5-1) stable; urgency=medium
+
+ * New upstream release
+
+ -- Ceph Release Team <ceph-maintainers@ceph.io> Mon, 17 Oct 2022 20:07:27 +0000
+
ceph (17.2.4-1) stable; urgency=medium
* New upstream release
-1353ed37dec8d74973edc3d5d5908c20ad5a7332
-17.2.4
+98318ae89f1a893a6ded3a640405cdbb33e08757
+17.2.5
'failed to determine if parent device {} is BlueStore. err: {}'.format(parent, e)))
continue
- bs_info = _get_bluestore_info(dev)
- if bs_info is None:
- # None is also returned in the rare event that there is an issue reading info from
- # a BlueStore disk, so be sure to log our assumption that it isn't bluestore
- logger.info('device {} does not have BlueStore information'.format(dev))
- continue
- uuid = bs_info['osd_uuid']
- if uuid not in result:
- result[uuid] = {}
- result[uuid].update(bs_info)
+ bs_info = _get_bluestore_info(dev)
+ if bs_info is None:
+ # None is also returned in the rare event that there is an issue reading info from
+ # a BlueStore disk, so be sure to log our assumption that it isn't bluestore
+ logger.info('device {} does not have BlueStore information'.format(dev))
+ continue
+ uuid = bs_info['osd_uuid']
+ if uuid not in result:
+ result[uuid] = {}
+ result[uuid].update(bs_info)
return result
was_min = true;
}
acting_readable_until_ub[i] = a.readable_until_ub;
+ break;
}
- break;
}
}
if (was_min) {
auto old_ru = readable_until;
recalc_readable_until();
- if (now >= old_ru) {
+ if (now < old_ru) {
pl->recheck_readable();
}
}
all_nfs = self.rook_cluster.get_resource("cephnfses")
nfs_pods = self.rook_cluster.describe_pods('nfs', None, None)
for nfs in all_nfs:
- if nfs['spec']['rados']['pool'] != NFS_POOL_NAME:
- continue
+ # Starting with V.17.2.0, the 'rados' spec part in 'cephnfs' resources does not contain the 'pool' item
+ if 'pool' in nfs['spec']['rados']:
+ if nfs['spec']['rados']['pool'] != NFS_POOL_NAME:
+ continue
nfs_name = nfs['metadata']['name']
svc = 'nfs.' + nfs_name
if svc in spec:
# Grab output from the "daemon.x heap stats" command
for daemon in daemons:
- daemon_type, daemon_id = daemon.split('.')
+ daemon_type, daemon_id = daemon.split('.', 1)
heap_stats = self.parse_heap_stats(daemon_type, daemon_id)
if heap_stats:
if (daemon_type != 'osd'):
# Grab output from the "dump_mempools" command
for daemon in daemons:
- daemon_type, daemon_id = daemon.split('.')
+ daemon_type, daemon_id = daemon.split('.', 1)
cmd_dict = {
'prefix': 'dump_mempools',
'format': 'json'