]> git.proxmox.com Git - mirror_zfs.git/commitdiff
Zpool can start allocating from metaslab before TRIMs have completed
authorJason King <jasonbking@users.noreply.github.com>
Thu, 12 Oct 2023 18:01:54 +0000 (13:01 -0500)
committerGitHub <noreply@github.com>
Thu, 12 Oct 2023 18:01:54 +0000 (11:01 -0700)
When doing a manual TRIM on a zpool, the metaslab being TRIMmed is
potentially re-enabled before all queued TRIM zios for that metaslab
have completed. Since TRIM zios have the lowest priority, it is
possible to get into a situation where allocations occur from the
just re-enabled metaslab and cut ahead of queued TRIMs to the same
metaslab.  If the ranges overlap, this will cause corruption.

We were able to trigger this pretty consistently with a small single
top-level vdev zpool (i.e. small number of metaslabs) with heavy
parallel write activity while performing a manual TRIM against a
somewhat 'slow' device (so TRIMs took a bit of time to complete).
With the patch, we've not been able to recreate it since. It was on
illumos, but inspection of the OpenZFS trim code looks like the
relevant pieces are largely unchanged and so it appears it would be
vulnerable to the same issue.

Reviewed-by: Igor Kozhukhov <igor@dilos.org>
Reviewed-by: Alexander Motin <mav@FreeBSD.org>
Reviewed-by: Brian Behlendorf <behlendorf1@llnl.gov>
Signed-off-by: Jason King <jking@racktopsystems.com>
Illumos-issue: https://www.illumos.org/issues/15939
Closes #15395

module/zfs/vdev_trim.c

index 0d71b94343420bdbabc35697b4a1dca089e7e143..03e17db024ea03306716e93134b5de8295c752d0 100644 (file)
@@ -23,6 +23,7 @@
  * Copyright (c) 2016 by Delphix. All rights reserved.
  * Copyright (c) 2019 by Lawrence Livermore National Security, LLC.
  * Copyright (c) 2021 Hewlett Packard Enterprise Development LP
+ * Copyright 2023 RackTop Systems, Inc.
  */
 
 #include <sys/spa.h>
@@ -591,6 +592,7 @@ vdev_trim_ranges(trim_args_t *ta)
        uint64_t extent_bytes_max = ta->trim_extent_bytes_max;
        uint64_t extent_bytes_min = ta->trim_extent_bytes_min;
        spa_t *spa = vd->vdev_spa;
+       int error = 0;
 
        ta->trim_start_time = gethrtime();
        ta->trim_bytes_done = 0;
@@ -610,19 +612,32 @@ vdev_trim_ranges(trim_args_t *ta)
                uint64_t writes_required = ((size - 1) / extent_bytes_max) + 1;
 
                for (uint64_t w = 0; w < writes_required; w++) {
-                       int error;
-
                        error = vdev_trim_range(ta, VDEV_LABEL_START_SIZE +
                            rs_get_start(rs, ta->trim_tree) +
                            (w *extent_bytes_max), MIN(size -
                            (w * extent_bytes_max), extent_bytes_max));
                        if (error != 0) {
-                               return (error);
+                               goto done;
                        }
                }
        }
 
-       return (0);
+done:
+       /*
+        * Make sure all TRIMs for this metaslab have completed before
+        * returning. TRIM zios have lower priority over regular or syncing
+        * zios, so all TRIM zios for this metaslab must complete before the
+        * metaslab is re-enabled. Otherwise it's possible write zios to
+        * this metaslab could cut ahead of still queued TRIM zios for this
+        * metaslab causing corruption if the ranges overlap.
+        */
+       mutex_enter(&vd->vdev_trim_io_lock);
+       while (vd->vdev_trim_inflight[0] > 0) {
+               cv_wait(&vd->vdev_trim_io_cv, &vd->vdev_trim_io_lock);
+       }
+       mutex_exit(&vd->vdev_trim_io_lock);
+
+       return (error);
 }
 
 static void
@@ -941,11 +956,6 @@ vdev_trim_thread(void *arg)
        }
 
        spa_config_exit(spa, SCL_CONFIG, FTAG);
-       mutex_enter(&vd->vdev_trim_io_lock);
-       while (vd->vdev_trim_inflight[0] > 0) {
-               cv_wait(&vd->vdev_trim_io_cv, &vd->vdev_trim_io_lock);
-       }
-       mutex_exit(&vd->vdev_trim_io_lock);
 
        range_tree_destroy(ta.trim_tree);