Print this page
3006 VERIFY[S,U,P] and ASSERT[S,U,P] frequently check if first argument is zero
Split |
Close |
Expand all |
Collapse all |
--- old/usr/src/uts/common/fs/zfs/dsl_deadlist.c
+++ new/usr/src/uts/common/fs/zfs/dsl_deadlist.c
1 1 /*
2 2 * CDDL HEADER START
3 3 *
4 4 * The contents of this file are subject to the terms of the
5 5 * Common Development and Distribution License (the "License").
6 6 * You may not use this file except in compliance with the License.
7 7 *
8 8 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
9 9 * or http://www.opensolaris.org/os/licensing.
10 10 * See the License for the specific language governing permissions
11 11 * and limitations under the License.
12 12 *
↓ open down ↓ |
12 lines elided |
↑ open up ↑ |
13 13 * When distributing Covered Code, include this CDDL HEADER in each
14 14 * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
15 15 * If applicable, add the following below this CDDL HEADER, with the
16 16 * fields enclosed by brackets "[]" replaced with your own identifying
17 17 * information: Portions Copyright [yyyy] [name of copyright owner]
18 18 *
19 19 * CDDL HEADER END
20 20 */
21 21 /*
22 22 * Copyright (c) 2010, Oracle and/or its affiliates. All rights reserved.
23 - * Copyright (c) 2011 by Delphix. All rights reserved.
23 + * Copyright (c) 2012 by Delphix. All rights reserved.
24 24 */
25 25
26 26 #include <sys/dsl_dataset.h>
27 27 #include <sys/dmu.h>
28 28 #include <sys/refcount.h>
29 29 #include <sys/zap.h>
30 30 #include <sys/zfs_context.h>
31 31 #include <sys/dsl_pool.h>
32 32
33 33 /*
34 34 * Deadlist concurrency:
35 35 *
36 36 * Deadlists can only be modified from the syncing thread.
37 37 *
38 38 * Except for dsl_deadlist_insert(), it can only be modified with the
39 39 * dp_config_rwlock held with RW_WRITER.
40 40 *
41 41 * The accessors (dsl_deadlist_space() and dsl_deadlist_space_range()) can
42 42 * be called concurrently, from open context, with the dl_config_rwlock held
43 43 * with RW_READER.
44 44 *
45 45 * Therefore, we only need to provide locking between dsl_deadlist_insert() and
46 46 * the accessors, protecting:
47 47 * dl_phys->dl_used,comp,uncomp
48 48 * and protecting the dl_tree from being loaded.
49 49 * The locking is provided by dl_lock. Note that locking on the bpobj_t
50 50 * provides its own locking, and dl_oldfmt is immutable.
51 51 */
52 52
53 53 static int
54 54 dsl_deadlist_compare(const void *arg1, const void *arg2)
55 55 {
56 56 const dsl_deadlist_entry_t *dle1 = arg1;
57 57 const dsl_deadlist_entry_t *dle2 = arg2;
58 58
59 59 if (dle1->dle_mintxg < dle2->dle_mintxg)
60 60 return (-1);
61 61 else if (dle1->dle_mintxg > dle2->dle_mintxg)
62 62 return (+1);
63 63 else
64 64 return (0);
65 65 }
66 66
67 67 static void
68 68 dsl_deadlist_load_tree(dsl_deadlist_t *dl)
69 69 {
70 70 zap_cursor_t zc;
71 71 zap_attribute_t za;
72 72
73 73 ASSERT(!dl->dl_oldfmt);
74 74 if (dl->dl_havetree)
↓ open down ↓ |
41 lines elided |
↑ open up ↑ |
75 75 return;
76 76
77 77 avl_create(&dl->dl_tree, dsl_deadlist_compare,
78 78 sizeof (dsl_deadlist_entry_t),
79 79 offsetof(dsl_deadlist_entry_t, dle_node));
80 80 for (zap_cursor_init(&zc, dl->dl_os, dl->dl_object);
81 81 zap_cursor_retrieve(&zc, &za) == 0;
82 82 zap_cursor_advance(&zc)) {
83 83 dsl_deadlist_entry_t *dle = kmem_alloc(sizeof (*dle), KM_SLEEP);
84 84 dle->dle_mintxg = strtonum(za.za_name, NULL);
85 - VERIFY3U(0, ==, bpobj_open(&dle->dle_bpobj, dl->dl_os,
85 + VERIFY0(bpobj_open(&dle->dle_bpobj, dl->dl_os,
86 86 za.za_first_integer));
87 87 avl_add(&dl->dl_tree, dle);
88 88 }
89 89 zap_cursor_fini(&zc);
90 90 dl->dl_havetree = B_TRUE;
91 91 }
92 92
93 93 void
94 94 dsl_deadlist_open(dsl_deadlist_t *dl, objset_t *os, uint64_t object)
95 95 {
96 96 dmu_object_info_t doi;
97 97
98 98 mutex_init(&dl->dl_lock, NULL, MUTEX_DEFAULT, NULL);
99 99 dl->dl_os = os;
100 100 dl->dl_object = object;
101 - VERIFY3U(0, ==, dmu_bonus_hold(os, object, dl, &dl->dl_dbuf));
101 + VERIFY0(dmu_bonus_hold(os, object, dl, &dl->dl_dbuf));
102 102 dmu_object_info_from_db(dl->dl_dbuf, &doi);
103 103 if (doi.doi_type == DMU_OT_BPOBJ) {
104 104 dmu_buf_rele(dl->dl_dbuf, dl);
105 105 dl->dl_dbuf = NULL;
106 106 dl->dl_oldfmt = B_TRUE;
107 - VERIFY3U(0, ==, bpobj_open(&dl->dl_bpobj, os, object));
107 + VERIFY0(bpobj_open(&dl->dl_bpobj, os, object));
108 108 return;
109 109 }
110 110
111 111 dl->dl_oldfmt = B_FALSE;
112 112 dl->dl_phys = dl->dl_dbuf->db_data;
113 113 dl->dl_havetree = B_FALSE;
114 114 }
115 115
116 116 void
117 117 dsl_deadlist_close(dsl_deadlist_t *dl)
118 118 {
119 119 void *cookie = NULL;
120 120 dsl_deadlist_entry_t *dle;
121 121
122 122 if (dl->dl_oldfmt) {
123 123 dl->dl_oldfmt = B_FALSE;
124 124 bpobj_close(&dl->dl_bpobj);
125 125 return;
126 126 }
127 127
128 128 if (dl->dl_havetree) {
129 129 while ((dle = avl_destroy_nodes(&dl->dl_tree, &cookie))
130 130 != NULL) {
131 131 bpobj_close(&dle->dle_bpobj);
132 132 kmem_free(dle, sizeof (*dle));
133 133 }
134 134 avl_destroy(&dl->dl_tree);
135 135 }
136 136 dmu_buf_rele(dl->dl_dbuf, dl);
137 137 mutex_destroy(&dl->dl_lock);
138 138 dl->dl_dbuf = NULL;
139 139 dl->dl_phys = NULL;
140 140 }
141 141
142 142 uint64_t
143 143 dsl_deadlist_alloc(objset_t *os, dmu_tx_t *tx)
144 144 {
145 145 if (spa_version(dmu_objset_spa(os)) < SPA_VERSION_DEADLISTS)
146 146 return (bpobj_alloc(os, SPA_MAXBLOCKSIZE, tx));
147 147 return (zap_create(os, DMU_OT_DEADLIST, DMU_OT_DEADLIST_HDR,
↓ open down ↓ |
30 lines elided |
↑ open up ↑ |
148 148 sizeof (dsl_deadlist_phys_t), tx));
149 149 }
150 150
151 151 void
152 152 dsl_deadlist_free(objset_t *os, uint64_t dlobj, dmu_tx_t *tx)
153 153 {
154 154 dmu_object_info_t doi;
155 155 zap_cursor_t zc;
156 156 zap_attribute_t za;
157 157
158 - VERIFY3U(0, ==, dmu_object_info(os, dlobj, &doi));
158 + VERIFY0(dmu_object_info(os, dlobj, &doi));
159 159 if (doi.doi_type == DMU_OT_BPOBJ) {
160 160 bpobj_free(os, dlobj, tx);
161 161 return;
162 162 }
163 163
164 164 for (zap_cursor_init(&zc, os, dlobj);
165 165 zap_cursor_retrieve(&zc, &za) == 0;
166 166 zap_cursor_advance(&zc))
167 167 bpobj_free(os, za.za_first_integer, tx);
168 168 zap_cursor_fini(&zc);
169 - VERIFY3U(0, ==, dmu_object_free(os, dlobj, tx));
169 + VERIFY0(dmu_object_free(os, dlobj, tx));
170 170 }
171 171
172 172 void
173 173 dsl_deadlist_insert(dsl_deadlist_t *dl, const blkptr_t *bp, dmu_tx_t *tx)
174 174 {
175 175 dsl_deadlist_entry_t dle_tofind;
176 176 dsl_deadlist_entry_t *dle;
177 177 avl_index_t where;
178 178
179 179 if (dl->dl_oldfmt) {
180 180 bpobj_enqueue(&dl->dl_bpobj, bp, tx);
181 181 return;
182 182 }
183 183
184 184 dsl_deadlist_load_tree(dl);
185 185
186 186 dmu_buf_will_dirty(dl->dl_dbuf, tx);
187 187 mutex_enter(&dl->dl_lock);
188 188 dl->dl_phys->dl_used +=
189 189 bp_get_dsize_sync(dmu_objset_spa(dl->dl_os), bp);
190 190 dl->dl_phys->dl_comp += BP_GET_PSIZE(bp);
191 191 dl->dl_phys->dl_uncomp += BP_GET_UCSIZE(bp);
192 192 mutex_exit(&dl->dl_lock);
193 193
194 194 dle_tofind.dle_mintxg = bp->blk_birth;
195 195 dle = avl_find(&dl->dl_tree, &dle_tofind, &where);
196 196 if (dle == NULL)
197 197 dle = avl_nearest(&dl->dl_tree, where, AVL_BEFORE);
198 198 else
199 199 dle = AVL_PREV(&dl->dl_tree, dle);
200 200 bpobj_enqueue(&dle->dle_bpobj, bp, tx);
201 201 }
202 202
203 203 /*
204 204 * Insert new key in deadlist, which must be > all current entries.
205 205 * mintxg is not inclusive.
206 206 */
207 207 void
208 208 dsl_deadlist_add_key(dsl_deadlist_t *dl, uint64_t mintxg, dmu_tx_t *tx)
209 209 {
210 210 uint64_t obj;
↓ open down ↓ |
31 lines elided |
↑ open up ↑ |
211 211 dsl_deadlist_entry_t *dle;
212 212
213 213 if (dl->dl_oldfmt)
214 214 return;
215 215
216 216 dsl_deadlist_load_tree(dl);
217 217
218 218 dle = kmem_alloc(sizeof (*dle), KM_SLEEP);
219 219 dle->dle_mintxg = mintxg;
220 220 obj = bpobj_alloc(dl->dl_os, SPA_MAXBLOCKSIZE, tx);
221 - VERIFY3U(0, ==, bpobj_open(&dle->dle_bpobj, dl->dl_os, obj));
221 + VERIFY0(bpobj_open(&dle->dle_bpobj, dl->dl_os, obj));
222 222 avl_add(&dl->dl_tree, dle);
223 223
224 - VERIFY3U(0, ==, zap_add_int_key(dl->dl_os, dl->dl_object,
224 + VERIFY0(zap_add_int_key(dl->dl_os, dl->dl_object,
225 225 mintxg, obj, tx));
226 226 }
227 227
228 228 /*
229 229 * Remove this key, merging its entries into the previous key.
230 230 */
231 231 void
232 232 dsl_deadlist_remove_key(dsl_deadlist_t *dl, uint64_t mintxg, dmu_tx_t *tx)
233 233 {
234 234 dsl_deadlist_entry_t dle_tofind;
235 235 dsl_deadlist_entry_t *dle, *dle_prev;
236 236
237 237 if (dl->dl_oldfmt)
238 238 return;
239 239
240 240 dsl_deadlist_load_tree(dl);
241 241
242 242 dle_tofind.dle_mintxg = mintxg;
↓ open down ↓ |
8 lines elided |
↑ open up ↑ |
243 243 dle = avl_find(&dl->dl_tree, &dle_tofind, NULL);
244 244 dle_prev = AVL_PREV(&dl->dl_tree, dle);
245 245
246 246 bpobj_enqueue_subobj(&dle_prev->dle_bpobj,
247 247 dle->dle_bpobj.bpo_object, tx);
248 248
249 249 avl_remove(&dl->dl_tree, dle);
250 250 bpobj_close(&dle->dle_bpobj);
251 251 kmem_free(dle, sizeof (*dle));
252 252
253 - VERIFY3U(0, ==, zap_remove_int(dl->dl_os, dl->dl_object, mintxg, tx));
253 + VERIFY0(zap_remove_int(dl->dl_os, dl->dl_object, mintxg, tx));
254 254 }
255 255
256 256 /*
257 257 * Walk ds's snapshots to regenerate generate ZAP & AVL.
258 258 */
259 259 static void
260 260 dsl_deadlist_regenerate(objset_t *os, uint64_t dlobj,
261 261 uint64_t mrs_obj, dmu_tx_t *tx)
262 262 {
263 263 dsl_deadlist_t dl;
264 264 dsl_pool_t *dp = dmu_objset_pool(os);
265 265
266 266 dsl_deadlist_open(&dl, os, dlobj);
267 267 if (dl.dl_oldfmt) {
268 268 dsl_deadlist_close(&dl);
269 269 return;
270 270 }
271 271
272 272 while (mrs_obj != 0) {
273 273 dsl_dataset_t *ds;
274 - VERIFY3U(0, ==, dsl_dataset_hold_obj(dp, mrs_obj, FTAG, &ds));
274 + VERIFY0(dsl_dataset_hold_obj(dp, mrs_obj, FTAG, &ds));
275 275 dsl_deadlist_add_key(&dl, ds->ds_phys->ds_prev_snap_txg, tx);
276 276 mrs_obj = ds->ds_phys->ds_prev_snap_obj;
277 277 dsl_dataset_rele(ds, FTAG);
278 278 }
279 279 dsl_deadlist_close(&dl);
280 280 }
281 281
282 282 uint64_t
283 283 dsl_deadlist_clone(dsl_deadlist_t *dl, uint64_t maxtxg,
284 284 uint64_t mrs_obj, dmu_tx_t *tx)
285 285 {
286 286 dsl_deadlist_entry_t *dle;
287 287 uint64_t newobj;
288 288
289 289 newobj = dsl_deadlist_alloc(dl->dl_os, tx);
290 290
291 291 if (dl->dl_oldfmt) {
292 292 dsl_deadlist_regenerate(dl->dl_os, newobj, mrs_obj, tx);
293 293 return (newobj);
294 294 }
295 295
↓ open down ↓ |
11 lines elided |
↑ open up ↑ |
296 296 dsl_deadlist_load_tree(dl);
297 297
298 298 for (dle = avl_first(&dl->dl_tree); dle;
299 299 dle = AVL_NEXT(&dl->dl_tree, dle)) {
300 300 uint64_t obj;
301 301
302 302 if (dle->dle_mintxg >= maxtxg)
303 303 break;
304 304
305 305 obj = bpobj_alloc(dl->dl_os, SPA_MAXBLOCKSIZE, tx);
306 - VERIFY3U(0, ==, zap_add_int_key(dl->dl_os, newobj,
306 + VERIFY0(zap_add_int_key(dl->dl_os, newobj,
307 307 dle->dle_mintxg, obj, tx));
308 308 }
309 309 return (newobj);
310 310 }
311 311
312 312 void
313 313 dsl_deadlist_space(dsl_deadlist_t *dl,
314 314 uint64_t *usedp, uint64_t *compp, uint64_t *uncompp)
315 315 {
316 316 if (dl->dl_oldfmt) {
317 - VERIFY3U(0, ==, bpobj_space(&dl->dl_bpobj,
317 + VERIFY0(bpobj_space(&dl->dl_bpobj,
318 318 usedp, compp, uncompp));
319 319 return;
320 320 }
321 321
322 322 mutex_enter(&dl->dl_lock);
323 323 *usedp = dl->dl_phys->dl_used;
324 324 *compp = dl->dl_phys->dl_comp;
325 325 *uncompp = dl->dl_phys->dl_uncomp;
326 326 mutex_exit(&dl->dl_lock);
327 327 }
328 328
329 329 /*
330 330 * return space used in the range (mintxg, maxtxg].
331 331 * Includes maxtxg, does not include mintxg.
332 332 * mintxg and maxtxg must both be keys in the deadlist (unless maxtxg is
333 333 * larger than any bp in the deadlist (eg. UINT64_MAX)).
↓ open down ↓ |
6 lines elided |
↑ open up ↑ |
334 334 */
335 335 void
336 336 dsl_deadlist_space_range(dsl_deadlist_t *dl, uint64_t mintxg, uint64_t maxtxg,
337 337 uint64_t *usedp, uint64_t *compp, uint64_t *uncompp)
338 338 {
339 339 dsl_deadlist_entry_t *dle;
340 340 dsl_deadlist_entry_t dle_tofind;
341 341 avl_index_t where;
342 342
343 343 if (dl->dl_oldfmt) {
344 - VERIFY3U(0, ==, bpobj_space_range(&dl->dl_bpobj,
344 + VERIFY0(bpobj_space_range(&dl->dl_bpobj,
345 345 mintxg, maxtxg, usedp, compp, uncompp));
346 346 return;
347 347 }
348 348
349 349 *usedp = *compp = *uncompp = 0;
350 350
351 351 mutex_enter(&dl->dl_lock);
352 352 dsl_deadlist_load_tree(dl);
353 353 dle_tofind.dle_mintxg = mintxg;
354 354 dle = avl_find(&dl->dl_tree, &dle_tofind, &where);
355 355 /*
↓ open down ↓ |
1 lines elided |
↑ open up ↑ |
356 356 * If we don't find this mintxg, there shouldn't be anything
357 357 * after it either.
358 358 */
359 359 ASSERT(dle != NULL ||
360 360 avl_nearest(&dl->dl_tree, where, AVL_AFTER) == NULL);
361 361
362 362 for (; dle && dle->dle_mintxg < maxtxg;
363 363 dle = AVL_NEXT(&dl->dl_tree, dle)) {
364 364 uint64_t used, comp, uncomp;
365 365
366 - VERIFY3U(0, ==, bpobj_space(&dle->dle_bpobj,
366 + VERIFY0(bpobj_space(&dle->dle_bpobj,
367 367 &used, &comp, &uncomp));
368 368
369 369 *usedp += used;
370 370 *compp += comp;
371 371 *uncompp += uncomp;
372 372 }
373 373 mutex_exit(&dl->dl_lock);
374 374 }
375 375
376 376 static void
377 377 dsl_deadlist_insert_bpobj(dsl_deadlist_t *dl, uint64_t obj, uint64_t birth,
378 378 dmu_tx_t *tx)
379 379 {
380 380 dsl_deadlist_entry_t dle_tofind;
381 381 dsl_deadlist_entry_t *dle;
382 382 avl_index_t where;
383 383 uint64_t used, comp, uncomp;
384 384 bpobj_t bpo;
385 385
386 - VERIFY3U(0, ==, bpobj_open(&bpo, dl->dl_os, obj));
387 - VERIFY3U(0, ==, bpobj_space(&bpo, &used, &comp, &uncomp));
386 + VERIFY0(bpobj_open(&bpo, dl->dl_os, obj));
387 + VERIFY0(bpobj_space(&bpo, &used, &comp, &uncomp));
388 388 bpobj_close(&bpo);
389 389
390 390 dsl_deadlist_load_tree(dl);
391 391
392 392 dmu_buf_will_dirty(dl->dl_dbuf, tx);
393 393 mutex_enter(&dl->dl_lock);
394 394 dl->dl_phys->dl_used += used;
395 395 dl->dl_phys->dl_comp += comp;
396 396 dl->dl_phys->dl_uncomp += uncomp;
397 397 mutex_exit(&dl->dl_lock);
398 398
399 399 dle_tofind.dle_mintxg = birth;
400 400 dle = avl_find(&dl->dl_tree, &dle_tofind, &where);
401 401 if (dle == NULL)
402 402 dle = avl_nearest(&dl->dl_tree, where, AVL_BEFORE);
403 403 bpobj_enqueue_subobj(&dle->dle_bpobj, obj, tx);
404 404 }
405 405
406 406 static int
407 407 dsl_deadlist_insert_cb(void *arg, const blkptr_t *bp, dmu_tx_t *tx)
408 408 {
409 409 dsl_deadlist_t *dl = arg;
410 410 dsl_deadlist_insert(dl, bp, tx);
411 411 return (0);
412 412 }
413 413
414 414 /*
415 415 * Merge the deadlist pointed to by 'obj' into dl. obj will be left as
416 416 * an empty deadlist.
↓ open down ↓ |
19 lines elided |
↑ open up ↑ |
417 417 */
418 418 void
419 419 dsl_deadlist_merge(dsl_deadlist_t *dl, uint64_t obj, dmu_tx_t *tx)
420 420 {
421 421 zap_cursor_t zc;
422 422 zap_attribute_t za;
423 423 dmu_buf_t *bonus;
424 424 dsl_deadlist_phys_t *dlp;
425 425 dmu_object_info_t doi;
426 426
427 - VERIFY3U(0, ==, dmu_object_info(dl->dl_os, obj, &doi));
427 + VERIFY0(dmu_object_info(dl->dl_os, obj, &doi));
428 428 if (doi.doi_type == DMU_OT_BPOBJ) {
429 429 bpobj_t bpo;
430 - VERIFY3U(0, ==, bpobj_open(&bpo, dl->dl_os, obj));
431 - VERIFY3U(0, ==, bpobj_iterate(&bpo,
430 + VERIFY0(bpobj_open(&bpo, dl->dl_os, obj));
431 + VERIFY0(bpobj_iterate(&bpo,
432 432 dsl_deadlist_insert_cb, dl, tx));
433 433 bpobj_close(&bpo);
434 434 return;
435 435 }
436 436
437 437 for (zap_cursor_init(&zc, dl->dl_os, obj);
438 438 zap_cursor_retrieve(&zc, &za) == 0;
439 439 zap_cursor_advance(&zc)) {
440 440 uint64_t mintxg = strtonum(za.za_name, NULL);
441 441 dsl_deadlist_insert_bpobj(dl, za.za_first_integer, mintxg, tx);
442 - VERIFY3U(0, ==, zap_remove_int(dl->dl_os, obj, mintxg, tx));
442 + VERIFY0(zap_remove_int(dl->dl_os, obj, mintxg, tx));
443 443 }
444 444 zap_cursor_fini(&zc);
445 445
446 - VERIFY3U(0, ==, dmu_bonus_hold(dl->dl_os, obj, FTAG, &bonus));
446 + VERIFY0(dmu_bonus_hold(dl->dl_os, obj, FTAG, &bonus));
447 447 dlp = bonus->db_data;
448 448 dmu_buf_will_dirty(bonus, tx);
449 449 bzero(dlp, sizeof (*dlp));
450 450 dmu_buf_rele(bonus, FTAG);
451 451 }
452 452
453 453 /*
454 454 * Remove entries on dl that are >= mintxg, and put them on the bpobj.
455 455 */
456 456 void
457 457 dsl_deadlist_move_bpobj(dsl_deadlist_t *dl, bpobj_t *bpo, uint64_t mintxg,
458 458 dmu_tx_t *tx)
459 459 {
460 460 dsl_deadlist_entry_t dle_tofind;
461 461 dsl_deadlist_entry_t *dle;
462 462 avl_index_t where;
463 463
464 464 ASSERT(!dl->dl_oldfmt);
465 465 dmu_buf_will_dirty(dl->dl_dbuf, tx);
466 466 dsl_deadlist_load_tree(dl);
467 467
↓ open down ↓ |
11 lines elided |
↑ open up ↑ |
468 468 dle_tofind.dle_mintxg = mintxg;
469 469 dle = avl_find(&dl->dl_tree, &dle_tofind, &where);
470 470 if (dle == NULL)
471 471 dle = avl_nearest(&dl->dl_tree, where, AVL_AFTER);
472 472 while (dle) {
473 473 uint64_t used, comp, uncomp;
474 474 dsl_deadlist_entry_t *dle_next;
475 475
476 476 bpobj_enqueue_subobj(bpo, dle->dle_bpobj.bpo_object, tx);
477 477
478 - VERIFY3U(0, ==, bpobj_space(&dle->dle_bpobj,
478 + VERIFY0(bpobj_space(&dle->dle_bpobj,
479 479 &used, &comp, &uncomp));
480 480 mutex_enter(&dl->dl_lock);
481 481 ASSERT3U(dl->dl_phys->dl_used, >=, used);
482 482 ASSERT3U(dl->dl_phys->dl_comp, >=, comp);
483 483 ASSERT3U(dl->dl_phys->dl_uncomp, >=, uncomp);
484 484 dl->dl_phys->dl_used -= used;
485 485 dl->dl_phys->dl_comp -= comp;
486 486 dl->dl_phys->dl_uncomp -= uncomp;
487 487 mutex_exit(&dl->dl_lock);
488 488
489 - VERIFY3U(0, ==, zap_remove_int(dl->dl_os, dl->dl_object,
489 + VERIFY0(zap_remove_int(dl->dl_os, dl->dl_object,
490 490 dle->dle_mintxg, tx));
491 491
492 492 dle_next = AVL_NEXT(&dl->dl_tree, dle);
493 493 avl_remove(&dl->dl_tree, dle);
494 494 bpobj_close(&dle->dle_bpobj);
495 495 kmem_free(dle, sizeof (*dle));
496 496 dle = dle_next;
497 497 }
498 498 }
XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX