Prepared by: | Steven Hartland (smh) on 2013-Apr-25 12:27 +0000 UTC |
---|---|
Workspace: | /home/smh/ws/illumos-gate (at 0b14a472c0a0) |
Compare against: | d8fa96c (at d8fa96c30212) |
Summary of changes: | 276 lines changed: 116 ins; 126 del; 34 mod; 16088 unchg |
Patch of changes: | illumos-3740.patch |
Printable review: | illumos-3740.pdf |
Cdiffs Udiffs Wdiffs Sdiffs Frames Old New Patch Raw usr/src/lib/libzfs/common/libzfs.h
Optimize creation and removal of temporary "user holds" placed on snapshots by a zfs send, by ensuring all the required holds and releases are done in a single dsl_sync_task. Creation now collates the required holds during a dry run and then uses a single lzc_hold call via zfs_hold_apply instead of processing each snapshot in turn. Defered (on exit) cleanup by the kernel is also now done in dsl_sync_task by reusing dsl_dataset_user_release. On a test with 11 volumes in a tree each with 8 snapshots on a single HDD zpool this reduces the time required to perform a full send from 20 seconds to under 0.8 seconds. For reference eliminating the hold entirely reduces this 0.15 seconds. While I'm here:- * Remove some unused structures * Fix nvlist_t leak in zfs_release_one3 lines changed: 3 ins; 0 del; 0 mod; 757 unchg
Cdiffs Udiffs Wdiffs Sdiffs Frames Old New Patch Raw usr/src/lib/libzfs/common/libzfs_dataset.c
Optimize creation and removal of temporary "user holds" placed on snapshots by a zfs send, by ensuring all the required holds and releases are done in a single dsl_sync_task. Creation now collates the required holds during a dry run and then uses a single lzc_hold call via zfs_hold_apply instead of processing each snapshot in turn. Defered (on exit) cleanup by the kernel is also now done in dsl_sync_task by reusing dsl_dataset_user_release. On a test with 11 volumes in a tree each with 8 snapshots on a single HDD zpool this reduces the time required to perform a full send from 20 seconds to under 0.8 seconds. For reference eliminating the hold entirely reduces this 0.15 seconds. While I'm here:- * Remove some unused structures * Fix nvlist_t leak in zfs_release_one58 lines changed: 46 ins; 11 del; 1 mod; 4455 unchg
Cdiffs Udiffs Wdiffs Sdiffs Frames Old New Patch Raw usr/src/lib/libzfs/common/libzfs_sendrecv.c
Optimize creation and removal of temporary "user holds" placed on snapshots by a zfs send, by ensuring all the required holds and releases are done in a single dsl_sync_task. Creation now collates the required holds during a dry run and then uses a single lzc_hold call via zfs_hold_apply instead of processing each snapshot in turn. Defered (on exit) cleanup by the kernel is also now done in dsl_sync_task by reusing dsl_dataset_user_release. On a test with 11 volumes in a tree each with 8 snapshots on a single HDD zpool this reduces the time required to perform a full send from 20 seconds to under 0.8 seconds. For reference eliminating the hold entirely reduces this 0.15 seconds. While I'm here:- * Remove some unused structures * Fix nvlist_t leak in zfs_release_one36 lines changed: 28 ins; 3 del; 5 mod; 3187 unchg
Cdiffs Udiffs Wdiffs Sdiffs Frames Old New Patch Raw usr/src/uts/common/fs/zfs/dsl_pool.c
Optimize creation and removal of temporary "user holds" placed on snapshots by a zfs send, by ensuring all the required holds and releases are done in a single dsl_sync_task. Creation now collates the required holds during a dry run and then uses a single lzc_hold call via zfs_hold_apply instead of processing each snapshot in turn. Defered (on exit) cleanup by the kernel is also now done in dsl_sync_task by reusing dsl_dataset_user_release. On a test with 11 volumes in a tree each with 8 snapshots on a single HDD zpool this reduces the time required to perform a full send from 20 seconds to under 0.8 seconds. For reference eliminating the hold entirely reduces this 0.15 seconds. While I'm here:- * Remove some unused structures * Fix nvlist_t leak in zfs_release_one26 lines changed: 22 ins; 3 del; 1 mod; 1027 unchg
Cdiffs Udiffs Wdiffs Sdiffs Frames Old New Patch Raw usr/src/uts/common/fs/zfs/dsl_userhold.c
Optimize creation and removal of temporary "user holds" placed on snapshots by a zfs send, by ensuring all the required holds and releases are done in a single dsl_sync_task. Creation now collates the required holds during a dry run and then uses a single lzc_hold call via zfs_hold_apply instead of processing each snapshot in turn. Defered (on exit) cleanup by the kernel is also now done in dsl_sync_task by reusing dsl_dataset_user_release. On a test with 11 volumes in a tree each with 8 snapshots on a single HDD zpool this reduces the time required to perform a full send from 20 seconds to under 0.8 seconds. For reference eliminating the hold entirely reduces this 0.15 seconds. While I'm here:- * Remove some unused structures * Fix nvlist_t leak in zfs_release_one139 lines changed: 17 ins; 96 del; 26 mod; 414 unchg
Cdiffs Udiffs Wdiffs Sdiffs Frames Old New Patch Raw usr/src/uts/common/fs/zfs/sys/dsl_dataset.h
Optimize creation and removal of temporary "user holds" placed on snapshots by a zfs send, by ensuring all the required holds and releases are done in a single dsl_sync_task. Creation now collates the required holds during a dry run and then uses a single lzc_hold call via zfs_hold_apply instead of processing each snapshot in turn. Defered (on exit) cleanup by the kernel is also now done in dsl_sync_task by reusing dsl_dataset_user_release. On a test with 11 volumes in a tree each with 8 snapshots on a single HDD zpool this reduces the time required to perform a full send from 20 seconds to under 0.8 seconds. For reference eliminating the hold entirely reduces this 0.15 seconds. While I'm here:- * Remove some unused structures * Fix nvlist_t leak in zfs_release_one2 lines changed: 0 ins; 1 del; 1 mod; 285 unchg
Cdiffs Udiffs Wdiffs Sdiffs Frames Old New Patch Raw usr/src/uts/common/fs/zfs/sys/dsl_userhold.h
Optimize creation and removal of temporary "user holds" placed on snapshots by a zfs send, by ensuring all the required holds and releases are done in a single dsl_sync_task. Creation now collates the required holds during a dry run and then uses a single lzc_hold call via zfs_hold_apply instead of processing each snapshot in turn. Defered (on exit) cleanup by the kernel is also now done in dsl_sync_task by reusing dsl_dataset_user_release. On a test with 11 volumes in a tree each with 8 snapshots on a single HDD zpool this reduces the time required to perform a full send from 20 seconds to under 0.8 seconds. For reference eliminating the hold entirely reduces this 0.15 seconds. While I'm here:- * Remove some unused structures * Fix nvlist_t leak in zfs_release_one2 lines changed: 0 ins; 2 del; 0 mod; 55 unchg
Cdiffs Udiffs Wdiffs Sdiffs Frames Old New Patch Raw usr/src/uts/common/fs/zfs/zfs_ioctl.c
Optimize creation and removal of temporary "user holds" placed on snapshots by a zfs send, by ensuring all the required holds and releases are done in a single dsl_sync_task. Creation now collates the required holds during a dry run and then uses a single lzc_hold call via zfs_hold_apply instead of processing each snapshot in turn. Defered (on exit) cleanup by the kernel is also now done in dsl_sync_task by reusing dsl_dataset_user_release. On a test with 11 volumes in a tree each with 8 snapshots on a single HDD zpool this reduces the time required to perform a full send from 20 seconds to under 0.8 seconds. For reference eliminating the hold entirely reduces this 0.15 seconds. While I'm here:- * Remove some unused structures * Fix nvlist_t leak in zfs_release_one10 lines changed: 0 ins; 10 del; 0 mod; 5908 unchg
This code review page was prepared using /opt/onbld/bin/webrev. Webrev is maintained by the illumos project. The latest version may be obtained here.