1 /*
   2  * CDDL HEADER START
   3  *
   4  * The contents of this file are subject to the terms of the
   5  * Common Development and Distribution License (the "License").
   6  * You may not use this file except in compliance with the License.
   7  *
   8  * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
   9  * or http://www.opensolaris.org/os/licensing.
  10  * See the License for the specific language governing permissions
  11  * and limitations under the License.
  12  *
  13  * When distributing Covered Code, include this CDDL HEADER in each
  14  * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
  15  * If applicable, add the following below this CDDL HEADER, with the
  16  * fields enclosed by brackets "[]" replaced with your own identifying
  17  * information: Portions Copyright [yyyy] [name of copyright owner]
  18  *
  19  * CDDL HEADER END
  20  */
  21 /*
  22  * Copyright 2009 Sun Microsystems, Inc.  All rights reserved.
  23  * Use is subject to license terms.
  24  */
  25 
  26 /*
  27  * Copyright (c) 2013 by Delphix. All rights reserved.
  28  */
  29 
  30 #ifndef _SYS_TXG_IMPL_H
  31 #define _SYS_TXG_IMPL_H
  32 
  33 #include <sys/spa.h>
  34 #include <sys/txg.h>
  35 
  36 #ifdef  __cplusplus
  37 extern "C" {
  38 #endif
  39 
  40 /*
  41  * The tx_cpu structure is a per-cpu structure that is used to track
  42  * the number of active transaction holds (tc_count). As transactions
  43  * are assigned into a transaction group the appropriate tc_count is
  44  * incremented to indicate that there are pending changes that have yet
  45  * to quiesce. Consumers evenutally call txg_rele_to_sync() to decrement
  46  * the tc_count. A transaction group is not considered quiesced until all
  47  * tx_cpu structures have reached a tc_count of zero.
  48  *
  49  * This structure is a per-cpu structure by design. Updates to this structure
  50  * are frequent and concurrent. Having a single structure would result in
  51  * heavy lock contention so a per-cpu design was implemented. With the fanned
  52  * out mutex design, consumers only need to lock the mutex associated with
  53  * thread's cpu.
  54  *
  55  * The tx_cpu contains two locks, the tc_lock and tc_open_lock.
  56  * The tc_lock is used to protect all members of the tx_cpu structure with
  57  * the exception of the tc_open_lock. This lock should only be held for a
  58  * short period of time, typically when updating the value of tc_count.
  59  *
  60  * The tc_open_lock protects the tx_open_txg member of the tx_state structure.
  61  * This lock is used to ensure that transactions are only assigned into
  62  * the current open transaction group. In order to move the current open
  63  * transaction group to the quiesce phase, the txg_quiesce thread must
  64  * grab all tc_open_locks, increment the tx_open_txg, and drop the locks.
  65  * The tc_open_lock is held until the transaction is assigned into the
  66  * transaction group. Typically, this is a short operation but if throttling
  67  * is occuring it may be held for longer periods of time.
  68  */
  69 struct tx_cpu {
  70         kmutex_t        tc_open_lock;   /* protects tx_open_txg */
  71         kmutex_t        tc_lock;        /* protects the rest of this struct */
  72         kcondvar_t      tc_cv[TXG_SIZE];
  73         uint64_t        tc_count[TXG_SIZE];     /* tx hold count on each txg */
  74         list_t          tc_callbacks[TXG_SIZE]; /* commit cb list */
  75         char            tc_pad[8];              /* pad to fill 3 cache lines */
  76 };
  77 
  78 /*
  79  * The tx_state structure maintains the state information about the different
  80  * stages of the pool's transcation groups. A per pool tx_state structure
  81  * is used to track this information. The tx_state structure also points to
  82  * an array of tx_cpu structures (described above). Although the tx_sync_lock
  83  * is used to protect the members of this structure, it is not used to
  84  * protect the tx_open_txg. Instead a special lock in the tx_cpu structure
  85  * is used. Readers of tx_open_txg must grab the per-cpu tc_open_lock.
  86  * Any thread wishing to update tx_open_txg must grab the tc_open_lock on
  87  * every cpu (see txg_quiesce()).
  88  */
  89 typedef struct tx_state {
  90         tx_cpu_t        *tx_cpu;        /* protects access to tx_open_txg */
  91         kmutex_t        tx_sync_lock;   /* protects the rest of this struct */
  92         uint64_t        tx_open_txg;    /* currently open txg id */
  93         uint64_t        tx_quiesced_txg; /* quiesced txg waiting for sync */
  94         uint64_t        tx_syncing_txg; /* currently syncing txg id */
  95         uint64_t        tx_synced_txg;  /* last synced txg id */
  96 
  97         uint64_t        tx_sync_txg_waiting; /* txg we're waiting to sync */
  98         uint64_t        tx_quiesce_txg_waiting; /* txg we're waiting to open */
  99 
 100         kcondvar_t      tx_sync_more_cv;
 101         kcondvar_t      tx_sync_done_cv;
 102         kcondvar_t      tx_quiesce_more_cv;
 103         kcondvar_t      tx_quiesce_done_cv;
 104         kcondvar_t      tx_timeout_cv;
 105         kcondvar_t      tx_exit_cv;     /* wait for all threads to exit */
 106 
 107         uint8_t         tx_threads;     /* number of threads */
 108         uint8_t         tx_exiting;     /* set when we're exiting */
 109 
 110         kthread_t       *tx_sync_thread;
 111         kthread_t       *tx_quiesce_thread;
 112 
 113         taskq_t         *tx_commit_cb_taskq; /* commit callback taskq */
 114 } tx_state_t;
 115 
 116 #ifdef  __cplusplus
 117 }
 118 #endif
 119 
 120 #endif  /* _SYS_TXG_IMPL_H */