Arne Jansen | da5c813 | 2011-09-13 12:29:12 +0200 | [diff] [blame] | 1 | /* |
| 2 | * Copyright (C) 2011 STRATO AG |
| 3 | * written by Arne Jansen <sensille@gmx.net> |
| 4 | * Distributed under the GNU GPL license version 2. |
| 5 | */ |
| 6 | |
| 7 | #include <linux/slab.h> |
Arne Jansen | da5c813 | 2011-09-13 12:29:12 +0200 | [diff] [blame] | 8 | #include "ulist.h" |
Wang Shilong | 4c7a6f7 | 2014-01-29 00:25:34 +0800 | [diff] [blame] | 9 | #include "ctree.h" |
Arne Jansen | da5c813 | 2011-09-13 12:29:12 +0200 | [diff] [blame] | 10 | |
| 11 | /* |
| 12 | * ulist is a generic data structure to hold a collection of unique u64 |
| 13 | * values. The only operations it supports is adding to the list and |
| 14 | * enumerating it. |
| 15 | * It is possible to store an auxiliary value along with the key. |
| 16 | * |
Arne Jansen | da5c813 | 2011-09-13 12:29:12 +0200 | [diff] [blame] | 17 | * A sample usage for ulists is the enumeration of directed graphs without |
| 18 | * visiting a node twice. The pseudo-code could look like this: |
| 19 | * |
| 20 | * ulist = ulist_alloc(); |
| 21 | * ulist_add(ulist, root); |
Jan Schmidt | cd1b413 | 2012-05-22 14:56:50 +0200 | [diff] [blame] | 22 | * ULIST_ITER_INIT(&uiter); |
Arne Jansen | da5c813 | 2011-09-13 12:29:12 +0200 | [diff] [blame] | 23 | * |
Jan Schmidt | cd1b413 | 2012-05-22 14:56:50 +0200 | [diff] [blame] | 24 | * while ((elem = ulist_next(ulist, &uiter)) { |
Arne Jansen | da5c813 | 2011-09-13 12:29:12 +0200 | [diff] [blame] | 25 | * for (all child nodes n in elem) |
| 26 | * ulist_add(ulist, n); |
| 27 | * do something useful with the node; |
| 28 | * } |
| 29 | * ulist_free(ulist); |
| 30 | * |
Nicholas D Steeves | 0132761 | 2016-05-19 21:18:45 -0400 | [diff] [blame] | 31 | * This assumes the graph nodes are addressable by u64. This stems from the |
Arne Jansen | da5c813 | 2011-09-13 12:29:12 +0200 | [diff] [blame] | 32 | * usage for tree enumeration in btrfs, where the logical addresses are |
| 33 | * 64 bit. |
| 34 | * |
| 35 | * It is also useful for tree enumeration which could be done elegantly |
| 36 | * recursively, but is not possible due to kernel stack limitations. The |
| 37 | * loop would be similar to the above. |
| 38 | */ |
| 39 | |
| 40 | /** |
| 41 | * ulist_init - freshly initialize a ulist |
| 42 | * @ulist: the ulist to initialize |
| 43 | * |
| 44 | * Note: don't use this function to init an already used ulist, use |
| 45 | * ulist_reinit instead. |
| 46 | */ |
| 47 | void ulist_init(struct ulist *ulist) |
| 48 | { |
Wang Shilong | 4c7a6f7 | 2014-01-29 00:25:34 +0800 | [diff] [blame] | 49 | INIT_LIST_HEAD(&ulist->nodes); |
Wang Shilong | f7f82b8 | 2013-04-12 12:12:17 +0000 | [diff] [blame] | 50 | ulist->root = RB_ROOT; |
Wang Shilong | 4c7a6f7 | 2014-01-29 00:25:34 +0800 | [diff] [blame] | 51 | ulist->nnodes = 0; |
Arne Jansen | da5c813 | 2011-09-13 12:29:12 +0200 | [diff] [blame] | 52 | } |
Arne Jansen | da5c813 | 2011-09-13 12:29:12 +0200 | [diff] [blame] | 53 | |
| 54 | /** |
| 55 | * ulist_fini - free up additionally allocated memory for the ulist |
| 56 | * @ulist: the ulist from which to free the additional memory |
| 57 | * |
| 58 | * This is useful in cases where the base 'struct ulist' has been statically |
| 59 | * allocated. |
| 60 | */ |
David Sterba | 9d037933 | 2017-02-13 13:40:16 +0100 | [diff] [blame] | 61 | void ulist_fini(struct ulist *ulist) |
Arne Jansen | da5c813 | 2011-09-13 12:29:12 +0200 | [diff] [blame] | 62 | { |
Wang Shilong | 4c7a6f7 | 2014-01-29 00:25:34 +0800 | [diff] [blame] | 63 | struct ulist_node *node; |
| 64 | struct ulist_node *next; |
| 65 | |
| 66 | list_for_each_entry_safe(node, next, &ulist->nodes, list) { |
| 67 | kfree(node); |
| 68 | } |
Wang Shilong | f7f82b8 | 2013-04-12 12:12:17 +0000 | [diff] [blame] | 69 | ulist->root = RB_ROOT; |
Wang Shilong | 4c7a6f7 | 2014-01-29 00:25:34 +0800 | [diff] [blame] | 70 | INIT_LIST_HEAD(&ulist->nodes); |
Arne Jansen | da5c813 | 2011-09-13 12:29:12 +0200 | [diff] [blame] | 71 | } |
Arne Jansen | da5c813 | 2011-09-13 12:29:12 +0200 | [diff] [blame] | 72 | |
| 73 | /** |
| 74 | * ulist_reinit - prepare a ulist for reuse |
| 75 | * @ulist: ulist to be reused |
| 76 | * |
| 77 | * Free up all additional memory allocated for the list elements and reinit |
| 78 | * the ulist. |
| 79 | */ |
| 80 | void ulist_reinit(struct ulist *ulist) |
| 81 | { |
| 82 | ulist_fini(ulist); |
| 83 | ulist_init(ulist); |
| 84 | } |
Arne Jansen | da5c813 | 2011-09-13 12:29:12 +0200 | [diff] [blame] | 85 | |
| 86 | /** |
| 87 | * ulist_alloc - dynamically allocate a ulist |
| 88 | * @gfp_mask: allocation flags to for base allocation |
| 89 | * |
| 90 | * The allocated ulist will be returned in an initialized state. |
| 91 | */ |
Daniel J Blueman | 2eec6c8 | 2012-04-26 00:37:14 +0800 | [diff] [blame] | 92 | struct ulist *ulist_alloc(gfp_t gfp_mask) |
Arne Jansen | da5c813 | 2011-09-13 12:29:12 +0200 | [diff] [blame] | 93 | { |
| 94 | struct ulist *ulist = kmalloc(sizeof(*ulist), gfp_mask); |
| 95 | |
| 96 | if (!ulist) |
| 97 | return NULL; |
| 98 | |
| 99 | ulist_init(ulist); |
| 100 | |
| 101 | return ulist; |
| 102 | } |
Arne Jansen | da5c813 | 2011-09-13 12:29:12 +0200 | [diff] [blame] | 103 | |
| 104 | /** |
| 105 | * ulist_free - free dynamically allocated ulist |
| 106 | * @ulist: ulist to free |
| 107 | * |
| 108 | * It is not necessary to call ulist_fini before. |
| 109 | */ |
| 110 | void ulist_free(struct ulist *ulist) |
| 111 | { |
| 112 | if (!ulist) |
| 113 | return; |
| 114 | ulist_fini(ulist); |
| 115 | kfree(ulist); |
| 116 | } |
Arne Jansen | da5c813 | 2011-09-13 12:29:12 +0200 | [diff] [blame] | 117 | |
Wang Shilong | f7f82b8 | 2013-04-12 12:12:17 +0000 | [diff] [blame] | 118 | static struct ulist_node *ulist_rbtree_search(struct ulist *ulist, u64 val) |
| 119 | { |
| 120 | struct rb_node *n = ulist->root.rb_node; |
| 121 | struct ulist_node *u = NULL; |
| 122 | |
| 123 | while (n) { |
| 124 | u = rb_entry(n, struct ulist_node, rb_node); |
| 125 | if (u->val < val) |
| 126 | n = n->rb_right; |
| 127 | else if (u->val > val) |
| 128 | n = n->rb_left; |
| 129 | else |
| 130 | return u; |
| 131 | } |
| 132 | return NULL; |
| 133 | } |
| 134 | |
Qu Wenruo | d4b8040 | 2015-04-20 09:26:02 +0800 | [diff] [blame] | 135 | static void ulist_rbtree_erase(struct ulist *ulist, struct ulist_node *node) |
| 136 | { |
| 137 | rb_erase(&node->rb_node, &ulist->root); |
| 138 | list_del(&node->list); |
| 139 | kfree(node); |
| 140 | BUG_ON(ulist->nnodes == 0); |
| 141 | ulist->nnodes--; |
| 142 | } |
| 143 | |
Wang Shilong | f7f82b8 | 2013-04-12 12:12:17 +0000 | [diff] [blame] | 144 | static int ulist_rbtree_insert(struct ulist *ulist, struct ulist_node *ins) |
| 145 | { |
| 146 | struct rb_node **p = &ulist->root.rb_node; |
| 147 | struct rb_node *parent = NULL; |
| 148 | struct ulist_node *cur = NULL; |
| 149 | |
| 150 | while (*p) { |
| 151 | parent = *p; |
| 152 | cur = rb_entry(parent, struct ulist_node, rb_node); |
| 153 | |
| 154 | if (cur->val < ins->val) |
| 155 | p = &(*p)->rb_right; |
| 156 | else if (cur->val > ins->val) |
| 157 | p = &(*p)->rb_left; |
| 158 | else |
| 159 | return -EEXIST; |
| 160 | } |
| 161 | rb_link_node(&ins->rb_node, parent, p); |
| 162 | rb_insert_color(&ins->rb_node, &ulist->root); |
| 163 | return 0; |
| 164 | } |
| 165 | |
Arne Jansen | da5c813 | 2011-09-13 12:29:12 +0200 | [diff] [blame] | 166 | /** |
| 167 | * ulist_add - add an element to the ulist |
| 168 | * @ulist: ulist to add the element to |
| 169 | * @val: value to add to ulist |
| 170 | * @aux: auxiliary value to store along with val |
| 171 | * @gfp_mask: flags to use for allocation |
| 172 | * |
| 173 | * Note: locking must be provided by the caller. In case of rwlocks write |
| 174 | * locking is needed |
| 175 | * |
| 176 | * Add an element to a ulist. The @val will only be added if it doesn't |
| 177 | * already exist. If it is added, the auxiliary value @aux is stored along with |
| 178 | * it. In case @val already exists in the ulist, @aux is ignored, even if |
| 179 | * it differs from the already stored value. |
| 180 | * |
| 181 | * ulist_add returns 0 if @val already exists in ulist and 1 if @val has been |
| 182 | * inserted. |
| 183 | * In case of allocation failure -ENOMEM is returned and the ulist stays |
| 184 | * unaltered. |
| 185 | */ |
Alexander Block | 34d73f5 | 2012-07-28 16:18:58 +0200 | [diff] [blame] | 186 | int ulist_add(struct ulist *ulist, u64 val, u64 aux, gfp_t gfp_mask) |
Arne Jansen | da5c813 | 2011-09-13 12:29:12 +0200 | [diff] [blame] | 187 | { |
Jan Schmidt | 3301958 | 2012-05-30 18:05:21 +0200 | [diff] [blame] | 188 | return ulist_add_merge(ulist, val, aux, NULL, gfp_mask); |
| 189 | } |
| 190 | |
Alexander Block | 34d73f5 | 2012-07-28 16:18:58 +0200 | [diff] [blame] | 191 | int ulist_add_merge(struct ulist *ulist, u64 val, u64 aux, |
| 192 | u64 *old_aux, gfp_t gfp_mask) |
Jan Schmidt | 3301958 | 2012-05-30 18:05:21 +0200 | [diff] [blame] | 193 | { |
Wang Shilong | 4c7a6f7 | 2014-01-29 00:25:34 +0800 | [diff] [blame] | 194 | int ret; |
| 195 | struct ulist_node *node; |
| 196 | |
Wang Shilong | f7f82b8 | 2013-04-12 12:12:17 +0000 | [diff] [blame] | 197 | node = ulist_rbtree_search(ulist, val); |
| 198 | if (node) { |
| 199 | if (old_aux) |
| 200 | *old_aux = node->aux; |
| 201 | return 0; |
Arne Jansen | da5c813 | 2011-09-13 12:29:12 +0200 | [diff] [blame] | 202 | } |
Wang Shilong | 4c7a6f7 | 2014-01-29 00:25:34 +0800 | [diff] [blame] | 203 | node = kmalloc(sizeof(*node), gfp_mask); |
| 204 | if (!node) |
| 205 | return -ENOMEM; |
Arne Jansen | da5c813 | 2011-09-13 12:29:12 +0200 | [diff] [blame] | 206 | |
Wang Shilong | 4c7a6f7 | 2014-01-29 00:25:34 +0800 | [diff] [blame] | 207 | node->val = val; |
| 208 | node->aux = aux; |
Liu Bo | 35f0399 | 2013-06-28 12:37:45 +0800 | [diff] [blame] | 209 | |
Wang Shilong | 4c7a6f7 | 2014-01-29 00:25:34 +0800 | [diff] [blame] | 210 | ret = ulist_rbtree_insert(ulist, node); |
| 211 | ASSERT(!ret); |
| 212 | list_add_tail(&node->list, &ulist->nodes); |
| 213 | ulist->nnodes++; |
Arne Jansen | da5c813 | 2011-09-13 12:29:12 +0200 | [diff] [blame] | 214 | |
| 215 | return 1; |
| 216 | } |
Arne Jansen | da5c813 | 2011-09-13 12:29:12 +0200 | [diff] [blame] | 217 | |
Qu Wenruo | d4b8040 | 2015-04-20 09:26:02 +0800 | [diff] [blame] | 218 | /* |
| 219 | * ulist_del - delete one node from ulist |
| 220 | * @ulist: ulist to remove node from |
| 221 | * @val: value to delete |
| 222 | * @aux: aux to delete |
| 223 | * |
| 224 | * The deletion will only be done when *BOTH* val and aux matches. |
| 225 | * Return 0 for successful delete. |
| 226 | * Return > 0 for not found. |
| 227 | */ |
| 228 | int ulist_del(struct ulist *ulist, u64 val, u64 aux) |
| 229 | { |
| 230 | struct ulist_node *node; |
| 231 | |
| 232 | node = ulist_rbtree_search(ulist, val); |
| 233 | /* Not found */ |
| 234 | if (!node) |
| 235 | return 1; |
| 236 | |
| 237 | if (node->aux != aux) |
| 238 | return 1; |
| 239 | |
| 240 | /* Found and delete */ |
| 241 | ulist_rbtree_erase(ulist, node); |
| 242 | return 0; |
| 243 | } |
| 244 | |
Arne Jansen | da5c813 | 2011-09-13 12:29:12 +0200 | [diff] [blame] | 245 | /** |
| 246 | * ulist_next - iterate ulist |
| 247 | * @ulist: ulist to iterate |
Jan Schmidt | cd1b413 | 2012-05-22 14:56:50 +0200 | [diff] [blame] | 248 | * @uiter: iterator variable, initialized with ULIST_ITER_INIT(&iterator) |
Arne Jansen | da5c813 | 2011-09-13 12:29:12 +0200 | [diff] [blame] | 249 | * |
| 250 | * Note: locking must be provided by the caller. In case of rwlocks only read |
| 251 | * locking is needed |
| 252 | * |
Jan Schmidt | cd1b413 | 2012-05-22 14:56:50 +0200 | [diff] [blame] | 253 | * This function is used to iterate an ulist. |
| 254 | * It returns the next element from the ulist or %NULL when the |
Arne Jansen | da5c813 | 2011-09-13 12:29:12 +0200 | [diff] [blame] | 255 | * end is reached. No guarantee is made with respect to the order in which |
| 256 | * the elements are returned. They might neither be returned in order of |
| 257 | * addition nor in ascending order. |
| 258 | * It is allowed to call ulist_add during an enumeration. Newly added items |
| 259 | * are guaranteed to show up in the running enumeration. |
| 260 | */ |
Jan Schmidt | cd1b413 | 2012-05-22 14:56:50 +0200 | [diff] [blame] | 261 | struct ulist_node *ulist_next(struct ulist *ulist, struct ulist_iterator *uiter) |
Arne Jansen | da5c813 | 2011-09-13 12:29:12 +0200 | [diff] [blame] | 262 | { |
Wang Shilong | 4c7a6f7 | 2014-01-29 00:25:34 +0800 | [diff] [blame] | 263 | struct ulist_node *node; |
Arne Jansen | da5c813 | 2011-09-13 12:29:12 +0200 | [diff] [blame] | 264 | |
Wang Shilong | 4c7a6f7 | 2014-01-29 00:25:34 +0800 | [diff] [blame] | 265 | if (list_empty(&ulist->nodes)) |
| 266 | return NULL; |
| 267 | if (uiter->cur_list && uiter->cur_list->next == &ulist->nodes) |
| 268 | return NULL; |
| 269 | if (uiter->cur_list) { |
| 270 | uiter->cur_list = uiter->cur_list->next; |
| 271 | } else { |
| 272 | uiter->cur_list = ulist->nodes.next; |
Wang Shilong | 4c7a6f7 | 2014-01-29 00:25:34 +0800 | [diff] [blame] | 273 | } |
| 274 | node = list_entry(uiter->cur_list, struct ulist_node, list); |
Wang Shilong | 4c7a6f7 | 2014-01-29 00:25:34 +0800 | [diff] [blame] | 275 | return node; |
Arne Jansen | da5c813 | 2011-09-13 12:29:12 +0200 | [diff] [blame] | 276 | } |