blob: 8dd0e8dfdaf407290e12d7ac3f4393f24e99db34 [file] [log] [blame]
Arne Jansenda5c8132011-09-13 12:29:12 +02001/*
2 * Copyright (C) 2011 STRATO AG
3 * written by Arne Jansen <sensille@gmx.net>
4 * Distributed under the GNU GPL license version 2.
5 */
6
7#include <linux/slab.h>
Paul Gortmaker180e0012013-02-14 13:50:15 -07008#include <linux/export.h>
Arne Jansenda5c8132011-09-13 12:29:12 +02009#include "ulist.h"
Wang Shilong4c7a6f72014-01-29 00:25:34 +080010#include "ctree.h"
Arne Jansenda5c8132011-09-13 12:29:12 +020011
12/*
13 * ulist is a generic data structure to hold a collection of unique u64
14 * values. The only operations it supports is adding to the list and
15 * enumerating it.
16 * It is possible to store an auxiliary value along with the key.
17 *
Arne Jansenda5c8132011-09-13 12:29:12 +020018 * A sample usage for ulists is the enumeration of directed graphs without
19 * visiting a node twice. The pseudo-code could look like this:
20 *
21 * ulist = ulist_alloc();
22 * ulist_add(ulist, root);
Jan Schmidtcd1b4132012-05-22 14:56:50 +020023 * ULIST_ITER_INIT(&uiter);
Arne Jansenda5c8132011-09-13 12:29:12 +020024 *
Jan Schmidtcd1b4132012-05-22 14:56:50 +020025 * while ((elem = ulist_next(ulist, &uiter)) {
Arne Jansenda5c8132011-09-13 12:29:12 +020026 * for (all child nodes n in elem)
27 * ulist_add(ulist, n);
28 * do something useful with the node;
29 * }
30 * ulist_free(ulist);
31 *
32 * This assumes the graph nodes are adressable by u64. This stems from the
33 * usage for tree enumeration in btrfs, where the logical addresses are
34 * 64 bit.
35 *
36 * It is also useful for tree enumeration which could be done elegantly
37 * recursively, but is not possible due to kernel stack limitations. The
38 * loop would be similar to the above.
39 */
40
41/**
42 * ulist_init - freshly initialize a ulist
43 * @ulist: the ulist to initialize
44 *
45 * Note: don't use this function to init an already used ulist, use
46 * ulist_reinit instead.
47 */
48void ulist_init(struct ulist *ulist)
49{
Wang Shilong4c7a6f72014-01-29 00:25:34 +080050 INIT_LIST_HEAD(&ulist->nodes);
Wang Shilongf7f82b82013-04-12 12:12:17 +000051 ulist->root = RB_ROOT;
Wang Shilong4c7a6f72014-01-29 00:25:34 +080052 ulist->nnodes = 0;
Arne Jansenda5c8132011-09-13 12:29:12 +020053}
54EXPORT_SYMBOL(ulist_init);
55
56/**
57 * ulist_fini - free up additionally allocated memory for the ulist
58 * @ulist: the ulist from which to free the additional memory
59 *
60 * This is useful in cases where the base 'struct ulist' has been statically
61 * allocated.
62 */
63void ulist_fini(struct ulist *ulist)
64{
Wang Shilong4c7a6f72014-01-29 00:25:34 +080065 struct ulist_node *node;
66 struct ulist_node *next;
67
68 list_for_each_entry_safe(node, next, &ulist->nodes, list) {
69 kfree(node);
70 }
Wang Shilongf7f82b82013-04-12 12:12:17 +000071 ulist->root = RB_ROOT;
Wang Shilong4c7a6f72014-01-29 00:25:34 +080072 INIT_LIST_HEAD(&ulist->nodes);
Arne Jansenda5c8132011-09-13 12:29:12 +020073}
74EXPORT_SYMBOL(ulist_fini);
75
76/**
77 * ulist_reinit - prepare a ulist for reuse
78 * @ulist: ulist to be reused
79 *
80 * Free up all additional memory allocated for the list elements and reinit
81 * the ulist.
82 */
83void ulist_reinit(struct ulist *ulist)
84{
85 ulist_fini(ulist);
86 ulist_init(ulist);
87}
88EXPORT_SYMBOL(ulist_reinit);
89
90/**
91 * ulist_alloc - dynamically allocate a ulist
92 * @gfp_mask: allocation flags to for base allocation
93 *
94 * The allocated ulist will be returned in an initialized state.
95 */
Daniel J Blueman2eec6c82012-04-26 00:37:14 +080096struct ulist *ulist_alloc(gfp_t gfp_mask)
Arne Jansenda5c8132011-09-13 12:29:12 +020097{
98 struct ulist *ulist = kmalloc(sizeof(*ulist), gfp_mask);
99
100 if (!ulist)
101 return NULL;
102
103 ulist_init(ulist);
104
105 return ulist;
106}
107EXPORT_SYMBOL(ulist_alloc);
108
109/**
110 * ulist_free - free dynamically allocated ulist
111 * @ulist: ulist to free
112 *
113 * It is not necessary to call ulist_fini before.
114 */
115void ulist_free(struct ulist *ulist)
116{
117 if (!ulist)
118 return;
119 ulist_fini(ulist);
120 kfree(ulist);
121}
122EXPORT_SYMBOL(ulist_free);
123
Wang Shilongf7f82b82013-04-12 12:12:17 +0000124static struct ulist_node *ulist_rbtree_search(struct ulist *ulist, u64 val)
125{
126 struct rb_node *n = ulist->root.rb_node;
127 struct ulist_node *u = NULL;
128
129 while (n) {
130 u = rb_entry(n, struct ulist_node, rb_node);
131 if (u->val < val)
132 n = n->rb_right;
133 else if (u->val > val)
134 n = n->rb_left;
135 else
136 return u;
137 }
138 return NULL;
139}
140
141static int ulist_rbtree_insert(struct ulist *ulist, struct ulist_node *ins)
142{
143 struct rb_node **p = &ulist->root.rb_node;
144 struct rb_node *parent = NULL;
145 struct ulist_node *cur = NULL;
146
147 while (*p) {
148 parent = *p;
149 cur = rb_entry(parent, struct ulist_node, rb_node);
150
151 if (cur->val < ins->val)
152 p = &(*p)->rb_right;
153 else if (cur->val > ins->val)
154 p = &(*p)->rb_left;
155 else
156 return -EEXIST;
157 }
158 rb_link_node(&ins->rb_node, parent, p);
159 rb_insert_color(&ins->rb_node, &ulist->root);
160 return 0;
161}
162
Arne Jansenda5c8132011-09-13 12:29:12 +0200163/**
164 * ulist_add - add an element to the ulist
165 * @ulist: ulist to add the element to
166 * @val: value to add to ulist
167 * @aux: auxiliary value to store along with val
168 * @gfp_mask: flags to use for allocation
169 *
170 * Note: locking must be provided by the caller. In case of rwlocks write
171 * locking is needed
172 *
173 * Add an element to a ulist. The @val will only be added if it doesn't
174 * already exist. If it is added, the auxiliary value @aux is stored along with
175 * it. In case @val already exists in the ulist, @aux is ignored, even if
176 * it differs from the already stored value.
177 *
178 * ulist_add returns 0 if @val already exists in ulist and 1 if @val has been
179 * inserted.
180 * In case of allocation failure -ENOMEM is returned and the ulist stays
181 * unaltered.
182 */
Alexander Block34d73f52012-07-28 16:18:58 +0200183int ulist_add(struct ulist *ulist, u64 val, u64 aux, gfp_t gfp_mask)
Arne Jansenda5c8132011-09-13 12:29:12 +0200184{
Jan Schmidt33019582012-05-30 18:05:21 +0200185 return ulist_add_merge(ulist, val, aux, NULL, gfp_mask);
186}
187
Alexander Block34d73f52012-07-28 16:18:58 +0200188int ulist_add_merge(struct ulist *ulist, u64 val, u64 aux,
189 u64 *old_aux, gfp_t gfp_mask)
Jan Schmidt33019582012-05-30 18:05:21 +0200190{
Wang Shilong4c7a6f72014-01-29 00:25:34 +0800191 int ret;
192 struct ulist_node *node;
193
Wang Shilongf7f82b82013-04-12 12:12:17 +0000194 node = ulist_rbtree_search(ulist, val);
195 if (node) {
196 if (old_aux)
197 *old_aux = node->aux;
198 return 0;
Arne Jansenda5c8132011-09-13 12:29:12 +0200199 }
Wang Shilong4c7a6f72014-01-29 00:25:34 +0800200 node = kmalloc(sizeof(*node), gfp_mask);
201 if (!node)
202 return -ENOMEM;
Arne Jansenda5c8132011-09-13 12:29:12 +0200203
Wang Shilong4c7a6f72014-01-29 00:25:34 +0800204 node->val = val;
205 node->aux = aux;
206#ifdef CONFIG_BTRFS_DEBUG
207 node->seqnum = ulist->nnodes;
208#endif
Liu Bo35f03992013-06-28 12:37:45 +0800209
Wang Shilong4c7a6f72014-01-29 00:25:34 +0800210 ret = ulist_rbtree_insert(ulist, node);
211 ASSERT(!ret);
212 list_add_tail(&node->list, &ulist->nodes);
213 ulist->nnodes++;
Arne Jansenda5c8132011-09-13 12:29:12 +0200214
215 return 1;
216}
217EXPORT_SYMBOL(ulist_add);
218
219/**
220 * ulist_next - iterate ulist
221 * @ulist: ulist to iterate
Jan Schmidtcd1b4132012-05-22 14:56:50 +0200222 * @uiter: iterator variable, initialized with ULIST_ITER_INIT(&iterator)
Arne Jansenda5c8132011-09-13 12:29:12 +0200223 *
224 * Note: locking must be provided by the caller. In case of rwlocks only read
225 * locking is needed
226 *
Jan Schmidtcd1b4132012-05-22 14:56:50 +0200227 * This function is used to iterate an ulist.
228 * It returns the next element from the ulist or %NULL when the
Arne Jansenda5c8132011-09-13 12:29:12 +0200229 * end is reached. No guarantee is made with respect to the order in which
230 * the elements are returned. They might neither be returned in order of
231 * addition nor in ascending order.
232 * It is allowed to call ulist_add during an enumeration. Newly added items
233 * are guaranteed to show up in the running enumeration.
234 */
Jan Schmidtcd1b4132012-05-22 14:56:50 +0200235struct ulist_node *ulist_next(struct ulist *ulist, struct ulist_iterator *uiter)
Arne Jansenda5c8132011-09-13 12:29:12 +0200236{
Wang Shilong4c7a6f72014-01-29 00:25:34 +0800237 struct ulist_node *node;
Arne Jansenda5c8132011-09-13 12:29:12 +0200238
Wang Shilong4c7a6f72014-01-29 00:25:34 +0800239 if (list_empty(&ulist->nodes))
240 return NULL;
241 if (uiter->cur_list && uiter->cur_list->next == &ulist->nodes)
242 return NULL;
243 if (uiter->cur_list) {
244 uiter->cur_list = uiter->cur_list->next;
245 } else {
246 uiter->cur_list = ulist->nodes.next;
247#ifdef CONFIG_BTRFS_DEBUG
248 uiter->i = 0;
249#endif
250 }
251 node = list_entry(uiter->cur_list, struct ulist_node, list);
252#ifdef CONFIG_BTRFS_DEBUG
253 ASSERT(node->seqnum == uiter->i);
254 ASSERT(uiter->i >= 0 && uiter->i < ulist->nnodes);
255 uiter->i++;
256#endif
257 return node;
Arne Jansenda5c8132011-09-13 12:29:12 +0200258}
259EXPORT_SYMBOL(ulist_next);