blob: 1227c593627a62b5a1e27a237a63175c4038dfe0 [file] [log] [blame]
Heiko J Schickfab97222006-09-22 15:22:22 -07001/*
2 * IBM eServer eHCA Infiniband device driver for Linux on POWER
3 *
4 * internal queue handling
5 *
6 * Authors: Waleri Fomin <fomin@de.ibm.com>
7 * Reinhard Ernst <rernst@de.ibm.com>
8 * Christoph Raisch <raisch@de.ibm.com>
9 *
10 * Copyright (c) 2005 IBM Corporation
11 *
12 * This source code is distributed under a dual license of GPL v2.0 and OpenIB
13 * BSD.
14 *
15 * OpenIB BSD License
16 *
17 * Redistribution and use in source and binary forms, with or without
18 * modification, are permitted provided that the following conditions are met:
19 *
20 * Redistributions of source code must retain the above copyright notice, this
21 * list of conditions and the following disclaimer.
22 *
23 * Redistributions in binary form must reproduce the above copyright notice,
24 * this list of conditions and the following disclaimer in the documentation
25 * and/or other materials
26 * provided with the distribution.
27 *
28 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
29 * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
30 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
31 * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
32 * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
33 * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
34 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR
35 * BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER
36 * IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
37 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
38 * POSSIBILITY OF SUCH DAMAGE.
39 */
40
41#include "ehca_tools.h"
42#include "ipz_pt_fn.h"
Stefan Roschere2f81da2007-07-20 16:04:17 +020043#include "ehca_classes.h"
44
45#define PAGES_PER_KPAGE (PAGE_SIZE >> EHCA_PAGESHIFT)
46
47struct kmem_cache *small_qp_cache;
Heiko J Schickfab97222006-09-22 15:22:22 -070048
49void *ipz_qpageit_get_inc(struct ipz_queue *queue)
50{
51 void *ret = ipz_qeit_get(queue);
52 queue->current_q_offset += queue->pagesize;
53 if (queue->current_q_offset > queue->queue_length) {
54 queue->current_q_offset -= queue->pagesize;
55 ret = NULL;
56 }
Stefan Roschere2f81da2007-07-20 16:04:17 +020057 if (((u64)ret) % queue->pagesize) {
Heiko J Schickfab97222006-09-22 15:22:22 -070058 ehca_gen_err("ERROR!! not at PAGE-Boundary");
59 return NULL;
60 }
61 return ret;
62}
63
64void *ipz_qeit_eq_get_inc(struct ipz_queue *queue)
65{
66 void *ret = ipz_qeit_get(queue);
67 u64 last_entry_in_q = queue->queue_length - queue->qe_size;
68
69 queue->current_q_offset += queue->qe_size;
70 if (queue->current_q_offset > last_entry_in_q) {
71 queue->current_q_offset = 0;
72 queue->toggle_state = (~queue->toggle_state) & 1;
73 }
74
75 return ret;
76}
77
Hoang-Nam Nguyen2771e9e2006-11-20 23:54:12 +010078int ipz_queue_abs_to_offset(struct ipz_queue *queue, u64 addr, u64 *q_offset)
79{
80 int i;
81 for (i = 0; i < queue->queue_length / queue->pagesize; i++) {
82 u64 page = (u64)virt_to_abs(queue->queue_pages[i]);
83 if (addr >= page && addr < page + queue->pagesize) {
84 *q_offset = addr - page + i * queue->pagesize;
85 return 0;
86 }
87 }
88 return -EINVAL;
89}
90
Stefan Roschere2f81da2007-07-20 16:04:17 +020091#if PAGE_SHIFT < EHCA_PAGESHIFT
92#error Kernel pages must be at least as large than eHCA pages (4K) !
93#endif
Heiko J Schickfab97222006-09-22 15:22:22 -070094
Stefan Roschere2f81da2007-07-20 16:04:17 +020095/*
96 * allocate pages for queue:
97 * outer loop allocates whole kernel pages (page aligned) and
98 * inner loop divides a kernel page into smaller hca queue pages
99 */
100static int alloc_queue_pages(struct ipz_queue *queue, const u32 nr_of_pages)
101{
102 int k, f = 0;
103 u8 *kpage;
104
Heiko J Schickfab97222006-09-22 15:22:22 -0700105 while (f < nr_of_pages) {
Stefan Roschere2f81da2007-07-20 16:04:17 +0200106 kpage = (u8 *)get_zeroed_page(GFP_KERNEL);
Heiko J Schickfab97222006-09-22 15:22:22 -0700107 if (!kpage)
Stefan Roschere2f81da2007-07-20 16:04:17 +0200108 goto out;
109
110 for (k = 0; k < PAGES_PER_KPAGE && f < nr_of_pages; k++) {
111 queue->queue_pages[f] = (struct ipz_page *)kpage;
Heiko J Schickfab97222006-09-22 15:22:22 -0700112 kpage += EHCA_PAGESIZE;
113 f++;
114 }
115 }
Heiko J Schickfab97222006-09-22 15:22:22 -0700116 return 1;
117
Stefan Roschere2f81da2007-07-20 16:04:17 +0200118out:
119 for (f = 0; f < nr_of_pages && queue->queue_pages[f];
120 f += PAGES_PER_KPAGE)
Heiko J Schickfab97222006-09-22 15:22:22 -0700121 free_page((unsigned long)(queue->queue_pages)[f]);
Heiko J Schickfab97222006-09-22 15:22:22 -0700122 return 0;
123}
124
Stefan Roschere2f81da2007-07-20 16:04:17 +0200125static int alloc_small_queue_page(struct ipz_queue *queue, struct ehca_pd *pd)
Heiko J Schickfab97222006-09-22 15:22:22 -0700126{
Stefan Roschere2f81da2007-07-20 16:04:17 +0200127 int order = ilog2(queue->pagesize) - 9;
128 struct ipz_small_queue_page *page;
129 unsigned long bit;
130
131 mutex_lock(&pd->lock);
132
133 if (!list_empty(&pd->free[order]))
134 page = list_entry(pd->free[order].next,
135 struct ipz_small_queue_page, list);
136 else {
137 page = kmem_cache_zalloc(small_qp_cache, GFP_KERNEL);
138 if (!page)
139 goto out;
140
141 page->page = get_zeroed_page(GFP_KERNEL);
142 if (!page->page) {
143 kmem_cache_free(small_qp_cache, page);
144 goto out;
145 }
146
147 list_add(&page->list, &pd->free[order]);
148 }
149
150 bit = find_first_zero_bit(page->bitmap, IPZ_SPAGE_PER_KPAGE >> order);
151 __set_bit(bit, page->bitmap);
152 page->fill++;
153
154 if (page->fill == IPZ_SPAGE_PER_KPAGE >> order)
155 list_move(&page->list, &pd->full[order]);
156
157 mutex_unlock(&pd->lock);
158
159 queue->queue_pages[0] = (void *)(page->page | (bit << (order + 9)));
160 queue->small_page = page;
Stefan Roscher441633b92007-09-11 15:26:33 +0200161 queue->offset = bit << (order + 9);
Stefan Roschere2f81da2007-07-20 16:04:17 +0200162 return 1;
163
164out:
165 ehca_err(pd->ib_pd.device, "failed to allocate small queue page");
Julia Lawall1a867c32008-07-22 14:18:10 -0700166 mutex_unlock(&pd->lock);
Stefan Roschere2f81da2007-07-20 16:04:17 +0200167 return 0;
168}
169
170static void free_small_queue_page(struct ipz_queue *queue, struct ehca_pd *pd)
171{
172 int order = ilog2(queue->pagesize) - 9;
173 struct ipz_small_queue_page *page = queue->small_page;
174 unsigned long bit;
175 int free_page = 0;
176
Stefan Roscherfecea0a2007-08-31 16:02:59 +0200177 bit = ((unsigned long)queue->queue_pages[0] & ~PAGE_MASK)
Stefan Roschere2f81da2007-07-20 16:04:17 +0200178 >> (order + 9);
179
180 mutex_lock(&pd->lock);
181
182 __clear_bit(bit, page->bitmap);
183 page->fill--;
184
185 if (page->fill == 0) {
186 list_del(&page->list);
187 free_page = 1;
188 }
189
190 if (page->fill == (IPZ_SPAGE_PER_KPAGE >> order) - 1)
191 /* the page was full until we freed the chunk */
192 list_move_tail(&page->list, &pd->free[order]);
193
194 mutex_unlock(&pd->lock);
195
196 if (free_page) {
197 free_page(page->page);
198 kmem_cache_free(small_qp_cache, page);
199 }
200}
201
202int ipz_queue_ctor(struct ehca_pd *pd, struct ipz_queue *queue,
203 const u32 nr_of_pages, const u32 pagesize,
204 const u32 qe_size, const u32 nr_of_sg,
205 int is_small)
206{
207 if (pagesize > PAGE_SIZE) {
208 ehca_gen_err("FATAL ERROR: pagesize=%x "
209 "is greater than kernel page size", pagesize);
210 return 0;
211 }
212
213 /* init queue fields */
214 queue->queue_length = nr_of_pages * pagesize;
215 queue->pagesize = pagesize;
216 queue->qe_size = qe_size;
217 queue->act_nr_of_sg = nr_of_sg;
218 queue->current_q_offset = 0;
219 queue->toggle_state = 1;
220 queue->small_page = NULL;
221
222 /* allocate queue page pointers */
Anton Blanchardbf31a1a2009-05-13 16:52:40 -0700223 queue->queue_pages = kmalloc(nr_of_pages * sizeof(void *), GFP_KERNEL);
Stefan Roschere2f81da2007-07-20 16:04:17 +0200224 if (!queue->queue_pages) {
Stefan Roscherc94f1562009-05-13 16:52:42 -0700225 queue->queue_pages = vmalloc(nr_of_pages * sizeof(void *));
226 if (!queue->queue_pages) {
227 ehca_gen_err("Couldn't allocate queue page list");
228 return 0;
229 }
Stefan Roschere2f81da2007-07-20 16:04:17 +0200230 }
231 memset(queue->queue_pages, 0, nr_of_pages * sizeof(void *));
232
233 /* allocate actual queue pages */
234 if (is_small) {
235 if (!alloc_small_queue_page(queue, pd))
236 goto ipz_queue_ctor_exit0;
237 } else
238 if (!alloc_queue_pages(queue, nr_of_pages))
239 goto ipz_queue_ctor_exit0;
240
241 return 1;
242
243ipz_queue_ctor_exit0:
244 ehca_gen_err("Couldn't alloc pages queue=%p "
245 "nr_of_pages=%x", queue, nr_of_pages);
Stefan Roscherc94f1562009-05-13 16:52:42 -0700246 if (is_vmalloc_addr(queue->queue_pages))
247 vfree(queue->queue_pages);
248 else
249 kfree(queue->queue_pages);
Stefan Roschere2f81da2007-07-20 16:04:17 +0200250
251 return 0;
252}
253
254int ipz_queue_dtor(struct ehca_pd *pd, struct ipz_queue *queue)
255{
256 int i, nr_pages;
Heiko J Schickfab97222006-09-22 15:22:22 -0700257
258 if (!queue || !queue->queue_pages) {
259 ehca_gen_dbg("queue or queue_pages is NULL");
260 return 0;
261 }
Stefan Roschere2f81da2007-07-20 16:04:17 +0200262
263 if (queue->small_page)
264 free_small_queue_page(queue, pd);
265 else {
266 nr_pages = queue->queue_length / queue->pagesize;
267 for (i = 0; i < nr_pages; i += PAGES_PER_KPAGE)
268 free_page((unsigned long)queue->queue_pages[i]);
269 }
270
Stefan Roscherc94f1562009-05-13 16:52:42 -0700271 if (is_vmalloc_addr(queue->queue_pages))
272 vfree(queue->queue_pages);
273 else
274 kfree(queue->queue_pages);
Heiko J Schickfab97222006-09-22 15:22:22 -0700275
276 return 1;
277}
Stefan Roschere2f81da2007-07-20 16:04:17 +0200278
279int ehca_init_small_qp_cache(void)
280{
281 small_qp_cache = kmem_cache_create("ehca_cache_small_qp",
282 sizeof(struct ipz_small_queue_page),
283 0, SLAB_HWCACHE_ALIGN, NULL);
284 if (!small_qp_cache)
285 return -ENOMEM;
286
287 return 0;
288}
289
290void ehca_cleanup_small_qp_cache(void)
291{
292 kmem_cache_destroy(small_qp_cache);
293}