| /* |
| * IBM eServer eHCA Infiniband device driver for Linux on POWER |
| * |
| * internal queue handling |
| * |
| * Authors: Waleri Fomin <fomin@de.ibm.com> |
| * Reinhard Ernst <rernst@de.ibm.com> |
| * Christoph Raisch <raisch@de.ibm.com> |
| * |
| * Copyright (c) 2005 IBM Corporation |
| * |
| * This source code is distributed under a dual license of GPL v2.0 and OpenIB |
| * BSD. |
| * |
| * OpenIB BSD License |
| * |
| * Redistribution and use in source and binary forms, with or without |
| * modification, are permitted provided that the following conditions are met: |
| * |
| * Redistributions of source code must retain the above copyright notice, this |
| * list of conditions and the following disclaimer. |
| * |
| * Redistributions in binary form must reproduce the above copyright notice, |
| * this list of conditions and the following disclaimer in the documentation |
| * and/or other materials |
| * provided with the distribution. |
| * |
| * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" |
| * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE |
| * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE |
| * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE |
| * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR |
| * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF |
| * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR |
| * BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER |
| * IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) |
| * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE |
| * POSSIBILITY OF SUCH DAMAGE. |
| */ |
| |
| #include "ehca_tools.h" |
| #include "ipz_pt_fn.h" |
| #include "ehca_classes.h" |
| |
| #define PAGES_PER_KPAGE (PAGE_SIZE >> EHCA_PAGESHIFT) |
| |
| struct kmem_cache *small_qp_cache; |
| |
| void *ipz_qpageit_get_inc(struct ipz_queue *queue) |
| { |
| void *ret = ipz_qeit_get(queue); |
| queue->current_q_offset += queue->pagesize; |
| if (queue->current_q_offset > queue->queue_length) { |
| queue->current_q_offset -= queue->pagesize; |
| ret = NULL; |
| } |
| if (((u64)ret) % queue->pagesize) { |
| ehca_gen_err("ERROR!! not at PAGE-Boundary"); |
| return NULL; |
| } |
| return ret; |
| } |
| |
| void *ipz_qeit_eq_get_inc(struct ipz_queue *queue) |
| { |
| void *ret = ipz_qeit_get(queue); |
| u64 last_entry_in_q = queue->queue_length - queue->qe_size; |
| |
| queue->current_q_offset += queue->qe_size; |
| if (queue->current_q_offset > last_entry_in_q) { |
| queue->current_q_offset = 0; |
| queue->toggle_state = (~queue->toggle_state) & 1; |
| } |
| |
| return ret; |
| } |
| |
| int ipz_queue_abs_to_offset(struct ipz_queue *queue, u64 addr, u64 *q_offset) |
| { |
| int i; |
| for (i = 0; i < queue->queue_length / queue->pagesize; i++) { |
| u64 page = (u64)virt_to_abs(queue->queue_pages[i]); |
| if (addr >= page && addr < page + queue->pagesize) { |
| *q_offset = addr - page + i * queue->pagesize; |
| return 0; |
| } |
| } |
| return -EINVAL; |
| } |
| |
| #if PAGE_SHIFT < EHCA_PAGESHIFT |
| #error Kernel pages must be at least as large than eHCA pages (4K) ! |
| #endif |
| |
| /* |
| * allocate pages for queue: |
| * outer loop allocates whole kernel pages (page aligned) and |
| * inner loop divides a kernel page into smaller hca queue pages |
| */ |
| static int alloc_queue_pages(struct ipz_queue *queue, const u32 nr_of_pages) |
| { |
| int k, f = 0; |
| u8 *kpage; |
| |
| while (f < nr_of_pages) { |
| kpage = (u8 *)get_zeroed_page(GFP_KERNEL); |
| if (!kpage) |
| goto out; |
| |
| for (k = 0; k < PAGES_PER_KPAGE && f < nr_of_pages; k++) { |
| queue->queue_pages[f] = (struct ipz_page *)kpage; |
| kpage += EHCA_PAGESIZE; |
| f++; |
| } |
| } |
| return 1; |
| |
| out: |
| for (f = 0; f < nr_of_pages && queue->queue_pages[f]; |
| f += PAGES_PER_KPAGE) |
| free_page((unsigned long)(queue->queue_pages)[f]); |
| return 0; |
| } |
| |
| static int alloc_small_queue_page(struct ipz_queue *queue, struct ehca_pd *pd) |
| { |
| int order = ilog2(queue->pagesize) - 9; |
| struct ipz_small_queue_page *page; |
| unsigned long bit; |
| |
| mutex_lock(&pd->lock); |
| |
| if (!list_empty(&pd->free[order])) |
| page = list_entry(pd->free[order].next, |
| struct ipz_small_queue_page, list); |
| else { |
| page = kmem_cache_zalloc(small_qp_cache, GFP_KERNEL); |
| if (!page) |
| goto out; |
| |
| page->page = get_zeroed_page(GFP_KERNEL); |
| if (!page->page) { |
| kmem_cache_free(small_qp_cache, page); |
| goto out; |
| } |
| |
| list_add(&page->list, &pd->free[order]); |
| } |
| |
| bit = find_first_zero_bit(page->bitmap, IPZ_SPAGE_PER_KPAGE >> order); |
| __set_bit(bit, page->bitmap); |
| page->fill++; |
| |
| if (page->fill == IPZ_SPAGE_PER_KPAGE >> order) |
| list_move(&page->list, &pd->full[order]); |
| |
| mutex_unlock(&pd->lock); |
| |
| queue->queue_pages[0] = (void *)(page->page | (bit << (order + 9))); |
| queue->small_page = page; |
| queue->offset = bit << (order + 9); |
| return 1; |
| |
| out: |
| ehca_err(pd->ib_pd.device, "failed to allocate small queue page"); |
| mutex_unlock(&pd->lock); |
| return 0; |
| } |
| |
| static void free_small_queue_page(struct ipz_queue *queue, struct ehca_pd *pd) |
| { |
| int order = ilog2(queue->pagesize) - 9; |
| struct ipz_small_queue_page *page = queue->small_page; |
| unsigned long bit; |
| int free_page = 0; |
| |
| bit = ((unsigned long)queue->queue_pages[0] & ~PAGE_MASK) |
| >> (order + 9); |
| |
| mutex_lock(&pd->lock); |
| |
| __clear_bit(bit, page->bitmap); |
| page->fill--; |
| |
| if (page->fill == 0) { |
| list_del(&page->list); |
| free_page = 1; |
| } |
| |
| if (page->fill == (IPZ_SPAGE_PER_KPAGE >> order) - 1) |
| /* the page was full until we freed the chunk */ |
| list_move_tail(&page->list, &pd->free[order]); |
| |
| mutex_unlock(&pd->lock); |
| |
| if (free_page) { |
| free_page(page->page); |
| kmem_cache_free(small_qp_cache, page); |
| } |
| } |
| |
| int ipz_queue_ctor(struct ehca_pd *pd, struct ipz_queue *queue, |
| const u32 nr_of_pages, const u32 pagesize, |
| const u32 qe_size, const u32 nr_of_sg, |
| int is_small) |
| { |
| if (pagesize > PAGE_SIZE) { |
| ehca_gen_err("FATAL ERROR: pagesize=%x " |
| "is greater than kernel page size", pagesize); |
| return 0; |
| } |
| |
| /* init queue fields */ |
| queue->queue_length = nr_of_pages * pagesize; |
| queue->pagesize = pagesize; |
| queue->qe_size = qe_size; |
| queue->act_nr_of_sg = nr_of_sg; |
| queue->current_q_offset = 0; |
| queue->toggle_state = 1; |
| queue->small_page = NULL; |
| |
| /* allocate queue page pointers */ |
| queue->queue_pages = vmalloc(nr_of_pages * sizeof(void *)); |
| if (!queue->queue_pages) { |
| ehca_gen_err("Couldn't allocate queue page list"); |
| return 0; |
| } |
| memset(queue->queue_pages, 0, nr_of_pages * sizeof(void *)); |
| |
| /* allocate actual queue pages */ |
| if (is_small) { |
| if (!alloc_small_queue_page(queue, pd)) |
| goto ipz_queue_ctor_exit0; |
| } else |
| if (!alloc_queue_pages(queue, nr_of_pages)) |
| goto ipz_queue_ctor_exit0; |
| |
| return 1; |
| |
| ipz_queue_ctor_exit0: |
| ehca_gen_err("Couldn't alloc pages queue=%p " |
| "nr_of_pages=%x", queue, nr_of_pages); |
| vfree(queue->queue_pages); |
| |
| return 0; |
| } |
| |
| int ipz_queue_dtor(struct ehca_pd *pd, struct ipz_queue *queue) |
| { |
| int i, nr_pages; |
| |
| if (!queue || !queue->queue_pages) { |
| ehca_gen_dbg("queue or queue_pages is NULL"); |
| return 0; |
| } |
| |
| if (queue->small_page) |
| free_small_queue_page(queue, pd); |
| else { |
| nr_pages = queue->queue_length / queue->pagesize; |
| for (i = 0; i < nr_pages; i += PAGES_PER_KPAGE) |
| free_page((unsigned long)queue->queue_pages[i]); |
| } |
| |
| vfree(queue->queue_pages); |
| |
| return 1; |
| } |
| |
| int ehca_init_small_qp_cache(void) |
| { |
| small_qp_cache = kmem_cache_create("ehca_cache_small_qp", |
| sizeof(struct ipz_small_queue_page), |
| 0, SLAB_HWCACHE_ALIGN, NULL); |
| if (!small_qp_cache) |
| return -ENOMEM; |
| |
| return 0; |
| } |
| |
| void ehca_cleanup_small_qp_cache(void) |
| { |
| kmem_cache_destroy(small_qp_cache); |
| } |