2  *  IBM eServer eHCA Infiniband device driver for Linux on POWER
 
   4  *  Completion queue handling
 
   6  *  Authors: Waleri Fomin <fomin@de.ibm.com>
 
   7  *           Khadija Souissi <souissi@de.ibm.com>
 
   8  *           Reinhard Ernst <rernst@de.ibm.com>
 
   9  *           Heiko J Schick <schickhj@de.ibm.com>
 
  10  *           Hoang-Nam Nguyen <hnguyen@de.ibm.com>
 
  13  *  Copyright (c) 2005 IBM Corporation
 
  15  *  All rights reserved.
 
  17  *  This source code is distributed under a dual license of GPL v2.0 and OpenIB
 
  22  * Redistribution and use in source and binary forms, with or without
 
  23  * modification, are permitted provided that the following conditions are met:
 
  25  * Redistributions of source code must retain the above copyright notice, this
 
  26  * list of conditions and the following disclaimer.
 
  28  * Redistributions in binary form must reproduce the above copyright notice,
 
  29  * this list of conditions and the following disclaimer in the documentation
 
  30  * and/or other materials
 
  31  * provided with the distribution.
 
  33  * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
 
  34  * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
 
  35  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
 
  36  * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
 
  37  * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
 
  38  * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
 
  39  * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR
 
  40  * BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER
 
  41  * IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
 
  42  * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
 
  43  * POSSIBILITY OF SUCH DAMAGE.
 
  46 #include <asm/current.h>
 
  48 #include "ehca_iverbs.h"
 
  49 #include "ehca_classes.h"
 
  53 static struct kmem_cache *cq_cache;
 
  55 int ehca_cq_assign_qp(struct ehca_cq *cq, struct ehca_qp *qp)
 
  57         unsigned int qp_num = qp->real_qp_num;
 
  58         unsigned int key = qp_num & (QP_HASHTAB_LEN-1);
 
  61         spin_lock_irqsave(&cq->spinlock, flags);
 
  62         hlist_add_head(&qp->list_entries, &cq->qp_hashtab[key]);
 
  63         spin_unlock_irqrestore(&cq->spinlock, flags);
 
  65         ehca_dbg(cq->ib_cq.device, "cq_num=%x real_qp_num=%x",
 
  66                  cq->cq_number, qp_num);
 
  71 int ehca_cq_unassign_qp(struct ehca_cq *cq, unsigned int real_qp_num)
 
  74         unsigned int key = real_qp_num & (QP_HASHTAB_LEN-1);
 
  75         struct hlist_node *iter;
 
  79         spin_lock_irqsave(&cq->spinlock, flags);
 
  80         hlist_for_each(iter, &cq->qp_hashtab[key]) {
 
  81                 qp = hlist_entry(iter, struct ehca_qp, list_entries);
 
  82                 if (qp->real_qp_num == real_qp_num) {
 
  84                         ehca_dbg(cq->ib_cq.device,
 
  85                                  "removed qp from cq .cq_num=%x real_qp_num=%x",
 
  86                                  cq->cq_number, real_qp_num);
 
  91         spin_unlock_irqrestore(&cq->spinlock, flags);
 
  93                 ehca_err(cq->ib_cq.device,
 
  94                          "qp not found cq_num=%x real_qp_num=%x",
 
  95                          cq->cq_number, real_qp_num);
 
 100 struct ehca_qp *ehca_cq_get_qp(struct ehca_cq *cq, int real_qp_num)
 
 102         struct ehca_qp *ret = NULL;
 
 103         unsigned int key = real_qp_num & (QP_HASHTAB_LEN-1);
 
 104         struct hlist_node *iter;
 
 106         hlist_for_each(iter, &cq->qp_hashtab[key]) {
 
 107                 qp = hlist_entry(iter, struct ehca_qp, list_entries);
 
 108                 if (qp->real_qp_num == real_qp_num) {
 
 116 struct ib_cq *ehca_create_cq(struct ib_device *device, int cqe, int comp_vector,
 
 117                              struct ib_ucontext *context,
 
 118                              struct ib_udata *udata)
 
 120         static const u32 additional_cqe = 20;
 
 122         struct ehca_cq *my_cq;
 
 123         struct ehca_shca *shca =
 
 124                 container_of(device, struct ehca_shca, ib_device);
 
 125         struct ipz_adapter_handle adapter_handle;
 
 126         struct ehca_alloc_cq_parms param; /* h_call's out parameters */
 
 130         u64 rpage, cqx_fec, h_ret;
 
 134         if (cqe >= 0xFFFFFFFF - 64 - additional_cqe)
 
 135                 return ERR_PTR(-EINVAL);
 
 137         my_cq = kmem_cache_zalloc(cq_cache, GFP_KERNEL);
 
 139                 ehca_err(device, "Out of memory for ehca_cq struct device=%p",
 
 141                 return ERR_PTR(-ENOMEM);
 
 144         memset(¶m, 0, sizeof(struct ehca_alloc_cq_parms));
 
 146         spin_lock_init(&my_cq->spinlock);
 
 147         spin_lock_init(&my_cq->cb_lock);
 
 148         spin_lock_init(&my_cq->task_lock);
 
 149         atomic_set(&my_cq->nr_events, 0);
 
 150         init_waitqueue_head(&my_cq->wait_completion);
 
 151         my_cq->ownpid = current->tgid;
 
 155         adapter_handle = shca->ipz_hca_handle;
 
 156         param.eq_handle = shca->eq.ipz_eq_handle;
 
 159                 if (!idr_pre_get(&ehca_cq_idr, GFP_KERNEL)) {
 
 160                         cq = ERR_PTR(-ENOMEM);
 
 161                         ehca_err(device, "Can't reserve idr nr. device=%p",
 
 163                         goto create_cq_exit1;
 
 166                 write_lock_irqsave(&ehca_cq_idr_lock, flags);
 
 167                 ret = idr_get_new(&ehca_cq_idr, my_cq, &my_cq->token);
 
 168                 write_unlock_irqrestore(&ehca_cq_idr_lock, flags);
 
 170         } while (ret == -EAGAIN);
 
 173                 cq = ERR_PTR(-ENOMEM);
 
 174                 ehca_err(device, "Can't allocate new idr entry. device=%p",
 
 176                 goto create_cq_exit1;
 
 180          * CQs maximum depth is 4GB-64, but we need additional 20 as buffer
 
 181          * for receiving errors CQEs.
 
 183         param.nr_cqe = cqe + additional_cqe;
 
 184         h_ret = hipz_h_alloc_resource_cq(adapter_handle, my_cq, ¶m);
 
 186         if (h_ret != H_SUCCESS) {
 
 187                 ehca_err(device, "hipz_h_alloc_resource_cq() failed "
 
 188                          "h_ret=%lx device=%p", h_ret, device);
 
 189                 cq = ERR_PTR(ehca2ib_return_code(h_ret));
 
 190                 goto create_cq_exit2;
 
 193         ipz_rc = ipz_queue_ctor(NULL, &my_cq->ipz_queue, param.act_pages,
 
 194                                 EHCA_PAGESIZE, sizeof(struct ehca_cqe), 0, 0);
 
 196                 ehca_err(device, "ipz_queue_ctor() failed ipz_rc=%x device=%p",
 
 198                 cq = ERR_PTR(-EINVAL);
 
 199                 goto create_cq_exit3;
 
 202         for (counter = 0; counter < param.act_pages; counter++) {
 
 203                 vpage = ipz_qpageit_get_inc(&my_cq->ipz_queue);
 
 205                         ehca_err(device, "ipz_qpageit_get_inc() "
 
 206                                  "returns NULL device=%p", device);
 
 207                         cq = ERR_PTR(-EAGAIN);
 
 208                         goto create_cq_exit4;
 
 210                 rpage = virt_to_abs(vpage);
 
 212                 h_ret = hipz_h_register_rpage_cq(adapter_handle,
 
 213                                                  my_cq->ipz_cq_handle,
 
 222                 if (h_ret < H_SUCCESS) {
 
 223                         ehca_err(device, "hipz_h_register_rpage_cq() failed "
 
 224                                  "ehca_cq=%p cq_num=%x h_ret=%lx counter=%i "
 
 225                                  "act_pages=%i", my_cq, my_cq->cq_number,
 
 226                                  h_ret, counter, param.act_pages);
 
 227                         cq = ERR_PTR(-EINVAL);
 
 228                         goto create_cq_exit4;
 
 231                 if (counter == (param.act_pages - 1)) {
 
 232                         vpage = ipz_qpageit_get_inc(&my_cq->ipz_queue);
 
 233                         if ((h_ret != H_SUCCESS) || vpage) {
 
 234                                 ehca_err(device, "Registration of pages not "
 
 235                                          "complete ehca_cq=%p cq_num=%x "
 
 236                                          "h_ret=%lx", my_cq, my_cq->cq_number,
 
 238                                 cq = ERR_PTR(-EAGAIN);
 
 239                                 goto create_cq_exit4;
 
 242                         if (h_ret != H_PAGE_REGISTERED) {
 
 243                                 ehca_err(device, "Registration of page failed "
 
 244                                          "ehca_cq=%p cq_num=%x h_ret=%lx"
 
 245                                          "counter=%i act_pages=%i",
 
 246                                          my_cq, my_cq->cq_number,
 
 247                                          h_ret, counter, param.act_pages);
 
 248                                 cq = ERR_PTR(-ENOMEM);
 
 249                                 goto create_cq_exit4;
 
 254         ipz_qeit_reset(&my_cq->ipz_queue);
 
 256         gal = my_cq->galpas.kernel;
 
 257         cqx_fec = hipz_galpa_load(gal, CQTEMM_OFFSET(cqx_fec));
 
 258         ehca_dbg(device, "ehca_cq=%p cq_num=%x CQX_FEC=%lx",
 
 259                  my_cq, my_cq->cq_number, cqx_fec);
 
 261         my_cq->ib_cq.cqe = my_cq->nr_of_entries =
 
 262                 param.act_nr_of_entries - additional_cqe;
 
 263         my_cq->cq_number = (my_cq->ipz_cq_handle.handle) & 0xffff;
 
 265         for (i = 0; i < QP_HASHTAB_LEN; i++)
 
 266                 INIT_HLIST_HEAD(&my_cq->qp_hashtab[i]);
 
 269                 struct ipz_queue *ipz_queue = &my_cq->ipz_queue;
 
 270                 struct ehca_create_cq_resp resp;
 
 271                 memset(&resp, 0, sizeof(resp));
 
 272                 resp.cq_number = my_cq->cq_number;
 
 273                 resp.token = my_cq->token;
 
 274                 resp.ipz_queue.qe_size = ipz_queue->qe_size;
 
 275                 resp.ipz_queue.act_nr_of_sg = ipz_queue->act_nr_of_sg;
 
 276                 resp.ipz_queue.queue_length = ipz_queue->queue_length;
 
 277                 resp.ipz_queue.pagesize = ipz_queue->pagesize;
 
 278                 resp.ipz_queue.toggle_state = ipz_queue->toggle_state;
 
 279                 if (ib_copy_to_udata(udata, &resp, sizeof(resp))) {
 
 280                         ehca_err(device, "Copy to udata failed.");
 
 281                         goto create_cq_exit4;
 
 288         ipz_queue_dtor(NULL, &my_cq->ipz_queue);
 
 291         h_ret = hipz_h_destroy_cq(adapter_handle, my_cq, 1);
 
 292         if (h_ret != H_SUCCESS)
 
 293                 ehca_err(device, "hipz_h_destroy_cq() failed ehca_cq=%p "
 
 294                          "cq_num=%x h_ret=%lx", my_cq, my_cq->cq_number, h_ret);
 
 297         write_lock_irqsave(&ehca_cq_idr_lock, flags);
 
 298         idr_remove(&ehca_cq_idr, my_cq->token);
 
 299         write_unlock_irqrestore(&ehca_cq_idr_lock, flags);
 
 302         kmem_cache_free(cq_cache, my_cq);
 
 307 int ehca_destroy_cq(struct ib_cq *cq)
 
 310         struct ehca_cq *my_cq = container_of(cq, struct ehca_cq, ib_cq);
 
 311         int cq_num = my_cq->cq_number;
 
 312         struct ib_device *device = cq->device;
 
 313         struct ehca_shca *shca = container_of(device, struct ehca_shca,
 
 315         struct ipz_adapter_handle adapter_handle = shca->ipz_hca_handle;
 
 316         u32 cur_pid = current->tgid;
 
 320                 if (my_cq->mm_count_galpa || my_cq->mm_count_queue) {
 
 321                         ehca_err(device, "Resources still referenced in "
 
 322                                  "user space cq_num=%x", my_cq->cq_number);
 
 325                 if (my_cq->ownpid != cur_pid) {
 
 326                         ehca_err(device, "Invalid caller pid=%x ownpid=%x "
 
 328                                  cur_pid, my_cq->ownpid, my_cq->cq_number);
 
 334          * remove the CQ from the idr first to make sure
 
 335          * no more interrupt tasklets will touch this CQ
 
 337         write_lock_irqsave(&ehca_cq_idr_lock, flags);
 
 338         idr_remove(&ehca_cq_idr, my_cq->token);
 
 339         write_unlock_irqrestore(&ehca_cq_idr_lock, flags);
 
 341         /* now wait until all pending events have completed */
 
 342         wait_event(my_cq->wait_completion, !atomic_read(&my_cq->nr_events));
 
 344         /* nobody's using our CQ any longer -- we can destroy it */
 
 345         h_ret = hipz_h_destroy_cq(adapter_handle, my_cq, 0);
 
 346         if (h_ret == H_R_STATE) {
 
 347                 /* cq in err: read err data and destroy it forcibly */
 
 348                 ehca_dbg(device, "ehca_cq=%p cq_num=%x ressource=%lx in err "
 
 349                          "state. Try to delete it forcibly.",
 
 350                          my_cq, cq_num, my_cq->ipz_cq_handle.handle);
 
 351                 ehca_error_data(shca, my_cq, my_cq->ipz_cq_handle.handle);
 
 352                 h_ret = hipz_h_destroy_cq(adapter_handle, my_cq, 1);
 
 353                 if (h_ret == H_SUCCESS)
 
 354                         ehca_dbg(device, "cq_num=%x deleted successfully.",
 
 357         if (h_ret != H_SUCCESS) {
 
 358                 ehca_err(device, "hipz_h_destroy_cq() failed h_ret=%lx "
 
 359                          "ehca_cq=%p cq_num=%x", h_ret, my_cq, cq_num);
 
 360                 return ehca2ib_return_code(h_ret);
 
 362         ipz_queue_dtor(NULL, &my_cq->ipz_queue);
 
 363         kmem_cache_free(cq_cache, my_cq);
 
 368 int ehca_resize_cq(struct ib_cq *cq, int cqe, struct ib_udata *udata)
 
 370         struct ehca_cq *my_cq = container_of(cq, struct ehca_cq, ib_cq);
 
 371         u32 cur_pid = current->tgid;
 
 373         if (cq->uobject && my_cq->ownpid != cur_pid) {
 
 374                 ehca_err(cq->device, "Invalid caller pid=%x ownpid=%x",
 
 375                          cur_pid, my_cq->ownpid);
 
 379         /* TODO: proper resize needs to be done */
 
 380         ehca_err(cq->device, "not implemented yet");
 
 385 int ehca_init_cq_cache(void)
 
 387         cq_cache = kmem_cache_create("ehca_cache_cq",
 
 388                                      sizeof(struct ehca_cq), 0,
 
 396 void ehca_cleanup_cq_cache(void)
 
 399                 kmem_cache_destroy(cq_cache);