| /* |
| * Copyright (c) 2005 Ammasso, Inc. All rights reserved. |
| * Copyright (c) 2005 Open Grid Computing, Inc. All rights reserved. |
| * |
| * This software is available to you under a choice of one of two |
| * licenses. You may choose to be licensed under the terms of the GNU |
| * General Public License (GPL) Version 2, available from the file |
| * COPYING in the main directory of this source tree, or the |
| * OpenIB.org BSD license below: |
| * |
| * Redistribution and use in source and binary forms, with or |
| * without modification, are permitted provided that the following |
| * conditions are met: |
| * |
| * - Redistributions of source code must retain the above |
| * copyright notice, this list of conditions and the following |
| * disclaimer. |
| * |
| * - Redistributions in binary form must reproduce the above |
| * copyright notice, this list of conditions and the following |
| * disclaimer in the documentation and/or other materials |
| * provided with the distribution. |
| * |
| * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, |
| * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF |
| * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND |
| * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS |
| * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN |
| * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN |
| * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE |
| * SOFTWARE. |
| */ |
| #include <linux/slab.h> |
| |
| #include "c2.h" |
| #include "c2_vq.h" |
| |
| #define PBL_VIRT 1 |
| #define PBL_PHYS 2 |
| |
| /* |
| * Send all the PBL messages to convey the remainder of the PBL |
| * Wait for the adapter's reply on the last one. |
| * This is indicated by setting the MEM_PBL_COMPLETE in the flags. |
| * |
| * NOTE: vq_req is _not_ freed by this function. The VQ Host |
| * Reply buffer _is_ freed by this function. |
| */ |
| static int |
| send_pbl_messages(struct c2_dev *c2dev, __be32 stag_index, |
| unsigned long va, u32 pbl_depth, |
| struct c2_vq_req *vq_req, int pbl_type) |
| { |
| u32 pbe_count; /* amt that fits in a PBL msg */ |
| u32 count; /* amt in this PBL MSG. */ |
| struct c2wr_nsmr_pbl_req *wr; /* PBL WR ptr */ |
| struct c2wr_nsmr_pbl_rep *reply; /* reply ptr */ |
| int err, pbl_virt, pbl_index, i; |
| |
| switch (pbl_type) { |
| case PBL_VIRT: |
| pbl_virt = 1; |
| break; |
| case PBL_PHYS: |
| pbl_virt = 0; |
| break; |
| default: |
| return -EINVAL; |
| break; |
| } |
| |
| pbe_count = (c2dev->req_vq.msg_size - |
| sizeof(struct c2wr_nsmr_pbl_req)) / sizeof(u64); |
| wr = kmalloc(c2dev->req_vq.msg_size, GFP_KERNEL); |
| if (!wr) { |
| return -ENOMEM; |
| } |
| c2_wr_set_id(wr, CCWR_NSMR_PBL); |
| |
| /* |
| * Only the last PBL message will generate a reply from the verbs, |
| * so we set the context to 0 indicating there is no kernel verbs |
| * handler blocked awaiting this reply. |
| */ |
| wr->hdr.context = 0; |
| wr->rnic_handle = c2dev->adapter_handle; |
| wr->stag_index = stag_index; /* already swapped */ |
| wr->flags = 0; |
| pbl_index = 0; |
| while (pbl_depth) { |
| count = min(pbe_count, pbl_depth); |
| wr->addrs_length = cpu_to_be32(count); |
| |
| /* |
| * If this is the last message, then reference the |
| * vq request struct cuz we're gonna wait for a reply. |
| * also make this PBL msg as the last one. |
| */ |
| if (count == pbl_depth) { |
| /* |
| * reference the request struct. dereferenced in the |
| * int handler. |
| */ |
| vq_req_get(c2dev, vq_req); |
| wr->flags = cpu_to_be32(MEM_PBL_COMPLETE); |
| |
| /* |
| * This is the last PBL message. |
| * Set the context to our VQ Request Object so we can |
| * wait for the reply. |
| */ |
| wr->hdr.context = (unsigned long) vq_req; |
| } |
| |
| /* |
| * If pbl_virt is set then va is a virtual address |
| * that describes a virtually contiguous memory |
| * allocation. The wr needs the start of each virtual page |
| * to be converted to the corresponding physical address |
| * of the page. If pbl_virt is not set then va is an array |
| * of physical addresses and there is no conversion to do. |
| * Just fill in the wr with what is in the array. |
| */ |
| for (i = 0; i < count; i++) { |
| if (pbl_virt) { |
| va += PAGE_SIZE; |
| } else { |
| wr->paddrs[i] = |
| cpu_to_be64(((u64 *)va)[pbl_index + i]); |
| } |
| } |
| |
| /* |
| * Send WR to adapter |
| */ |
| err = vq_send_wr(c2dev, (union c2wr *) wr); |
| if (err) { |
| if (count <= pbe_count) { |
| vq_req_put(c2dev, vq_req); |
| } |
| goto bail0; |
| } |
| pbl_depth -= count; |
| pbl_index += count; |
| } |
| |
| /* |
| * Now wait for the reply... |
| */ |
| err = vq_wait_for_reply(c2dev, vq_req); |
| if (err) { |
| goto bail0; |
| } |
| |
| /* |
| * Process reply |
| */ |
| reply = (struct c2wr_nsmr_pbl_rep *) (unsigned long) vq_req->reply_msg; |
| if (!reply) { |
| err = -ENOMEM; |
| goto bail0; |
| } |
| |
| err = c2_errno(reply); |
| |
| vq_repbuf_free(c2dev, reply); |
| bail0: |
| kfree(wr); |
| return err; |
| } |
| |
| #define C2_PBL_MAX_DEPTH 131072 |
| int |
| c2_nsmr_register_phys_kern(struct c2_dev *c2dev, u64 *addr_list, |
| int page_size, int pbl_depth, u32 length, |
| u32 offset, u64 *va, enum c2_acf acf, |
| struct c2_mr *mr) |
| { |
| struct c2_vq_req *vq_req; |
| struct c2wr_nsmr_register_req *wr; |
| struct c2wr_nsmr_register_rep *reply; |
| u16 flags; |
| int i, pbe_count, count; |
| int err; |
| |
| if (!va || !length || !addr_list || !pbl_depth) |
| return -EINTR; |
| |
| /* |
| * Verify PBL depth is within rnic max |
| */ |
| if (pbl_depth > C2_PBL_MAX_DEPTH) { |
| return -EINTR; |
| } |
| |
| /* |
| * allocate verbs request object |
| */ |
| vq_req = vq_req_alloc(c2dev); |
| if (!vq_req) |
| return -ENOMEM; |
| |
| wr = kmalloc(c2dev->req_vq.msg_size, GFP_KERNEL); |
| if (!wr) { |
| err = -ENOMEM; |
| goto bail0; |
| } |
| |
| /* |
| * build the WR |
| */ |
| c2_wr_set_id(wr, CCWR_NSMR_REGISTER); |
| wr->hdr.context = (unsigned long) vq_req; |
| wr->rnic_handle = c2dev->adapter_handle; |
| |
| flags = (acf | MEM_VA_BASED | MEM_REMOTE); |
| |
| /* |
| * compute how many pbes can fit in the message |
| */ |
| pbe_count = (c2dev->req_vq.msg_size - |
| sizeof(struct c2wr_nsmr_register_req)) / sizeof(u64); |
| |
| if (pbl_depth <= pbe_count) { |
| flags |= MEM_PBL_COMPLETE; |
| } |
| wr->flags = cpu_to_be16(flags); |
| wr->stag_key = 0; //stag_key; |
| wr->va = cpu_to_be64(*va); |
| wr->pd_id = mr->pd->pd_id; |
| wr->pbe_size = cpu_to_be32(page_size); |
| wr->length = cpu_to_be32(length); |
| wr->pbl_depth = cpu_to_be32(pbl_depth); |
| wr->fbo = cpu_to_be32(offset); |
| count = min(pbl_depth, pbe_count); |
| wr->addrs_length = cpu_to_be32(count); |
| |
| /* |
| * fill out the PBL for this message |
| */ |
| for (i = 0; i < count; i++) { |
| wr->paddrs[i] = cpu_to_be64(addr_list[i]); |
| } |
| |
| /* |
| * regerence the request struct |
| */ |
| vq_req_get(c2dev, vq_req); |
| |
| /* |
| * send the WR to the adapter |
| */ |
| err = vq_send_wr(c2dev, (union c2wr *) wr); |
| if (err) { |
| vq_req_put(c2dev, vq_req); |
| goto bail1; |
| } |
| |
| /* |
| * wait for reply from adapter |
| */ |
| err = vq_wait_for_reply(c2dev, vq_req); |
| if (err) { |
| goto bail1; |
| } |
| |
| /* |
| * process reply |
| */ |
| reply = |
| (struct c2wr_nsmr_register_rep *) (unsigned long) (vq_req->reply_msg); |
| if (!reply) { |
| err = -ENOMEM; |
| goto bail1; |
| } |
| if ((err = c2_errno(reply))) { |
| goto bail2; |
| } |
| //*p_pb_entries = be32_to_cpu(reply->pbl_depth); |
| mr->ibmr.lkey = mr->ibmr.rkey = be32_to_cpu(reply->stag_index); |
| vq_repbuf_free(c2dev, reply); |
| |
| /* |
| * if there are still more PBEs we need to send them to |
| * the adapter and wait for a reply on the final one. |
| * reuse vq_req for this purpose. |
| */ |
| pbl_depth -= count; |
| if (pbl_depth) { |
| |
| vq_req->reply_msg = (unsigned long) NULL; |
| atomic_set(&vq_req->reply_ready, 0); |
| err = send_pbl_messages(c2dev, |
| cpu_to_be32(mr->ibmr.lkey), |
| (unsigned long) &addr_list[i], |
| pbl_depth, vq_req, PBL_PHYS); |
| if (err) { |
| goto bail1; |
| } |
| } |
| |
| vq_req_free(c2dev, vq_req); |
| kfree(wr); |
| |
| return err; |
| |
| bail2: |
| vq_repbuf_free(c2dev, reply); |
| bail1: |
| kfree(wr); |
| bail0: |
| vq_req_free(c2dev, vq_req); |
| return err; |
| } |
| |
| int c2_stag_dealloc(struct c2_dev *c2dev, u32 stag_index) |
| { |
| struct c2_vq_req *vq_req; /* verbs request object */ |
| struct c2wr_stag_dealloc_req wr; /* work request */ |
| struct c2wr_stag_dealloc_rep *reply; /* WR reply */ |
| int err; |
| |
| |
| /* |
| * allocate verbs request object |
| */ |
| vq_req = vq_req_alloc(c2dev); |
| if (!vq_req) { |
| return -ENOMEM; |
| } |
| |
| /* |
| * Build the WR |
| */ |
| c2_wr_set_id(&wr, CCWR_STAG_DEALLOC); |
| wr.hdr.context = (u64) (unsigned long) vq_req; |
| wr.rnic_handle = c2dev->adapter_handle; |
| wr.stag_index = cpu_to_be32(stag_index); |
| |
| /* |
| * reference the request struct. dereferenced in the int handler. |
| */ |
| vq_req_get(c2dev, vq_req); |
| |
| /* |
| * Send WR to adapter |
| */ |
| err = vq_send_wr(c2dev, (union c2wr *) & wr); |
| if (err) { |
| vq_req_put(c2dev, vq_req); |
| goto bail0; |
| } |
| |
| /* |
| * Wait for reply from adapter |
| */ |
| err = vq_wait_for_reply(c2dev, vq_req); |
| if (err) { |
| goto bail0; |
| } |
| |
| /* |
| * Process reply |
| */ |
| reply = (struct c2wr_stag_dealloc_rep *) (unsigned long) vq_req->reply_msg; |
| if (!reply) { |
| err = -ENOMEM; |
| goto bail0; |
| } |
| |
| err = c2_errno(reply); |
| |
| vq_repbuf_free(c2dev, reply); |
| bail0: |
| vq_req_free(c2dev, vq_req); |
| return err; |
| } |