 dc544bc9cb
			
		
	
	
	dc544bc9cb
	
	
	
		
			
			Signed-off-by: Roland Dreier <rolandd@cisco.com> Acked-by: Steve Wise <swise@opengridcomputing.com>
		
			
				
	
	
		
			375 lines
		
	
	
	
		
			8.7 KiB
			
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			375 lines
		
	
	
	
		
			8.7 KiB
			
		
	
	
	
		
			C
		
	
	
	
	
	
| /*
 | |
|  * Copyright (c) 2005 Ammasso, Inc. All rights reserved.
 | |
|  * Copyright (c) 2005 Open Grid Computing, Inc. All rights reserved.
 | |
|  *
 | |
|  * This software is available to you under a choice of one of two
 | |
|  * licenses.  You may choose to be licensed under the terms of the GNU
 | |
|  * General Public License (GPL) Version 2, available from the file
 | |
|  * COPYING in the main directory of this source tree, or the
 | |
|  * OpenIB.org BSD license below:
 | |
|  *
 | |
|  *     Redistribution and use in source and binary forms, with or
 | |
|  *     without modification, are permitted provided that the following
 | |
|  *     conditions are met:
 | |
|  *
 | |
|  *      - Redistributions of source code must retain the above
 | |
|  *        copyright notice, this list of conditions and the following
 | |
|  *        disclaimer.
 | |
|  *
 | |
|  *      - Redistributions in binary form must reproduce the above
 | |
|  *        copyright notice, this list of conditions and the following
 | |
|  *        disclaimer in the documentation and/or other materials
 | |
|  *        provided with the distribution.
 | |
|  *
 | |
|  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
 | |
|  * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
 | |
|  * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
 | |
|  * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
 | |
|  * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
 | |
|  * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
 | |
|  * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
 | |
|  * SOFTWARE.
 | |
|  */
 | |
| #include "c2.h"
 | |
| #include "c2_vq.h"
 | |
| 
 | |
| #define PBL_VIRT 1
 | |
| #define PBL_PHYS 2
 | |
| 
 | |
| /*
 | |
|  * Send all the PBL messages to convey the remainder of the PBL
 | |
|  * Wait for the adapter's reply on the last one.
 | |
|  * This is indicated by setting the MEM_PBL_COMPLETE in the flags.
 | |
|  *
 | |
|  * NOTE:  vq_req is _not_ freed by this function.  The VQ Host
 | |
|  *	  Reply buffer _is_ freed by this function.
 | |
|  */
 | |
| static int
 | |
| send_pbl_messages(struct c2_dev *c2dev, __be32 stag_index,
 | |
| 		  unsigned long va, u32 pbl_depth,
 | |
| 		  struct c2_vq_req *vq_req, int pbl_type)
 | |
| {
 | |
| 	u32 pbe_count;		/* amt that fits in a PBL msg */
 | |
| 	u32 count;		/* amt in this PBL MSG. */
 | |
| 	struct c2wr_nsmr_pbl_req *wr;	/* PBL WR ptr */
 | |
| 	struct c2wr_nsmr_pbl_rep *reply;	/* reply ptr */
 | |
|  	int err, pbl_virt, pbl_index, i;
 | |
| 
 | |
| 	switch (pbl_type) {
 | |
| 	case PBL_VIRT:
 | |
| 		pbl_virt = 1;
 | |
| 		break;
 | |
| 	case PBL_PHYS:
 | |
| 		pbl_virt = 0;
 | |
| 		break;
 | |
| 	default:
 | |
| 		return -EINVAL;
 | |
| 		break;
 | |
| 	}
 | |
| 
 | |
| 	pbe_count = (c2dev->req_vq.msg_size -
 | |
| 		     sizeof(struct c2wr_nsmr_pbl_req)) / sizeof(u64);
 | |
| 	wr = kmalloc(c2dev->req_vq.msg_size, GFP_KERNEL);
 | |
| 	if (!wr) {
 | |
| 		return -ENOMEM;
 | |
| 	}
 | |
| 	c2_wr_set_id(wr, CCWR_NSMR_PBL);
 | |
| 
 | |
| 	/*
 | |
| 	 * Only the last PBL message will generate a reply from the verbs,
 | |
| 	 * so we set the context to 0 indicating there is no kernel verbs
 | |
| 	 * handler blocked awaiting this reply.
 | |
| 	 */
 | |
| 	wr->hdr.context = 0;
 | |
| 	wr->rnic_handle = c2dev->adapter_handle;
 | |
| 	wr->stag_index = stag_index;	/* already swapped */
 | |
| 	wr->flags = 0;
 | |
| 	pbl_index = 0;
 | |
| 	while (pbl_depth) {
 | |
| 		count = min(pbe_count, pbl_depth);
 | |
| 		wr->addrs_length = cpu_to_be32(count);
 | |
| 
 | |
| 		/*
 | |
| 		 *  If this is the last message, then reference the
 | |
| 		 *  vq request struct cuz we're gonna wait for a reply.
 | |
| 		 *  also make this PBL msg as the last one.
 | |
| 		 */
 | |
| 		if (count == pbl_depth) {
 | |
| 			/*
 | |
| 			 * reference the request struct.  dereferenced in the
 | |
| 			 * int handler.
 | |
| 			 */
 | |
| 			vq_req_get(c2dev, vq_req);
 | |
| 			wr->flags = cpu_to_be32(MEM_PBL_COMPLETE);
 | |
| 
 | |
| 			/*
 | |
| 			 * This is the last PBL message.
 | |
| 			 * Set the context to our VQ Request Object so we can
 | |
| 			 * wait for the reply.
 | |
| 			 */
 | |
| 			wr->hdr.context = (unsigned long) vq_req;
 | |
| 		}
 | |
| 
 | |
| 		/*
 | |
| 		 * If pbl_virt is set then va is a virtual address
 | |
| 		 * that describes a virtually contiguous memory
 | |
| 		 * allocation. The wr needs the start of each virtual page
 | |
| 		 * to be converted to the corresponding physical address
 | |
| 		 * of the page. If pbl_virt is not set then va is an array
 | |
| 		 * of physical addresses and there is no conversion to do.
 | |
| 		 * Just fill in the wr with what is in the array.
 | |
| 		 */
 | |
| 		for (i = 0; i < count; i++) {
 | |
| 			if (pbl_virt) {
 | |
| 				va += PAGE_SIZE;
 | |
| 			} else {
 | |
|  				wr->paddrs[i] =
 | |
| 				    cpu_to_be64(((u64 *)va)[pbl_index + i]);
 | |
| 			}
 | |
| 		}
 | |
| 
 | |
| 		/*
 | |
| 		 * Send WR to adapter
 | |
| 		 */
 | |
| 		err = vq_send_wr(c2dev, (union c2wr *) wr);
 | |
| 		if (err) {
 | |
| 			if (count <= pbe_count) {
 | |
| 				vq_req_put(c2dev, vq_req);
 | |
| 			}
 | |
| 			goto bail0;
 | |
| 		}
 | |
| 		pbl_depth -= count;
 | |
| 		pbl_index += count;
 | |
| 	}
 | |
| 
 | |
| 	/*
 | |
| 	 *  Now wait for the reply...
 | |
| 	 */
 | |
| 	err = vq_wait_for_reply(c2dev, vq_req);
 | |
| 	if (err) {
 | |
| 		goto bail0;
 | |
| 	}
 | |
| 
 | |
| 	/*
 | |
| 	 * Process reply
 | |
| 	 */
 | |
| 	reply = (struct c2wr_nsmr_pbl_rep *) (unsigned long) vq_req->reply_msg;
 | |
| 	if (!reply) {
 | |
| 		err = -ENOMEM;
 | |
| 		goto bail0;
 | |
| 	}
 | |
| 
 | |
| 	err = c2_errno(reply);
 | |
| 
 | |
| 	vq_repbuf_free(c2dev, reply);
 | |
|       bail0:
 | |
| 	kfree(wr);
 | |
| 	return err;
 | |
| }
 | |
| 
 | |
| #define C2_PBL_MAX_DEPTH 131072
 | |
| int
 | |
| c2_nsmr_register_phys_kern(struct c2_dev *c2dev, u64 *addr_list,
 | |
|  			   int page_size, int pbl_depth, u32 length,
 | |
|  			   u32 offset, u64 *va, enum c2_acf acf,
 | |
| 			   struct c2_mr *mr)
 | |
| {
 | |
| 	struct c2_vq_req *vq_req;
 | |
| 	struct c2wr_nsmr_register_req *wr;
 | |
| 	struct c2wr_nsmr_register_rep *reply;
 | |
| 	u16 flags;
 | |
| 	int i, pbe_count, count;
 | |
| 	int err;
 | |
| 
 | |
| 	if (!va || !length || !addr_list || !pbl_depth)
 | |
| 		return -EINTR;
 | |
| 
 | |
| 	/*
 | |
| 	 * Verify PBL depth is within rnic max
 | |
| 	 */
 | |
| 	if (pbl_depth > C2_PBL_MAX_DEPTH) {
 | |
| 		return -EINTR;
 | |
| 	}
 | |
| 
 | |
| 	/*
 | |
| 	 * allocate verbs request object
 | |
| 	 */
 | |
| 	vq_req = vq_req_alloc(c2dev);
 | |
| 	if (!vq_req)
 | |
| 		return -ENOMEM;
 | |
| 
 | |
| 	wr = kmalloc(c2dev->req_vq.msg_size, GFP_KERNEL);
 | |
| 	if (!wr) {
 | |
| 		err = -ENOMEM;
 | |
| 		goto bail0;
 | |
| 	}
 | |
| 
 | |
| 	/*
 | |
| 	 * build the WR
 | |
| 	 */
 | |
| 	c2_wr_set_id(wr, CCWR_NSMR_REGISTER);
 | |
| 	wr->hdr.context = (unsigned long) vq_req;
 | |
| 	wr->rnic_handle = c2dev->adapter_handle;
 | |
| 
 | |
| 	flags = (acf | MEM_VA_BASED | MEM_REMOTE);
 | |
| 
 | |
| 	/*
 | |
| 	 * compute how many pbes can fit in the message
 | |
| 	 */
 | |
| 	pbe_count = (c2dev->req_vq.msg_size -
 | |
| 		     sizeof(struct c2wr_nsmr_register_req)) / sizeof(u64);
 | |
| 
 | |
| 	if (pbl_depth <= pbe_count) {
 | |
| 		flags |= MEM_PBL_COMPLETE;
 | |
| 	}
 | |
| 	wr->flags = cpu_to_be16(flags);
 | |
| 	wr->stag_key = 0;	//stag_key;
 | |
| 	wr->va = cpu_to_be64(*va);
 | |
| 	wr->pd_id = mr->pd->pd_id;
 | |
| 	wr->pbe_size = cpu_to_be32(page_size);
 | |
| 	wr->length = cpu_to_be32(length);
 | |
| 	wr->pbl_depth = cpu_to_be32(pbl_depth);
 | |
| 	wr->fbo = cpu_to_be32(offset);
 | |
| 	count = min(pbl_depth, pbe_count);
 | |
| 	wr->addrs_length = cpu_to_be32(count);
 | |
| 
 | |
| 	/*
 | |
| 	 * fill out the PBL for this message
 | |
| 	 */
 | |
| 	for (i = 0; i < count; i++) {
 | |
| 		wr->paddrs[i] = cpu_to_be64(addr_list[i]);
 | |
| 	}
 | |
| 
 | |
| 	/*
 | |
| 	 * regerence the request struct
 | |
| 	 */
 | |
| 	vq_req_get(c2dev, vq_req);
 | |
| 
 | |
| 	/*
 | |
| 	 * send the WR to the adapter
 | |
| 	 */
 | |
| 	err = vq_send_wr(c2dev, (union c2wr *) wr);
 | |
| 	if (err) {
 | |
| 		vq_req_put(c2dev, vq_req);
 | |
| 		goto bail1;
 | |
| 	}
 | |
| 
 | |
| 	/*
 | |
| 	 * wait for reply from adapter
 | |
| 	 */
 | |
| 	err = vq_wait_for_reply(c2dev, vq_req);
 | |
| 	if (err) {
 | |
| 		goto bail1;
 | |
| 	}
 | |
| 
 | |
| 	/*
 | |
| 	 * process reply
 | |
| 	 */
 | |
| 	reply =
 | |
| 	    (struct c2wr_nsmr_register_rep *) (unsigned long) (vq_req->reply_msg);
 | |
| 	if (!reply) {
 | |
| 		err = -ENOMEM;
 | |
| 		goto bail1;
 | |
| 	}
 | |
| 	if ((err = c2_errno(reply))) {
 | |
| 		goto bail2;
 | |
| 	}
 | |
| 	//*p_pb_entries = be32_to_cpu(reply->pbl_depth);
 | |
| 	mr->ibmr.lkey = mr->ibmr.rkey = be32_to_cpu(reply->stag_index);
 | |
| 	vq_repbuf_free(c2dev, reply);
 | |
| 
 | |
| 	/*
 | |
| 	 * if there are still more PBEs we need to send them to
 | |
| 	 * the adapter and wait for a reply on the final one.
 | |
| 	 * reuse vq_req for this purpose.
 | |
| 	 */
 | |
| 	pbl_depth -= count;
 | |
| 	if (pbl_depth) {
 | |
| 
 | |
| 		vq_req->reply_msg = (unsigned long) NULL;
 | |
| 		atomic_set(&vq_req->reply_ready, 0);
 | |
| 		err = send_pbl_messages(c2dev,
 | |
| 					cpu_to_be32(mr->ibmr.lkey),
 | |
| 					(unsigned long) &addr_list[i],
 | |
| 					pbl_depth, vq_req, PBL_PHYS);
 | |
| 		if (err) {
 | |
| 			goto bail1;
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	vq_req_free(c2dev, vq_req);
 | |
| 	kfree(wr);
 | |
| 
 | |
| 	return err;
 | |
| 
 | |
|       bail2:
 | |
| 	vq_repbuf_free(c2dev, reply);
 | |
|       bail1:
 | |
| 	kfree(wr);
 | |
|       bail0:
 | |
| 	vq_req_free(c2dev, vq_req);
 | |
| 	return err;
 | |
| }
 | |
| 
 | |
| int c2_stag_dealloc(struct c2_dev *c2dev, u32 stag_index)
 | |
| {
 | |
| 	struct c2_vq_req *vq_req;	/* verbs request object */
 | |
| 	struct c2wr_stag_dealloc_req wr;	/* work request */
 | |
| 	struct c2wr_stag_dealloc_rep *reply;	/* WR reply  */
 | |
| 	int err;
 | |
| 
 | |
| 
 | |
| 	/*
 | |
| 	 * allocate verbs request object
 | |
| 	 */
 | |
| 	vq_req = vq_req_alloc(c2dev);
 | |
| 	if (!vq_req) {
 | |
| 		return -ENOMEM;
 | |
| 	}
 | |
| 
 | |
| 	/*
 | |
| 	 * Build the WR
 | |
| 	 */
 | |
| 	c2_wr_set_id(&wr, CCWR_STAG_DEALLOC);
 | |
| 	wr.hdr.context = (u64) (unsigned long) vq_req;
 | |
| 	wr.rnic_handle = c2dev->adapter_handle;
 | |
| 	wr.stag_index = cpu_to_be32(stag_index);
 | |
| 
 | |
| 	/*
 | |
| 	 * reference the request struct.  dereferenced in the int handler.
 | |
| 	 */
 | |
| 	vq_req_get(c2dev, vq_req);
 | |
| 
 | |
| 	/*
 | |
| 	 * Send WR to adapter
 | |
| 	 */
 | |
| 	err = vq_send_wr(c2dev, (union c2wr *) & wr);
 | |
| 	if (err) {
 | |
| 		vq_req_put(c2dev, vq_req);
 | |
| 		goto bail0;
 | |
| 	}
 | |
| 
 | |
| 	/*
 | |
| 	 * Wait for reply from adapter
 | |
| 	 */
 | |
| 	err = vq_wait_for_reply(c2dev, vq_req);
 | |
| 	if (err) {
 | |
| 		goto bail0;
 | |
| 	}
 | |
| 
 | |
| 	/*
 | |
| 	 * Process reply
 | |
| 	 */
 | |
| 	reply = (struct c2wr_stag_dealloc_rep *) (unsigned long) vq_req->reply_msg;
 | |
| 	if (!reply) {
 | |
| 		err = -ENOMEM;
 | |
| 		goto bail0;
 | |
| 	}
 | |
| 
 | |
| 	err = c2_errno(reply);
 | |
| 
 | |
| 	vq_repbuf_free(c2dev, reply);
 | |
|       bail0:
 | |
| 	vq_req_free(c2dev, vq_req);
 | |
| 	return err;
 | |
| }
 |