Linux Kernel  3.7.1
 All Data Structures Namespaces Files Functions Variables Typedefs Enumerations Enumerator Macros Groups Pages
c2_mm.c
Go to the documentation of this file.
1 /*
2  * Copyright (c) 2005 Ammasso, Inc. All rights reserved.
3  * Copyright (c) 2005 Open Grid Computing, Inc. All rights reserved.
4  *
5  * This software is available to you under a choice of one of two
6  * licenses. You may choose to be licensed under the terms of the GNU
7  * General Public License (GPL) Version 2, available from the file
8  * COPYING in the main directory of this source tree, or the
9  * OpenIB.org BSD license below:
10  *
11  * Redistribution and use in source and binary forms, with or
12  * without modification, are permitted provided that the following
13  * conditions are met:
14  *
15  * - Redistributions of source code must retain the above
16  * copyright notice, this list of conditions and the following
17  * disclaimer.
18  *
19  * - Redistributions in binary form must reproduce the above
20  * copyright notice, this list of conditions and the following
21  * disclaimer in the documentation and/or other materials
22  * provided with the distribution.
23  *
24  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
25  * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
26  * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
27  * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
28  * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
29  * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
30  * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
31  * SOFTWARE.
32  */
33 #include <linux/slab.h>
34 
35 #include "c2.h"
36 #include "c2_vq.h"
37 
38 #define PBL_VIRT 1
39 #define PBL_PHYS 2
40 
41 /*
42  * Send all the PBL messages to convey the remainder of the PBL
43  * Wait for the adapter's reply on the last one.
44  * This is indicated by setting the MEM_PBL_COMPLETE in the flags.
45  *
46  * NOTE: vq_req is _not_ freed by this function. The VQ Host
47  * Reply buffer _is_ freed by this function.
48  */
49 static int
50 send_pbl_messages(struct c2_dev *c2dev, __be32 stag_index,
51  unsigned long va, u32 pbl_depth,
52  struct c2_vq_req *vq_req, int pbl_type)
53 {
54  u32 pbe_count; /* amt that fits in a PBL msg */
55  u32 count; /* amt in this PBL MSG. */
56  struct c2wr_nsmr_pbl_req *wr; /* PBL WR ptr */
57  struct c2wr_nsmr_pbl_rep *reply; /* reply ptr */
58  int err, pbl_virt, pbl_index, i;
59 
60  switch (pbl_type) {
61  case PBL_VIRT:
62  pbl_virt = 1;
63  break;
64  case PBL_PHYS:
65  pbl_virt = 0;
66  break;
67  default:
68  return -EINVAL;
69  break;
70  }
71 
72  pbe_count = (c2dev->req_vq.msg_size -
73  sizeof(struct c2wr_nsmr_pbl_req)) / sizeof(u64);
74  wr = kmalloc(c2dev->req_vq.msg_size, GFP_KERNEL);
75  if (!wr) {
76  return -ENOMEM;
77  }
78  c2_wr_set_id(wr, CCWR_NSMR_PBL);
79 
80  /*
81  * Only the last PBL message will generate a reply from the verbs,
82  * so we set the context to 0 indicating there is no kernel verbs
83  * handler blocked awaiting this reply.
84  */
85  wr->hdr.context = 0;
86  wr->rnic_handle = c2dev->adapter_handle;
87  wr->stag_index = stag_index; /* already swapped */
88  wr->flags = 0;
89  pbl_index = 0;
90  while (pbl_depth) {
91  count = min(pbe_count, pbl_depth);
92  wr->addrs_length = cpu_to_be32(count);
93 
94  /*
95  * If this is the last message, then reference the
96  * vq request struct cuz we're gonna wait for a reply.
97  * also make this PBL msg as the last one.
98  */
99  if (count == pbl_depth) {
100  /*
101  * reference the request struct. dereferenced in the
102  * int handler.
103  */
104  vq_req_get(c2dev, vq_req);
105  wr->flags = cpu_to_be32(MEM_PBL_COMPLETE);
106 
107  /*
108  * This is the last PBL message.
109  * Set the context to our VQ Request Object so we can
110  * wait for the reply.
111  */
112  wr->hdr.context = (unsigned long) vq_req;
113  }
114 
115  /*
116  * If pbl_virt is set then va is a virtual address
117  * that describes a virtually contiguous memory
118  * allocation. The wr needs the start of each virtual page
119  * to be converted to the corresponding physical address
120  * of the page. If pbl_virt is not set then va is an array
121  * of physical addresses and there is no conversion to do.
122  * Just fill in the wr with what is in the array.
123  */
124  for (i = 0; i < count; i++) {
125  if (pbl_virt) {
126  va += PAGE_SIZE;
127  } else {
128  wr->paddrs[i] =
129  cpu_to_be64(((u64 *)va)[pbl_index + i]);
130  }
131  }
132 
133  /*
134  * Send WR to adapter
135  */
136  err = vq_send_wr(c2dev, (union c2wr *) wr);
137  if (err) {
138  if (count <= pbe_count) {
139  vq_req_put(c2dev, vq_req);
140  }
141  goto bail0;
142  }
143  pbl_depth -= count;
144  pbl_index += count;
145  }
146 
147  /*
148  * Now wait for the reply...
149  */
150  err = vq_wait_for_reply(c2dev, vq_req);
151  if (err) {
152  goto bail0;
153  }
154 
155  /*
156  * Process reply
157  */
158  reply = (struct c2wr_nsmr_pbl_rep *) (unsigned long) vq_req->reply_msg;
159  if (!reply) {
160  err = -ENOMEM;
161  goto bail0;
162  }
163 
164  err = c2_errno(reply);
165 
166  vq_repbuf_free(c2dev, reply);
167  bail0:
168  kfree(wr);
169  return err;
170 }
171 
172 #define C2_PBL_MAX_DEPTH 131072
173 int
174 c2_nsmr_register_phys_kern(struct c2_dev *c2dev, u64 *addr_list,
175  int page_size, int pbl_depth, u32 length,
176  u32 offset, u64 *va, enum c2_acf acf,
177  struct c2_mr *mr)
178 {
179  struct c2_vq_req *vq_req;
180  struct c2wr_nsmr_register_req *wr;
181  struct c2wr_nsmr_register_rep *reply;
182  u16 flags;
183  int i, pbe_count, count;
184  int err;
185 
186  if (!va || !length || !addr_list || !pbl_depth)
187  return -EINTR;
188 
189  /*
190  * Verify PBL depth is within rnic max
191  */
192  if (pbl_depth > C2_PBL_MAX_DEPTH) {
193  return -EINTR;
194  }
195 
196  /*
197  * allocate verbs request object
198  */
199  vq_req = vq_req_alloc(c2dev);
200  if (!vq_req)
201  return -ENOMEM;
202 
203  wr = kmalloc(c2dev->req_vq.msg_size, GFP_KERNEL);
204  if (!wr) {
205  err = -ENOMEM;
206  goto bail0;
207  }
208 
209  /*
210  * build the WR
211  */
212  c2_wr_set_id(wr, CCWR_NSMR_REGISTER);
213  wr->hdr.context = (unsigned long) vq_req;
214  wr->rnic_handle = c2dev->adapter_handle;
215 
216  flags = (acf | MEM_VA_BASED | MEM_REMOTE);
217 
218  /*
219  * compute how many pbes can fit in the message
220  */
221  pbe_count = (c2dev->req_vq.msg_size -
222  sizeof(struct c2wr_nsmr_register_req)) / sizeof(u64);
223 
224  if (pbl_depth <= pbe_count) {
225  flags |= MEM_PBL_COMPLETE;
226  }
227  wr->flags = cpu_to_be16(flags);
228  wr->stag_key = 0; //stag_key;
229  wr->va = cpu_to_be64(*va);
230  wr->pd_id = mr->pd->pd_id;
231  wr->pbe_size = cpu_to_be32(page_size);
232  wr->length = cpu_to_be32(length);
233  wr->pbl_depth = cpu_to_be32(pbl_depth);
234  wr->fbo = cpu_to_be32(offset);
235  count = min(pbl_depth, pbe_count);
236  wr->addrs_length = cpu_to_be32(count);
237 
238  /*
239  * fill out the PBL for this message
240  */
241  for (i = 0; i < count; i++) {
242  wr->paddrs[i] = cpu_to_be64(addr_list[i]);
243  }
244 
245  /*
246  * regerence the request struct
247  */
248  vq_req_get(c2dev, vq_req);
249 
250  /*
251  * send the WR to the adapter
252  */
253  err = vq_send_wr(c2dev, (union c2wr *) wr);
254  if (err) {
255  vq_req_put(c2dev, vq_req);
256  goto bail1;
257  }
258 
259  /*
260  * wait for reply from adapter
261  */
262  err = vq_wait_for_reply(c2dev, vq_req);
263  if (err) {
264  goto bail1;
265  }
266 
267  /*
268  * process reply
269  */
270  reply =
271  (struct c2wr_nsmr_register_rep *) (unsigned long) (vq_req->reply_msg);
272  if (!reply) {
273  err = -ENOMEM;
274  goto bail1;
275  }
276  if ((err = c2_errno(reply))) {
277  goto bail2;
278  }
279  //*p_pb_entries = be32_to_cpu(reply->pbl_depth);
280  mr->ibmr.lkey = mr->ibmr.rkey = be32_to_cpu(reply->stag_index);
281  vq_repbuf_free(c2dev, reply);
282 
283  /*
284  * if there are still more PBEs we need to send them to
285  * the adapter and wait for a reply on the final one.
286  * reuse vq_req for this purpose.
287  */
288  pbl_depth -= count;
289  if (pbl_depth) {
290 
291  vq_req->reply_msg = (unsigned long) NULL;
292  atomic_set(&vq_req->reply_ready, 0);
293  err = send_pbl_messages(c2dev,
294  cpu_to_be32(mr->ibmr.lkey),
295  (unsigned long) &addr_list[i],
296  pbl_depth, vq_req, PBL_PHYS);
297  if (err) {
298  goto bail1;
299  }
300  }
301 
302  vq_req_free(c2dev, vq_req);
303  kfree(wr);
304 
305  return err;
306 
307  bail2:
308  vq_repbuf_free(c2dev, reply);
309  bail1:
310  kfree(wr);
311  bail0:
312  vq_req_free(c2dev, vq_req);
313  return err;
314 }
315 
317 {
318  struct c2_vq_req *vq_req; /* verbs request object */
319  struct c2wr_stag_dealloc_req wr; /* work request */
320  struct c2wr_stag_dealloc_rep *reply; /* WR reply */
321  int err;
322 
323 
324  /*
325  * allocate verbs request object
326  */
327  vq_req = vq_req_alloc(c2dev);
328  if (!vq_req) {
329  return -ENOMEM;
330  }
331 
332  /*
333  * Build the WR
334  */
335  c2_wr_set_id(&wr, CCWR_STAG_DEALLOC);
336  wr.hdr.context = (u64) (unsigned long) vq_req;
337  wr.rnic_handle = c2dev->adapter_handle;
338  wr.stag_index = cpu_to_be32(stag_index);
339 
340  /*
341  * reference the request struct. dereferenced in the int handler.
342  */
343  vq_req_get(c2dev, vq_req);
344 
345  /*
346  * Send WR to adapter
347  */
348  err = vq_send_wr(c2dev, (union c2wr *) & wr);
349  if (err) {
350  vq_req_put(c2dev, vq_req);
351  goto bail0;
352  }
353 
354  /*
355  * Wait for reply from adapter
356  */
357  err = vq_wait_for_reply(c2dev, vq_req);
358  if (err) {
359  goto bail0;
360  }
361 
362  /*
363  * Process reply
364  */
365  reply = (struct c2wr_stag_dealloc_rep *) (unsigned long) vq_req->reply_msg;
366  if (!reply) {
367  err = -ENOMEM;
368  goto bail0;
369  }
370 
371  err = c2_errno(reply);
372 
373  vq_repbuf_free(c2dev, reply);
374  bail0:
375  vq_req_free(c2dev, vq_req);
376  return err;
377 }