2 * Copyright (c) 2006, 2007 Cisco Systems, Inc. All rights reserved.
3 * Copyright (c) 2005 Mellanox Technologies. All rights reserved.
5 * This software is available to you under a choice of one of two
6 * licenses. You may choose to be licensed under the terms of the GNU
7 * General Public License (GPL) Version 2, available from the file
8 * COPYING in the main directory of this source tree, or the
9 * OpenIB.org BSD license below:
11 * Redistribution and use in source and binary forms, with or
12 * without modification, are permitted provided that the following
15 * - Redistributions of source code must retain the above
16 * copyright notice, this list of conditions and the following
19 * - Redistributions in binary form must reproduce the above
20 * copyright notice, this list of conditions and the following
21 * disclaimer in the documentation and/or other materials
22 * provided with the distribution.
24 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
25 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
26 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
27 * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
28 * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
29 * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
30 * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
34 #include <linux/errno.h>
35 #include <linux/export.h>
36 #include <linux/io-mapping.h>
44 MLX4_NUM_RESERVED_UARS
= 8
47 int mlx4_pd_alloc(struct mlx4_dev
*dev
, u32
*pdn
)
49 struct mlx4_priv
*priv
= mlx4_priv(dev
);
51 *pdn
= mlx4_bitmap_alloc(&priv
->pd_bitmap
);
57 EXPORT_SYMBOL_GPL(mlx4_pd_alloc
);
59 void mlx4_pd_free(struct mlx4_dev
*dev
, u32 pdn
)
61 mlx4_bitmap_free(&mlx4_priv(dev
)->pd_bitmap
, pdn
, MLX4_USE_RR
);
63 EXPORT_SYMBOL_GPL(mlx4_pd_free
);
65 int __mlx4_xrcd_alloc(struct mlx4_dev
*dev
, u32
*xrcdn
)
67 struct mlx4_priv
*priv
= mlx4_priv(dev
);
69 *xrcdn
= mlx4_bitmap_alloc(&priv
->xrcd_bitmap
);
76 int mlx4_xrcd_alloc(struct mlx4_dev
*dev
, u32
*xrcdn
)
81 if (mlx4_is_mfunc(dev
)) {
82 err
= mlx4_cmd_imm(dev
, 0, &out_param
,
83 RES_XRCD
, RES_OP_RESERVE
,
85 MLX4_CMD_TIME_CLASS_A
, MLX4_CMD_WRAPPED
);
89 *xrcdn
= get_param_l(&out_param
);
92 return __mlx4_xrcd_alloc(dev
, xrcdn
);
94 EXPORT_SYMBOL_GPL(mlx4_xrcd_alloc
);
96 void __mlx4_xrcd_free(struct mlx4_dev
*dev
, u32 xrcdn
)
98 mlx4_bitmap_free(&mlx4_priv(dev
)->xrcd_bitmap
, xrcdn
, MLX4_USE_RR
);
101 void mlx4_xrcd_free(struct mlx4_dev
*dev
, u32 xrcdn
)
106 if (mlx4_is_mfunc(dev
)) {
107 set_param_l(&in_param
, xrcdn
);
108 err
= mlx4_cmd(dev
, in_param
, RES_XRCD
,
109 RES_OP_RESERVE
, MLX4_CMD_FREE_RES
,
110 MLX4_CMD_TIME_CLASS_A
, MLX4_CMD_WRAPPED
);
112 mlx4_warn(dev
, "Failed to release xrcdn %d\n", xrcdn
);
114 __mlx4_xrcd_free(dev
, xrcdn
);
116 EXPORT_SYMBOL_GPL(mlx4_xrcd_free
);
118 int mlx4_init_pd_table(struct mlx4_dev
*dev
)
120 struct mlx4_priv
*priv
= mlx4_priv(dev
);
122 return mlx4_bitmap_init(&priv
->pd_bitmap
, dev
->caps
.num_pds
,
123 (1 << NOT_MASKED_PD_BITS
) - 1,
124 dev
->caps
.reserved_pds
, 0);
127 void mlx4_cleanup_pd_table(struct mlx4_dev
*dev
)
129 mlx4_bitmap_cleanup(&mlx4_priv(dev
)->pd_bitmap
);
132 int mlx4_init_xrcd_table(struct mlx4_dev
*dev
)
134 struct mlx4_priv
*priv
= mlx4_priv(dev
);
136 return mlx4_bitmap_init(&priv
->xrcd_bitmap
, (1 << 16),
137 (1 << 16) - 1, dev
->caps
.reserved_xrcds
+ 1, 0);
140 void mlx4_cleanup_xrcd_table(struct mlx4_dev
*dev
)
142 mlx4_bitmap_cleanup(&mlx4_priv(dev
)->xrcd_bitmap
);
145 int mlx4_uar_alloc(struct mlx4_dev
*dev
, struct mlx4_uar
*uar
)
149 uar
->index
= mlx4_bitmap_alloc(&mlx4_priv(dev
)->uar_table
.bitmap
);
150 if (uar
->index
== -1)
153 if (mlx4_is_slave(dev
))
154 offset
= uar
->index
% ((int)pci_resource_len(dev
->persist
->pdev
,
156 dev
->caps
.uar_page_size
);
159 uar
->pfn
= (pci_resource_start(dev
->persist
->pdev
, 2) >> PAGE_SHIFT
)
164 EXPORT_SYMBOL_GPL(mlx4_uar_alloc
);
166 void mlx4_uar_free(struct mlx4_dev
*dev
, struct mlx4_uar
*uar
)
168 mlx4_bitmap_free(&mlx4_priv(dev
)->uar_table
.bitmap
, uar
->index
, MLX4_USE_RR
);
170 EXPORT_SYMBOL_GPL(mlx4_uar_free
);
172 int mlx4_bf_alloc(struct mlx4_dev
*dev
, struct mlx4_bf
*bf
, int node
)
174 struct mlx4_priv
*priv
= mlx4_priv(dev
);
175 struct mlx4_uar
*uar
;
179 if (!priv
->bf_mapping
)
182 mutex_lock(&priv
->bf_mutex
);
183 if (!list_empty(&priv
->bf_list
))
184 uar
= list_entry(priv
->bf_list
.next
, struct mlx4_uar
, bf_list
);
186 if (mlx4_bitmap_avail(&priv
->uar_table
.bitmap
) < MLX4_NUM_RESERVED_UARS
) {
190 uar
= kmalloc_node(sizeof(*uar
), GFP_KERNEL
, node
);
192 uar
= kmalloc(sizeof(*uar
), GFP_KERNEL
);
198 err
= mlx4_uar_alloc(dev
, uar
);
202 uar
->map
= ioremap(uar
->pfn
<< PAGE_SHIFT
, PAGE_SIZE
);
208 uar
->bf_map
= io_mapping_map_wc(priv
->bf_mapping
,
209 uar
->index
<< PAGE_SHIFT
,
215 uar
->free_bf_bmap
= 0;
216 list_add(&uar
->bf_list
, &priv
->bf_list
);
219 idx
= ffz(uar
->free_bf_bmap
);
220 uar
->free_bf_bmap
|= 1 << idx
;
223 bf
->buf_size
= dev
->caps
.bf_reg_size
/ 2;
224 bf
->reg
= uar
->bf_map
+ idx
* dev
->caps
.bf_reg_size
;
225 if (uar
->free_bf_bmap
== (1 << dev
->caps
.bf_regs_per_page
) - 1)
226 list_del_init(&uar
->bf_list
);
235 mlx4_uar_free(dev
, uar
);
241 mutex_unlock(&priv
->bf_mutex
);
244 EXPORT_SYMBOL_GPL(mlx4_bf_alloc
);
246 void mlx4_bf_free(struct mlx4_dev
*dev
, struct mlx4_bf
*bf
)
248 struct mlx4_priv
*priv
= mlx4_priv(dev
);
251 if (!bf
->uar
|| !bf
->uar
->bf_map
)
254 mutex_lock(&priv
->bf_mutex
);
255 idx
= (bf
->reg
- bf
->uar
->bf_map
) / dev
->caps
.bf_reg_size
;
256 bf
->uar
->free_bf_bmap
&= ~(1 << idx
);
257 if (!bf
->uar
->free_bf_bmap
) {
258 if (!list_empty(&bf
->uar
->bf_list
))
259 list_del(&bf
->uar
->bf_list
);
261 io_mapping_unmap(bf
->uar
->bf_map
);
262 iounmap(bf
->uar
->map
);
263 mlx4_uar_free(dev
, bf
->uar
);
265 } else if (list_empty(&bf
->uar
->bf_list
))
266 list_add(&bf
->uar
->bf_list
, &priv
->bf_list
);
268 mutex_unlock(&priv
->bf_mutex
);
270 EXPORT_SYMBOL_GPL(mlx4_bf_free
);
272 int mlx4_init_uar_table(struct mlx4_dev
*dev
)
274 int num_reserved_uar
= mlx4_get_num_reserved_uar(dev
);
276 mlx4_dbg(dev
, "uar_page_shift = %d", dev
->uar_page_shift
);
277 mlx4_dbg(dev
, "Effective reserved_uars=%d", dev
->caps
.reserved_uars
);
279 if (dev
->caps
.num_uars
<= num_reserved_uar
) {
281 dev
, "Only %d UAR pages (need more than %d)\n",
282 dev
->caps
.num_uars
, num_reserved_uar
);
283 mlx4_err(dev
, "Increase firmware log2_uar_bar_megabytes?\n");
287 return mlx4_bitmap_init(&mlx4_priv(dev
)->uar_table
.bitmap
,
288 dev
->caps
.num_uars
, dev
->caps
.num_uars
- 1,
289 dev
->caps
.reserved_uars
, 0);
292 void mlx4_cleanup_uar_table(struct mlx4_dev
*dev
)
294 mlx4_bitmap_cleanup(&mlx4_priv(dev
)->uar_table
.bitmap
);