2 * Copyright (c) 2013-2015, Mellanox Technologies. All rights reserved.
4 * This software is available to you under a choice of one of two
5 * licenses. You may choose to be licensed under the terms of the GNU
6 * General Public License (GPL) Version 2, available from the file
7 * COPYING in the main directory of this source tree, or the
8 * OpenIB.org BSD license below:
10 * Redistribution and use in source and binary forms, with or
11 * without modification, are permitted provided that the following
14 * - Redistributions of source code must retain the above
15 * copyright notice, this list of conditions and the following
18 * - Redistributions in binary form must reproduce the above
19 * copyright notice, this list of conditions and the following
20 * disclaimer in the documentation and/or other materials
21 * provided with the distribution.
23 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
24 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
25 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
26 * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
27 * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
28 * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
29 * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
33 #ifndef MLX5_CORE_CQ_H
34 #define MLX5_CORE_CQ_H
36 #include <rdma/ib_verbs.h>
37 #include <linux/mlx5/driver.h>
38 #include <linux/refcount.h>
45 struct mlx5_uars_page
*uar
;
47 struct completion free
;
50 void (*comp
) (struct mlx5_core_cq
*);
51 void (*event
) (struct mlx5_core_cq
*, enum mlx5_event
);
54 struct mlx5_rsc_debug
*dbg
;
57 struct list_head list
;
58 void (*comp
)(struct mlx5_core_cq
*);
61 int reset_notify_added
;
62 struct list_head reset_notify
;
67 MLX5_CQE_SYNDROME_LOCAL_LENGTH_ERR
= 0x01,
68 MLX5_CQE_SYNDROME_LOCAL_QP_OP_ERR
= 0x02,
69 MLX5_CQE_SYNDROME_LOCAL_PROT_ERR
= 0x04,
70 MLX5_CQE_SYNDROME_WR_FLUSH_ERR
= 0x05,
71 MLX5_CQE_SYNDROME_MW_BIND_ERR
= 0x06,
72 MLX5_CQE_SYNDROME_BAD_RESP_ERR
= 0x10,
73 MLX5_CQE_SYNDROME_LOCAL_ACCESS_ERR
= 0x11,
74 MLX5_CQE_SYNDROME_REMOTE_INVAL_REQ_ERR
= 0x12,
75 MLX5_CQE_SYNDROME_REMOTE_ACCESS_ERR
= 0x13,
76 MLX5_CQE_SYNDROME_REMOTE_OP_ERR
= 0x14,
77 MLX5_CQE_SYNDROME_TRANSPORT_RETRY_EXC_ERR
= 0x15,
78 MLX5_CQE_SYNDROME_RNR_RETRY_EXC_ERR
= 0x16,
79 MLX5_CQE_SYNDROME_REMOTE_ABORTED_ERR
= 0x22,
83 MLX5_CQE_OWNER_MASK
= 1,
85 MLX5_CQE_RESP_WR_IMM
= 1,
86 MLX5_CQE_RESP_SEND
= 2,
87 MLX5_CQE_RESP_SEND_IMM
= 3,
88 MLX5_CQE_RESP_SEND_INV
= 4,
89 MLX5_CQE_RESIZE_CQ
= 5,
90 MLX5_CQE_SIG_ERR
= 12,
91 MLX5_CQE_REQ_ERR
= 13,
92 MLX5_CQE_RESP_ERR
= 14,
93 MLX5_CQE_INVALID
= 15,
97 MLX5_CQ_MODIFY_PERIOD
= 1 << 0,
98 MLX5_CQ_MODIFY_COUNT
= 1 << 1,
99 MLX5_CQ_MODIFY_OVERRUN
= 1 << 2,
103 MLX5_CQ_OPMOD_RESIZE
= 1,
104 MLX5_MODIFY_CQ_MASK_LOG_SIZE
= 1 << 0,
105 MLX5_MODIFY_CQ_MASK_PG_OFFSET
= 1 << 1,
106 MLX5_MODIFY_CQ_MASK_PG_SIZE
= 1 << 2,
109 struct mlx5_cq_modify_params
{
128 CQE_SIZE_128_PAD
= 2,
131 #define MLX5_MAX_CQ_PERIOD (BIT(__mlx5_bit_sz(cqc, cq_period)) - 1)
132 #define MLX5_MAX_CQ_COUNT (BIT(__mlx5_bit_sz(cqc, cq_max_count)) - 1)
134 static inline int cqe_sz_to_mlx_sz(u8 size
, int padding_128_en
)
136 return padding_128_en
? CQE_SIZE_128_PAD
:
137 size
== 64 ? CQE_SIZE_64
: CQE_SIZE_128
;
140 static inline void mlx5_cq_set_ci(struct mlx5_core_cq
*cq
)
142 *cq
->set_ci_db
= cpu_to_be32(cq
->cons_index
& 0xffffff);
146 MLX5_CQ_DB_REQ_NOT_SOL
= 1 << 24,
147 MLX5_CQ_DB_REQ_NOT
= 0 << 24
150 static inline void mlx5_cq_arm(struct mlx5_core_cq
*cq
, u32 cmd
,
151 void __iomem
*uar_page
,
159 ci
= cons_index
& 0xffffff;
161 *cq
->arm_db
= cpu_to_be32(sn
<< 28 | cmd
| ci
);
163 /* Make sure that the doorbell record in host memory is
164 * written before ringing the doorbell via PCI MMIO.
168 doorbell
[0] = cpu_to_be32(sn
<< 28 | cmd
| ci
);
169 doorbell
[1] = cpu_to_be32(cq
->cqn
);
171 mlx5_write64(doorbell
, uar_page
+ MLX5_CQ_DOORBELL
, NULL
);
174 int mlx5_init_cq_table(struct mlx5_core_dev
*dev
);
175 void mlx5_cleanup_cq_table(struct mlx5_core_dev
*dev
);
176 int mlx5_core_create_cq(struct mlx5_core_dev
*dev
, struct mlx5_core_cq
*cq
,
178 int mlx5_core_destroy_cq(struct mlx5_core_dev
*dev
, struct mlx5_core_cq
*cq
);
179 int mlx5_core_query_cq(struct mlx5_core_dev
*dev
, struct mlx5_core_cq
*cq
,
180 u32
*out
, int outlen
);
181 int mlx5_core_modify_cq(struct mlx5_core_dev
*dev
, struct mlx5_core_cq
*cq
,
183 int mlx5_core_modify_cq_moderation(struct mlx5_core_dev
*dev
,
184 struct mlx5_core_cq
*cq
, u16 cq_period
,
186 int mlx5_debug_cq_add(struct mlx5_core_dev
*dev
, struct mlx5_core_cq
*cq
);
187 void mlx5_debug_cq_remove(struct mlx5_core_dev
*dev
, struct mlx5_core_cq
*cq
);
189 #endif /* MLX5_CORE_CQ_H */