1 /* SPDX-License-Identifier: ((GPL-2.0 WITH Linux-syscall-note) OR BSD-2-Clause) */
3 * Copyright (c) 2012-2016 VMware, Inc. All rights reserved.
5 * This program is free software; you can redistribute it and/or
6 * modify it under the terms of EITHER the GNU General Public License
7 * version 2 as published by the Free Software Foundation or the BSD
8 * 2-Clause License. This program is distributed in the hope that it
9 * will be useful, but WITHOUT ANY WARRANTY; WITHOUT EVEN THE IMPLIED
10 * WARRANTY OF MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE.
11 * See the GNU General Public License version 2 for more details at
12 * http://www.gnu.org/licenses/old-licenses/gpl-2.0.en.html.
14 * You should have received a copy of the GNU General Public License
15 * along with this program available in the file COPYING in the main
16 * directory of this source tree.
18 * The BSD 2-Clause License
20 * Redistribution and use in source and binary forms, with or
21 * without modification, are permitted provided that the following
24 * - Redistributions of source code must retain the above
25 * copyright notice, this list of conditions and the following
28 * - Redistributions in binary form must reproduce the above
29 * copyright notice, this list of conditions and the following
30 * disclaimer in the documentation and/or other materials
31 * provided with the distribution.
33 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
34 * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
35 * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS
36 * FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE
37 * COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT,
38 * INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
39 * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
40 * SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
41 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT,
42 * STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
43 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED
44 * OF THE POSSIBILITY OF SUCH DAMAGE.
47 #ifndef __VMW_PVRDMA_ABI_H__
48 #define __VMW_PVRDMA_ABI_H__
50 #include <linux/types.h>
52 #define PVRDMA_UVERBS_ABI_VERSION 3 /* ABI Version. */
53 #define PVRDMA_UAR_HANDLE_MASK 0x00FFFFFF /* Bottom 24 bits. */
54 #define PVRDMA_UAR_QP_OFFSET 0 /* QP doorbell. */
55 #define PVRDMA_UAR_QP_SEND (1 << 30) /* Send bit. */
56 #define PVRDMA_UAR_QP_RECV (1 << 31) /* Recv bit. */
57 #define PVRDMA_UAR_CQ_OFFSET 4 /* CQ doorbell. */
58 #define PVRDMA_UAR_CQ_ARM_SOL (1 << 29) /* Arm solicited bit. */
59 #define PVRDMA_UAR_CQ_ARM (1 << 30) /* Arm bit. */
60 #define PVRDMA_UAR_CQ_POLL (1 << 31) /* Poll bit. */
61 #define PVRDMA_UAR_SRQ_OFFSET 8 /* SRQ doorbell. */
62 #define PVRDMA_UAR_SRQ_RECV (1 << 30) /* Recv bit. */
64 enum pvrdma_wr_opcode
{
66 PVRDMA_WR_RDMA_WRITE_WITH_IMM
,
68 PVRDMA_WR_SEND_WITH_IMM
,
70 PVRDMA_WR_ATOMIC_CMP_AND_SWP
,
71 PVRDMA_WR_ATOMIC_FETCH_AND_ADD
,
73 PVRDMA_WR_SEND_WITH_INV
,
74 PVRDMA_WR_RDMA_READ_WITH_INV
,
76 PVRDMA_WR_FAST_REG_MR
,
77 PVRDMA_WR_MASKED_ATOMIC_CMP_AND_SWP
,
78 PVRDMA_WR_MASKED_ATOMIC_FETCH_AND_ADD
,
84 enum pvrdma_wc_status
{
86 PVRDMA_WC_LOC_LEN_ERR
,
87 PVRDMA_WC_LOC_QP_OP_ERR
,
88 PVRDMA_WC_LOC_EEC_OP_ERR
,
89 PVRDMA_WC_LOC_PROT_ERR
,
90 PVRDMA_WC_WR_FLUSH_ERR
,
91 PVRDMA_WC_MW_BIND_ERR
,
92 PVRDMA_WC_BAD_RESP_ERR
,
93 PVRDMA_WC_LOC_ACCESS_ERR
,
94 PVRDMA_WC_REM_INV_REQ_ERR
,
95 PVRDMA_WC_REM_ACCESS_ERR
,
97 PVRDMA_WC_RETRY_EXC_ERR
,
98 PVRDMA_WC_RNR_RETRY_EXC_ERR
,
99 PVRDMA_WC_LOC_RDD_VIOL_ERR
,
100 PVRDMA_WC_REM_INV_RD_REQ_ERR
,
101 PVRDMA_WC_REM_ABORT_ERR
,
102 PVRDMA_WC_INV_EECN_ERR
,
103 PVRDMA_WC_INV_EEC_STATE_ERR
,
105 PVRDMA_WC_RESP_TIMEOUT_ERR
,
106 PVRDMA_WC_GENERAL_ERR
,
109 enum pvrdma_wc_opcode
{
111 PVRDMA_WC_RDMA_WRITE
,
118 PVRDMA_WC_FAST_REG_MR
,
119 PVRDMA_WC_MASKED_COMP_SWAP
,
120 PVRDMA_WC_MASKED_FETCH_ADD
,
121 PVRDMA_WC_RECV
= 1 << 7,
122 PVRDMA_WC_RECV_RDMA_WITH_IMM
,
125 enum pvrdma_wc_flags
{
126 PVRDMA_WC_GRH
= 1 << 0,
127 PVRDMA_WC_WITH_IMM
= 1 << 1,
128 PVRDMA_WC_WITH_INVALIDATE
= 1 << 2,
129 PVRDMA_WC_IP_CSUM_OK
= 1 << 3,
130 PVRDMA_WC_WITH_SMAC
= 1 << 4,
131 PVRDMA_WC_WITH_VLAN
= 1 << 5,
132 PVRDMA_WC_WITH_NETWORK_HDR_TYPE
= 1 << 6,
133 PVRDMA_WC_FLAGS_MAX
= PVRDMA_WC_WITH_NETWORK_HDR_TYPE
,
136 struct pvrdma_alloc_ucontext_resp
{
141 struct pvrdma_alloc_pd_resp
{
146 struct pvrdma_create_cq
{
147 __aligned_u64 buf_addr
;
152 struct pvrdma_create_cq_resp
{
157 struct pvrdma_resize_cq
{
158 __aligned_u64 buf_addr
;
163 struct pvrdma_create_srq
{
164 __aligned_u64 buf_addr
;
169 struct pvrdma_create_srq_resp
{
174 struct pvrdma_create_qp
{
175 __aligned_u64 rbuf_addr
;
176 __aligned_u64 sbuf_addr
;
179 __aligned_u64 qp_addr
;
182 struct pvrdma_create_qp_resp
{
187 /* PVRDMA masked atomic compare and swap */
188 struct pvrdma_ex_cmp_swap
{
189 __aligned_u64 swap_val
;
190 __aligned_u64 compare_val
;
191 __aligned_u64 swap_mask
;
192 __aligned_u64 compare_mask
;
195 /* PVRDMA masked atomic fetch and add */
196 struct pvrdma_ex_fetch_add
{
197 __aligned_u64 add_val
;
198 __aligned_u64 field_boundary
;
201 /* PVRDMA address vector. */
204 __u32 sl_tclass_flowlabel
;
214 /* PVRDMA scatter/gather entry */
221 /* PVRDMA receive queue work request */
222 struct pvrdma_rq_wqe_hdr
{
223 __aligned_u64 wr_id
; /* wr id */
224 __u32 num_sge
; /* size of s/g array */
225 __u32 total_len
; /* reserved */
227 /* Use pvrdma_sge (ib_sge) for receive queue s/g array elements. */
229 /* PVRDMA send queue work request */
230 struct pvrdma_sq_wqe_hdr
{
231 __aligned_u64 wr_id
; /* wr id */
232 __u32 num_sge
; /* size of s/g array */
233 __u32 total_len
; /* reserved */
234 __u32 opcode
; /* operation type */
235 __u32 send_flags
; /* wr flags */
238 __u32 invalidate_rkey
;
243 __aligned_u64 remote_addr
;
248 __aligned_u64 remote_addr
;
249 __aligned_u64 compare_add
;
255 __aligned_u64 remote_addr
;
259 struct pvrdma_ex_cmp_swap cmp_swap
;
260 struct pvrdma_ex_fetch_add fetch_add
;
264 __aligned_u64 iova_start
;
265 __aligned_u64 pl_pdir_dma
;
280 /* Use pvrdma_sge (ib_sge) for send queue s/g array elements. */
282 /* Completion queue element. */
299 __u8 network_hdr_type
;
300 __u8 reserved2
[6]; /* Pad to next power of 2 (64). */
303 #endif /* __VMW_PVRDMA_ABI_H__ */