1 /***********************license start***************
2 * Author: Cavium Networks
4 * Contact: support@caviumnetworks.com
5 * This file is part of the OCTEON SDK
7 * Copyright (c) 2003-2008 Cavium Networks
9 * This file is free software; you can redistribute it and/or modify
10 * it under the terms of the GNU General Public License, Version 2, as
11 * published by the Free Software Foundation.
13 * This file is distributed in the hope that it will be useful, but
14 * AS-IS and WITHOUT ANY WARRANTY; without even the implied warranty
15 * of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE, TITLE, or
16 * NONINFRINGEMENT. See the GNU General Public License for more
19 * You should have received a copy of the GNU General Public License
20 * along with this file; if not, write to the Free Software
21 * Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA
22 * or visit http://www.gnu.org/licenses/.
24 * This file may also be available under a different license from Cavium.
25 * Contact Cavium Networks for more information
26 ***********************license end**************************************/
29 * Support functions for managing command queues used for
30 * various hardware blocks.
33 #include <linux/kernel.h>
35 #include <asm/octeon/octeon.h>
37 #include <asm/octeon/cvmx-config.h>
38 #include <asm/octeon/cvmx-fpa.h>
39 #include <asm/octeon/cvmx-cmd-queue.h>
41 #include <asm/octeon/cvmx-npei-defs.h>
42 #include <asm/octeon/cvmx-pexp-defs.h>
43 #include <asm/octeon/cvmx-pko-defs.h>
46 * This application uses this pointer to access the global queue
47 * state. It points to a bootmem named block.
49 __cvmx_cmd_queue_all_state_t
*__cvmx_cmd_queue_state_ptr
;
50 EXPORT_SYMBOL_GPL(__cvmx_cmd_queue_state_ptr
);
53 * Initialize the Global queue state pointer.
55 * Returns CVMX_CMD_QUEUE_SUCCESS or a failure code
57 static cvmx_cmd_queue_result_t
__cvmx_cmd_queue_init_state_ptr(void)
59 char *alloc_name
= "cvmx_cmd_queues";
60 #if defined(CONFIG_CAVIUM_RESERVE32) && CONFIG_CAVIUM_RESERVE32
61 extern uint64_t octeon_reserve32_memory
;
64 if (likely(__cvmx_cmd_queue_state_ptr
))
65 return CVMX_CMD_QUEUE_SUCCESS
;
67 #if defined(CONFIG_CAVIUM_RESERVE32) && CONFIG_CAVIUM_RESERVE32
68 if (octeon_reserve32_memory
)
69 __cvmx_cmd_queue_state_ptr
=
70 cvmx_bootmem_alloc_named_range(sizeof(*__cvmx_cmd_queue_state_ptr
),
71 octeon_reserve32_memory
,
72 octeon_reserve32_memory
+
73 (CONFIG_CAVIUM_RESERVE32
<<
74 20) - 1, 128, alloc_name
);
77 __cvmx_cmd_queue_state_ptr
=
78 cvmx_bootmem_alloc_named(sizeof(*__cvmx_cmd_queue_state_ptr
),
81 if (__cvmx_cmd_queue_state_ptr
)
82 memset(__cvmx_cmd_queue_state_ptr
, 0,
83 sizeof(*__cvmx_cmd_queue_state_ptr
));
85 struct cvmx_bootmem_named_block_desc
*block_desc
=
86 cvmx_bootmem_find_named_block(alloc_name
);
88 __cvmx_cmd_queue_state_ptr
=
89 cvmx_phys_to_ptr(block_desc
->base_addr
);
92 ("ERROR: cvmx_cmd_queue_initialize: Unable to get named block %s.\n",
94 return CVMX_CMD_QUEUE_NO_MEMORY
;
97 return CVMX_CMD_QUEUE_SUCCESS
;
101 * Initialize a command queue for use. The initial FPA buffer is
102 * allocated and the hardware unit is configured to point to the
105 * @queue_id: Hardware command queue to initialize.
106 * @max_depth: Maximum outstanding commands that can be queued.
107 * @fpa_pool: FPA pool the command queues should come from.
108 * @pool_size: Size of each buffer in the FPA pool (bytes)
110 * Returns CVMX_CMD_QUEUE_SUCCESS or a failure code
112 cvmx_cmd_queue_result_t
cvmx_cmd_queue_initialize(cvmx_cmd_queue_id_t queue_id
,
113 int max_depth
, int fpa_pool
,
116 __cvmx_cmd_queue_state_t
*qstate
;
117 cvmx_cmd_queue_result_t result
= __cvmx_cmd_queue_init_state_ptr();
118 if (result
!= CVMX_CMD_QUEUE_SUCCESS
)
121 qstate
= __cvmx_cmd_queue_get_state(queue_id
);
123 return CVMX_CMD_QUEUE_INVALID_PARAM
;
126 * We artificially limit max_depth to 1<<20 words. It is an
129 if (CVMX_CMD_QUEUE_ENABLE_MAX_DEPTH
) {
130 if ((max_depth
< 0) || (max_depth
> 1 << 20))
131 return CVMX_CMD_QUEUE_INVALID_PARAM
;
132 } else if (max_depth
!= 0)
133 return CVMX_CMD_QUEUE_INVALID_PARAM
;
135 if ((fpa_pool
< 0) || (fpa_pool
> 7))
136 return CVMX_CMD_QUEUE_INVALID_PARAM
;
137 if ((pool_size
< 128) || (pool_size
> 65536))
138 return CVMX_CMD_QUEUE_INVALID_PARAM
;
140 /* See if someone else has already initialized the queue */
141 if (qstate
->base_ptr_div128
) {
142 if (max_depth
!= (int)qstate
->max_depth
) {
143 cvmx_dprintf("ERROR: cvmx_cmd_queue_initialize: "
144 "Queue already initialized with different "
146 (int)qstate
->max_depth
);
147 return CVMX_CMD_QUEUE_INVALID_PARAM
;
149 if (fpa_pool
!= qstate
->fpa_pool
) {
150 cvmx_dprintf("ERROR: cvmx_cmd_queue_initialize: "
151 "Queue already initialized with different "
154 return CVMX_CMD_QUEUE_INVALID_PARAM
;
156 if ((pool_size
>> 3) - 1 != qstate
->pool_size_m1
) {
157 cvmx_dprintf("ERROR: cvmx_cmd_queue_initialize: "
158 "Queue already initialized with different "
159 "FPA pool size (%u).\n",
160 (qstate
->pool_size_m1
+ 1) << 3);
161 return CVMX_CMD_QUEUE_INVALID_PARAM
;
164 return CVMX_CMD_QUEUE_ALREADY_SETUP
;
166 union cvmx_fpa_ctl_status status
;
169 status
.u64
= cvmx_read_csr(CVMX_FPA_CTL_STATUS
);
171 cvmx_dprintf("ERROR: cvmx_cmd_queue_initialize: "
172 "FPA is not enabled.\n");
173 return CVMX_CMD_QUEUE_NO_MEMORY
;
175 buffer
= cvmx_fpa_alloc(fpa_pool
);
176 if (buffer
== NULL
) {
177 cvmx_dprintf("ERROR: cvmx_cmd_queue_initialize: "
178 "Unable to allocate initial buffer.\n");
179 return CVMX_CMD_QUEUE_NO_MEMORY
;
182 memset(qstate
, 0, sizeof(*qstate
));
183 qstate
->max_depth
= max_depth
;
184 qstate
->fpa_pool
= fpa_pool
;
185 qstate
->pool_size_m1
= (pool_size
>> 3) - 1;
186 qstate
->base_ptr_div128
= cvmx_ptr_to_phys(buffer
) / 128;
188 * We zeroed the now serving field so we need to also
191 __cvmx_cmd_queue_state_ptr
->
192 ticket
[__cvmx_cmd_queue_get_index(queue_id
)] = 0;
194 return CVMX_CMD_QUEUE_SUCCESS
;
199 * Shutdown a queue a free it's command buffers to the FPA. The
200 * hardware connected to the queue must be stopped before this
201 * function is called.
203 * @queue_id: Queue to shutdown
205 * Returns CVMX_CMD_QUEUE_SUCCESS or a failure code
207 cvmx_cmd_queue_result_t
cvmx_cmd_queue_shutdown(cvmx_cmd_queue_id_t queue_id
)
209 __cvmx_cmd_queue_state_t
*qptr
= __cvmx_cmd_queue_get_state(queue_id
);
211 cvmx_dprintf("ERROR: cvmx_cmd_queue_shutdown: Unable to "
212 "get queue information.\n");
213 return CVMX_CMD_QUEUE_INVALID_PARAM
;
216 if (cvmx_cmd_queue_length(queue_id
) > 0) {
217 cvmx_dprintf("ERROR: cvmx_cmd_queue_shutdown: Queue still "
218 "has data in it.\n");
219 return CVMX_CMD_QUEUE_FULL
;
222 __cvmx_cmd_queue_lock(queue_id
, qptr
);
223 if (qptr
->base_ptr_div128
) {
224 cvmx_fpa_free(cvmx_phys_to_ptr
225 ((uint64_t) qptr
->base_ptr_div128
<< 7),
227 qptr
->base_ptr_div128
= 0;
229 __cvmx_cmd_queue_unlock(qptr
);
231 return CVMX_CMD_QUEUE_SUCCESS
;
235 * Return the number of command words pending in the queue. This
236 * function may be relatively slow for some hardware units.
238 * @queue_id: Hardware command queue to query
240 * Returns Number of outstanding commands
242 int cvmx_cmd_queue_length(cvmx_cmd_queue_id_t queue_id
)
244 if (CVMX_ENABLE_PARAMETER_CHECKING
) {
245 if (__cvmx_cmd_queue_get_state(queue_id
) == NULL
)
246 return CVMX_CMD_QUEUE_INVALID_PARAM
;
250 * The cast is here so gcc with check that all values in the
251 * cvmx_cmd_queue_id_t enumeration are here.
253 switch ((cvmx_cmd_queue_id_t
) (queue_id
& 0xff0000)) {
254 case CVMX_CMD_QUEUE_PKO_BASE
:
256 * FIXME: Need atomic lock on
257 * CVMX_PKO_REG_READ_IDX. Right now we are normally
258 * called with the queue lock, so that is a SLIGHT
259 * amount of protection.
261 cvmx_write_csr(CVMX_PKO_REG_READ_IDX
, queue_id
& 0xffff);
262 if (OCTEON_IS_MODEL(OCTEON_CN3XXX
)) {
263 union cvmx_pko_mem_debug9 debug9
;
264 debug9
.u64
= cvmx_read_csr(CVMX_PKO_MEM_DEBUG9
);
265 return debug9
.cn38xx
.doorbell
;
267 union cvmx_pko_mem_debug8 debug8
;
268 debug8
.u64
= cvmx_read_csr(CVMX_PKO_MEM_DEBUG8
);
269 return debug8
.cn50xx
.doorbell
;
271 case CVMX_CMD_QUEUE_ZIP
:
272 case CVMX_CMD_QUEUE_DFA
:
273 case CVMX_CMD_QUEUE_RAID
:
274 /* FIXME: Implement other lengths */
276 case CVMX_CMD_QUEUE_DMA_BASE
:
278 union cvmx_npei_dmax_counts dmax_counts
;
280 cvmx_read_csr(CVMX_PEXP_NPEI_DMAX_COUNTS
282 return dmax_counts
.s
.dbell
;
284 case CVMX_CMD_QUEUE_END
:
285 return CVMX_CMD_QUEUE_INVALID_PARAM
;
287 return CVMX_CMD_QUEUE_INVALID_PARAM
;
291 * Return the command buffer to be written to. The purpose of this
292 * function is to allow CVMX routine access t othe low level buffer
293 * for initial hardware setup. User applications should not call this
296 * @queue_id: Command queue to query
298 * Returns Command buffer or NULL on failure
300 void *cvmx_cmd_queue_buffer(cvmx_cmd_queue_id_t queue_id
)
302 __cvmx_cmd_queue_state_t
*qptr
= __cvmx_cmd_queue_get_state(queue_id
);
303 if (qptr
&& qptr
->base_ptr_div128
)
304 return cvmx_phys_to_ptr((uint64_t) qptr
->base_ptr_div128
<< 7);