1 // SPDX-License-Identifier: GPL-2.0
2 #include <linux/delay.h>
4 #include "nitrox_dev.h"
5 #include "nitrox_csr.h"
8 #define MAX_CSR_RETRIES 10
11 * emu_enable_cores - Enable EMU cluster cores.
12 * @ndev: NITROX device
14 static void emu_enable_cores(struct nitrox_device
*ndev
)
16 union emu_se_enable emu_se
;
17 union emu_ae_enable emu_ae
;
20 /* AE cores 20 per cluster */
22 emu_ae
.s
.enable
= 0xfffff;
24 /* SE cores 16 per cluster */
26 emu_se
.s
.enable
= 0xffff;
28 /* enable per cluster cores */
29 for (i
= 0; i
< NR_CLUSTERS
; i
++) {
30 nitrox_write_csr(ndev
, EMU_AE_ENABLEX(i
), emu_ae
.value
);
31 nitrox_write_csr(ndev
, EMU_SE_ENABLEX(i
), emu_se
.value
);
36 * nitrox_config_emu_unit - configure EMU unit.
37 * @ndev: NITROX device
39 void nitrox_config_emu_unit(struct nitrox_device
*ndev
)
41 union emu_wd_int_ena_w1s emu_wd_int
;
42 union emu_ge_int_ena_w1s emu_ge_int
;
47 emu_enable_cores(ndev
);
49 /* enable general error and watch dog interrupts */
51 emu_ge_int
.s
.se_ge
= 0xffff;
52 emu_ge_int
.s
.ae_ge
= 0xfffff;
54 emu_wd_int
.s
.se_wd
= 1;
56 for (i
= 0; i
< NR_CLUSTERS
; i
++) {
57 offset
= EMU_WD_INT_ENA_W1SX(i
);
58 nitrox_write_csr(ndev
, offset
, emu_wd_int
.value
);
59 offset
= EMU_GE_INT_ENA_W1SX(i
);
60 nitrox_write_csr(ndev
, offset
, emu_ge_int
.value
);
64 static void reset_pkt_input_ring(struct nitrox_device
*ndev
, int ring
)
66 union nps_pkt_in_instr_ctl pkt_in_ctl
;
67 union nps_pkt_in_done_cnts pkt_in_cnts
;
68 int max_retries
= MAX_CSR_RETRIES
;
71 /* step 1: disable the ring, clear enable bit */
72 offset
= NPS_PKT_IN_INSTR_CTLX(ring
);
73 pkt_in_ctl
.value
= nitrox_read_csr(ndev
, offset
);
75 nitrox_write_csr(ndev
, offset
, pkt_in_ctl
.value
);
77 /* step 2: wait to clear [ENB] */
78 usleep_range(100, 150);
80 pkt_in_ctl
.value
= nitrox_read_csr(ndev
, offset
);
81 if (!pkt_in_ctl
.s
.enb
)
84 } while (max_retries
--);
86 /* step 3: clear done counts */
87 offset
= NPS_PKT_IN_DONE_CNTSX(ring
);
88 pkt_in_cnts
.value
= nitrox_read_csr(ndev
, offset
);
89 nitrox_write_csr(ndev
, offset
, pkt_in_cnts
.value
);
90 usleep_range(50, 100);
93 void enable_pkt_input_ring(struct nitrox_device
*ndev
, int ring
)
95 union nps_pkt_in_instr_ctl pkt_in_ctl
;
96 int max_retries
= MAX_CSR_RETRIES
;
99 /* 64-byte instruction size */
100 offset
= NPS_PKT_IN_INSTR_CTLX(ring
);
101 pkt_in_ctl
.value
= nitrox_read_csr(ndev
, offset
);
102 pkt_in_ctl
.s
.is64b
= 1;
103 pkt_in_ctl
.s
.enb
= 1;
104 nitrox_write_csr(ndev
, offset
, pkt_in_ctl
.value
);
106 /* wait for set [ENB] */
108 pkt_in_ctl
.value
= nitrox_read_csr(ndev
, offset
);
109 if (pkt_in_ctl
.s
.enb
)
112 } while (max_retries
--);
116 * nitrox_config_pkt_input_rings - configure Packet Input Rings
117 * @ndev: NITROX device
119 void nitrox_config_pkt_input_rings(struct nitrox_device
*ndev
)
123 for (i
= 0; i
< ndev
->nr_queues
; i
++) {
124 struct nitrox_cmdq
*cmdq
= &ndev
->pkt_inq
[i
];
125 union nps_pkt_in_instr_rsize pkt_in_rsize
;
126 union nps_pkt_in_instr_baoff_dbell pkt_in_dbell
;
129 reset_pkt_input_ring(ndev
, i
);
133 * configure ring base address 16-byte aligned,
134 * size and interrupt threshold.
136 offset
= NPS_PKT_IN_INSTR_BADDRX(i
);
137 nitrox_write_csr(ndev
, offset
, cmdq
->dma
);
139 /* configure ring size */
140 offset
= NPS_PKT_IN_INSTR_RSIZEX(i
);
141 pkt_in_rsize
.value
= 0;
142 pkt_in_rsize
.s
.rsize
= ndev
->qlen
;
143 nitrox_write_csr(ndev
, offset
, pkt_in_rsize
.value
);
145 /* set high threshold for pkt input ring interrupts */
146 offset
= NPS_PKT_IN_INT_LEVELSX(i
);
147 nitrox_write_csr(ndev
, offset
, 0xffffffff);
149 /* step 5: clear off door bell counts */
150 offset
= NPS_PKT_IN_INSTR_BAOFF_DBELLX(i
);
151 pkt_in_dbell
.value
= 0;
152 pkt_in_dbell
.s
.dbell
= 0xffffffff;
153 nitrox_write_csr(ndev
, offset
, pkt_in_dbell
.value
);
155 /* enable the ring */
156 enable_pkt_input_ring(ndev
, i
);
160 static void reset_pkt_solicit_port(struct nitrox_device
*ndev
, int port
)
162 union nps_pkt_slc_ctl pkt_slc_ctl
;
163 union nps_pkt_slc_cnts pkt_slc_cnts
;
164 int max_retries
= MAX_CSR_RETRIES
;
167 /* step 1: disable slc port */
168 offset
= NPS_PKT_SLC_CTLX(port
);
169 pkt_slc_ctl
.value
= nitrox_read_csr(ndev
, offset
);
170 pkt_slc_ctl
.s
.enb
= 0;
171 nitrox_write_csr(ndev
, offset
, pkt_slc_ctl
.value
);
174 usleep_range(100, 150);
175 /* wait to clear [ENB] */
177 pkt_slc_ctl
.value
= nitrox_read_csr(ndev
, offset
);
178 if (!pkt_slc_ctl
.s
.enb
)
181 } while (max_retries
--);
183 /* step 3: clear slc counters */
184 offset
= NPS_PKT_SLC_CNTSX(port
);
185 pkt_slc_cnts
.value
= nitrox_read_csr(ndev
, offset
);
186 nitrox_write_csr(ndev
, offset
, pkt_slc_cnts
.value
);
187 usleep_range(50, 100);
190 void enable_pkt_solicit_port(struct nitrox_device
*ndev
, int port
)
192 union nps_pkt_slc_ctl pkt_slc_ctl
;
193 int max_retries
= MAX_CSR_RETRIES
;
196 offset
= NPS_PKT_SLC_CTLX(port
);
197 pkt_slc_ctl
.value
= 0;
198 pkt_slc_ctl
.s
.enb
= 1;
200 * 8 trailing 0x00 bytes will be added
201 * to the end of the outgoing packet.
204 /* enable response header */
205 pkt_slc_ctl
.s
.rh
= 1;
206 nitrox_write_csr(ndev
, offset
, pkt_slc_ctl
.value
);
208 /* wait to set [ENB] */
210 pkt_slc_ctl
.value
= nitrox_read_csr(ndev
, offset
);
211 if (pkt_slc_ctl
.s
.enb
)
214 } while (max_retries
--);
217 static void config_pkt_solicit_port(struct nitrox_device
*ndev
, int port
)
219 union nps_pkt_slc_int_levels pkt_slc_int
;
222 reset_pkt_solicit_port(ndev
, port
);
224 /* step 4: configure interrupt levels */
225 offset
= NPS_PKT_SLC_INT_LEVELSX(port
);
226 pkt_slc_int
.value
= 0;
227 /* time interrupt threshold */
228 pkt_slc_int
.s
.timet
= 0x3fffff;
229 nitrox_write_csr(ndev
, offset
, pkt_slc_int
.value
);
231 /* enable the solicit port */
232 enable_pkt_solicit_port(ndev
, port
);
235 void nitrox_config_pkt_solicit_ports(struct nitrox_device
*ndev
)
239 for (i
= 0; i
< ndev
->nr_queues
; i
++)
240 config_pkt_solicit_port(ndev
, i
);
244 * enable_nps_core_interrupts - enable NPS core interrutps
245 * @ndev: NITROX device.
247 * This includes NPS core interrupts.
249 static void enable_nps_core_interrupts(struct nitrox_device
*ndev
)
251 union nps_core_int_ena_w1s core_int
;
253 /* NPS core interrutps */
255 core_int
.s
.host_wr_err
= 1;
256 core_int
.s
.host_wr_timeout
= 1;
257 core_int
.s
.exec_wr_timeout
= 1;
258 core_int
.s
.npco_dma_malform
= 1;
259 core_int
.s
.host_nps_wr_err
= 1;
260 nitrox_write_csr(ndev
, NPS_CORE_INT_ENA_W1S
, core_int
.value
);
263 void nitrox_config_nps_core_unit(struct nitrox_device
*ndev
)
265 union nps_core_gbl_vfcfg core_gbl_vfcfg
;
267 /* endian control information */
268 nitrox_write_csr(ndev
, NPS_CORE_CONTROL
, 1ULL);
270 /* disable ILK interface */
271 core_gbl_vfcfg
.value
= 0;
272 core_gbl_vfcfg
.s
.ilk_disable
= 1;
273 core_gbl_vfcfg
.s
.cfg
= __NDEV_MODE_PF
;
274 nitrox_write_csr(ndev
, NPS_CORE_GBL_VFCFG
, core_gbl_vfcfg
.value
);
276 /* enable nps core interrupts */
277 enable_nps_core_interrupts(ndev
);
281 * enable_nps_pkt_interrupts - enable NPS packet interrutps
282 * @ndev: NITROX device.
284 * This includes NPS packet in and slc interrupts.
286 static void enable_nps_pkt_interrupts(struct nitrox_device
*ndev
)
288 /* NPS packet in ring interrupts */
289 nitrox_write_csr(ndev
, NPS_PKT_IN_RERR_LO_ENA_W1S
, (~0ULL));
290 nitrox_write_csr(ndev
, NPS_PKT_IN_RERR_HI_ENA_W1S
, (~0ULL));
291 nitrox_write_csr(ndev
, NPS_PKT_IN_ERR_TYPE_ENA_W1S
, (~0ULL));
292 /* NPS packet slc port interrupts */
293 nitrox_write_csr(ndev
, NPS_PKT_SLC_RERR_HI_ENA_W1S
, (~0ULL));
294 nitrox_write_csr(ndev
, NPS_PKT_SLC_RERR_LO_ENA_W1S
, (~0ULL));
295 nitrox_write_csr(ndev
, NPS_PKT_SLC_ERR_TYPE_ENA_W1S
, (~0uLL));
298 void nitrox_config_nps_pkt_unit(struct nitrox_device
*ndev
)
300 /* config input and solicit ports */
301 nitrox_config_pkt_input_rings(ndev
);
302 nitrox_config_pkt_solicit_ports(ndev
);
304 /* enable nps packet interrupts */
305 enable_nps_pkt_interrupts(ndev
);
308 static void reset_aqm_ring(struct nitrox_device
*ndev
, int ring
)
310 union aqmq_en aqmq_en_reg
;
311 union aqmq_activity_stat activity_stat
;
312 union aqmq_cmp_cnt cmp_cnt
;
313 int max_retries
= MAX_CSR_RETRIES
;
316 /* step 1: disable the queue */
317 offset
= AQMQ_ENX(ring
);
318 aqmq_en_reg
.value
= 0;
319 aqmq_en_reg
.queue_enable
= 0;
320 nitrox_write_csr(ndev
, offset
, aqmq_en_reg
.value
);
322 /* step 2: wait for AQMQ_ACTIVITY_STATX[QUEUE_ACTIVE] to clear */
323 usleep_range(100, 150);
324 offset
= AQMQ_ACTIVITY_STATX(ring
);
326 activity_stat
.value
= nitrox_read_csr(ndev
, offset
);
327 if (!activity_stat
.queue_active
)
330 } while (max_retries
--);
332 /* step 3: clear commands completed count */
333 offset
= AQMQ_CMP_CNTX(ring
);
334 cmp_cnt
.value
= nitrox_read_csr(ndev
, offset
);
335 nitrox_write_csr(ndev
, offset
, cmp_cnt
.value
);
336 usleep_range(50, 100);
339 void enable_aqm_ring(struct nitrox_device
*ndev
, int ring
)
341 union aqmq_en aqmq_en_reg
;
344 offset
= AQMQ_ENX(ring
);
345 aqmq_en_reg
.value
= 0;
346 aqmq_en_reg
.queue_enable
= 1;
347 nitrox_write_csr(ndev
, offset
, aqmq_en_reg
.value
);
348 usleep_range(50, 100);
351 void nitrox_config_aqm_rings(struct nitrox_device
*ndev
)
355 for (ring
= 0; ring
< ndev
->nr_queues
; ring
++) {
356 struct nitrox_cmdq
*cmdq
= ndev
->aqmq
[ring
];
357 union aqmq_drbl drbl
;
358 union aqmq_qsz qsize
;
359 union aqmq_cmp_thr cmp_thr
;
363 reset_aqm_ring(ndev
, ring
);
365 /* step 4: clear doorbell count of ring */
366 offset
= AQMQ_DRBLX(ring
);
368 drbl
.dbell_count
= 0xFFFFFFFF;
369 nitrox_write_csr(ndev
, offset
, drbl
.value
);
371 /* step 5: configure host ring details */
373 /* set host address for next command of ring */
374 offset
= AQMQ_NXT_CMDX(ring
);
375 nitrox_write_csr(ndev
, offset
, 0ULL);
377 /* set host address of ring base */
378 offset
= AQMQ_BADRX(ring
);
379 nitrox_write_csr(ndev
, offset
, cmdq
->dma
);
382 offset
= AQMQ_QSZX(ring
);
384 qsize
.host_queue_size
= ndev
->qlen
;
385 nitrox_write_csr(ndev
, offset
, qsize
.value
);
387 /* set command completion threshold */
388 offset
= AQMQ_CMP_THRX(ring
);
390 cmp_thr
.commands_completed_threshold
= 1;
391 nitrox_write_csr(ndev
, offset
, cmp_thr
.value
);
393 /* step 6: enable the queue */
394 enable_aqm_ring(ndev
, ring
);
398 static void enable_aqm_interrupts(struct nitrox_device
*ndev
)
400 /* clear interrupt enable bits */
401 nitrox_write_csr(ndev
, AQM_DBELL_OVF_LO_ENA_W1S
, (~0ULL));
402 nitrox_write_csr(ndev
, AQM_DBELL_OVF_HI_ENA_W1S
, (~0ULL));
403 nitrox_write_csr(ndev
, AQM_DMA_RD_ERR_LO_ENA_W1S
, (~0ULL));
404 nitrox_write_csr(ndev
, AQM_DMA_RD_ERR_HI_ENA_W1S
, (~0ULL));
405 nitrox_write_csr(ndev
, AQM_EXEC_NA_LO_ENA_W1S
, (~0ULL));
406 nitrox_write_csr(ndev
, AQM_EXEC_NA_HI_ENA_W1S
, (~0ULL));
407 nitrox_write_csr(ndev
, AQM_EXEC_ERR_LO_ENA_W1S
, (~0ULL));
408 nitrox_write_csr(ndev
, AQM_EXEC_ERR_HI_ENA_W1S
, (~0ULL));
411 void nitrox_config_aqm_unit(struct nitrox_device
*ndev
)
413 /* config aqm command queues */
414 nitrox_config_aqm_rings(ndev
);
416 /* enable aqm interrupts */
417 enable_aqm_interrupts(ndev
);
420 void nitrox_config_pom_unit(struct nitrox_device
*ndev
)
422 union pom_int_ena_w1s pom_int
;
425 /* enable pom interrupts */
427 pom_int
.s
.illegal_dport
= 1;
428 nitrox_write_csr(ndev
, POM_INT_ENA_W1S
, pom_int
.value
);
430 /* enable perf counters */
431 for (i
= 0; i
< ndev
->hw
.se_cores
; i
++)
432 nitrox_write_csr(ndev
, POM_PERF_CTL
, BIT_ULL(i
));
436 * nitrox_config_rand_unit - enable NITROX random number unit
437 * @ndev: NITROX device
439 void nitrox_config_rand_unit(struct nitrox_device
*ndev
)
441 union efl_rnm_ctl_status efl_rnm_ctl
;
444 offset
= EFL_RNM_CTL_STATUS
;
445 efl_rnm_ctl
.value
= nitrox_read_csr(ndev
, offset
);
446 efl_rnm_ctl
.s
.ent_en
= 1;
447 efl_rnm_ctl
.s
.rng_en
= 1;
448 nitrox_write_csr(ndev
, offset
, efl_rnm_ctl
.value
);
451 void nitrox_config_efl_unit(struct nitrox_device
*ndev
)
455 for (i
= 0; i
< NR_CLUSTERS
; i
++) {
456 union efl_core_int_ena_w1s efl_core_int
;
459 /* EFL core interrupts */
460 offset
= EFL_CORE_INT_ENA_W1SX(i
);
461 efl_core_int
.value
= 0;
462 efl_core_int
.s
.len_ovr
= 1;
463 efl_core_int
.s
.d_left
= 1;
464 efl_core_int
.s
.epci_decode_err
= 1;
465 nitrox_write_csr(ndev
, offset
, efl_core_int
.value
);
467 offset
= EFL_CORE_VF_ERR_INT0_ENA_W1SX(i
);
468 nitrox_write_csr(ndev
, offset
, (~0ULL));
469 offset
= EFL_CORE_VF_ERR_INT1_ENA_W1SX(i
);
470 nitrox_write_csr(ndev
, offset
, (~0ULL));
474 void nitrox_config_bmi_unit(struct nitrox_device
*ndev
)
476 union bmi_ctl bmi_ctl
;
477 union bmi_int_ena_w1s bmi_int_ena
;
480 /* no threshold limits for PCIe */
482 bmi_ctl
.value
= nitrox_read_csr(ndev
, offset
);
483 bmi_ctl
.s
.max_pkt_len
= 0xff;
484 bmi_ctl
.s
.nps_free_thrsh
= 0xff;
485 bmi_ctl
.s
.nps_hdrq_thrsh
= 0x7a;
486 nitrox_write_csr(ndev
, offset
, bmi_ctl
.value
);
488 /* enable interrupts */
489 offset
= BMI_INT_ENA_W1S
;
490 bmi_int_ena
.value
= 0;
491 bmi_int_ena
.s
.max_len_err_nps
= 1;
492 bmi_int_ena
.s
.pkt_rcv_err_nps
= 1;
493 bmi_int_ena
.s
.fpf_undrrn
= 1;
494 nitrox_write_csr(ndev
, offset
, bmi_int_ena
.value
);
497 void nitrox_config_bmo_unit(struct nitrox_device
*ndev
)
499 union bmo_ctl2 bmo_ctl2
;
502 /* no threshold limits for PCIe */
504 bmo_ctl2
.value
= nitrox_read_csr(ndev
, offset
);
505 bmo_ctl2
.s
.nps_slc_buf_thrsh
= 0xff;
506 nitrox_write_csr(ndev
, offset
, bmo_ctl2
.value
);
509 void invalidate_lbc(struct nitrox_device
*ndev
)
511 union lbc_inval_ctl lbc_ctl
;
512 union lbc_inval_status lbc_stat
;
513 int max_retries
= MAX_CSR_RETRIES
;
517 offset
= LBC_INVAL_CTL
;
518 lbc_ctl
.value
= nitrox_read_csr(ndev
, offset
);
519 lbc_ctl
.s
.cam_inval_start
= 1;
520 nitrox_write_csr(ndev
, offset
, lbc_ctl
.value
);
522 offset
= LBC_INVAL_STATUS
;
524 lbc_stat
.value
= nitrox_read_csr(ndev
, offset
);
528 } while (max_retries
--);
531 void nitrox_config_lbc_unit(struct nitrox_device
*ndev
)
533 union lbc_int_ena_w1s lbc_int_ena
;
536 invalidate_lbc(ndev
);
538 /* enable interrupts */
539 offset
= LBC_INT_ENA_W1S
;
540 lbc_int_ena
.value
= 0;
541 lbc_int_ena
.s
.dma_rd_err
= 1;
542 lbc_int_ena
.s
.over_fetch_err
= 1;
543 lbc_int_ena
.s
.cam_inval_abort
= 1;
544 lbc_int_ena
.s
.cam_hard_err
= 1;
545 nitrox_write_csr(ndev
, offset
, lbc_int_ena
.value
);
547 offset
= LBC_PLM_VF1_64_INT_ENA_W1S
;
548 nitrox_write_csr(ndev
, offset
, (~0ULL));
549 offset
= LBC_PLM_VF65_128_INT_ENA_W1S
;
550 nitrox_write_csr(ndev
, offset
, (~0ULL));
552 offset
= LBC_ELM_VF1_64_INT_ENA_W1S
;
553 nitrox_write_csr(ndev
, offset
, (~0ULL));
554 offset
= LBC_ELM_VF65_128_INT_ENA_W1S
;
555 nitrox_write_csr(ndev
, offset
, (~0ULL));
558 void config_nps_core_vfcfg_mode(struct nitrox_device
*ndev
, enum vf_mode mode
)
560 union nps_core_gbl_vfcfg vfcfg
;
562 vfcfg
.value
= nitrox_read_csr(ndev
, NPS_CORE_GBL_VFCFG
);
563 vfcfg
.s
.cfg
= mode
& 0x7;
565 nitrox_write_csr(ndev
, NPS_CORE_GBL_VFCFG
, vfcfg
.value
);
568 static const char *get_core_option(u8 se_cores
, u8 ae_cores
)
570 const char *option
= "";
572 if (ae_cores
== AE_MAX_CORES
) {
581 } else if (ae_cores
== (AE_MAX_CORES
/ 2)) {
590 static const char *get_feature_option(u8 zip_cores
, int core_freq
)
594 else if (zip_cores
< ZIP_MAX_CORES
)
597 if (core_freq
>= 850)
599 else if (core_freq
>= 750)
601 else if (core_freq
>= 550)
607 void nitrox_get_hwinfo(struct nitrox_device
*ndev
)
609 union emu_fuse_map emu_fuse
;
610 union rst_boot rst_boot
;
611 union fus_dat1 fus_dat1
;
612 unsigned char name
[IFNAMSIZ
* 2] = {};
616 /* get core frequency */
618 rst_boot
.value
= nitrox_read_csr(ndev
, offset
);
619 ndev
->hw
.freq
= (rst_boot
.pnr_mul
+ 3) * PLL_REF_CLK
;
621 for (i
= 0; i
< NR_CLUSTERS
; i
++) {
622 offset
= EMU_FUSE_MAPX(i
);
623 emu_fuse
.value
= nitrox_read_csr(ndev
, offset
);
624 if (emu_fuse
.s
.valid
) {
625 dead_cores
= hweight32(emu_fuse
.s
.ae_fuse
);
626 ndev
->hw
.ae_cores
+= AE_CORES_PER_CLUSTER
- dead_cores
;
627 dead_cores
= hweight16(emu_fuse
.s
.se_fuse
);
628 ndev
->hw
.se_cores
+= SE_CORES_PER_CLUSTER
- dead_cores
;
631 /* find zip hardware availability */
633 fus_dat1
.value
= nitrox_read_csr(ndev
, offset
);
634 if (!fus_dat1
.nozip
) {
635 dead_cores
= hweight8(fus_dat1
.zip_info
);
636 ndev
->hw
.zip_cores
= ZIP_MAX_CORES
- dead_cores
;
639 /* determine the partname
640 * CNN55<core option>-<freq><pincount>-<feature option>-<rev>
642 snprintf(name
, sizeof(name
), "CNN55%s-%3dBG676%s-1.%u",
643 get_core_option(ndev
->hw
.se_cores
, ndev
->hw
.ae_cores
),
645 get_feature_option(ndev
->hw
.zip_cores
, ndev
->hw
.freq
),
646 ndev
->hw
.revision_id
);
649 strncpy(ndev
->hw
.partname
, name
, sizeof(ndev
->hw
.partname
));
652 void enable_pf2vf_mbox_interrupts(struct nitrox_device
*ndev
)
657 /* Mailbox interrupt low enable set register */
658 reg_addr
= NPS_PKT_MBOX_INT_LO_ENA_W1S
;
659 nitrox_write_csr(ndev
, reg_addr
, value
);
661 /* Mailbox interrupt high enable set register */
662 reg_addr
= NPS_PKT_MBOX_INT_HI_ENA_W1S
;
663 nitrox_write_csr(ndev
, reg_addr
, value
);
666 void disable_pf2vf_mbox_interrupts(struct nitrox_device
*ndev
)
671 /* Mailbox interrupt low enable clear register */
672 reg_addr
= NPS_PKT_MBOX_INT_LO_ENA_W1C
;
673 nitrox_write_csr(ndev
, reg_addr
, value
);
675 /* Mailbox interrupt high enable clear register */
676 reg_addr
= NPS_PKT_MBOX_INT_HI_ENA_W1C
;
677 nitrox_write_csr(ndev
, reg_addr
, value
);