1 /* SPDX-License-Identifier: GPL-2.0 */
5 #include <asm/byteorder.h>
6 #include <linux/types.h>
10 #define AE_CORES_PER_CLUSTER 20
11 #define SE_CORES_PER_CLUSTER 16
14 #define EMU_BIST_STATUSX(_i) (0x1402700 + ((_i) * 0x40000))
15 #define UCD_BIST_STATUS 0x12C0070
16 #define NPS_CORE_BIST_REG 0x10000E8
17 #define NPS_CORE_NPC_BIST_REG 0x1000128
18 #define NPS_PKT_SLC_BIST_REG 0x1040088
19 #define NPS_PKT_IN_BIST_REG 0x1040100
20 #define POM_BIST_REG 0x11C0100
21 #define BMI_BIST_REG 0x1140080
22 #define EFL_CORE_BIST_REGX(_i) (0x1240100 + ((_i) * 0x400))
23 #define EFL_TOP_BIST_STAT 0x1241090
24 #define BMO_BIST_REG 0x1180080
25 #define LBC_BIST_STATUS 0x1200020
26 #define PEM_BIST_STATUSX(_i) (0x1080468 | ((_i) << 18))
29 #define EMU_SE_ENABLEX(_i) (0x1400000 + ((_i) * 0x40000))
30 #define EMU_AE_ENABLEX(_i) (0x1400008 + ((_i) * 0x40000))
31 #define EMU_WD_INT_ENA_W1SX(_i) (0x1402318 + ((_i) * 0x40000))
32 #define EMU_GE_INT_ENA_W1SX(_i) (0x1402518 + ((_i) * 0x40000))
33 #define EMU_FUSE_MAPX(_i) (0x1402708 + ((_i) * 0x40000))
36 #define UCD_UCODE_LOAD_BLOCK_NUM 0x12C0010
37 #define UCD_UCODE_LOAD_IDX_DATAX(_i) (0x12C0018 + ((_i) * 0x20))
38 #define UCD_SE_EID_UCODE_BLOCK_NUMX(_i) (0x12C0000 + ((_i) * 0x1000))
40 /* NPS core registers */
41 #define NPS_CORE_GBL_VFCFG 0x1000000
42 #define NPS_CORE_CONTROL 0x1000008
43 #define NPS_CORE_INT_ACTIVE 0x1000080
44 #define NPS_CORE_INT 0x10000A0
45 #define NPS_CORE_INT_ENA_W1S 0x10000B8
46 #define NPS_STATS_PKT_DMA_RD_CNT 0x1000180
47 #define NPS_STATS_PKT_DMA_WR_CNT 0x1000190
49 /* NPS packet registers */
50 #define NPS_PKT_INT 0x1040018
51 #define NPS_PKT_IN_RERR_HI 0x1040108
52 #define NPS_PKT_IN_RERR_HI_ENA_W1S 0x1040120
53 #define NPS_PKT_IN_RERR_LO 0x1040128
54 #define NPS_PKT_IN_RERR_LO_ENA_W1S 0x1040140
55 #define NPS_PKT_IN_ERR_TYPE 0x1040148
56 #define NPS_PKT_IN_ERR_TYPE_ENA_W1S 0x1040160
57 #define NPS_PKT_IN_INSTR_CTLX(_i) (0x10060 + ((_i) * 0x40000))
58 #define NPS_PKT_IN_INSTR_BADDRX(_i) (0x10068 + ((_i) * 0x40000))
59 #define NPS_PKT_IN_INSTR_RSIZEX(_i) (0x10070 + ((_i) * 0x40000))
60 #define NPS_PKT_IN_DONE_CNTSX(_i) (0x10080 + ((_i) * 0x40000))
61 #define NPS_PKT_IN_INSTR_BAOFF_DBELLX(_i) (0x10078 + ((_i) * 0x40000))
62 #define NPS_PKT_IN_INT_LEVELSX(_i) (0x10088 + ((_i) * 0x40000))
64 #define NPS_PKT_SLC_RERR_HI 0x1040208
65 #define NPS_PKT_SLC_RERR_HI_ENA_W1S 0x1040220
66 #define NPS_PKT_SLC_RERR_LO 0x1040228
67 #define NPS_PKT_SLC_RERR_LO_ENA_W1S 0x1040240
68 #define NPS_PKT_SLC_ERR_TYPE 0x1040248
69 #define NPS_PKT_SLC_ERR_TYPE_ENA_W1S 0x1040260
70 #define NPS_PKT_SLC_CTLX(_i) (0x10000 + ((_i) * 0x40000))
71 #define NPS_PKT_SLC_CNTSX(_i) (0x10008 + ((_i) * 0x40000))
72 #define NPS_PKT_SLC_INT_LEVELSX(_i) (0x10010 + ((_i) * 0x40000))
75 #define POM_INT_ENA_W1S 0x11C0018
76 #define POM_GRP_EXECMASKX(_i) (0x11C1100 | ((_i) * 8))
77 #define POM_INT 0x11C0000
78 #define POM_PERF_CTL 0x11CC400
81 #define BMI_INT 0x1140000
82 #define BMI_CTL 0x1140020
83 #define BMI_INT_ENA_W1S 0x1140018
84 #define BMI_NPS_PKT_CNT 0x1140070
87 #define EFL_CORE_INT_ENA_W1SX(_i) (0x1240018 + ((_i) * 0x400))
88 #define EFL_CORE_VF_ERR_INT0X(_i) (0x1240050 + ((_i) * 0x400))
89 #define EFL_CORE_VF_ERR_INT0_ENA_W1SX(_i) (0x1240068 + ((_i) * 0x400))
90 #define EFL_CORE_VF_ERR_INT1X(_i) (0x1240070 + ((_i) * 0x400))
91 #define EFL_CORE_VF_ERR_INT1_ENA_W1SX(_i) (0x1240088 + ((_i) * 0x400))
92 #define EFL_CORE_SE_ERR_INTX(_i) (0x12400A0 + ((_i) * 0x400))
93 #define EFL_RNM_CTL_STATUS 0x1241800
94 #define EFL_CORE_INTX(_i) (0x1240000 + ((_i) * 0x400))
97 #define BMO_CTL2 0x1180028
98 #define BMO_NPS_SLC_PKT_CNT 0x1180078
101 #define LBC_INT 0x1200000
102 #define LBC_INVAL_CTL 0x1201010
103 #define LBC_PLM_VF1_64_INT 0x1202008
104 #define LBC_INVAL_STATUS 0x1202010
105 #define LBC_INT_ENA_W1S 0x1203000
106 #define LBC_PLM_VF1_64_INT_ENA_W1S 0x1205008
107 #define LBC_PLM_VF65_128_INT 0x1206008
108 #define LBC_ELM_VF1_64_INT 0x1208000
109 #define LBC_PLM_VF65_128_INT_ENA_W1S 0x1209008
110 #define LBC_ELM_VF1_64_INT_ENA_W1S 0x120B000
111 #define LBC_ELM_VF65_128_INT 0x120C000
112 #define LBC_ELM_VF65_128_INT_ENA_W1S 0x120F000
115 #define PEM0_INT 0x1080428
118 * struct emu_fuse_map - EMU Fuse Map Registers
119 * @ae_fuse: Fuse settings for AE 19..0
120 * @se_fuse: Fuse settings for SE 15..0
122 * A set bit indicates the unit is fuse disabled.
127 #if (defined(__BIG_ENDIAN_BITFIELD))
144 * struct emu_se_enable - Symmetric Engine Enable Registers
145 * @enable: Individual enables for each of the clusters
146 * 16 symmetric engines.
148 union emu_se_enable
{
151 #if (defined(__BIG_ENDIAN_BITFIELD))
162 * struct emu_ae_enable - EMU Asymmetric engines.
163 * @enable: Individual enables for each of the cluster's
164 * 20 Asymmetric Engines.
166 union emu_ae_enable
{
169 #if (defined(__BIG_ENDIAN_BITFIELD))
180 * struct emu_wd_int_ena_w1s - EMU Interrupt Enable Registers
181 * @ae_wd: Reads or sets enable for EMU(0..3)_WD_INT[AE_WD]
182 * @se_wd: Reads or sets enable for EMU(0..3)_WD_INT[SE_WD]
184 union emu_wd_int_ena_w1s
{
187 #if (defined(__BIG_ENDIAN_BITFIELD))
202 * struct emu_ge_int_ena_w1s - EMU Interrupt Enable set registers
203 * @ae_ge: Reads or sets enable for EMU(0..3)_GE_INT[AE_GE]
204 * @se_ge: Reads or sets enable for EMU(0..3)_GE_INT[SE_GE]
206 union emu_ge_int_ena_w1s
{
209 #if (defined(__BIG_ENDIAN_BITFIELD))
224 * struct nps_pkt_slc_ctl - Solicited Packet Out Control Registers
225 * @rh: Indicates whether to remove or include the response header
226 * 1 = Include, 0 = Remove
227 * @z: If set, 8 trailing 0x00 bytes will be added to the end of the
229 * @enb: Enable for this port.
231 union nps_pkt_slc_ctl
{
234 #if defined(__BIG_ENDIAN_BITFIELD)
249 * struct nps_pkt_slc_cnts - Solicited Packet Out Count Registers
250 * @slc_int: Returns a 1 when:
251 * NPS_PKT_SLC(i)_CNTS[CNT] > NPS_PKT_SLC(i)_INT_LEVELS[CNT], or
252 * NPS_PKT_SLC(i)_CNTS[TIMER] > NPS_PKT_SLC(i)_INT_LEVELS[TIMET].
253 * To clear the bit, the CNTS register must be written to clear.
254 * @in_int: Returns a 1 when:
255 * NPS_PKT_IN(i)_DONE_CNTS[CNT] > NPS_PKT_IN(i)_INT_LEVELS[CNT].
256 * To clear the bit, the DONE_CNTS register must be written to clear.
257 * @mbox_int: Returns a 1 when:
258 * NPS_PKT_MBOX_PF_VF(i)_INT[INTR] is set. To clear the bit,
259 * write NPS_PKT_MBOX_PF_VF(i)_INT[INTR] with 1.
260 * @timer: Timer, incremented every 2048 coprocessor clock cycles
261 * when [CNT] is not zero. The hardware clears both [TIMER] and
262 * [INT] when [CNT] goes to 0.
263 * @cnt: Packet counter. Hardware adds to [CNT] as it sends packets out.
264 * On a write to this CSR, hardware subtracts the amount written to the
265 * [CNT] field from [CNT].
267 union nps_pkt_slc_cnts
{
270 #if defined(__BIG_ENDIAN_BITFIELD)
293 * struct nps_pkt_slc_int_levels - Solicited Packet Out Interrupt Levels
295 * @bmode: Determines whether NPS_PKT_SLC_CNTS[CNT] is a byte or
297 * @timet: Output port counter time interrupt threshold.
298 * @cnt: Output port counter interrupt threshold.
300 union nps_pkt_slc_int_levels
{
303 #if defined(__BIG_ENDIAN_BITFIELD)
318 * struct nps_pkt_inst - NPS Packet Interrupt Register
319 * @in_err: Set when any NPS_PKT_IN_RERR_HI/LO bit and
320 * corresponding NPS_PKT_IN_RERR_*_ENA_* bit are bot set.
321 * @uns_err: Set when any NSP_PKT_UNS_RERR_HI/LO bit and
322 * corresponding NPS_PKT_UNS_RERR_*_ENA_* bit are both set.
323 * @slc_er: Set when any NSP_PKT_SLC_RERR_HI/LO bit and
324 * corresponding NPS_PKT_SLC_RERR_*_ENA_* bit are both set.
329 #if defined(__BIG_ENDIAN_BITFIELD)
358 * struct nps_pkt_in_done_cnts - Input instruction ring counts registers
359 * @slc_cnt: Returns a 1 when:
360 * NPS_PKT_SLC(i)_CNTS[CNT] > NPS_PKT_SLC(i)_INT_LEVELS[CNT], or
361 * NPS_PKT_SLC(i)_CNTS[TIMER] > NPS_PKT_SCL(i)_INT_LEVELS[TIMET]
362 * To clear the bit, the CNTS register must be
363 * written to clear the underlying condition
364 * @uns_int: Return a 1 when:
365 * NPS_PKT_UNS(i)_CNTS[CNT] > NPS_PKT_UNS(i)_INT_LEVELS[CNT], or
366 * NPS_PKT_UNS(i)_CNTS[TIMER] > NPS_PKT_UNS(i)_INT_LEVELS[TIMET]
367 * To clear the bit, the CNTS register must be
368 * written to clear the underlying condition
369 * @in_int: Returns a 1 when:
370 * NPS_PKT_IN(i)_DONE_CNTS[CNT] > NPS_PKT_IN(i)_INT_LEVELS[CNT]
371 * To clear the bit, the DONE_CNTS register
372 * must be written to clear the underlying condition
373 * @mbox_int: Returns a 1 when:
374 * NPS_PKT_MBOX_PF_VF(i)_INT[INTR] is set.
375 * To clear the bit, write NPS_PKT_MBOX_PF_VF(i)_INT[INTR]
377 * @resend: A write of 1 will resend an MSI-X interrupt message if any
378 * of the following conditions are true for this ring "i".
379 * NPS_PKT_SLC(i)_CNTS[CNT] > NPS_PKT_SLC(i)_INT_LEVELS[CNT]
380 * NPS_PKT_SLC(i)_CNTS[TIMER] > NPS_PKT_SLC(i)_INT_LEVELS[TIMET]
381 * NPS_PKT_UNS(i)_CNTS[CNT] > NPS_PKT_UNS(i)_INT_LEVELS[CNT]
382 * NPS_PKT_UNS(i)_CNTS[TIMER] > NPS_PKT_UNS(i)_INT_LEVELS[TIMET]
383 * NPS_PKT_IN(i)_DONE_CNTS[CNT] > NPS_PKT_IN(i)_INT_LEVELS[CNT]
384 * NPS_PKT_MBOX_PF_VF(i)_INT[INTR] is set
385 * @cnt: Packet counter. Hardware adds to [CNT] as it reads
386 * packets. On a write to this CSR, hardware substracts the
387 * amount written to the [CNT] field from [CNT], which will
388 * clear PKT_IN(i)_INT_STATUS[INTR] if [CNT] becomes <=
389 * NPS_PKT_IN(i)_INT_LEVELS[CNT]. This register should be
390 * cleared before enabling a ring by reading the current
391 * value and writing it back.
393 union nps_pkt_in_done_cnts
{
396 #if defined(__BIG_ENDIAN_BITFIELD)
417 * struct nps_pkt_in_instr_ctl - Input Instruction Ring Control Registers.
418 * @is64b: If 1, the ring uses 64-byte instructions. If 0, the
419 * ring uses 32-byte instructions.
420 * @enb: Enable for the input ring.
422 union nps_pkt_in_instr_ctl
{
425 #if (defined(__BIG_ENDIAN_BITFIELD))
438 * struct nps_pkt_in_instr_rsize - Input instruction ring size registers
439 * @rsize: Ring size (number of instructions)
441 union nps_pkt_in_instr_rsize
{
444 #if (defined(__BIG_ENDIAN_BITFIELD))
455 * struct nps_pkt_in_instr_baoff_dbell - Input instruction ring
456 * base address offset and doorbell registers
457 * @aoff: Address offset. The offset from the NPS_PKT_IN_INSTR_BADDR
458 * where the next pointer is read.
459 * @dbell: Pointer list doorbell count. Write operations to this field
460 * increments the present value here. Read operations return the
463 union nps_pkt_in_instr_baoff_dbell
{
466 #if (defined(__BIG_ENDIAN_BITFIELD))
477 * struct nps_core_int_ena_w1s - NPS core interrupt enable set register
478 * @host_nps_wr_err: Reads or sets enable for
479 * NPS_CORE_INT[HOST_NPS_WR_ERR].
480 * @npco_dma_malform: Reads or sets enable for
481 * NPS_CORE_INT[NPCO_DMA_MALFORM].
482 * @exec_wr_timeout: Reads or sets enable for
483 * NPS_CORE_INT[EXEC_WR_TIMEOUT].
484 * @host_wr_timeout: Reads or sets enable for
485 * NPS_CORE_INT[HOST_WR_TIMEOUT].
486 * @host_wr_err: Reads or sets enable for
487 * NPS_CORE_INT[HOST_WR_ERR]
489 union nps_core_int_ena_w1s
{
492 #if (defined(__BIG_ENDIAN_BITFIELD))
494 u64 host_nps_wr_err
: 1;
495 u64 npco_dma_malform
: 1;
496 u64 exec_wr_timeout
: 1;
497 u64 host_wr_timeout
: 1;
509 u64 host_wr_timeout
: 1;
510 u64 exec_wr_timeout
: 1;
511 u64 npco_dma_malform
: 1;
512 u64 host_nps_wr_err
: 1;
519 * struct nps_core_gbl_vfcfg - Global VF Configuration Register.
520 * @ilk_disable: When set, this bit indicates that the ILK interface has
522 * @obaf: BMO allocation control
523 * 0 = allocate per queue
524 * 1 = allocate per VF
525 * @ibaf: BMI allocation control
526 * 0 = allocate per queue
527 * 1 = allocate per VF
528 * @zaf: ZIP allocation control
529 * 0 = allocate per queue
530 * 1 = allocate per VF
531 * @aeaf: AE allocation control
532 * 0 = allocate per queue
533 * 1 = allocate per VF
534 * @seaf: SE allocation control
535 * 0 = allocation per queue
536 * 1 = allocate per VF
539 union nps_core_gbl_vfcfg
{
542 #if (defined(__BIG_ENDIAN_BITFIELD))
565 * struct nps_core_int_active - NPS Core Interrupt Active Register
566 * @resend: Resend MSI-X interrupt if needs to handle interrupts
567 * Sofware can set this bit and then exit the ISR.
568 * @ocla: Set when any OCLA(0)_INT and corresponding OCLA(0_INT_ENA_W1C
570 * @mbox: Set when any NPS_PKT_MBOX_INT_LO/HI and corresponding
571 * NPS_PKT_MBOX_INT_LO_ENA_W1C/HI_ENA_W1C bits are set
572 * @emu: bit i is set in [EMU] when any EMU(i)_INT bit is set
573 * @bmo: Set when any BMO_INT bit is set
574 * @bmi: Set when any BMI_INT bit is set or when any non-RO
575 * BMI_INT and corresponding BMI_INT_ENA_W1C bits are both set
576 * @aqm: Set when any AQM_INT bit is set
577 * @zqm: Set when any ZQM_INT bit is set
578 * @efl: Set when any EFL_INT RO bit is set or when any non-RO EFL_INT
579 * and corresponding EFL_INT_ENA_W1C bits are both set
580 * @ilk: Set when any ILK_INT bit is set
581 * @lbc: Set when any LBC_INT RO bit is set or when any non-RO LBC_INT
582 * and corresponding LBC_INT_ENA_W1C bits are bot set
583 * @pem: Set when any PEM(0)_INT RO bit is set or when any non-RO
584 * PEM(0)_INT and corresponding PEM(0)_INT_ENA_W1C bit are both set
585 * @ucd: Set when any UCD_INT bit is set
586 * @zctl: Set when any ZIP_INT RO bit is set or when any non-RO ZIP_INT
587 * and corresponding ZIP_INT_ENA_W1C bits are both set
588 * @lbm: Set when any LBM_INT bit is set
589 * @nps_pkt: Set when any NPS_PKT_INT bit is set
590 * @nps_core: Set when any NPS_CORE_INT RO bit is set or when non-RO
591 * NPS_CORE_INT and corresponding NSP_CORE_INT_ENA_W1C bits are both set
593 union nps_core_int_active
{
596 #if (defined(__BIG_ENDIAN_BITFIELD))
641 * struct efl_core_int - EFL Interrupt Registers
642 * @epci_decode_err: EPCI decoded a transacation that was unknown
643 * This error should only occurred when there is a micrcode/SE error
644 * and should be considered fatal
645 * @ae_err: An AE uncorrectable error occurred.
646 * See EFL_CORE(0..3)_AE_ERR_INT
647 * @se_err: An SE uncorrectable error occurred.
648 * See EFL_CORE(0..3)_SE_ERR_INT
649 * @dbe: Double-bit error occurred in EFL
650 * @sbe: Single-bit error occurred in EFL
651 * @d_left: Asserted when new POM-Header-BMI-data is
652 * being sent to an Exec, and that Exec has Not read all BMI
653 * data associated with the previous POM header
654 * @len_ovr: Asserted when an Exec-Read is issued that is more than
655 * 14 greater in length that the BMI data left to be read
660 #if (defined(__BIG_ENDIAN_BITFIELD))
662 u64 epci_decode_err
: 1;
676 u64 epci_decode_err
: 1;
683 * struct efl_core_int_ena_w1s - EFL core interrupt enable set register
684 * @epci_decode_err: Reads or sets enable for
685 * EFL_CORE(0..3)_INT[EPCI_DECODE_ERR].
686 * @d_left: Reads or sets enable for
687 * EFL_CORE(0..3)_INT[D_LEFT].
688 * @len_ovr: Reads or sets enable for
689 * EFL_CORE(0..3)_INT[LEN_OVR].
691 union efl_core_int_ena_w1s
{
694 #if (defined(__BIG_ENDIAN_BITFIELD))
696 u64 epci_decode_err
: 1;
704 u64 epci_decode_err
: 1;
711 * struct efl_rnm_ctl_status - RNM Control and Status Register
712 * @ent_sel: Select input to RNM FIFO
713 * @exp_ent: Exported entropy enable for random number generator
714 * @rng_rst: Reset to RNG. Setting this bit to 1 cancels the generation
715 * of the current random number.
716 * @rnm_rst: Reset the RNM. Setting this bit to 1 clears all sorted numbers
717 * in the random number memory.
718 * @rng_en: Enabled the output of the RNG.
719 * @ent_en: Entropy enable for random number generator.
721 union efl_rnm_ctl_status
{
724 #if (defined(__BIG_ENDIAN_BITFIELD))
745 * struct bmi_ctl - BMI control register
746 * @ilk_hdrq_thrsh: Maximum number of header queue locations
747 * that ILK packets may consume. When the threshold is
748 * exceeded ILK_XOFF is sent to the BMI_X2P_ARB.
749 * @nps_hdrq_thrsh: Maximum number of header queue locations
750 * that NPS packets may consume. When the threshold is
751 * exceeded NPS_XOFF is sent to the BMI_X2P_ARB.
752 * @totl_hdrq_thrsh: Maximum number of header queue locations
753 * that the sum of ILK and NPS packets may consume.
754 * @ilk_free_thrsh: Maximum number of buffers that ILK packet
755 * flows may consume before ILK_XOFF is sent to the BMI_X2P_ARB.
756 * @nps_free_thrsh: Maximum number of buffers that NPS packet
757 * flows may consume before NPS XOFF is sent to the BMI_X2p_ARB.
758 * @totl_free_thrsh: Maximum number of buffers that bot ILK and NPS
759 * packet flows may consume before both NPS_XOFF and ILK_XOFF
760 * are asserted to the BMI_X2P_ARB.
761 * @max_pkt_len: Maximum packet length, integral number of 256B
767 #if (defined(__BIG_ENDIAN_BITFIELD))
769 u64 ilk_hdrq_thrsh
: 8;
770 u64 nps_hdrq_thrsh
: 8;
771 u64 totl_hdrq_thrsh
: 8;
772 u64 ilk_free_thrsh
: 8;
773 u64 nps_free_thrsh
: 8;
774 u64 totl_free_thrsh
: 8;
778 u64 totl_free_thrsh
: 8;
779 u64 nps_free_thrsh
: 8;
780 u64 ilk_free_thrsh
: 8;
781 u64 totl_hdrq_thrsh
: 8;
782 u64 nps_hdrq_thrsh
: 8;
783 u64 ilk_hdrq_thrsh
: 8;
790 * struct bmi_int_ena_w1s - BMI interrupt enable set register
791 * @ilk_req_oflw: Reads or sets enable for
792 * BMI_INT[ILK_REQ_OFLW].
793 * @nps_req_oflw: Reads or sets enable for
794 * BMI_INT[NPS_REQ_OFLW].
795 * @fpf_undrrn: Reads or sets enable for
796 * BMI_INT[FPF_UNDRRN].
797 * @eop_err_ilk: Reads or sets enable for
798 * BMI_INT[EOP_ERR_ILK].
799 * @eop_err_nps: Reads or sets enable for
800 * BMI_INT[EOP_ERR_NPS].
801 * @sop_err_ilk: Reads or sets enable for
802 * BMI_INT[SOP_ERR_ILK].
803 * @sop_err_nps: Reads or sets enable for
804 * BMI_INT[SOP_ERR_NPS].
805 * @pkt_rcv_err_ilk: Reads or sets enable for
806 * BMI_INT[PKT_RCV_ERR_ILK].
807 * @pkt_rcv_err_nps: Reads or sets enable for
808 * BMI_INT[PKT_RCV_ERR_NPS].
809 * @max_len_err_ilk: Reads or sets enable for
810 * BMI_INT[MAX_LEN_ERR_ILK].
811 * @max_len_err_nps: Reads or sets enable for
812 * BMI_INT[MAX_LEN_ERR_NPS].
814 union bmi_int_ena_w1s
{
817 #if (defined(__BIG_ENDIAN_BITFIELD))
819 u64 ilk_req_oflw
: 1;
820 u64 nps_req_oflw
: 1;
828 u64 pkt_rcv_err_ilk
: 1;
829 u64 pkt_rcv_err_nps
: 1;
830 u64 max_len_err_ilk
: 1;
831 u64 max_len_err_nps
: 1;
833 u64 max_len_err_nps
: 1;
834 u64 max_len_err_ilk
: 1;
835 u64 pkt_rcv_err_nps
: 1;
836 u64 pkt_rcv_err_ilk
: 1;
844 u64 nps_req_oflw
: 1;
845 u64 ilk_req_oflw
: 1;
852 * struct bmo_ctl2 - BMO Control2 Register
853 * @arb_sel: Determines P2X Arbitration
854 * @ilk_buf_thrsh: Maximum number of buffers that the
855 * ILK packet flows may consume before ILK XOFF is
856 * asserted to the POM.
857 * @nps_slc_buf_thrsh: Maximum number of buffers that the
858 * NPS_SLC packet flow may consume before NPS_SLC XOFF is
859 * asserted to the POM.
860 * @nps_uns_buf_thrsh: Maximum number of buffers that the
861 * NPS_UNS packet flow may consume before NPS_UNS XOFF is
862 * asserted to the POM.
863 * @totl_buf_thrsh: Maximum number of buffers that ILK, NPS_UNS and
864 * NPS_SLC packet flows may consume before NPS_UNS XOFF, NSP_SLC and
865 * ILK_XOFF are all asserted POM.
870 #if (defined(__BIG_ENDIAN_BITFIELD))
873 u64 ilk_buf_thrsh
: 8;
874 u64 nps_slc_buf_thrsh
: 8;
875 u64 nps_uns_buf_thrsh
: 8;
876 u64 totl_buf_thrsh
: 8;
878 u64 totl_buf_thrsh
: 8;
879 u64 nps_uns_buf_thrsh
: 8;
880 u64 nps_slc_buf_thrsh
: 8;
881 u64 ilk_buf_thrsh
: 8;
889 * struct pom_int_ena_w1s - POM interrupt enable set register
890 * @illegal_intf: Reads or sets enable for POM_INT[ILLEGAL_INTF].
891 * @illegal_dport: Reads or sets enable for POM_INT[ILLEGAL_DPORT].
893 union pom_int_ena_w1s
{
896 #if (defined(__BIG_ENDIAN_BITFIELD))
898 u64 illegal_intf
: 1;
899 u64 illegal_dport
: 1;
905 u64 illegal_dport
: 1;
906 u64 illegal_intf
: 1;
913 * struct lbc_inval_ctl - LBC invalidation control register
914 * @wait_timer: Wait timer for wait state. [WAIT_TIMER] must
915 * always be written with its reset value.
916 * @cam_inval_start: Software should write [CAM_INVAL_START]=1
917 * to initiate an LBC cache invalidation. After this, software
918 * should read LBC_INVAL_STATUS until LBC_INVAL_STATUS[DONE] is set.
919 * LBC hardware clears [CAVM_INVAL_START] before software can
920 * observed LBC_INVAL_STATUS[DONE] to be set
922 union lbc_inval_ctl
{
925 #if (defined(__BIG_ENDIAN_BITFIELD))
929 u64 cam_inval_start
: 1;
933 u64 cam_inval_start
: 1;
942 * struct lbc_int_ena_w1s - LBC interrupt enable set register
943 * @cam_hard_err: Reads or sets enable for LBC_INT[CAM_HARD_ERR].
944 * @cam_inval_abort: Reads or sets enable for LBC_INT[CAM_INVAL_ABORT].
945 * @over_fetch_err: Reads or sets enable for LBC_INT[OVER_FETCH_ERR].
946 * @cache_line_to_err: Reads or sets enable for
947 * LBC_INT[CACHE_LINE_TO_ERR].
948 * @cam_soft_err: Reads or sets enable for
949 * LBC_INT[CAM_SOFT_ERR].
950 * @dma_rd_err: Reads or sets enable for
951 * LBC_INT[DMA_RD_ERR].
953 union lbc_int_ena_w1s
{
956 #if (defined(__BIG_ENDIAN_BITFIELD))
958 u64 cam_hard_err
: 1;
959 u64 cam_inval_abort
: 1;
960 u64 over_fetch_err
: 1;
961 u64 cache_line_to_err
: 1;
963 u64 cam_soft_err
: 1;
967 u64 cam_soft_err
: 1;
969 u64 cache_line_to_err
: 1;
970 u64 over_fetch_err
: 1;
971 u64 cam_inval_abort
: 1;
972 u64 cam_hard_err
: 1;
979 * struct lbc_int - LBC interrupt summary register
980 * @cam_hard_err: indicates a fatal hardware error.
981 * It requires system reset.
982 * When [CAM_HARD_ERR] is set, LBC stops logging any new information in
983 * LBC_POM_MISS_INFO_LOG,
984 * LBC_POM_MISS_ADDR_LOG,
985 * LBC_EFL_MISS_INFO_LOG, and
986 * LBC_EFL_MISS_ADDR_LOG.
987 * Software should sample them.
988 * @cam_inval_abort: indicates a fatal hardware error.
989 * System reset is required.
990 * @over_fetch_err: indicates a fatal hardware error
991 * System reset is required
992 * @cache_line_to_err: is a debug feature.
993 * This timeout interrupt bit tells the software that
994 * a cacheline in LBC has non-zero usage and the context
995 * has not been used for greater than the
996 * LBC_TO_CNT[TO_CNT] time interval.
997 * @sbe: Memory SBE error. This is recoverable via ECC.
998 * See LBC_ECC_INT for more details.
999 * @dbe: Memory DBE error. This is a fatal and requires a
1001 * @pref_dat_len_mismatch_err: Summary bit for context length
1003 * @rd_dat_len_mismatch_err: Summary bit for SE read data length
1004 * greater than data prefect length errors.
1005 * @cam_soft_err: is recoverable. Software must complete a
1006 * LBC_INVAL_CTL[CAM_INVAL_START] invalidation sequence and
1007 * then clear [CAM_SOFT_ERR].
1008 * @dma_rd_err: A context prefect read of host memory returned with
1014 #if (defined(__BIG_ENDIAN_BITFIELD))
1016 u64 cam_hard_err
: 1;
1017 u64 cam_inval_abort
: 1;
1018 u64 over_fetch_err
: 1;
1019 u64 cache_line_to_err
: 1;
1022 u64 pref_dat_len_mismatch_err
: 1;
1023 u64 rd_dat_len_mismatch_err
: 1;
1024 u64 cam_soft_err
: 1;
1028 u64 cam_soft_err
: 1;
1029 u64 rd_dat_len_mismatch_err
: 1;
1030 u64 pref_dat_len_mismatch_err
: 1;
1033 u64 cache_line_to_err
: 1;
1034 u64 over_fetch_err
: 1;
1035 u64 cam_inval_abort
: 1;
1036 u64 cam_hard_err
: 1;
1043 * struct lbc_inval_status: LBC Invalidation status register
1044 * @cam_clean_entry_complete_cnt: The number of entries that are
1045 * cleaned up successfully.
1046 * @cam_clean_entry_cnt: The number of entries that have the CAM
1047 * inval command issued.
1048 * @cam_inval_state: cam invalidation FSM state
1049 * @cam_inval_abort: cam invalidation abort
1050 * @cam_rst_rdy: lbc_cam reset ready
1051 * @done: LBC clears [DONE] when
1052 * LBC_INVAL_CTL[CAM_INVAL_START] is written with a one,
1053 * and sets [DONE] when it completes the invalidation
1056 union lbc_inval_status
{
1059 #if (defined(__BIG_ENDIAN_BITFIELD))
1061 u64 cam_clean_entry_complete_cnt
: 9;
1063 u64 cam_clean_entry_cnt
: 9;
1065 u64 cam_inval_state
: 3;
1067 u64 cam_inval_abort
: 1;
1068 u64 cam_rst_rdy
: 1;
1072 u64 cam_rst_rdy
: 1;
1073 u64 cam_inval_abort
: 1;
1075 u64 cam_inval_state
: 3;
1077 u64 cam_clean_entry_cnt
: 9;
1079 u64 cam_clean_entry_complete_cnt
: 9;
1085 #endif /* __NITROX_CSR_H */