1 /* SPDX-License-Identifier: GPL-2.0
3 * Copyright 2020-2022 HabanaLabs, Ltd.
11 #include <uapi/drm/habanalabs_accel.h>
12 #include "../common/habanalabs.h"
13 #include <linux/habanalabs/hl_boot_if.h>
14 #include "../include/gaudi2/gaudi2.h"
15 #include "../include/gaudi2/gaudi2_packets.h"
16 #include "../include/gaudi2/gaudi2_fw_if.h"
17 #include "../include/gaudi2/gaudi2_async_events.h"
19 #define GAUDI2_LINUX_FW_FILE "habanalabs/gaudi2/gaudi2-fit.itb"
20 #define GAUDI2_BOOT_FIT_FILE "habanalabs/gaudi2/gaudi2-boot-fit.itb"
22 #define GAUDI2_CPU_TIMEOUT_USEC 30000000 /* 30s */
24 #define NUMBER_OF_PDMA_QUEUES 2
25 #define NUMBER_OF_EDMA_QUEUES 8
26 #define NUMBER_OF_MME_QUEUES 4
27 #define NUMBER_OF_TPC_QUEUES 25
28 #define NUMBER_OF_NIC_QUEUES 24
29 #define NUMBER_OF_ROT_QUEUES 2
30 #define NUMBER_OF_CPU_QUEUES 1
32 #define NUMBER_OF_HW_QUEUES ((NUMBER_OF_PDMA_QUEUES + \
33 NUMBER_OF_EDMA_QUEUES + \
34 NUMBER_OF_MME_QUEUES + \
35 NUMBER_OF_TPC_QUEUES + \
36 NUMBER_OF_NIC_QUEUES + \
37 NUMBER_OF_ROT_QUEUES + \
38 NUMBER_OF_CPU_QUEUES) * \
41 #define NUMBER_OF_QUEUES (NUMBER_OF_CPU_QUEUES + NUMBER_OF_HW_QUEUES)
43 #define DCORE_NUM_OF_SOB \
44 (((mmDCORE0_SYNC_MNGR_OBJS_SOB_OBJ_8191 - \
45 mmDCORE0_SYNC_MNGR_OBJS_SOB_OBJ_0) + 4) >> 2)
47 #define DCORE_NUM_OF_MONITORS \
48 (((mmDCORE0_SYNC_MNGR_OBJS_MON_STATUS_2047 - \
49 mmDCORE0_SYNC_MNGR_OBJS_MON_STATUS_0) + 4) >> 2)
51 #define NUMBER_OF_DEC ((NUM_OF_DEC_PER_DCORE * NUM_OF_DCORES) + NUMBER_OF_PCIE_DEC)
53 /* Map all arcs dccm + arc schedulers acp blocks */
54 #define NUM_OF_USER_ACP_BLOCKS (NUM_OF_SCHEDULER_ARC + 2)
55 #define NUM_OF_USER_NIC_UMR_BLOCKS 15
56 #define NUM_OF_EXPOSED_SM_BLOCKS ((NUM_OF_DCORES - 1) * 2)
57 #define NUM_USER_MAPPED_BLOCKS \
58 (NUM_ARC_CPUS + NUM_OF_USER_ACP_BLOCKS + NUMBER_OF_DEC + \
59 NUM_OF_EXPOSED_SM_BLOCKS + \
60 (NIC_NUMBER_OF_ENGINES * NUM_OF_USER_NIC_UMR_BLOCKS))
62 /* Within the user mapped array, decoder entries start post all the ARC related
65 #define USR_MAPPED_BLK_DEC_START_IDX \
66 (NUM_ARC_CPUS + NUM_OF_USER_ACP_BLOCKS + \
67 (NIC_NUMBER_OF_ENGINES * NUM_OF_USER_NIC_UMR_BLOCKS))
69 #define USR_MAPPED_BLK_SM_START_IDX \
70 (NUM_ARC_CPUS + NUM_OF_USER_ACP_BLOCKS + NUMBER_OF_DEC + \
71 (NIC_NUMBER_OF_ENGINES * NUM_OF_USER_NIC_UMR_BLOCKS))
73 #define SM_OBJS_BLOCK_SIZE (mmDCORE0_SYNC_MNGR_OBJS_SM_SEC_0 - \
74 mmDCORE0_SYNC_MNGR_OBJS_SOB_OBJ_0)
76 #define GAUDI2_MAX_PENDING_CS 64
78 #if !IS_MAX_PENDING_CS_VALID(GAUDI2_MAX_PENDING_CS)
79 #error "GAUDI2_MAX_PENDING_CS must be power of 2 and greater than 1"
82 #define CORESIGHT_TIMEOUT_USEC 100000 /* 100 ms */
84 #define GAUDI2_PREBOOT_REQ_TIMEOUT_USEC 25000000 /* 25s */
85 #define GAUDI2_PREBOOT_EXTENDED_REQ_TIMEOUT_USEC 85000000 /* 85s */
87 #define GAUDI2_BOOT_FIT_REQ_TIMEOUT_USEC 10000000 /* 10s */
89 #define GAUDI2_NIC_CLK_FREQ 450000000ull /* 450 MHz */
91 #define DC_POWER_DEFAULT 60000 /* 60W */
93 #define GAUDI2_HBM_NUM 6
95 #define DMA_MAX_TRANSFER_SIZE U32_MAX
97 #define GAUDI2_DEFAULT_CARD_NAME "HL225"
99 #define QMAN_STREAMS 4
101 #define NUM_OF_MME_SBTE_PORTS 5
102 #define NUM_OF_MME_WB_PORTS 2
104 #define GAUDI2_ENGINE_ID_DCORE_OFFSET \
105 (GAUDI2_DCORE1_ENGINE_ID_EDMA_0 - GAUDI2_DCORE0_ENGINE_ID_EDMA_0)
107 /* DRAM Memory Map */
109 #define CPU_FW_IMAGE_SIZE 0x10000000 /* 256MB */
110 #define CPU_FW_IMAGE_ADDR DRAM_PHYS_BASE
111 #define PMMU_PAGE_TABLES_SIZE 0x10000000 /* 256MB */
112 #define EDMA_PQS_SIZE SZ_2M
113 #define EDMA_SCRATCHPAD_SIZE SZ_1M
114 #define HMMU_PAGE_TABLES_SIZE SZ_1M
116 #define NIC_NUMBER_OF_PORTS NIC_NUMBER_OF_ENGINES
118 #define NUMBER_OF_PCIE_DEC 2
119 #define PCIE_DEC_SHIFT 8
121 #define SRAM_USER_BASE_OFFSET 0
123 /* cluster binning */
124 #define MAX_FAULTY_HBMS 1
125 #define GAUDI2_XBAR_EDGE_FULL_MASK 0xF
126 #define GAUDI2_EDMA_FULL_MASK 0xFF
127 #define GAUDI2_DRAM_FULL_MASK 0x3F
129 /* Host virtual address space. */
131 #define VA_HOST_SPACE_PAGE_START 0xFFF0000000000000ull
132 #define VA_HOST_SPACE_PAGE_END 0xFFF0800000000000ull /* 140TB */
134 #define VA_HOST_SPACE_HPAGE_START 0xFFF0800000000000ull
135 #define VA_HOST_SPACE_HPAGE_END 0xFFF1000000000000ull /* 140TB */
138 #define VA_HOST_SPACE_PAGE_SIZE (VA_HOST_SPACE_PAGE_END - VA_HOST_SPACE_PAGE_START)
141 #define VA_HOST_SPACE_HPAGE_SIZE (VA_HOST_SPACE_HPAGE_END - VA_HOST_SPACE_HPAGE_START)
143 #define VA_HOST_SPACE_SIZE (VA_HOST_SPACE_PAGE_SIZE + VA_HOST_SPACE_HPAGE_SIZE)
145 #define HOST_SPACE_INTERNAL_CB_SZ SZ_2M
148 * HBM virtual address space
149 * Gaudi2 has 6 HBM devices, each supporting 16GB total of 96GB at most.
150 * No core separation is supported so we can have one chunk of virtual address
151 * space just above the physical ones.
152 * The virtual address space starts immediately after the end of the physical
153 * address space which is determined at run-time.
155 #define VA_HBM_SPACE_END 0x1002000000000000ull
157 #define HW_CAP_PLL BIT_ULL(0)
158 #define HW_CAP_DRAM BIT_ULL(1)
159 #define HW_CAP_PMMU BIT_ULL(2)
160 #define HW_CAP_CPU BIT_ULL(3)
161 #define HW_CAP_MSIX BIT_ULL(4)
163 #define HW_CAP_CPU_Q BIT_ULL(5)
164 #define HW_CAP_CPU_Q_SHIFT 5
166 #define HW_CAP_CLK_GATE BIT_ULL(6)
167 #define HW_CAP_KDMA BIT_ULL(7)
168 #define HW_CAP_SRAM_SCRAMBLER BIT_ULL(8)
170 #define HW_CAP_DCORE0_DMMU0 BIT_ULL(9)
171 #define HW_CAP_DCORE0_DMMU1 BIT_ULL(10)
172 #define HW_CAP_DCORE0_DMMU2 BIT_ULL(11)
173 #define HW_CAP_DCORE0_DMMU3 BIT_ULL(12)
174 #define HW_CAP_DCORE1_DMMU0 BIT_ULL(13)
175 #define HW_CAP_DCORE1_DMMU1 BIT_ULL(14)
176 #define HW_CAP_DCORE1_DMMU2 BIT_ULL(15)
177 #define HW_CAP_DCORE1_DMMU3 BIT_ULL(16)
178 #define HW_CAP_DCORE2_DMMU0 BIT_ULL(17)
179 #define HW_CAP_DCORE2_DMMU1 BIT_ULL(18)
180 #define HW_CAP_DCORE2_DMMU2 BIT_ULL(19)
181 #define HW_CAP_DCORE2_DMMU3 BIT_ULL(20)
182 #define HW_CAP_DCORE3_DMMU0 BIT_ULL(21)
183 #define HW_CAP_DCORE3_DMMU1 BIT_ULL(22)
184 #define HW_CAP_DCORE3_DMMU2 BIT_ULL(23)
185 #define HW_CAP_DCORE3_DMMU3 BIT_ULL(24)
186 #define HW_CAP_DMMU_MASK GENMASK_ULL(24, 9)
187 #define HW_CAP_DMMU_SHIFT 9
188 #define HW_CAP_PDMA_MASK BIT_ULL(26)
189 #define HW_CAP_EDMA_MASK GENMASK_ULL(34, 27)
190 #define HW_CAP_EDMA_SHIFT 27
191 #define HW_CAP_MME_MASK GENMASK_ULL(38, 35)
192 #define HW_CAP_MME_SHIFT 35
193 #define HW_CAP_ROT_MASK GENMASK_ULL(40, 39)
194 #define HW_CAP_ROT_SHIFT 39
195 #define HW_CAP_HBM_SCRAMBLER_HW_RESET BIT_ULL(41)
196 #define HW_CAP_HBM_SCRAMBLER_SW_RESET BIT_ULL(42)
197 #define HW_CAP_HBM_SCRAMBLER_MASK (HW_CAP_HBM_SCRAMBLER_HW_RESET | \
198 HW_CAP_HBM_SCRAMBLER_SW_RESET)
199 #define HW_CAP_HBM_SCRAMBLER_SHIFT 41
200 #define HW_CAP_RESERVED BIT(43)
201 #define HW_CAP_MMU_MASK (HW_CAP_PMMU | HW_CAP_DMMU_MASK)
203 /* Range Registers */
204 #define RR_TYPE_SHORT 0
205 #define RR_TYPE_LONG 1
206 #define RR_TYPE_SHORT_PRIV 2
207 #define RR_TYPE_LONG_PRIV 3
208 #define NUM_SHORT_LBW_RR 14
209 #define NUM_LONG_LBW_RR 4
210 #define NUM_SHORT_HBW_RR 6
211 #define NUM_LONG_HBW_RR 4
213 /* RAZWI initiator coordinates- X- 5 bits, Y- 4 bits */
214 #define RAZWI_INITIATOR_X_SHIFT 0
215 #define RAZWI_INITIATOR_X_MASK 0x1F
216 #define RAZWI_INITIATOR_Y_SHIFT 5
217 #define RAZWI_INITIATOR_Y_MASK 0xF
219 #define RTR_ID_X_Y(x, y) \
220 ((((y) & RAZWI_INITIATOR_Y_MASK) << RAZWI_INITIATOR_Y_SHIFT) | \
221 (((x) & RAZWI_INITIATOR_X_MASK) << RAZWI_INITIATOR_X_SHIFT))
223 /* decoders have separate mask */
224 #define HW_CAP_DEC_SHIFT 0
225 #define HW_CAP_DEC_MASK GENMASK_ULL(9, 0)
227 /* TPCs have separate mask */
228 #define HW_CAP_TPC_SHIFT 0
229 #define HW_CAP_TPC_MASK GENMASK_ULL(24, 0)
231 /* nics have separate mask */
232 #define HW_CAP_NIC_SHIFT 0
233 #define HW_CAP_NIC_MASK GENMASK_ULL(NIC_NUMBER_OF_ENGINES - 1, 0)
235 #define GAUDI2_ARC_PCI_MSB_ADDR(addr) (((addr) & GENMASK_ULL(49, 28)) >> 28)
237 #define GAUDI2_SOB_INCREMENT_BY_ONE (FIELD_PREP(DCORE0_SYNC_MNGR_OBJS_SOB_OBJ_VAL_MASK, 1) | \
238 FIELD_PREP(DCORE0_SYNC_MNGR_OBJS_SOB_OBJ_INC_MASK, 1))
240 #define GAUDI2_NUM_TESTED_QS (GAUDI2_QUEUE_ID_CPU_PQ - GAUDI2_QUEUE_ID_PDMA_0_0)
243 enum gaudi2_reserved_sob_id
{
244 GAUDI2_RESERVED_SOB_CS_COMPLETION_FIRST
,
245 GAUDI2_RESERVED_SOB_CS_COMPLETION_LAST
=
246 GAUDI2_RESERVED_SOB_CS_COMPLETION_FIRST
+ GAUDI2_MAX_PENDING_CS
- 1,
247 GAUDI2_RESERVED_SOB_KDMA_COMPLETION
,
248 GAUDI2_RESERVED_SOB_DEC_NRM_FIRST
,
249 GAUDI2_RESERVED_SOB_DEC_NRM_LAST
=
250 GAUDI2_RESERVED_SOB_DEC_NRM_FIRST
+ NUMBER_OF_DEC
- 1,
251 GAUDI2_RESERVED_SOB_DEC_ABNRM_FIRST
,
252 GAUDI2_RESERVED_SOB_DEC_ABNRM_LAST
=
253 GAUDI2_RESERVED_SOB_DEC_ABNRM_FIRST
+ NUMBER_OF_DEC
- 1,
254 GAUDI2_RESERVED_SOB_NUMBER
257 enum gaudi2_reserved_mon_id
{
258 GAUDI2_RESERVED_MON_CS_COMPLETION_FIRST
,
259 GAUDI2_RESERVED_MON_CS_COMPLETION_LAST
=
260 GAUDI2_RESERVED_MON_CS_COMPLETION_FIRST
+ GAUDI2_MAX_PENDING_CS
- 1,
261 GAUDI2_RESERVED_MON_KDMA_COMPLETION
,
262 GAUDI2_RESERVED_MON_DEC_NRM_FIRST
,
263 GAUDI2_RESERVED_MON_DEC_NRM_LAST
=
264 GAUDI2_RESERVED_MON_DEC_NRM_FIRST
+ 3 * NUMBER_OF_DEC
- 1,
265 GAUDI2_RESERVED_MON_DEC_ABNRM_FIRST
,
266 GAUDI2_RESERVED_MON_DEC_ABNRM_LAST
=
267 GAUDI2_RESERVED_MON_DEC_ABNRM_FIRST
+ 3 * NUMBER_OF_DEC
- 1,
268 GAUDI2_RESERVED_MON_NUMBER
271 enum gaudi2_reserved_cq_id
{
272 GAUDI2_RESERVED_CQ_CS_COMPLETION
,
273 GAUDI2_RESERVED_CQ_KDMA_COMPLETION
,
274 GAUDI2_RESERVED_CQ_NUMBER
278 * Gaudi2 subtitute TPCs Numbering
279 * At most- two faulty TPCs are allowed
280 * First replacement to a faulty TPC will be TPC24, second- TPC23
282 enum substitude_tpc
{
283 FAULTY_TPC_SUBTS_1_TPC_24
,
284 FAULTY_TPC_SUBTS_2_TPC_23
,
288 enum gaudi2_dma_core_id
{
289 DMA_CORE_ID_PDMA0
, /* Dcore 0 */
290 DMA_CORE_ID_PDMA1
, /* Dcore 0 */
291 DMA_CORE_ID_EDMA0
, /* Dcore 0 */
292 DMA_CORE_ID_EDMA1
, /* Dcore 0 */
293 DMA_CORE_ID_EDMA2
, /* Dcore 1 */
294 DMA_CORE_ID_EDMA3
, /* Dcore 1 */
295 DMA_CORE_ID_EDMA4
, /* Dcore 2 */
296 DMA_CORE_ID_EDMA5
, /* Dcore 2 */
297 DMA_CORE_ID_EDMA6
, /* Dcore 3 */
298 DMA_CORE_ID_EDMA7
, /* Dcore 3 */
299 DMA_CORE_ID_KDMA
, /* Dcore 0 */
303 enum gaudi2_rotator_id
{
342 /* the PCI TPC is placed last (mapped liked HW) */
371 /* specific EDMA enumeration */
372 enum gaudi2_edma_id
{
373 EDMA_ID_DCORE0_INSTANCE0
,
374 EDMA_ID_DCORE0_INSTANCE1
,
375 EDMA_ID_DCORE1_INSTANCE0
,
376 EDMA_ID_DCORE1_INSTANCE1
,
377 EDMA_ID_DCORE2_INSTANCE0
,
378 EDMA_ID_DCORE2_INSTANCE1
,
379 EDMA_ID_DCORE3_INSTANCE0
,
380 EDMA_ID_DCORE3_INSTANCE1
,
384 /* User interrupt count is aligned with HW CQ count.
385 * We have 64 CQ's per dcore, CQ0 in dcore 0 is reserved for legacy mode
387 #define GAUDI2_NUM_USER_INTERRUPTS 64
388 #define GAUDI2_NUM_RESERVED_INTERRUPTS 1
389 #define GAUDI2_TOTAL_USER_INTERRUPTS (GAUDI2_NUM_USER_INTERRUPTS + GAUDI2_NUM_RESERVED_INTERRUPTS)
391 enum gaudi2_irq_num
{
392 GAUDI2_IRQ_NUM_EVENT_QUEUE
= GAUDI2_EVENT_QUEUE_MSIX_IDX
,
393 GAUDI2_IRQ_NUM_DCORE0_DEC0_NRM
,
394 GAUDI2_IRQ_NUM_DCORE0_DEC0_ABNRM
,
395 GAUDI2_IRQ_NUM_DCORE0_DEC1_NRM
,
396 GAUDI2_IRQ_NUM_DCORE0_DEC1_ABNRM
,
397 GAUDI2_IRQ_NUM_DCORE1_DEC0_NRM
,
398 GAUDI2_IRQ_NUM_DCORE1_DEC0_ABNRM
,
399 GAUDI2_IRQ_NUM_DCORE1_DEC1_NRM
,
400 GAUDI2_IRQ_NUM_DCORE1_DEC1_ABNRM
,
401 GAUDI2_IRQ_NUM_DCORE2_DEC0_NRM
,
402 GAUDI2_IRQ_NUM_DCORE2_DEC0_ABNRM
,
403 GAUDI2_IRQ_NUM_DCORE2_DEC1_NRM
,
404 GAUDI2_IRQ_NUM_DCORE2_DEC1_ABNRM
,
405 GAUDI2_IRQ_NUM_DCORE3_DEC0_NRM
,
406 GAUDI2_IRQ_NUM_DCORE3_DEC0_ABNRM
,
407 GAUDI2_IRQ_NUM_DCORE3_DEC1_NRM
,
408 GAUDI2_IRQ_NUM_DCORE3_DEC1_ABNRM
,
409 GAUDI2_IRQ_NUM_SHARED_DEC0_NRM
,
410 GAUDI2_IRQ_NUM_SHARED_DEC0_ABNRM
,
411 GAUDI2_IRQ_NUM_SHARED_DEC1_NRM
,
412 GAUDI2_IRQ_NUM_SHARED_DEC1_ABNRM
,
413 GAUDI2_IRQ_NUM_DEC_LAST
= GAUDI2_IRQ_NUM_SHARED_DEC1_ABNRM
,
414 GAUDI2_IRQ_NUM_COMPLETION
,
415 GAUDI2_IRQ_NUM_NIC_PORT_FIRST
,
416 GAUDI2_IRQ_NUM_NIC_PORT_LAST
= (GAUDI2_IRQ_NUM_NIC_PORT_FIRST
+ NIC_NUMBER_OF_PORTS
- 1),
417 GAUDI2_IRQ_NUM_TPC_ASSERT
,
418 GAUDI2_IRQ_NUM_EQ_ERROR
,
419 GAUDI2_IRQ_NUM_USER_FIRST
,
420 GAUDI2_IRQ_NUM_USER_LAST
= (GAUDI2_IRQ_NUM_USER_FIRST
+ GAUDI2_NUM_USER_INTERRUPTS
- 1),
421 GAUDI2_IRQ_NUM_RESERVED_FIRST
,
422 GAUDI2_IRQ_NUM_RESERVED_LAST
= (GAUDI2_MSIX_ENTRIES
- GAUDI2_NUM_RESERVED_INTERRUPTS
- 1),
423 GAUDI2_IRQ_NUM_UNEXPECTED_ERROR
= RESERVED_MSIX_UNEXPECTED_USER_ERROR_INTERRUPT
,
424 GAUDI2_IRQ_NUM_LAST
= (GAUDI2_MSIX_ENTRIES
- 1)
427 static_assert(GAUDI2_IRQ_NUM_USER_FIRST
> GAUDI2_IRQ_NUM_SHARED_DEC1_ABNRM
);
430 * struct dup_block_ctx - context to initialize unit instances across multiple
431 * blocks where block can be either a dcore of duplicated
432 * common module. this code relies on constant offsets
433 * of blocks and unit instances in a block.
434 * @instance_cfg_fn: instance specific configuration function.
435 * @data: private configuration data.
436 * @base: base address of the first instance in the first block.
437 * @block_off: subsequent blocks address spacing.
438 * @instance_off: subsequent block's instances address spacing.
439 * @enabled_mask: mask of enabled instances (1- enabled, 0- disabled).
440 * @blocks: number of blocks.
441 * @instances: unit instances per block.
443 struct dup_block_ctx
{
444 void (*instance_cfg_fn
)(struct hl_device
*hdev
, u64 base
, void *data
);
451 unsigned int instances
;
455 * struct gaudi2_queues_test_info - Holds the address of a the messages used for testing the
457 * @dma_addr: the address used by the HW for accessing the message.
458 * @kern_addr: The address used by the driver for accessing the message.
460 struct gaudi2_queues_test_info
{
466 * struct gaudi2_device - ASIC specific manage structure.
467 * @cpucp_info_get: get information on device from CPU-CP
468 * @mapped_blocks: array that holds the base address and size of all blocks
470 * @lfsr_rand_seeds: array of MME ACC random seeds to set.
471 * @hw_queues_lock: protects the H/W queues from concurrent access.
472 * @scratchpad_kernel_address: general purpose PAGE_SIZE contiguous memory,
473 * this memory region should be write-only.
474 * currently used for HBW QMAN writes which is
476 * @scratchpad_bus_address: scratchpad bus address
477 * @virt_msix_db_cpu_addr: host memory page for the virtual MSI-X doorbell.
478 * @virt_msix_db_dma_addr: bus address of the page for the virtual MSI-X doorbell.
479 * @dram_bar_cur_addr: current address of DRAM PCI bar.
480 * @hw_cap_initialized: This field contains a bit per H/W engine. When that
481 * engine is initialized, that bit is set by the driver to
482 * signal we can use this engine in later code paths.
483 * Each bit is cleared upon reset of its corresponding H/W
485 * @active_hw_arc: This field contains a bit per ARC of an H/W engine with
486 * exception of TPC and NIC engines. Once an engine arc is
487 * initialized, its respective bit is set. Driver can uniquely
488 * identify each initialized ARC and use this information in
489 * later code paths. Each respective bit is cleared upon reset
490 * of its corresponding ARC of the H/W engine.
491 * @dec_hw_cap_initialized: This field contains a bit per decoder H/W engine.
492 * When that engine is initialized, that bit is set by
493 * the driver to signal we can use this engine in later
495 * Each bit is cleared upon reset of its corresponding H/W
497 * @tpc_hw_cap_initialized: This field contains a bit per TPC H/W engine.
498 * When that engine is initialized, that bit is set by
499 * the driver to signal we can use this engine in later
501 * Each bit is cleared upon reset of its corresponding H/W
503 * @active_tpc_arc: This field contains a bit per ARC of the TPC engines.
504 * Once an engine arc is initialized, its respective bit is
505 * set. Each respective bit is cleared upon reset of its
506 * corresponding ARC of the TPC engine.
507 * @nic_hw_cap_initialized: This field contains a bit per nic H/W engine.
508 * @active_nic_arc: This field contains a bit per ARC of the NIC engines.
509 * Once an engine arc is initialized, its respective bit is
510 * set. Each respective bit is cleared upon reset of its
511 * corresponding ARC of the NIC engine.
512 * @hw_events: array that holds all H/W events that are defined valid.
513 * @events_stat: array that holds histogram of all received events.
514 * @events_stat_aggregate: same as events_stat but doesn't get cleared on reset.
515 * @num_of_valid_hw_events: used to hold the number of valid H/W events.
516 * @nic_ports: array that holds all NIC ports manage structures.
517 * @nic_macros: array that holds all NIC macro manage structures.
518 * @core_info: core info to be used by the Ethernet driver.
519 * @aux_ops: functions for core <-> aux drivers communication.
520 * @flush_db_fifo: flag to force flush DB FIFO after a write.
521 * @hbm_cfg: HBM subsystem settings
522 * @hw_queues_lock_mutex: used by simulator instead of hw_queues_lock.
523 * @queues_test_info: information used by the driver when testing the HW queues.
525 struct gaudi2_device
{
526 int (*cpucp_info_get
)(struct hl_device
*hdev
);
528 struct user_mapped_block mapped_blocks
[NUM_USER_MAPPED_BLOCKS
];
529 int lfsr_rand_seeds
[MME_NUM_OF_LFSR_SEEDS
];
531 spinlock_t hw_queues_lock
;
533 void *scratchpad_kernel_address
;
534 dma_addr_t scratchpad_bus_address
;
536 void *virt_msix_db_cpu_addr
;
537 dma_addr_t virt_msix_db_dma_addr
;
539 u64 dram_bar_cur_addr
;
540 u64 hw_cap_initialized
;
542 u64 dec_hw_cap_initialized
;
543 u64 tpc_hw_cap_initialized
;
545 u64 nic_hw_cap_initialized
;
547 u32 hw_events
[GAUDI2_EVENT_SIZE
];
548 u32 events_stat
[GAUDI2_EVENT_SIZE
];
549 u32 events_stat_aggregate
[GAUDI2_EVENT_SIZE
];
550 u32 num_of_valid_hw_events
;
553 struct gaudi2_queues_test_info queues_test_info
[GAUDI2_NUM_TESTED_QS
];
557 * Types of the Gaudi2 IP blocks, used by special blocks iterator.
558 * Required for scenarios where only particular block types can be
559 * addressed (e.g., special PLDM images).
561 enum gaudi2_block_types
{
562 GAUDI2_BLOCK_TYPE_PLL
,
563 GAUDI2_BLOCK_TYPE_RTR
,
564 GAUDI2_BLOCK_TYPE_CPU
,
565 GAUDI2_BLOCK_TYPE_HIF
,
566 GAUDI2_BLOCK_TYPE_HBM
,
567 GAUDI2_BLOCK_TYPE_NIC
,
568 GAUDI2_BLOCK_TYPE_PCIE
,
569 GAUDI2_BLOCK_TYPE_PCIE_PMA
,
570 GAUDI2_BLOCK_TYPE_PDMA
,
571 GAUDI2_BLOCK_TYPE_EDMA
,
572 GAUDI2_BLOCK_TYPE_PMMU
,
573 GAUDI2_BLOCK_TYPE_PSOC
,
574 GAUDI2_BLOCK_TYPE_ROT
,
575 GAUDI2_BLOCK_TYPE_ARC_FARM
,
576 GAUDI2_BLOCK_TYPE_DEC
,
577 GAUDI2_BLOCK_TYPE_MME
,
578 GAUDI2_BLOCK_TYPE_EU_BIST
,
579 GAUDI2_BLOCK_TYPE_SYNC_MNGR
,
580 GAUDI2_BLOCK_TYPE_STLB
,
581 GAUDI2_BLOCK_TYPE_TPC
,
582 GAUDI2_BLOCK_TYPE_HMMU
,
583 GAUDI2_BLOCK_TYPE_SRAM
,
584 GAUDI2_BLOCK_TYPE_XBAR
,
585 GAUDI2_BLOCK_TYPE_KDMA
,
586 GAUDI2_BLOCK_TYPE_XDMA
,
587 GAUDI2_BLOCK_TYPE_XFT
,
588 GAUDI2_BLOCK_TYPE_MAX
591 extern const u32 gaudi2_dma_core_blocks_bases
[DMA_CORE_ID_SIZE
];
592 extern const u32 gaudi2_qm_blocks_bases
[GAUDI2_QUEUE_ID_SIZE
];
593 extern const u32 gaudi2_mme_acc_blocks_bases
[MME_ID_SIZE
];
594 extern const u32 gaudi2_mme_ctrl_lo_blocks_bases
[MME_ID_SIZE
];
595 extern const u32 edma_stream_base
[NUM_OF_EDMA_PER_DCORE
* NUM_OF_DCORES
];
596 extern const u32 gaudi2_rot_blocks_bases
[ROTATOR_ID_SIZE
];
598 void gaudi2_iterate_tpcs(struct hl_device
*hdev
, struct iterate_module_ctx
*ctx
);
599 int gaudi2_coresight_init(struct hl_device
*hdev
);
600 int gaudi2_debug_coresight(struct hl_device
*hdev
, struct hl_ctx
*ctx
, void *data
);
601 void gaudi2_halt_coresight(struct hl_device
*hdev
, struct hl_ctx
*ctx
);
602 void gaudi2_init_blocks(struct hl_device
*hdev
, struct dup_block_ctx
*cfg_ctx
);
603 bool gaudi2_is_hmmu_enabled(struct hl_device
*hdev
, int dcore_id
, int hmmu_id
);
604 void gaudi2_write_rr_to_all_lbw_rtrs(struct hl_device
*hdev
, u8 rr_type
, u32 rr_index
, u64 min_val
,
606 void gaudi2_pb_print_security_errors(struct hl_device
*hdev
, u32 block_addr
, u32 cause
,
608 int gaudi2_init_security(struct hl_device
*hdev
);
609 void gaudi2_ack_protection_bits_errors(struct hl_device
*hdev
);
610 int gaudi2_send_device_activity(struct hl_device
*hdev
, bool open
);
612 #endif /* GAUDI2P_H_ */