2 * Copyright 2014 IBM Corp.
4 * This program is free software; you can redistribute it and/or
5 * modify it under the terms of the GNU General Public License
6 * as published by the Free Software Foundation; either version
7 * 2 of the License, or (at your option) any later version.
10 #include <linux/spinlock.h>
11 #include <linux/sched.h>
12 #include <linux/slab.h>
13 #include <linux/sched.h>
14 #include <linux/mutex.h>
16 #include <linux/uaccess.h>
17 #include <asm/synch.h>
23 static int afu_control(struct cxl_afu
*afu
, u64 command
,
24 u64 result
, u64 mask
, bool enabled
)
26 u64 AFU_Cntl
= cxl_p2n_read(afu
, CXL_AFU_Cntl_An
);
27 unsigned long timeout
= jiffies
+ (HZ
* CXL_TIMEOUT
);
30 spin_lock(&afu
->afu_cntl_lock
);
31 pr_devel("AFU command starting: %llx\n", command
);
33 trace_cxl_afu_ctrl(afu
, command
);
35 cxl_p2n_write(afu
, CXL_AFU_Cntl_An
, AFU_Cntl
| command
);
37 AFU_Cntl
= cxl_p2n_read(afu
, CXL_AFU_Cntl_An
);
38 while ((AFU_Cntl
& mask
) != result
) {
39 if (time_after_eq(jiffies
, timeout
)) {
40 dev_warn(&afu
->dev
, "WARNING: AFU control timed out!\n");
44 pr_devel_ratelimited("AFU control... (0x%.16llx)\n",
47 AFU_Cntl
= cxl_p2n_read(afu
, CXL_AFU_Cntl_An
);
49 pr_devel("AFU command complete: %llx\n", command
);
50 afu
->enabled
= enabled
;
52 trace_cxl_afu_ctrl_done(afu
, command
, rc
);
53 spin_unlock(&afu
->afu_cntl_lock
);
58 static int afu_enable(struct cxl_afu
*afu
)
60 pr_devel("AFU enable request\n");
62 return afu_control(afu
, CXL_AFU_Cntl_An_E
,
63 CXL_AFU_Cntl_An_ES_Enabled
,
64 CXL_AFU_Cntl_An_ES_MASK
, true);
67 int cxl_afu_disable(struct cxl_afu
*afu
)
69 pr_devel("AFU disable request\n");
71 return afu_control(afu
, 0, CXL_AFU_Cntl_An_ES_Disabled
,
72 CXL_AFU_Cntl_An_ES_MASK
, false);
75 /* This will disable as well as reset */
76 int cxl_afu_reset(struct cxl_afu
*afu
)
78 pr_devel("AFU reset request\n");
80 return afu_control(afu
, CXL_AFU_Cntl_An_RA
,
81 CXL_AFU_Cntl_An_RS_Complete
| CXL_AFU_Cntl_An_ES_Disabled
,
82 CXL_AFU_Cntl_An_RS_MASK
| CXL_AFU_Cntl_An_ES_MASK
,
86 static int afu_check_and_enable(struct cxl_afu
*afu
)
90 return afu_enable(afu
);
93 int cxl_psl_purge(struct cxl_afu
*afu
)
95 u64 PSL_CNTL
= cxl_p1n_read(afu
, CXL_PSL_SCNTL_An
);
96 u64 AFU_Cntl
= cxl_p2n_read(afu
, CXL_AFU_Cntl_An
);
99 unsigned long timeout
= jiffies
+ (HZ
* CXL_TIMEOUT
);
102 trace_cxl_psl_ctrl(afu
, CXL_PSL_SCNTL_An_Pc
);
104 pr_devel("PSL purge request\n");
106 if ((AFU_Cntl
& CXL_AFU_Cntl_An_ES_MASK
) != CXL_AFU_Cntl_An_ES_Disabled
) {
107 WARN(1, "psl_purge request while AFU not disabled!\n");
108 cxl_afu_disable(afu
);
111 cxl_p1n_write(afu
, CXL_PSL_SCNTL_An
,
112 PSL_CNTL
| CXL_PSL_SCNTL_An_Pc
);
113 start
= local_clock();
114 PSL_CNTL
= cxl_p1n_read(afu
, CXL_PSL_SCNTL_An
);
115 while ((PSL_CNTL
& CXL_PSL_SCNTL_An_Ps_MASK
)
116 == CXL_PSL_SCNTL_An_Ps_Pending
) {
117 if (time_after_eq(jiffies
, timeout
)) {
118 dev_warn(&afu
->dev
, "WARNING: PSL Purge timed out!\n");
122 dsisr
= cxl_p2n_read(afu
, CXL_PSL_DSISR_An
);
123 pr_devel_ratelimited("PSL purging... PSL_CNTL: 0x%.16llx PSL_DSISR: 0x%.16llx\n", PSL_CNTL
, dsisr
);
124 if (dsisr
& CXL_PSL_DSISR_TRANS
) {
125 dar
= cxl_p2n_read(afu
, CXL_PSL_DAR_An
);
126 dev_notice(&afu
->dev
, "PSL purge terminating pending translation, DSISR: 0x%.16llx, DAR: 0x%.16llx\n", dsisr
, dar
);
127 cxl_p2n_write(afu
, CXL_PSL_TFC_An
, CXL_PSL_TFC_An_AE
);
129 dev_notice(&afu
->dev
, "PSL purge acknowledging pending non-translation fault, DSISR: 0x%.16llx\n", dsisr
);
130 cxl_p2n_write(afu
, CXL_PSL_TFC_An
, CXL_PSL_TFC_An_A
);
134 PSL_CNTL
= cxl_p1n_read(afu
, CXL_PSL_SCNTL_An
);
137 pr_devel("PSL purged in %lld ns\n", end
- start
);
139 cxl_p1n_write(afu
, CXL_PSL_SCNTL_An
,
140 PSL_CNTL
& ~CXL_PSL_SCNTL_An_Pc
);
142 trace_cxl_psl_ctrl_done(afu
, CXL_PSL_SCNTL_An_Pc
, rc
);
146 static int spa_max_procs(int spa_size
)
150 * end_of_SPA_area = SPA_Base + ((n+4) * 128) + (( ((n*8) + 127) >> 7) * 128) + 255
151 * Most of that junk is really just an overly-complicated way of saying
152 * the last 256 bytes are __aligned(128), so it's really:
153 * end_of_SPA_area = end_of_PSL_queue_area + __aligned(128) 255
155 * end_of_PSL_queue_area = SPA_Base + ((n+4) * 128) + (n*8) - 1
157 * sizeof(SPA) = ((n+4) * 128) + (n*8) + __aligned(128) 256
158 * Ignore the alignment (which is safe in this case as long as we are
159 * careful with our rounding) and solve for n:
161 return ((spa_size
/ 8) - 96) / 17;
164 static int alloc_spa(struct cxl_afu
*afu
)
168 /* Work out how many pages to allocate */
172 afu
->spa_size
= (1 << afu
->spa_order
) * PAGE_SIZE
;
173 afu
->spa_max_procs
= spa_max_procs(afu
->spa_size
);
174 } while (afu
->spa_max_procs
< afu
->num_procs
);
176 WARN_ON(afu
->spa_size
> 0x100000); /* Max size supported by the hardware */
178 if (!(afu
->spa
= (struct cxl_process_element
*)
179 __get_free_pages(GFP_KERNEL
| __GFP_ZERO
, afu
->spa_order
))) {
180 pr_err("cxl_alloc_spa: Unable to allocate scheduled process area\n");
183 pr_devel("spa pages: %i afu->spa_max_procs: %i afu->num_procs: %i\n",
184 1<<afu
->spa_order
, afu
->spa_max_procs
, afu
->num_procs
);
186 afu
->sw_command_status
= (__be64
*)((char *)afu
->spa
+
187 ((afu
->spa_max_procs
+ 3) * 128));
189 spap
= virt_to_phys(afu
->spa
) & CXL_PSL_SPAP_Addr
;
190 spap
|= ((afu
->spa_size
>> (12 - CXL_PSL_SPAP_Size_Shift
)) - 1) & CXL_PSL_SPAP_Size
;
191 spap
|= CXL_PSL_SPAP_V
;
192 pr_devel("cxl: SPA allocated at 0x%p. Max processes: %i, sw_command_status: 0x%p CXL_PSL_SPAP_An=0x%016llx\n", afu
->spa
, afu
->spa_max_procs
, afu
->sw_command_status
, spap
);
193 cxl_p1n_write(afu
, CXL_PSL_SPAP_An
, spap
);
198 static void release_spa(struct cxl_afu
*afu
)
200 cxl_p1n_write(afu
, CXL_PSL_SPAP_An
, 0);
201 free_pages((unsigned long) afu
->spa
, afu
->spa_order
);
204 int cxl_tlb_slb_invalidate(struct cxl
*adapter
)
206 unsigned long timeout
= jiffies
+ (HZ
* CXL_TIMEOUT
);
208 pr_devel("CXL adapter wide TLBIA & SLBIA\n");
210 cxl_p1_write(adapter
, CXL_PSL_AFUSEL
, CXL_PSL_AFUSEL_A
);
212 cxl_p1_write(adapter
, CXL_PSL_TLBIA
, CXL_TLB_SLB_IQ_ALL
);
213 while (cxl_p1_read(adapter
, CXL_PSL_TLBIA
) & CXL_TLB_SLB_P
) {
214 if (time_after_eq(jiffies
, timeout
)) {
215 dev_warn(&adapter
->dev
, "WARNING: CXL adapter wide TLBIA timed out!\n");
221 cxl_p1_write(adapter
, CXL_PSL_SLBIA
, CXL_TLB_SLB_IQ_ALL
);
222 while (cxl_p1_read(adapter
, CXL_PSL_SLBIA
) & CXL_TLB_SLB_P
) {
223 if (time_after_eq(jiffies
, timeout
)) {
224 dev_warn(&adapter
->dev
, "WARNING: CXL adapter wide SLBIA timed out!\n");
232 int cxl_afu_slbia(struct cxl_afu
*afu
)
234 unsigned long timeout
= jiffies
+ (HZ
* CXL_TIMEOUT
);
236 pr_devel("cxl_afu_slbia issuing SLBIA command\n");
237 cxl_p2n_write(afu
, CXL_SLBIA_An
, CXL_TLB_SLB_IQ_ALL
);
238 while (cxl_p2n_read(afu
, CXL_SLBIA_An
) & CXL_TLB_SLB_P
) {
239 if (time_after_eq(jiffies
, timeout
)) {
240 dev_warn(&afu
->dev
, "WARNING: CXL AFU SLBIA timed out!\n");
248 static int cxl_write_sstp(struct cxl_afu
*afu
, u64 sstp0
, u64 sstp1
)
252 /* 1. Disable SSTP by writing 0 to SSTP1[V] */
253 cxl_p2n_write(afu
, CXL_SSTP1_An
, 0);
255 /* 2. Invalidate all SLB entries */
256 if ((rc
= cxl_afu_slbia(afu
)))
259 /* 3. Set SSTP0_An */
260 cxl_p2n_write(afu
, CXL_SSTP0_An
, sstp0
);
262 /* 4. Set SSTP1_An */
263 cxl_p2n_write(afu
, CXL_SSTP1_An
, sstp1
);
268 /* Using per slice version may improve performance here. (ie. SLBIA_An) */
269 static void slb_invalid(struct cxl_context
*ctx
)
271 struct cxl
*adapter
= ctx
->afu
->adapter
;
274 WARN_ON(!mutex_is_locked(&ctx
->afu
->spa_mutex
));
276 cxl_p1_write(adapter
, CXL_PSL_LBISEL
,
277 ((u64
)be32_to_cpu(ctx
->elem
->common
.pid
) << 32) |
278 be32_to_cpu(ctx
->elem
->lpid
));
279 cxl_p1_write(adapter
, CXL_PSL_SLBIA
, CXL_TLB_SLB_IQ_LPIDPID
);
282 slbia
= cxl_p1_read(adapter
, CXL_PSL_SLBIA
);
283 if (!(slbia
& CXL_TLB_SLB_P
))
289 static int do_process_element_cmd(struct cxl_context
*ctx
,
290 u64 cmd
, u64 pe_state
)
293 unsigned long timeout
= jiffies
+ (HZ
* CXL_TIMEOUT
);
296 trace_cxl_llcmd(ctx
, cmd
);
298 WARN_ON(!ctx
->afu
->enabled
);
300 ctx
->elem
->software_state
= cpu_to_be32(pe_state
);
302 *(ctx
->afu
->sw_command_status
) = cpu_to_be64(cmd
| 0 | ctx
->pe
);
304 cxl_p1n_write(ctx
->afu
, CXL_PSL_LLCMD_An
, cmd
| ctx
->pe
);
306 if (time_after_eq(jiffies
, timeout
)) {
307 dev_warn(&ctx
->afu
->dev
, "WARNING: Process Element Command timed out!\n");
311 state
= be64_to_cpup(ctx
->afu
->sw_command_status
);
312 if (state
== ~0ULL) {
313 pr_err("cxl: Error adding process element to AFU\n");
317 if ((state
& (CXL_SPA_SW_CMD_MASK
| CXL_SPA_SW_STATE_MASK
| CXL_SPA_SW_LINK_MASK
)) ==
318 (cmd
| (cmd
>> 16) | ctx
->pe
))
321 * The command won't finish in the PSL if there are
322 * outstanding DSIs. Hence we need to yield here in
323 * case there are outstanding DSIs that we need to
324 * service. Tuning possiblity: we could wait for a
331 trace_cxl_llcmd_done(ctx
, cmd
, rc
);
335 static int add_process_element(struct cxl_context
*ctx
)
339 mutex_lock(&ctx
->afu
->spa_mutex
);
340 pr_devel("%s Adding pe: %i started\n", __func__
, ctx
->pe
);
341 if (!(rc
= do_process_element_cmd(ctx
, CXL_SPA_SW_CMD_ADD
, CXL_PE_SOFTWARE_STATE_V
)))
342 ctx
->pe_inserted
= true;
343 pr_devel("%s Adding pe: %i finished\n", __func__
, ctx
->pe
);
344 mutex_unlock(&ctx
->afu
->spa_mutex
);
348 static int terminate_process_element(struct cxl_context
*ctx
)
352 /* fast path terminate if it's already invalid */
353 if (!(ctx
->elem
->software_state
& cpu_to_be32(CXL_PE_SOFTWARE_STATE_V
)))
356 mutex_lock(&ctx
->afu
->spa_mutex
);
357 pr_devel("%s Terminate pe: %i started\n", __func__
, ctx
->pe
);
358 rc
= do_process_element_cmd(ctx
, CXL_SPA_SW_CMD_TERMINATE
,
359 CXL_PE_SOFTWARE_STATE_V
| CXL_PE_SOFTWARE_STATE_T
);
360 ctx
->elem
->software_state
= 0; /* Remove Valid bit */
361 pr_devel("%s Terminate pe: %i finished\n", __func__
, ctx
->pe
);
362 mutex_unlock(&ctx
->afu
->spa_mutex
);
366 static int remove_process_element(struct cxl_context
*ctx
)
370 mutex_lock(&ctx
->afu
->spa_mutex
);
371 pr_devel("%s Remove pe: %i started\n", __func__
, ctx
->pe
);
372 if (!(rc
= do_process_element_cmd(ctx
, CXL_SPA_SW_CMD_REMOVE
, 0)))
373 ctx
->pe_inserted
= false;
375 pr_devel("%s Remove pe: %i finished\n", __func__
, ctx
->pe
);
376 mutex_unlock(&ctx
->afu
->spa_mutex
);
382 static void assign_psn_space(struct cxl_context
*ctx
)
384 if (!ctx
->afu
->pp_size
|| ctx
->master
) {
385 ctx
->psn_phys
= ctx
->afu
->psn_phys
;
386 ctx
->psn_size
= ctx
->afu
->adapter
->ps_size
;
388 ctx
->psn_phys
= ctx
->afu
->psn_phys
+
389 (ctx
->afu
->pp_offset
+ ctx
->afu
->pp_size
* ctx
->pe
);
390 ctx
->psn_size
= ctx
->afu
->pp_size
;
394 static int activate_afu_directed(struct cxl_afu
*afu
)
398 dev_info(&afu
->dev
, "Activating AFU directed mode\n");
403 cxl_p1n_write(afu
, CXL_PSL_SCNTL_An
, CXL_PSL_SCNTL_An_PM_AFU
);
404 cxl_p1n_write(afu
, CXL_PSL_AMOR_An
, 0xFFFFFFFFFFFFFFFFULL
);
405 cxl_p1n_write(afu
, CXL_PSL_ID_An
, CXL_PSL_ID_An_F
| CXL_PSL_ID_An_L
);
407 afu
->current_mode
= CXL_MODE_DIRECTED
;
408 afu
->num_procs
= afu
->max_procs_virtualised
;
410 if ((rc
= cxl_chardev_m_afu_add(afu
)))
413 if ((rc
= cxl_sysfs_afu_m_add(afu
)))
416 if ((rc
= cxl_chardev_s_afu_add(afu
)))
421 cxl_sysfs_afu_m_remove(afu
);
423 cxl_chardev_afu_remove(afu
);
427 #ifdef CONFIG_CPU_LITTLE_ENDIAN
428 #define set_endian(sr) ((sr) |= CXL_PSL_SR_An_LE)
430 #define set_endian(sr) ((sr) &= ~(CXL_PSL_SR_An_LE))
433 static int attach_afu_directed(struct cxl_context
*ctx
, u64 wed
, u64 amr
)
438 assign_psn_space(ctx
);
440 ctx
->elem
->ctxtime
= 0; /* disable */
441 ctx
->elem
->lpid
= cpu_to_be32(mfspr(SPRN_LPID
));
442 ctx
->elem
->haurp
= 0; /* disable */
443 ctx
->elem
->sdr
= cpu_to_be64(mfspr(SPRN_SDR1
));
447 sr
|= CXL_PSL_SR_An_MP
;
448 if (mfspr(SPRN_LPCR
) & LPCR_TC
)
449 sr
|= CXL_PSL_SR_An_TC
;
450 /* HV=0, PR=1, R=1 for userspace
451 * For kernel contexts: this would need to change
453 sr
|= CXL_PSL_SR_An_PR
| CXL_PSL_SR_An_R
;
455 sr
&= ~(CXL_PSL_SR_An_HV
);
456 if (!test_tsk_thread_flag(current
, TIF_32BIT
))
457 sr
|= CXL_PSL_SR_An_SF
;
458 ctx
->elem
->common
.pid
= cpu_to_be32(current
->pid
);
459 ctx
->elem
->common
.tid
= 0;
460 ctx
->elem
->sr
= cpu_to_be64(sr
);
462 ctx
->elem
->common
.csrp
= 0; /* disable */
463 ctx
->elem
->common
.aurp0
= 0; /* disable */
464 ctx
->elem
->common
.aurp1
= 0; /* disable */
466 cxl_prefault(ctx
, wed
);
468 ctx
->elem
->common
.sstp0
= cpu_to_be64(ctx
->sstp0
);
469 ctx
->elem
->common
.sstp1
= cpu_to_be64(ctx
->sstp1
);
471 for (r
= 0; r
< CXL_IRQ_RANGES
; r
++) {
472 ctx
->elem
->ivte_offsets
[r
] = cpu_to_be16(ctx
->irqs
.offset
[r
]);
473 ctx
->elem
->ivte_ranges
[r
] = cpu_to_be16(ctx
->irqs
.range
[r
]);
476 ctx
->elem
->common
.amr
= cpu_to_be64(amr
);
477 ctx
->elem
->common
.wed
= cpu_to_be64(wed
);
479 /* first guy needs to enable */
480 if ((result
= afu_check_and_enable(ctx
->afu
)))
483 add_process_element(ctx
);
488 static int deactivate_afu_directed(struct cxl_afu
*afu
)
490 dev_info(&afu
->dev
, "Deactivating AFU directed mode\n");
492 afu
->current_mode
= 0;
495 cxl_sysfs_afu_m_remove(afu
);
496 cxl_chardev_afu_remove(afu
);
499 cxl_afu_disable(afu
);
507 static int activate_dedicated_process(struct cxl_afu
*afu
)
509 dev_info(&afu
->dev
, "Activating dedicated process mode\n");
511 cxl_p1n_write(afu
, CXL_PSL_SCNTL_An
, CXL_PSL_SCNTL_An_PM_Process
);
513 cxl_p1n_write(afu
, CXL_PSL_CtxTime_An
, 0); /* disable */
514 cxl_p1n_write(afu
, CXL_PSL_SPAP_An
, 0); /* disable */
515 cxl_p1n_write(afu
, CXL_PSL_AMOR_An
, 0xFFFFFFFFFFFFFFFFULL
);
516 cxl_p1n_write(afu
, CXL_PSL_LPID_An
, mfspr(SPRN_LPID
));
517 cxl_p1n_write(afu
, CXL_HAURP_An
, 0); /* disable */
518 cxl_p1n_write(afu
, CXL_PSL_SDR_An
, mfspr(SPRN_SDR1
));
520 cxl_p2n_write(afu
, CXL_CSRP_An
, 0); /* disable */
521 cxl_p2n_write(afu
, CXL_AURP0_An
, 0); /* disable */
522 cxl_p2n_write(afu
, CXL_AURP1_An
, 0); /* disable */
524 afu
->current_mode
= CXL_MODE_DEDICATED
;
527 return cxl_chardev_d_afu_add(afu
);
530 static int attach_dedicated(struct cxl_context
*ctx
, u64 wed
, u64 amr
)
532 struct cxl_afu
*afu
= ctx
->afu
;
539 sr
|= CXL_PSL_SR_An_MP
;
540 if (mfspr(SPRN_LPCR
) & LPCR_TC
)
541 sr
|= CXL_PSL_SR_An_TC
;
542 sr
|= CXL_PSL_SR_An_PR
| CXL_PSL_SR_An_R
;
543 if (!test_tsk_thread_flag(current
, TIF_32BIT
))
544 sr
|= CXL_PSL_SR_An_SF
;
545 cxl_p2n_write(afu
, CXL_PSL_PID_TID_An
, (u64
)current
->pid
<< 32);
546 cxl_p1n_write(afu
, CXL_PSL_SR_An
, sr
);
548 if ((rc
= cxl_write_sstp(afu
, ctx
->sstp0
, ctx
->sstp1
)))
551 cxl_prefault(ctx
, wed
);
553 cxl_p1n_write(afu
, CXL_PSL_IVTE_Offset_An
,
554 (((u64
)ctx
->irqs
.offset
[0] & 0xffff) << 48) |
555 (((u64
)ctx
->irqs
.offset
[1] & 0xffff) << 32) |
556 (((u64
)ctx
->irqs
.offset
[2] & 0xffff) << 16) |
557 ((u64
)ctx
->irqs
.offset
[3] & 0xffff));
558 cxl_p1n_write(afu
, CXL_PSL_IVTE_Limit_An
, (u64
)
559 (((u64
)ctx
->irqs
.range
[0] & 0xffff) << 48) |
560 (((u64
)ctx
->irqs
.range
[1] & 0xffff) << 32) |
561 (((u64
)ctx
->irqs
.range
[2] & 0xffff) << 16) |
562 ((u64
)ctx
->irqs
.range
[3] & 0xffff));
564 cxl_p2n_write(afu
, CXL_PSL_AMR_An
, amr
);
566 /* master only context for dedicated */
567 assign_psn_space(ctx
);
569 if ((rc
= cxl_afu_reset(afu
)))
572 cxl_p2n_write(afu
, CXL_PSL_WED_An
, wed
);
574 return afu_enable(afu
);
577 static int deactivate_dedicated_process(struct cxl_afu
*afu
)
579 dev_info(&afu
->dev
, "Deactivating dedicated process mode\n");
581 afu
->current_mode
= 0;
584 cxl_chardev_afu_remove(afu
);
589 int _cxl_afu_deactivate_mode(struct cxl_afu
*afu
, int mode
)
591 if (mode
== CXL_MODE_DIRECTED
)
592 return deactivate_afu_directed(afu
);
593 if (mode
== CXL_MODE_DEDICATED
)
594 return deactivate_dedicated_process(afu
);
598 int cxl_afu_deactivate_mode(struct cxl_afu
*afu
)
600 return _cxl_afu_deactivate_mode(afu
, afu
->current_mode
);
603 int cxl_afu_activate_mode(struct cxl_afu
*afu
, int mode
)
607 if (!(mode
& afu
->modes_supported
))
610 if (mode
== CXL_MODE_DIRECTED
)
611 return activate_afu_directed(afu
);
612 if (mode
== CXL_MODE_DEDICATED
)
613 return activate_dedicated_process(afu
);
618 int cxl_attach_process(struct cxl_context
*ctx
, bool kernel
, u64 wed
, u64 amr
)
620 ctx
->kernel
= kernel
;
621 if (ctx
->afu
->current_mode
== CXL_MODE_DIRECTED
)
622 return attach_afu_directed(ctx
, wed
, amr
);
624 if (ctx
->afu
->current_mode
== CXL_MODE_DEDICATED
)
625 return attach_dedicated(ctx
, wed
, amr
);
630 static inline int detach_process_native_dedicated(struct cxl_context
*ctx
)
632 cxl_afu_reset(ctx
->afu
);
633 cxl_afu_disable(ctx
->afu
);
634 cxl_psl_purge(ctx
->afu
);
638 static inline int detach_process_native_afu_directed(struct cxl_context
*ctx
)
640 if (!ctx
->pe_inserted
)
642 if (terminate_process_element(ctx
))
644 if (remove_process_element(ctx
))
650 int cxl_detach_process(struct cxl_context
*ctx
)
652 trace_cxl_detach(ctx
);
654 if (ctx
->afu
->current_mode
== CXL_MODE_DEDICATED
)
655 return detach_process_native_dedicated(ctx
);
657 return detach_process_native_afu_directed(ctx
);
660 int cxl_get_irq(struct cxl_afu
*afu
, struct cxl_irq_info
*info
)
664 info
->dsisr
= cxl_p2n_read(afu
, CXL_PSL_DSISR_An
);
665 info
->dar
= cxl_p2n_read(afu
, CXL_PSL_DAR_An
);
666 info
->dsr
= cxl_p2n_read(afu
, CXL_PSL_DSR_An
);
667 pidtid
= cxl_p2n_read(afu
, CXL_PSL_PID_TID_An
);
668 info
->pid
= pidtid
>> 32;
669 info
->tid
= pidtid
& 0xffffffff;
670 info
->afu_err
= cxl_p2n_read(afu
, CXL_AFU_ERR_An
);
671 info
->errstat
= cxl_p2n_read(afu
, CXL_PSL_ErrStat_An
);
676 static void recover_psl_err(struct cxl_afu
*afu
, u64 errstat
)
680 pr_devel("RECOVERING FROM PSL ERROR... (0x%.16llx)\n", errstat
);
682 /* Clear PSL_DSISR[PE] */
683 dsisr
= cxl_p2n_read(afu
, CXL_PSL_DSISR_An
);
684 cxl_p2n_write(afu
, CXL_PSL_DSISR_An
, dsisr
& ~CXL_PSL_DSISR_An_PE
);
686 /* Write 1s to clear error status bits */
687 cxl_p2n_write(afu
, CXL_PSL_ErrStat_An
, errstat
);
690 int cxl_ack_irq(struct cxl_context
*ctx
, u64 tfc
, u64 psl_reset_mask
)
692 trace_cxl_psl_irq_ack(ctx
, tfc
);
694 cxl_p2n_write(ctx
->afu
, CXL_PSL_TFC_An
, tfc
);
696 recover_psl_err(ctx
->afu
, psl_reset_mask
);
701 int cxl_check_error(struct cxl_afu
*afu
)
703 return (cxl_p1n_read(afu
, CXL_PSL_SCNTL_An
) == ~0ULL);