2 * Copyright 2018 Advanced Micro Devices, Inc.
4 * Permission is hereby granted, free of charge, to any person obtaining a
5 * copy of this software and associated documentation files (the "Software"),
6 * to deal in the Software without restriction, including without limitation
7 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
8 * and/or sell copies of the Software, and to permit persons to whom the
9 * Software is furnished to do so, subject to the following conditions:
11 * The above copyright notice and this permission notice shall be included in
12 * all copies or substantial portions of the Software.
14 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
15 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
16 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
17 * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
18 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
19 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
20 * OTHER DEALINGS IN THE SOFTWARE.
23 #include <linux/printk.h>
24 #include <linux/device.h>
25 #include <linux/slab.h>
26 #include <linux/pci.h>
27 #include <linux/amd-iommu.h>
29 #include "kfd_dbgmgr.h"
30 #include "kfd_topology.h"
31 #include "kfd_iommu.h"
33 static const u32 required_iommu_flags
= AMD_IOMMU_DEVICE_FLAG_ATS_SUP
|
34 AMD_IOMMU_DEVICE_FLAG_PRI_SUP
|
35 AMD_IOMMU_DEVICE_FLAG_PASID_SUP
;
37 /** kfd_iommu_check_device - Check whether IOMMU is available for device
39 int kfd_iommu_check_device(struct kfd_dev
*kfd
)
41 struct amd_iommu_device_info iommu_info
;
44 if (!kfd
->device_info
->needs_iommu_device
)
48 err
= amd_iommu_device_info(kfd
->pdev
, &iommu_info
);
52 if ((iommu_info
.flags
& required_iommu_flags
) != required_iommu_flags
)
58 /** kfd_iommu_device_init - Initialize IOMMU for device
60 int kfd_iommu_device_init(struct kfd_dev
*kfd
)
62 struct amd_iommu_device_info iommu_info
;
63 unsigned int pasid_limit
;
65 struct kfd_topology_device
*top_dev
;
67 top_dev
= kfd_topology_device_by_id(kfd
->id
);
70 * Overwrite ATS capability according to needs_iommu_device to fix
71 * potential missing corresponding bit in CRAT of BIOS.
73 if (!kfd
->device_info
->needs_iommu_device
) {
74 top_dev
->node_props
.capability
&= ~HSA_CAP_ATS_PRESENT
;
78 top_dev
->node_props
.capability
|= HSA_CAP_ATS_PRESENT
;
81 err
= amd_iommu_device_info(kfd
->pdev
, &iommu_info
);
84 "error getting iommu info. is the iommu enabled?\n");
88 if ((iommu_info
.flags
& required_iommu_flags
) != required_iommu_flags
) {
90 "error required iommu flags ats %i, pri %i, pasid %i\n",
91 (iommu_info
.flags
& AMD_IOMMU_DEVICE_FLAG_ATS_SUP
) != 0,
92 (iommu_info
.flags
& AMD_IOMMU_DEVICE_FLAG_PRI_SUP
) != 0,
93 (iommu_info
.flags
& AMD_IOMMU_DEVICE_FLAG_PASID_SUP
)
98 pasid_limit
= min_t(unsigned int,
99 (unsigned int)(1 << kfd
->device_info
->max_pasid_bits
),
100 iommu_info
.max_pasids
);
102 if (!kfd_set_pasid_limit(pasid_limit
)) {
103 dev_err(kfd_device
, "error setting pasid limit\n");
110 /** kfd_iommu_bind_process_to_device - Have the IOMMU bind a process
112 * Binds the given process to the given device using its PASID. This
113 * enables IOMMUv2 address translation for the process on the device.
115 * This function assumes that the process mutex is held.
117 int kfd_iommu_bind_process_to_device(struct kfd_process_device
*pdd
)
119 struct kfd_dev
*dev
= pdd
->dev
;
120 struct kfd_process
*p
= pdd
->process
;
123 if (!dev
->device_info
->needs_iommu_device
|| pdd
->bound
== PDD_BOUND
)
126 if (unlikely(pdd
->bound
== PDD_BOUND_SUSPENDED
)) {
127 pr_err("Binding PDD_BOUND_SUSPENDED pdd is unexpected!\n");
131 err
= amd_iommu_bind_pasid(dev
->pdev
, p
->pasid
, p
->lead_thread
);
133 pdd
->bound
= PDD_BOUND
;
138 /** kfd_iommu_unbind_process - Unbind process from all devices
140 * This removes all IOMMU device bindings of the process. To be used
141 * before process termination.
143 void kfd_iommu_unbind_process(struct kfd_process
*p
)
145 struct kfd_process_device
*pdd
;
147 list_for_each_entry(pdd
, &p
->per_device_data
, per_device_list
)
148 if (pdd
->bound
== PDD_BOUND
)
149 amd_iommu_unbind_pasid(pdd
->dev
->pdev
, p
->pasid
);
152 /* Callback for process shutdown invoked by the IOMMU driver */
153 static void iommu_pasid_shutdown_callback(struct pci_dev
*pdev
, int pasid
)
155 struct kfd_dev
*dev
= kfd_device_by_pci_dev(pdev
);
156 struct kfd_process
*p
;
157 struct kfd_process_device
*pdd
;
163 * Look for the process that matches the pasid. If there is no such
164 * process, we either released it in amdkfd's own notifier, or there
165 * is a bug. Unfortunately, there is no way to tell...
167 p
= kfd_lookup_process_by_pasid(pasid
);
171 pr_debug("Unbinding process %d from IOMMU\n", pasid
);
173 mutex_lock(kfd_get_dbgmgr_mutex());
175 if (dev
->dbgmgr
&& dev
->dbgmgr
->pasid
== p
->pasid
) {
176 if (!kfd_dbgmgr_unregister(dev
->dbgmgr
, p
)) {
177 kfd_dbgmgr_destroy(dev
->dbgmgr
);
182 mutex_unlock(kfd_get_dbgmgr_mutex());
184 mutex_lock(&p
->mutex
);
186 pdd
= kfd_get_process_device_data(dev
, p
);
188 /* For GPU relying on IOMMU, we need to dequeue here
189 * when PASID is still bound.
191 kfd_process_dequeue_from_device(pdd
);
193 mutex_unlock(&p
->mutex
);
195 kfd_unref_process(p
);
198 /* This function called by IOMMU driver on PPR failure */
199 static int iommu_invalid_ppr_cb(struct pci_dev
*pdev
, int pasid
,
200 unsigned long address
, u16 flags
)
204 dev_warn_ratelimited(kfd_device
,
205 "Invalid PPR device %x:%x.%x pasid %d address 0x%lX flags 0x%X",
206 PCI_BUS_NUM(pdev
->devfn
),
207 PCI_SLOT(pdev
->devfn
),
208 PCI_FUNC(pdev
->devfn
),
213 dev
= kfd_device_by_pci_dev(pdev
);
215 kfd_signal_iommu_event(dev
, pasid
, address
,
216 flags
& PPR_FAULT_WRITE
, flags
& PPR_FAULT_EXEC
);
218 return AMD_IOMMU_INV_PRI_RSP_INVALID
;
222 * Bind processes do the device that have been temporarily unbound
223 * (PDD_BOUND_SUSPENDED) in kfd_unbind_processes_from_device.
225 static int kfd_bind_processes_to_device(struct kfd_dev
*kfd
)
227 struct kfd_process_device
*pdd
;
228 struct kfd_process
*p
;
232 int idx
= srcu_read_lock(&kfd_processes_srcu
);
234 hash_for_each_rcu(kfd_processes_table
, temp
, p
, kfd_processes
) {
235 mutex_lock(&p
->mutex
);
236 pdd
= kfd_get_process_device_data(kfd
, p
);
238 if (WARN_ON(!pdd
) || pdd
->bound
!= PDD_BOUND_SUSPENDED
) {
239 mutex_unlock(&p
->mutex
);
243 err
= amd_iommu_bind_pasid(kfd
->pdev
, p
->pasid
,
246 pr_err("Unexpected pasid %d binding failure\n",
248 mutex_unlock(&p
->mutex
);
252 pdd
->bound
= PDD_BOUND
;
253 mutex_unlock(&p
->mutex
);
256 srcu_read_unlock(&kfd_processes_srcu
, idx
);
262 * Mark currently bound processes as PDD_BOUND_SUSPENDED. These
263 * processes will be restored to PDD_BOUND state in
264 * kfd_bind_processes_to_device.
266 static void kfd_unbind_processes_from_device(struct kfd_dev
*kfd
)
268 struct kfd_process_device
*pdd
;
269 struct kfd_process
*p
;
272 int idx
= srcu_read_lock(&kfd_processes_srcu
);
274 hash_for_each_rcu(kfd_processes_table
, temp
, p
, kfd_processes
) {
275 mutex_lock(&p
->mutex
);
276 pdd
= kfd_get_process_device_data(kfd
, p
);
279 mutex_unlock(&p
->mutex
);
283 if (pdd
->bound
== PDD_BOUND
)
284 pdd
->bound
= PDD_BOUND_SUSPENDED
;
285 mutex_unlock(&p
->mutex
);
288 srcu_read_unlock(&kfd_processes_srcu
, idx
);
291 /** kfd_iommu_suspend - Prepare IOMMU for suspend
293 * This unbinds processes from the device and disables the IOMMU for
296 void kfd_iommu_suspend(struct kfd_dev
*kfd
)
298 if (!kfd
->device_info
->needs_iommu_device
)
301 kfd_unbind_processes_from_device(kfd
);
303 amd_iommu_set_invalidate_ctx_cb(kfd
->pdev
, NULL
);
304 amd_iommu_set_invalid_ppr_cb(kfd
->pdev
, NULL
);
305 amd_iommu_free_device(kfd
->pdev
);
308 /** kfd_iommu_resume - Restore IOMMU after resume
310 * This reinitializes the IOMMU for the device and re-binds previously
311 * suspended processes to the device.
313 int kfd_iommu_resume(struct kfd_dev
*kfd
)
315 unsigned int pasid_limit
;
318 if (!kfd
->device_info
->needs_iommu_device
)
321 pasid_limit
= kfd_get_pasid_limit();
323 err
= amd_iommu_init_device(kfd
->pdev
, pasid_limit
);
327 amd_iommu_set_invalidate_ctx_cb(kfd
->pdev
,
328 iommu_pasid_shutdown_callback
);
329 amd_iommu_set_invalid_ppr_cb(kfd
->pdev
,
330 iommu_invalid_ppr_cb
);
332 err
= kfd_bind_processes_to_device(kfd
);
334 amd_iommu_set_invalidate_ctx_cb(kfd
->pdev
, NULL
);
335 amd_iommu_set_invalid_ppr_cb(kfd
->pdev
, NULL
);
336 amd_iommu_free_device(kfd
->pdev
);
343 extern bool amd_iommu_pc_supported(void);
344 extern u8
amd_iommu_pc_get_max_banks(u16 devid
);
345 extern u8
amd_iommu_pc_get_max_counters(u16 devid
);
347 /** kfd_iommu_add_perf_counters - Add IOMMU performance counters to topology
349 int kfd_iommu_add_perf_counters(struct kfd_topology_device
*kdev
)
351 struct kfd_perf_properties
*props
;
353 if (!(kdev
->node_props
.capability
& HSA_CAP_ATS_PRESENT
))
356 if (!amd_iommu_pc_supported())
359 props
= kfd_alloc_struct(props
);
362 strcpy(props
->block_name
, "iommu");
363 props
->max_concurrent
= amd_iommu_pc_get_max_banks(0) *
364 amd_iommu_pc_get_max_counters(0); /* assume one iommu */
365 list_add_tail(&props
->list
, &kdev
->perf_props
);