2 * ICSWX and ACOP Management
4 * Copyright (C) 2011 Anton Blanchard, IBM Corp. <anton@samba.org>
6 * This program is free software; you can redistribute it and/or
7 * modify it under the terms of the GNU General Public License
8 * as published by the Free Software Foundation; either version
9 * 2 of the License, or (at your option) any later version.
13 #include <linux/sched.h>
14 #include <linux/kernel.h>
15 #include <linux/errno.h>
16 #include <linux/types.h>
18 #include <linux/spinlock.h>
19 #include <linux/module.h>
20 #include <linux/uaccess.h>
25 * The processor and its L2 cache cause the icswx instruction to
26 * generate a COP_REQ transaction on PowerBus. The transaction has no
27 * address, and the processor does not perform an MMU access to
28 * authenticate the transaction. The command portion of the PowerBus
29 * COP_REQ transaction includes the LPAR_ID (LPID) and the coprocessor
30 * Process ID (PID), which the coprocessor compares to the authorized
31 * LPID and PID held in the coprocessor, to determine if the process
32 * is authorized to generate the transaction. The data of the COP_REQ
33 * transaction is 128-byte or less in size and is placed in cacheable
34 * memory on a 128-byte cache line boundary.
36 * The task to use a coprocessor should use use_cop() to mark the use
37 * of the Coprocessor Type (CT) and context switching. On a server
38 * class processor, the PID register is used only for coprocessor
39 * management + * and so a coprocessor PID is allocated before
40 * executing icswx + * instruction. Drop_cop() is used to free the
44 * Host Fabric Interface (HFI) is a PowerPC network coprocessor.
45 * Each HFI have multiple windows. Each HFI window serves as a
46 * network device sending to and receiving from HFI network.
47 * HFI immediate send function uses icswx instruction. The immediate
48 * send function allows small (single cache-line) packets be sent
49 * without using the regular HFI send FIFO and doorbell, which are
50 * much slower than immediate send.
52 * For each task intending to use HFI immediate send, the HFI driver
53 * calls use_cop() to obtain a coprocessor PID for the task.
54 * The HFI driver then allocate a free HFI window and save the
55 * coprocessor PID to the HFI window to allow the task to use the
58 * The HFI driver repeatedly creates immediate send packets and
59 * issues icswx instruction to send data through the HFI window.
60 * The HFI compares the coprocessor PID in the CPU PID register
61 * to the PID held in the HFI window to determine if the transaction
64 * When the task to release the HFI window, the HFI driver calls
65 * drop_cop() to release the coprocessor PID.
68 void switch_cop(struct mm_struct
*next
)
70 #ifdef CONFIG_PPC_ICSWX_PID
71 mtspr(SPRN_PID
, next
->context
.cop_pid
);
73 mtspr(SPRN_ACOP
, next
->context
.acop
);
77 * Start using a coprocessor.
78 * @acop: mask of coprocessor to be used.
79 * @mm: The mm the coprocessor to associate with. Most likely current mm.
81 * Return a positive PID if successful. Negative errno otherwise.
82 * The returned PID will be fed to the coprocessor to determine if an
83 * icswx transaction is authenticated.
85 int use_cop(unsigned long acop
, struct mm_struct
*mm
)
89 if (!cpu_has_feature(CPU_FTR_ICSWX
))
95 /* The page_table_lock ensures mm_users won't change under us */
96 spin_lock(&mm
->page_table_lock
);
97 spin_lock(mm
->context
.cop_lockp
);
99 ret
= get_cop_pid(mm
);
104 mm
->context
.acop
|= acop
;
109 * If this is a threaded process then there might be other threads
110 * running. We need to send an IPI to force them to pick up any
111 * change in PID and ACOP.
113 if (atomic_read(&mm
->mm_users
) > 1)
114 smp_call_function(sync_cop
, mm
, 1);
117 spin_unlock(mm
->context
.cop_lockp
);
118 spin_unlock(&mm
->page_table_lock
);
122 EXPORT_SYMBOL_GPL(use_cop
);
125 * Stop using a coprocessor.
126 * @acop: mask of coprocessor to be stopped.
127 * @mm: The mm the coprocessor associated with.
129 void drop_cop(unsigned long acop
, struct mm_struct
*mm
)
133 if (!cpu_has_feature(CPU_FTR_ICSWX
))
136 if (WARN_ON_ONCE(!mm
))
139 /* The page_table_lock ensures mm_users won't change under us */
140 spin_lock(&mm
->page_table_lock
);
141 spin_lock(mm
->context
.cop_lockp
);
143 mm
->context
.acop
&= ~acop
;
145 free_pid
= disable_cop_pid(mm
);
149 * If this is a threaded process then there might be other threads
150 * running. We need to send an IPI to force them to pick up any
151 * change in PID and ACOP.
153 if (atomic_read(&mm
->mm_users
) > 1)
154 smp_call_function(sync_cop
, mm
, 1);
156 if (free_pid
!= COP_PID_NONE
)
157 free_cop_pid(free_pid
);
159 spin_unlock(mm
->context
.cop_lockp
);
160 spin_unlock(&mm
->page_table_lock
);
162 EXPORT_SYMBOL_GPL(drop_cop
);
164 static int acop_use_cop(int ct
)
166 /* There is no alternate policy, yet */
171 * Get the instruction word at the NIP
173 static u32
acop_get_inst(struct pt_regs
*regs
)
178 p
= (u32 __user
*)regs
->nip
;
179 if (!access_ok(VERIFY_READ
, p
, sizeof(*p
)))
182 if (__get_user(inst
, p
))
189 * @regs: regsiters at time of interrupt
190 * @address: storage address
191 * @error_code: Fault code, usually the DSISR or ESR depending on
194 * Return 0 if we are able to resolve the data storage fault that
195 * results from a CT miss in the ACOP register.
197 int acop_handle_fault(struct pt_regs
*regs
, unsigned long address
,
198 unsigned long error_code
)
203 if (!cpu_has_feature(CPU_FTR_ICSWX
)) {
204 pr_info("No coprocessors available");
205 _exception(SIGILL
, regs
, ILL_ILLOPN
, address
);
208 if (!user_mode(regs
)) {
209 /* this could happen if the HV denies the
210 * kernel access, for now we just die */
211 die("ICSWX from kernel failed", regs
, SIGSEGV
);
214 /* Some implementations leave us a hint for the CT */
215 ct
= ICSWX_GET_CT_HINT(error_code
);
217 /* we have to peek at the instruction word to figure out CT */
221 inst
= acop_get_inst(regs
);
225 rs
= (inst
>> (31 - 10)) & 0x1f;
227 ct
= (ccw
>> 16) & 0x3f;
231 * We could be here because another thread has enabled acop
232 * but the ACOP register has yet to be updated.
234 * This should have been taken care of by the IPI to sync all
235 * the threads (see smp_call_function(sync_cop, mm, 1)), but
236 * that could take forever if there are a significant amount
239 * Given the number of threads on some of these systems,
240 * perhaps this is the best way to sync ACOP rather than whack
241 * every thread with an IPI.
243 if ((acop_copro_type_bit(ct
) & current
->active_mm
->context
.acop
) != 0) {
244 sync_cop(current
->active_mm
);
248 /* check for alternate policy */
249 if (!acop_use_cop(ct
))
252 /* at this point the CT is unknown to the system */
253 pr_warn("%s[%d]: Coprocessor %d is unavailable\n",
254 current
->comm
, current
->pid
, ct
);
256 /* get inst if we don't already have it */
258 inst
= acop_get_inst(regs
);
263 /* Check if the instruction is the "record form" */
266 * the instruction is "record" form so we can reject
269 regs
->ccr
&= ~(0xful
<< 28);
270 regs
->ccr
|= ICSWX_RC_NOT_FOUND
<< 28;
272 /* Move on to the next instruction */
276 * There is no architected mechanism to report a bad
277 * CT so we could either SIGILL or report nothing.
278 * Since the non-record version should only bu used
279 * for "hints" or "don't care" we should probably do
280 * nothing. However, I could see how some people
281 * might want an SIGILL so it here if you want it.
283 #ifdef CONFIG_PPC_ICSWX_USE_SIGILL
284 _exception(SIGILL
, regs
, ILL_ILLOPN
, address
);
292 EXPORT_SYMBOL_GPL(acop_handle_fault
);