2 * This file is subject to the terms and conditions of the GNU General Public
3 * License. See the file "COPYING" in the main directory of this archive
6 * Copyright (c) 2004-2005 Silicon Graphics, Inc. All Rights Reserved.
11 * Cross Partition (XP) base.
13 * XP provides a base from which its users can interact
14 * with XPC, yet not be dependent on XPC.
19 #include <linux/kernel.h>
20 #include <linux/interrupt.h>
21 #include <linux/module.h>
22 #include <linux/mutex.h>
23 #include <asm/sn/intr.h>
24 #include <asm/sn/sn_sal.h>
25 #include <asm/sn/xp.h>
29 * Target of nofault PIO read.
31 u64 xp_nofault_PIOR_target
;
35 * xpc_registrations[] keeps track of xpc_connect()'s done by the kernel-level
38 struct xpc_registration xpc_registrations
[XPC_NCHANNELS
];
42 * Initialize the XPC interface to indicate that XPC isn't loaded.
44 static enum xpc_retval
xpc_notloaded(void) { return xpcNotLoaded
; }
46 struct xpc_interface xpc_interface
= {
47 (void (*)(int)) xpc_notloaded
,
48 (void (*)(int)) xpc_notloaded
,
49 (enum xpc_retval (*)(partid_t
, int, u32
, void **)) xpc_notloaded
,
50 (enum xpc_retval (*)(partid_t
, int, void *)) xpc_notloaded
,
51 (enum xpc_retval (*)(partid_t
, int, void *, xpc_notify_func
, void *))
53 (void (*)(partid_t
, int, void *)) xpc_notloaded
,
54 (enum xpc_retval (*)(partid_t
, void *)) xpc_notloaded
59 * XPC calls this when it (the XPC module) has been loaded.
62 xpc_set_interface(void (*connect
)(int),
63 void (*disconnect
)(int),
64 enum xpc_retval (*allocate
)(partid_t
, int, u32
, void **),
65 enum xpc_retval (*send
)(partid_t
, int, void *),
66 enum xpc_retval (*send_notify
)(partid_t
, int, void *,
67 xpc_notify_func
, void *),
68 void (*received
)(partid_t
, int, void *),
69 enum xpc_retval (*partid_to_nasids
)(partid_t
, void *))
71 xpc_interface
.connect
= connect
;
72 xpc_interface
.disconnect
= disconnect
;
73 xpc_interface
.allocate
= allocate
;
74 xpc_interface
.send
= send
;
75 xpc_interface
.send_notify
= send_notify
;
76 xpc_interface
.received
= received
;
77 xpc_interface
.partid_to_nasids
= partid_to_nasids
;
82 * XPC calls this when it (the XPC module) is being unloaded.
85 xpc_clear_interface(void)
87 xpc_interface
.connect
= (void (*)(int)) xpc_notloaded
;
88 xpc_interface
.disconnect
= (void (*)(int)) xpc_notloaded
;
89 xpc_interface
.allocate
= (enum xpc_retval (*)(partid_t
, int, u32
,
90 void **)) xpc_notloaded
;
91 xpc_interface
.send
= (enum xpc_retval (*)(partid_t
, int, void *))
93 xpc_interface
.send_notify
= (enum xpc_retval (*)(partid_t
, int, void *,
94 xpc_notify_func
, void *)) xpc_notloaded
;
95 xpc_interface
.received
= (void (*)(partid_t
, int, void *))
97 xpc_interface
.partid_to_nasids
= (enum xpc_retval (*)(partid_t
, void *))
103 * Register for automatic establishment of a channel connection whenever
104 * a partition comes up.
108 * ch_number - channel # to register for connection.
109 * func - function to call for asynchronous notification of channel
110 * state changes (i.e., connection, disconnection, error) and
111 * the arrival of incoming messages.
112 * key - pointer to optional user-defined value that gets passed back
113 * to the user on any callouts made to func.
114 * payload_size - size in bytes of the XPC message's payload area which
115 * contains a user-defined message. The user should make
116 * this large enough to hold their largest message.
117 * nentries - max #of XPC message entries a message queue can contain.
118 * The actual number, which is determined when a connection
119 * is established and may be less then requested, will be
120 * passed to the user via the xpcConnected callout.
121 * assigned_limit - max number of kthreads allowed to be processing
122 * messages (per connection) at any given instant.
123 * idle_limit - max number of kthreads allowed to be idle at any given
127 xpc_connect(int ch_number
, xpc_channel_func func
, void *key
, u16 payload_size
,
128 u16 nentries
, u32 assigned_limit
, u32 idle_limit
)
130 struct xpc_registration
*registration
;
133 DBUG_ON(ch_number
< 0 || ch_number
>= XPC_NCHANNELS
);
134 DBUG_ON(payload_size
== 0 || nentries
== 0);
135 DBUG_ON(func
== NULL
);
136 DBUG_ON(assigned_limit
== 0 || idle_limit
> assigned_limit
);
138 registration
= &xpc_registrations
[ch_number
];
140 if (mutex_lock_interruptible(®istration
->mutex
) != 0) {
141 return xpcInterrupted
;
144 /* if XPC_CHANNEL_REGISTERED(ch_number) */
145 if (registration
->func
!= NULL
) {
146 mutex_unlock(®istration
->mutex
);
147 return xpcAlreadyRegistered
;
150 /* register the channel for connection */
151 registration
->msg_size
= XPC_MSG_SIZE(payload_size
);
152 registration
->nentries
= nentries
;
153 registration
->assigned_limit
= assigned_limit
;
154 registration
->idle_limit
= idle_limit
;
155 registration
->key
= key
;
156 registration
->func
= func
;
158 mutex_unlock(®istration
->mutex
);
160 xpc_interface
.connect(ch_number
);
167 * Remove the registration for automatic connection of the specified channel
168 * when a partition comes up.
170 * Before returning this xpc_disconnect() will wait for all connections on the
171 * specified channel have been closed/torndown. So the caller can be assured
172 * that they will not be receiving any more callouts from XPC to their
173 * function registered via xpc_connect().
177 * ch_number - channel # to unregister.
180 xpc_disconnect(int ch_number
)
182 struct xpc_registration
*registration
;
185 DBUG_ON(ch_number
< 0 || ch_number
>= XPC_NCHANNELS
);
187 registration
= &xpc_registrations
[ch_number
];
190 * We've decided not to make this a down_interruptible(), since we
191 * figured XPC's users will just turn around and call xpc_disconnect()
192 * again anyways, so we might as well wait, if need be.
194 mutex_lock(®istration
->mutex
);
196 /* if !XPC_CHANNEL_REGISTERED(ch_number) */
197 if (registration
->func
== NULL
) {
198 mutex_unlock(®istration
->mutex
);
202 /* remove the connection registration for the specified channel */
203 registration
->func
= NULL
;
204 registration
->key
= NULL
;
205 registration
->nentries
= 0;
206 registration
->msg_size
= 0;
207 registration
->assigned_limit
= 0;
208 registration
->idle_limit
= 0;
210 xpc_interface
.disconnect(ch_number
);
212 mutex_unlock(®istration
->mutex
);
222 u64 func_addr
= *(u64
*) xp_nofault_PIOR
;
223 u64 err_func_addr
= *(u64
*) xp_error_PIOR
;
226 if (!ia64_platform_is("sn2")) {
231 * Register a nofault code region which performs a cross-partition
232 * PIO read. If the PIO read times out, the MCA handler will consume
233 * the error and return to a kernel-provided instruction to indicate
234 * an error. This PIO read exists because it is guaranteed to timeout
235 * if the destination is down (AMO operations do not timeout on at
236 * least some CPUs on Shubs <= v1.2, which unfortunately we have to
239 if ((ret
= sn_register_nofault_code(func_addr
, err_func_addr
,
240 err_func_addr
, 1, 1)) != 0) {
241 printk(KERN_ERR
"XP: can't register nofault code, error=%d\n",
245 * Setup the nofault PIO read target. (There is no special reason why
246 * SH_IPI_ACCESS was selected.)
249 xp_nofault_PIOR_target
= SH2_IPI_ACCESS0
;
251 xp_nofault_PIOR_target
= SH1_IPI_ACCESS
;
254 /* initialize the connection registration mutex */
255 for (ch_number
= 0; ch_number
< XPC_NCHANNELS
; ch_number
++) {
256 mutex_init(&xpc_registrations
[ch_number
].mutex
);
261 module_init(xp_init
);
267 u64 func_addr
= *(u64
*) xp_nofault_PIOR
;
268 u64 err_func_addr
= *(u64
*) xp_error_PIOR
;
271 /* unregister the PIO read nofault code region */
272 (void) sn_register_nofault_code(func_addr
, err_func_addr
,
273 err_func_addr
, 1, 0);
275 module_exit(xp_exit
);
278 MODULE_AUTHOR("Silicon Graphics, Inc.");
279 MODULE_DESCRIPTION("Cross Partition (XP) base");
280 MODULE_LICENSE("GPL");
282 EXPORT_SYMBOL(xp_nofault_PIOR
);
283 EXPORT_SYMBOL(xp_nofault_PIOR_target
);
284 EXPORT_SYMBOL(xpc_registrations
);
285 EXPORT_SYMBOL(xpc_interface
);
286 EXPORT_SYMBOL(xpc_clear_interface
);
287 EXPORT_SYMBOL(xpc_set_interface
);
288 EXPORT_SYMBOL(xpc_connect
);
289 EXPORT_SYMBOL(xpc_disconnect
);