1 // SPDX-License-Identifier: GPL-2.0
3 * Copyright IBM Corp. 2012
6 * Jan Glauber <jang@linux.vnet.ibm.com>
9 #define KMSG_COMPONENT "zpci"
10 #define pr_fmt(fmt) KMSG_COMPONENT ": " fmt
12 #include <linux/compat.h>
13 #include <linux/kernel.h>
14 #include <linux/miscdevice.h>
15 #include <linux/slab.h>
16 #include <linux/err.h>
17 #include <linux/delay.h>
18 #include <linux/pci.h>
19 #include <linux/uaccess.h>
20 #include <asm/pci_debug.h>
21 #include <asm/pci_clp.h>
22 #include <asm/compat.h>
24 #include <uapi/asm/clp.h>
28 static void update_uid_checking(bool new)
30 if (zpci_unique_uid
!= new)
31 zpci_dbg(1, "uid checking:%d\n", new);
33 zpci_unique_uid
= new;
36 static inline void zpci_err_clp(unsigned int rsp
, int rc
)
41 } __packed data
= {rsp
, rc
};
43 zpci_err_hex(&data
, sizeof(data
));
47 * Call Logical Processor with c=1, lps=0 and command 1
48 * to get the bit mask of installed logical processors
50 static inline int clp_get_ilp(unsigned long *ilp
)
56 " .insn rrf,0xb9a00000,%[mask],%[cmd],8,0\n"
61 : [cc
] "+d" (cc
), [mask
] "=d" (mask
) : [cmd
] "a" (1)
68 * Call Logical Processor with c=0, the give constant lps and an lpcb request.
70 static inline int clp_req(void *data
, unsigned int lps
)
72 struct { u8 _
[CLP_BLK_SIZE
]; } *req
= data
;
77 " .insn rrf,0xb9a00000,%[ign],%[req],0,%[lps]\n"
82 : [cc
] "+d" (cc
), [ign
] "=d" (ignored
), "+m" (*req
)
83 : [req
] "a" (req
), [lps
] "i" (lps
)
88 static void *clp_alloc_block(gfp_t gfp_mask
)
90 return (void *) __get_free_pages(gfp_mask
, get_order(CLP_BLK_SIZE
));
93 static void clp_free_block(void *ptr
)
95 free_pages((unsigned long) ptr
, get_order(CLP_BLK_SIZE
));
98 static void clp_store_query_pci_fngrp(struct zpci_dev
*zdev
,
99 struct clp_rsp_query_pci_grp
*response
)
101 zdev
->tlb_refresh
= response
->refresh
;
102 zdev
->dma_mask
= response
->dasm
;
103 zdev
->msi_addr
= response
->msia
;
104 zdev
->max_msi
= response
->noi
;
105 zdev
->fmb_update
= response
->mui
;
107 switch (response
->version
) {
109 zdev
->max_bus_speed
= PCIE_SPEED_5_0GT
;
112 zdev
->max_bus_speed
= PCI_SPEED_UNKNOWN
;
117 static int clp_query_pci_fngrp(struct zpci_dev
*zdev
, u8 pfgid
)
119 struct clp_req_rsp_query_pci_grp
*rrb
;
122 rrb
= clp_alloc_block(GFP_KERNEL
);
126 memset(rrb
, 0, sizeof(*rrb
));
127 rrb
->request
.hdr
.len
= sizeof(rrb
->request
);
128 rrb
->request
.hdr
.cmd
= CLP_QUERY_PCI_FNGRP
;
129 rrb
->response
.hdr
.len
= sizeof(rrb
->response
);
130 rrb
->request
.pfgid
= pfgid
;
132 rc
= clp_req(rrb
, CLP_LPS_PCI
);
133 if (!rc
&& rrb
->response
.hdr
.rsp
== CLP_RC_OK
)
134 clp_store_query_pci_fngrp(zdev
, &rrb
->response
);
136 zpci_err("Q PCI FGRP:\n");
137 zpci_err_clp(rrb
->response
.hdr
.rsp
, rc
);
144 static int clp_store_query_pci_fn(struct zpci_dev
*zdev
,
145 struct clp_rsp_query_pci
*response
)
149 for (i
= 0; i
< PCI_BAR_COUNT
; i
++) {
150 zdev
->bars
[i
].val
= le32_to_cpu(response
->bar
[i
]);
151 zdev
->bars
[i
].size
= response
->bar_size
[i
];
153 zdev
->start_dma
= response
->sdma
;
154 zdev
->end_dma
= response
->edma
;
155 zdev
->pchid
= response
->pchid
;
156 zdev
->pfgid
= response
->pfgid
;
157 zdev
->pft
= response
->pft
;
158 zdev
->vfn
= response
->vfn
;
159 zdev
->uid
= response
->uid
;
160 zdev
->fmb_length
= sizeof(u32
) * response
->fmb_len
;
162 memcpy(zdev
->pfip
, response
->pfip
, sizeof(zdev
->pfip
));
163 if (response
->util_str_avail
) {
164 memcpy(zdev
->util_str
, response
->util_str
,
165 sizeof(zdev
->util_str
));
171 static int clp_query_pci_fn(struct zpci_dev
*zdev
, u32 fh
)
173 struct clp_req_rsp_query_pci
*rrb
;
176 rrb
= clp_alloc_block(GFP_KERNEL
);
180 memset(rrb
, 0, sizeof(*rrb
));
181 rrb
->request
.hdr
.len
= sizeof(rrb
->request
);
182 rrb
->request
.hdr
.cmd
= CLP_QUERY_PCI_FN
;
183 rrb
->response
.hdr
.len
= sizeof(rrb
->response
);
184 rrb
->request
.fh
= fh
;
186 rc
= clp_req(rrb
, CLP_LPS_PCI
);
187 if (!rc
&& rrb
->response
.hdr
.rsp
== CLP_RC_OK
) {
188 rc
= clp_store_query_pci_fn(zdev
, &rrb
->response
);
191 rc
= clp_query_pci_fngrp(zdev
, rrb
->response
.pfgid
);
193 zpci_err("Q PCI FN:\n");
194 zpci_err_clp(rrb
->response
.hdr
.rsp
, rc
);
202 int clp_add_pci_device(u32 fid
, u32 fh
, int configured
)
204 struct zpci_dev
*zdev
;
207 zpci_dbg(3, "add fid:%x, fh:%x, c:%d\n", fid
, fh
, configured
);
208 zdev
= kzalloc(sizeof(*zdev
), GFP_KERNEL
);
215 /* Query function properties and update zdev */
216 rc
= clp_query_pci_fn(zdev
, fh
);
221 zdev
->state
= ZPCI_FN_STATE_CONFIGURED
;
223 zdev
->state
= ZPCI_FN_STATE_STANDBY
;
225 rc
= zpci_create_device(zdev
);
231 zpci_dbg(0, "add fid:%x, rc:%d\n", fid
, rc
);
237 * Enable/Disable a given PCI function defined by its function handle.
239 static int clp_set_pci_fn(u32
*fh
, u8 nr_dma_as
, u8 command
)
241 struct clp_req_rsp_set_pci
*rrb
;
242 int rc
, retries
= 100;
244 rrb
= clp_alloc_block(GFP_KERNEL
);
249 memset(rrb
, 0, sizeof(*rrb
));
250 rrb
->request
.hdr
.len
= sizeof(rrb
->request
);
251 rrb
->request
.hdr
.cmd
= CLP_SET_PCI_FN
;
252 rrb
->response
.hdr
.len
= sizeof(rrb
->response
);
253 rrb
->request
.fh
= *fh
;
254 rrb
->request
.oc
= command
;
255 rrb
->request
.ndas
= nr_dma_as
;
257 rc
= clp_req(rrb
, CLP_LPS_PCI
);
258 if (rrb
->response
.hdr
.rsp
== CLP_RC_SETPCIFN_BUSY
) {
264 } while (rrb
->response
.hdr
.rsp
== CLP_RC_SETPCIFN_BUSY
);
266 if (!rc
&& rrb
->response
.hdr
.rsp
== CLP_RC_OK
)
267 *fh
= rrb
->response
.fh
;
269 zpci_err("Set PCI FN:\n");
270 zpci_err_clp(rrb
->response
.hdr
.rsp
, rc
);
277 int clp_enable_fh(struct zpci_dev
*zdev
, u8 nr_dma_as
)
282 rc
= clp_set_pci_fn(&fh
, nr_dma_as
, CLP_SET_ENABLE_PCI_FN
);
284 /* Success -> store enabled handle in zdev */
287 zpci_dbg(3, "ena fid:%x, fh:%x, rc:%d\n", zdev
->fid
, zdev
->fh
, rc
);
291 int clp_disable_fh(struct zpci_dev
*zdev
)
296 if (!zdev_enabled(zdev
))
299 rc
= clp_set_pci_fn(&fh
, 0, CLP_SET_DISABLE_PCI_FN
);
301 /* Success -> store disabled handle in zdev */
304 zpci_dbg(3, "dis fid:%x, fh:%x, rc:%d\n", zdev
->fid
, zdev
->fh
, rc
);
308 static int clp_list_pci(struct clp_req_rsp_list_pci
*rrb
, void *data
,
309 void (*cb
)(struct clp_fh_list_entry
*, void *))
311 u64 resume_token
= 0;
315 memset(rrb
, 0, sizeof(*rrb
));
316 rrb
->request
.hdr
.len
= sizeof(rrb
->request
);
317 rrb
->request
.hdr
.cmd
= CLP_LIST_PCI
;
318 /* store as many entries as possible */
319 rrb
->response
.hdr
.len
= CLP_BLK_SIZE
- LIST_PCI_HDR_LEN
;
320 rrb
->request
.resume_token
= resume_token
;
322 /* Get PCI function handle list */
323 rc
= clp_req(rrb
, CLP_LPS_PCI
);
324 if (rc
|| rrb
->response
.hdr
.rsp
!= CLP_RC_OK
) {
325 zpci_err("List PCI FN:\n");
326 zpci_err_clp(rrb
->response
.hdr
.rsp
, rc
);
331 update_uid_checking(rrb
->response
.uid_checking
);
332 WARN_ON_ONCE(rrb
->response
.entry_size
!=
333 sizeof(struct clp_fh_list_entry
));
335 entries
= (rrb
->response
.hdr
.len
- LIST_PCI_HDR_LEN
) /
336 rrb
->response
.entry_size
;
338 resume_token
= rrb
->response
.resume_token
;
339 for (i
= 0; i
< entries
; i
++)
340 cb(&rrb
->response
.fh_list
[i
], data
);
341 } while (resume_token
);
346 static void __clp_add(struct clp_fh_list_entry
*entry
, void *data
)
348 struct zpci_dev
*zdev
;
350 if (!entry
->vendor_id
)
353 zdev
= get_zdev_by_fid(entry
->fid
);
355 clp_add_pci_device(entry
->fid
, entry
->fh
, entry
->config_state
);
358 static void __clp_update(struct clp_fh_list_entry
*entry
, void *data
)
360 struct zpci_dev
*zdev
;
362 if (!entry
->vendor_id
)
365 zdev
= get_zdev_by_fid(entry
->fid
);
369 zdev
->fh
= entry
->fh
;
372 int clp_scan_pci_devices(void)
374 struct clp_req_rsp_list_pci
*rrb
;
377 rrb
= clp_alloc_block(GFP_KERNEL
);
381 rc
= clp_list_pci(rrb
, NULL
, __clp_add
);
387 int clp_rescan_pci_devices(void)
389 struct clp_req_rsp_list_pci
*rrb
;
392 zpci_remove_reserved_devices();
394 rrb
= clp_alloc_block(GFP_KERNEL
);
398 rc
= clp_list_pci(rrb
, NULL
, __clp_add
);
404 int clp_rescan_pci_devices_simple(void)
406 struct clp_req_rsp_list_pci
*rrb
;
409 rrb
= clp_alloc_block(GFP_NOWAIT
);
413 rc
= clp_list_pci(rrb
, NULL
, __clp_update
);
419 struct clp_state_data
{
421 enum zpci_state state
;
424 static void __clp_get_state(struct clp_fh_list_entry
*entry
, void *data
)
426 struct clp_state_data
*sd
= data
;
428 if (entry
->fid
!= sd
->fid
)
431 sd
->state
= entry
->config_state
;
434 int clp_get_state(u32 fid
, enum zpci_state
*state
)
436 struct clp_req_rsp_list_pci
*rrb
;
437 struct clp_state_data sd
= {fid
, ZPCI_FN_STATE_RESERVED
};
440 rrb
= clp_alloc_block(GFP_ATOMIC
);
444 rc
= clp_list_pci(rrb
, &sd
, __clp_get_state
);
452 static int clp_base_slpc(struct clp_req
*req
, struct clp_req_rsp_slpc
*lpcb
)
454 unsigned long limit
= PAGE_SIZE
- sizeof(lpcb
->request
);
456 if (lpcb
->request
.hdr
.len
!= sizeof(lpcb
->request
) ||
457 lpcb
->response
.hdr
.len
> limit
)
459 return clp_req(lpcb
, CLP_LPS_BASE
) ? -EOPNOTSUPP
: 0;
462 static int clp_base_command(struct clp_req
*req
, struct clp_req_hdr
*lpcb
)
465 case 0x0001: /* store logical-processor characteristics */
466 return clp_base_slpc(req
, (void *) lpcb
);
472 static int clp_pci_slpc(struct clp_req
*req
, struct clp_req_rsp_slpc
*lpcb
)
474 unsigned long limit
= PAGE_SIZE
- sizeof(lpcb
->request
);
476 if (lpcb
->request
.hdr
.len
!= sizeof(lpcb
->request
) ||
477 lpcb
->response
.hdr
.len
> limit
)
479 return clp_req(lpcb
, CLP_LPS_PCI
) ? -EOPNOTSUPP
: 0;
482 static int clp_pci_list(struct clp_req
*req
, struct clp_req_rsp_list_pci
*lpcb
)
484 unsigned long limit
= PAGE_SIZE
- sizeof(lpcb
->request
);
486 if (lpcb
->request
.hdr
.len
!= sizeof(lpcb
->request
) ||
487 lpcb
->response
.hdr
.len
> limit
)
489 if (lpcb
->request
.reserved2
!= 0)
491 return clp_req(lpcb
, CLP_LPS_PCI
) ? -EOPNOTSUPP
: 0;
494 static int clp_pci_query(struct clp_req
*req
,
495 struct clp_req_rsp_query_pci
*lpcb
)
497 unsigned long limit
= PAGE_SIZE
- sizeof(lpcb
->request
);
499 if (lpcb
->request
.hdr
.len
!= sizeof(lpcb
->request
) ||
500 lpcb
->response
.hdr
.len
> limit
)
502 if (lpcb
->request
.reserved2
!= 0 || lpcb
->request
.reserved3
!= 0)
504 return clp_req(lpcb
, CLP_LPS_PCI
) ? -EOPNOTSUPP
: 0;
507 static int clp_pci_query_grp(struct clp_req
*req
,
508 struct clp_req_rsp_query_pci_grp
*lpcb
)
510 unsigned long limit
= PAGE_SIZE
- sizeof(lpcb
->request
);
512 if (lpcb
->request
.hdr
.len
!= sizeof(lpcb
->request
) ||
513 lpcb
->response
.hdr
.len
> limit
)
515 if (lpcb
->request
.reserved2
!= 0 || lpcb
->request
.reserved3
!= 0 ||
516 lpcb
->request
.reserved4
!= 0)
518 return clp_req(lpcb
, CLP_LPS_PCI
) ? -EOPNOTSUPP
: 0;
521 static int clp_pci_command(struct clp_req
*req
, struct clp_req_hdr
*lpcb
)
524 case 0x0001: /* store logical-processor characteristics */
525 return clp_pci_slpc(req
, (void *) lpcb
);
526 case 0x0002: /* list PCI functions */
527 return clp_pci_list(req
, (void *) lpcb
);
528 case 0x0003: /* query PCI function */
529 return clp_pci_query(req
, (void *) lpcb
);
530 case 0x0004: /* query PCI function group */
531 return clp_pci_query_grp(req
, (void *) lpcb
);
537 static int clp_normal_command(struct clp_req
*req
)
539 struct clp_req_hdr
*lpcb
;
544 if (req
->lps
!= 0 && req
->lps
!= 2)
548 lpcb
= clp_alloc_block(GFP_KERNEL
);
553 uptr
= (void __force __user
*)(unsigned long) req
->data_p
;
554 if (copy_from_user(lpcb
, uptr
, PAGE_SIZE
) != 0)
558 if (lpcb
->fmt
!= 0 || lpcb
->reserved1
!= 0 || lpcb
->reserved2
!= 0)
563 rc
= clp_base_command(req
, lpcb
);
566 rc
= clp_pci_command(req
, lpcb
);
573 if (copy_to_user(uptr
, lpcb
, PAGE_SIZE
) != 0)
579 clp_free_block(lpcb
);
584 static int clp_immediate_command(struct clp_req
*req
)
590 if (req
->cmd
> 1 || clp_get_ilp(&ilp
) != 0)
593 uptr
= (void __force __user
*)(unsigned long) req
->data_p
;
595 /* Command code 0: test for a specific processor */
596 exists
= test_bit_inv(req
->lps
, &ilp
);
597 return put_user(exists
, (int __user
*) uptr
);
599 /* Command code 1: return bit mask of installed processors */
600 return put_user(ilp
, (unsigned long __user
*) uptr
);
603 static long clp_misc_ioctl(struct file
*filp
, unsigned int cmd
,
612 argp
= is_compat_task() ? compat_ptr(arg
) : (void __user
*) arg
;
613 if (copy_from_user(&req
, argp
, sizeof(req
)))
617 return req
.c
? clp_immediate_command(&req
) : clp_normal_command(&req
);
620 static int clp_misc_release(struct inode
*inode
, struct file
*filp
)
625 static const struct file_operations clp_misc_fops
= {
626 .owner
= THIS_MODULE
,
627 .open
= nonseekable_open
,
628 .release
= clp_misc_release
,
629 .unlocked_ioctl
= clp_misc_ioctl
,
630 .compat_ioctl
= clp_misc_ioctl
,
634 static struct miscdevice clp_misc_device
= {
635 .minor
= MISC_DYNAMIC_MINOR
,
637 .fops
= &clp_misc_fops
,
640 static int __init
clp_misc_init(void)
642 return misc_register(&clp_misc_device
);
645 device_initcall(clp_misc_init
);