1 // SPDX-License-Identifier: GPL-2.0
3 * Copyright IBM Corp. 2012
6 * Jan Glauber <jang@linux.vnet.ibm.com>
9 #define KMSG_COMPONENT "zpci"
10 #define pr_fmt(fmt) KMSG_COMPONENT ": " fmt
12 #include <linux/compat.h>
13 #include <linux/kernel.h>
14 #include <linux/miscdevice.h>
15 #include <linux/slab.h>
16 #include <linux/err.h>
17 #include <linux/delay.h>
18 #include <linux/pci.h>
19 #include <linux/uaccess.h>
20 #include <asm/pci_debug.h>
21 #include <asm/pci_clp.h>
23 #include <uapi/asm/clp.h>
27 void update_uid_checking(bool new)
29 if (zpci_unique_uid
!= new)
30 zpci_dbg(1, "uid checking:%d\n", new);
32 zpci_unique_uid
= new;
35 static inline void zpci_err_clp(unsigned int rsp
, int rc
)
40 } __packed data
= {rsp
, rc
};
42 zpci_err_hex(&data
, sizeof(data
));
46 * Call Logical Processor with c=1, lps=0 and command 1
47 * to get the bit mask of installed logical processors
49 static inline int clp_get_ilp(unsigned long *ilp
)
55 " .insn rrf,0xb9a00000,%[mask],%[cmd],8,0\n"
60 : [cc
] "+d" (cc
), [mask
] "=d" (mask
) : [cmd
] "a" (1)
67 * Call Logical Processor with c=0, the give constant lps and an lpcb request.
69 static __always_inline
int clp_req(void *data
, unsigned int lps
)
71 struct { u8 _
[CLP_BLK_SIZE
]; } *req
= data
;
76 " .insn rrf,0xb9a00000,%[ign],%[req],0,%[lps]\n"
81 : [cc
] "+d" (cc
), [ign
] "=d" (ignored
), "+m" (*req
)
82 : [req
] "a" (req
), [lps
] "i" (lps
)
87 static void *clp_alloc_block(gfp_t gfp_mask
)
89 return (void *) __get_free_pages(gfp_mask
, get_order(CLP_BLK_SIZE
));
92 static void clp_free_block(void *ptr
)
94 free_pages((unsigned long) ptr
, get_order(CLP_BLK_SIZE
));
97 static void clp_store_query_pci_fngrp(struct zpci_dev
*zdev
,
98 struct clp_rsp_query_pci_grp
*response
)
100 zdev
->tlb_refresh
= response
->refresh
;
101 zdev
->dma_mask
= response
->dasm
;
102 zdev
->msi_addr
= response
->msia
;
103 zdev
->max_msi
= response
->noi
;
104 zdev
->fmb_update
= response
->mui
;
106 switch (response
->version
) {
108 zdev
->max_bus_speed
= PCIE_SPEED_5_0GT
;
111 zdev
->max_bus_speed
= PCI_SPEED_UNKNOWN
;
116 static int clp_query_pci_fngrp(struct zpci_dev
*zdev
, u8 pfgid
)
118 struct clp_req_rsp_query_pci_grp
*rrb
;
121 rrb
= clp_alloc_block(GFP_KERNEL
);
125 memset(rrb
, 0, sizeof(*rrb
));
126 rrb
->request
.hdr
.len
= sizeof(rrb
->request
);
127 rrb
->request
.hdr
.cmd
= CLP_QUERY_PCI_FNGRP
;
128 rrb
->response
.hdr
.len
= sizeof(rrb
->response
);
129 rrb
->request
.pfgid
= pfgid
;
131 rc
= clp_req(rrb
, CLP_LPS_PCI
);
132 if (!rc
&& rrb
->response
.hdr
.rsp
== CLP_RC_OK
)
133 clp_store_query_pci_fngrp(zdev
, &rrb
->response
);
135 zpci_err("Q PCI FGRP:\n");
136 zpci_err_clp(rrb
->response
.hdr
.rsp
, rc
);
143 static int clp_store_query_pci_fn(struct zpci_dev
*zdev
,
144 struct clp_rsp_query_pci
*response
)
148 for (i
= 0; i
< PCI_STD_NUM_BARS
; i
++) {
149 zdev
->bars
[i
].val
= le32_to_cpu(response
->bar
[i
]);
150 zdev
->bars
[i
].size
= response
->bar_size
[i
];
152 zdev
->start_dma
= response
->sdma
;
153 zdev
->end_dma
= response
->edma
;
154 zdev
->pchid
= response
->pchid
;
155 zdev
->pfgid
= response
->pfgid
;
156 zdev
->pft
= response
->pft
;
157 zdev
->vfn
= response
->vfn
;
158 zdev
->uid
= response
->uid
;
159 zdev
->fmb_length
= sizeof(u32
) * response
->fmb_len
;
161 memcpy(zdev
->pfip
, response
->pfip
, sizeof(zdev
->pfip
));
162 if (response
->util_str_avail
) {
163 memcpy(zdev
->util_str
, response
->util_str
,
164 sizeof(zdev
->util_str
));
166 zdev
->mio_capable
= response
->mio_addr_avail
;
167 for (i
= 0; i
< PCI_STD_NUM_BARS
; i
++) {
168 if (!(response
->mio
.valid
& (1 << (PCI_STD_NUM_BARS
- i
- 1))))
171 zdev
->bars
[i
].mio_wb
= (void __iomem
*) response
->mio
.addr
[i
].wb
;
172 zdev
->bars
[i
].mio_wt
= (void __iomem
*) response
->mio
.addr
[i
].wt
;
177 static int clp_query_pci_fn(struct zpci_dev
*zdev
, u32 fh
)
179 struct clp_req_rsp_query_pci
*rrb
;
182 rrb
= clp_alloc_block(GFP_KERNEL
);
186 memset(rrb
, 0, sizeof(*rrb
));
187 rrb
->request
.hdr
.len
= sizeof(rrb
->request
);
188 rrb
->request
.hdr
.cmd
= CLP_QUERY_PCI_FN
;
189 rrb
->response
.hdr
.len
= sizeof(rrb
->response
);
190 rrb
->request
.fh
= fh
;
192 rc
= clp_req(rrb
, CLP_LPS_PCI
);
193 if (!rc
&& rrb
->response
.hdr
.rsp
== CLP_RC_OK
) {
194 rc
= clp_store_query_pci_fn(zdev
, &rrb
->response
);
197 rc
= clp_query_pci_fngrp(zdev
, rrb
->response
.pfgid
);
199 zpci_err("Q PCI FN:\n");
200 zpci_err_clp(rrb
->response
.hdr
.rsp
, rc
);
208 int clp_add_pci_device(u32 fid
, u32 fh
, int configured
)
210 struct zpci_dev
*zdev
;
213 zpci_dbg(3, "add fid:%x, fh:%x, c:%d\n", fid
, fh
, configured
);
214 zdev
= kzalloc(sizeof(*zdev
), GFP_KERNEL
);
221 /* Query function properties and update zdev */
222 rc
= clp_query_pci_fn(zdev
, fh
);
227 zdev
->state
= ZPCI_FN_STATE_CONFIGURED
;
229 zdev
->state
= ZPCI_FN_STATE_STANDBY
;
231 rc
= zpci_create_device(zdev
);
237 zpci_dbg(0, "add fid:%x, rc:%d\n", fid
, rc
);
243 * Enable/Disable a given PCI function and update its function handle if
246 static int clp_set_pci_fn(struct zpci_dev
*zdev
, u8 nr_dma_as
, u8 command
)
248 struct clp_req_rsp_set_pci
*rrb
;
249 int rc
, retries
= 100;
252 rrb
= clp_alloc_block(GFP_KERNEL
);
257 memset(rrb
, 0, sizeof(*rrb
));
258 rrb
->request
.hdr
.len
= sizeof(rrb
->request
);
259 rrb
->request
.hdr
.cmd
= CLP_SET_PCI_FN
;
260 rrb
->response
.hdr
.len
= sizeof(rrb
->response
);
261 rrb
->request
.fh
= zdev
->fh
;
262 rrb
->request
.oc
= command
;
263 rrb
->request
.ndas
= nr_dma_as
;
265 rc
= clp_req(rrb
, CLP_LPS_PCI
);
266 if (rrb
->response
.hdr
.rsp
== CLP_RC_SETPCIFN_BUSY
) {
272 } while (rrb
->response
.hdr
.rsp
== CLP_RC_SETPCIFN_BUSY
);
274 if (rc
|| rrb
->response
.hdr
.rsp
!= CLP_RC_OK
) {
275 zpci_err("Set PCI FN:\n");
276 zpci_err_clp(rrb
->response
.hdr
.rsp
, rc
);
279 if (!rc
&& rrb
->response
.hdr
.rsp
== CLP_RC_OK
) {
280 zdev
->fh
= rrb
->response
.fh
;
281 } else if (!rc
&& rrb
->response
.hdr
.rsp
== CLP_RC_SETPCIFN_ALRDY
&&
282 rrb
->response
.fh
== 0) {
283 /* Function is already in desired state - update handle */
284 rc
= clp_rescan_pci_devices_simple(&fid
);
290 int clp_enable_fh(struct zpci_dev
*zdev
, u8 nr_dma_as
)
294 rc
= clp_set_pci_fn(zdev
, nr_dma_as
, CLP_SET_ENABLE_PCI_FN
);
295 zpci_dbg(3, "ena fid:%x, fh:%x, rc:%d\n", zdev
->fid
, zdev
->fh
, rc
);
299 if (zpci_use_mio(zdev
)) {
300 rc
= clp_set_pci_fn(zdev
, nr_dma_as
, CLP_SET_ENABLE_MIO
);
301 zpci_dbg(3, "ena mio fid:%x, fh:%x, rc:%d\n",
302 zdev
->fid
, zdev
->fh
, rc
);
304 clp_disable_fh(zdev
);
310 int clp_disable_fh(struct zpci_dev
*zdev
)
314 if (!zdev_enabled(zdev
))
317 rc
= clp_set_pci_fn(zdev
, 0, CLP_SET_DISABLE_PCI_FN
);
318 zpci_dbg(3, "dis fid:%x, fh:%x, rc:%d\n", zdev
->fid
, zdev
->fh
, rc
);
322 static int clp_list_pci(struct clp_req_rsp_list_pci
*rrb
, void *data
,
323 void (*cb
)(struct clp_fh_list_entry
*, void *))
325 u64 resume_token
= 0;
329 memset(rrb
, 0, sizeof(*rrb
));
330 rrb
->request
.hdr
.len
= sizeof(rrb
->request
);
331 rrb
->request
.hdr
.cmd
= CLP_LIST_PCI
;
332 /* store as many entries as possible */
333 rrb
->response
.hdr
.len
= CLP_BLK_SIZE
- LIST_PCI_HDR_LEN
;
334 rrb
->request
.resume_token
= resume_token
;
336 /* Get PCI function handle list */
337 rc
= clp_req(rrb
, CLP_LPS_PCI
);
338 if (rc
|| rrb
->response
.hdr
.rsp
!= CLP_RC_OK
) {
339 zpci_err("List PCI FN:\n");
340 zpci_err_clp(rrb
->response
.hdr
.rsp
, rc
);
345 update_uid_checking(rrb
->response
.uid_checking
);
346 WARN_ON_ONCE(rrb
->response
.entry_size
!=
347 sizeof(struct clp_fh_list_entry
));
349 entries
= (rrb
->response
.hdr
.len
- LIST_PCI_HDR_LEN
) /
350 rrb
->response
.entry_size
;
352 resume_token
= rrb
->response
.resume_token
;
353 for (i
= 0; i
< entries
; i
++)
354 cb(&rrb
->response
.fh_list
[i
], data
);
355 } while (resume_token
);
360 static void __clp_add(struct clp_fh_list_entry
*entry
, void *data
)
362 struct zpci_dev
*zdev
;
364 if (!entry
->vendor_id
)
367 zdev
= get_zdev_by_fid(entry
->fid
);
369 clp_add_pci_device(entry
->fid
, entry
->fh
, entry
->config_state
);
372 static void __clp_update(struct clp_fh_list_entry
*entry
, void *data
)
374 struct zpci_dev
*zdev
;
377 if (!entry
->vendor_id
)
380 if (fid
&& *fid
!= entry
->fid
)
383 zdev
= get_zdev_by_fid(entry
->fid
);
387 zdev
->fh
= entry
->fh
;
390 int clp_scan_pci_devices(void)
392 struct clp_req_rsp_list_pci
*rrb
;
395 rrb
= clp_alloc_block(GFP_KERNEL
);
399 rc
= clp_list_pci(rrb
, NULL
, __clp_add
);
405 int clp_rescan_pci_devices(void)
407 struct clp_req_rsp_list_pci
*rrb
;
410 zpci_remove_reserved_devices();
412 rrb
= clp_alloc_block(GFP_KERNEL
);
416 rc
= clp_list_pci(rrb
, NULL
, __clp_add
);
422 /* Rescan PCI functions and refresh function handles. If fid is non-NULL only
423 * refresh the handle of the function matching @fid
425 int clp_rescan_pci_devices_simple(u32
*fid
)
427 struct clp_req_rsp_list_pci
*rrb
;
430 rrb
= clp_alloc_block(GFP_NOWAIT
);
434 rc
= clp_list_pci(rrb
, fid
, __clp_update
);
440 struct clp_state_data
{
442 enum zpci_state state
;
445 static void __clp_get_state(struct clp_fh_list_entry
*entry
, void *data
)
447 struct clp_state_data
*sd
= data
;
449 if (entry
->fid
!= sd
->fid
)
452 sd
->state
= entry
->config_state
;
455 int clp_get_state(u32 fid
, enum zpci_state
*state
)
457 struct clp_req_rsp_list_pci
*rrb
;
458 struct clp_state_data sd
= {fid
, ZPCI_FN_STATE_RESERVED
};
461 rrb
= clp_alloc_block(GFP_ATOMIC
);
465 rc
= clp_list_pci(rrb
, &sd
, __clp_get_state
);
473 static int clp_base_slpc(struct clp_req
*req
, struct clp_req_rsp_slpc
*lpcb
)
475 unsigned long limit
= PAGE_SIZE
- sizeof(lpcb
->request
);
477 if (lpcb
->request
.hdr
.len
!= sizeof(lpcb
->request
) ||
478 lpcb
->response
.hdr
.len
> limit
)
480 return clp_req(lpcb
, CLP_LPS_BASE
) ? -EOPNOTSUPP
: 0;
483 static int clp_base_command(struct clp_req
*req
, struct clp_req_hdr
*lpcb
)
486 case 0x0001: /* store logical-processor characteristics */
487 return clp_base_slpc(req
, (void *) lpcb
);
493 static int clp_pci_slpc(struct clp_req
*req
, struct clp_req_rsp_slpc
*lpcb
)
495 unsigned long limit
= PAGE_SIZE
- sizeof(lpcb
->request
);
497 if (lpcb
->request
.hdr
.len
!= sizeof(lpcb
->request
) ||
498 lpcb
->response
.hdr
.len
> limit
)
500 return clp_req(lpcb
, CLP_LPS_PCI
) ? -EOPNOTSUPP
: 0;
503 static int clp_pci_list(struct clp_req
*req
, struct clp_req_rsp_list_pci
*lpcb
)
505 unsigned long limit
= PAGE_SIZE
- sizeof(lpcb
->request
);
507 if (lpcb
->request
.hdr
.len
!= sizeof(lpcb
->request
) ||
508 lpcb
->response
.hdr
.len
> limit
)
510 if (lpcb
->request
.reserved2
!= 0)
512 return clp_req(lpcb
, CLP_LPS_PCI
) ? -EOPNOTSUPP
: 0;
515 static int clp_pci_query(struct clp_req
*req
,
516 struct clp_req_rsp_query_pci
*lpcb
)
518 unsigned long limit
= PAGE_SIZE
- sizeof(lpcb
->request
);
520 if (lpcb
->request
.hdr
.len
!= sizeof(lpcb
->request
) ||
521 lpcb
->response
.hdr
.len
> limit
)
523 if (lpcb
->request
.reserved2
!= 0 || lpcb
->request
.reserved3
!= 0)
525 return clp_req(lpcb
, CLP_LPS_PCI
) ? -EOPNOTSUPP
: 0;
528 static int clp_pci_query_grp(struct clp_req
*req
,
529 struct clp_req_rsp_query_pci_grp
*lpcb
)
531 unsigned long limit
= PAGE_SIZE
- sizeof(lpcb
->request
);
533 if (lpcb
->request
.hdr
.len
!= sizeof(lpcb
->request
) ||
534 lpcb
->response
.hdr
.len
> limit
)
536 if (lpcb
->request
.reserved2
!= 0 || lpcb
->request
.reserved3
!= 0 ||
537 lpcb
->request
.reserved4
!= 0)
539 return clp_req(lpcb
, CLP_LPS_PCI
) ? -EOPNOTSUPP
: 0;
542 static int clp_pci_command(struct clp_req
*req
, struct clp_req_hdr
*lpcb
)
545 case 0x0001: /* store logical-processor characteristics */
546 return clp_pci_slpc(req
, (void *) lpcb
);
547 case 0x0002: /* list PCI functions */
548 return clp_pci_list(req
, (void *) lpcb
);
549 case 0x0003: /* query PCI function */
550 return clp_pci_query(req
, (void *) lpcb
);
551 case 0x0004: /* query PCI function group */
552 return clp_pci_query_grp(req
, (void *) lpcb
);
558 static int clp_normal_command(struct clp_req
*req
)
560 struct clp_req_hdr
*lpcb
;
565 if (req
->lps
!= 0 && req
->lps
!= 2)
569 lpcb
= clp_alloc_block(GFP_KERNEL
);
574 uptr
= (void __force __user
*)(unsigned long) req
->data_p
;
575 if (copy_from_user(lpcb
, uptr
, PAGE_SIZE
) != 0)
579 if (lpcb
->fmt
!= 0 || lpcb
->reserved1
!= 0 || lpcb
->reserved2
!= 0)
584 rc
= clp_base_command(req
, lpcb
);
587 rc
= clp_pci_command(req
, lpcb
);
594 if (copy_to_user(uptr
, lpcb
, PAGE_SIZE
) != 0)
600 clp_free_block(lpcb
);
605 static int clp_immediate_command(struct clp_req
*req
)
611 if (req
->cmd
> 1 || clp_get_ilp(&ilp
) != 0)
614 uptr
= (void __force __user
*)(unsigned long) req
->data_p
;
616 /* Command code 0: test for a specific processor */
617 exists
= test_bit_inv(req
->lps
, &ilp
);
618 return put_user(exists
, (int __user
*) uptr
);
620 /* Command code 1: return bit mask of installed processors */
621 return put_user(ilp
, (unsigned long __user
*) uptr
);
624 static long clp_misc_ioctl(struct file
*filp
, unsigned int cmd
,
633 argp
= is_compat_task() ? compat_ptr(arg
) : (void __user
*) arg
;
634 if (copy_from_user(&req
, argp
, sizeof(req
)))
638 return req
.c
? clp_immediate_command(&req
) : clp_normal_command(&req
);
641 static int clp_misc_release(struct inode
*inode
, struct file
*filp
)
646 static const struct file_operations clp_misc_fops
= {
647 .owner
= THIS_MODULE
,
648 .open
= nonseekable_open
,
649 .release
= clp_misc_release
,
650 .unlocked_ioctl
= clp_misc_ioctl
,
651 .compat_ioctl
= clp_misc_ioctl
,
655 static struct miscdevice clp_misc_device
= {
656 .minor
= MISC_DYNAMIC_MINOR
,
658 .fops
= &clp_misc_fops
,
661 static int __init
clp_misc_init(void)
663 return misc_register(&clp_misc_device
);
666 device_initcall(clp_misc_init
);