1 // SPDX-License-Identifier: GPL-2.0
3 * Microsemi Switchtec(tm) PCIe Management Driver
4 * Copyright (c) 2017, Microsemi Corporation
7 #include <linux/switchtec.h>
8 #include <linux/switchtec_ioctl.h>
10 #include <linux/interrupt.h>
11 #include <linux/module.h>
13 #include <linux/uaccess.h>
14 #include <linux/poll.h>
15 #include <linux/wait.h>
17 MODULE_DESCRIPTION("Microsemi Switchtec(tm) PCIe Management Driver");
18 MODULE_VERSION("0.1");
19 MODULE_LICENSE("GPL");
20 MODULE_AUTHOR("Microsemi Corporation");
22 static int max_devices
= 16;
23 module_param(max_devices
, int, 0644);
24 MODULE_PARM_DESC(max_devices
, "max number of switchtec device instances");
26 static dev_t switchtec_devt
;
27 static DEFINE_IDA(switchtec_minor_ida
);
29 struct class *switchtec_class
;
30 EXPORT_SYMBOL_GPL(switchtec_class
);
39 struct switchtec_user
{
40 struct switchtec_dev
*stdev
;
42 enum mrpc_state state
;
44 struct completion comp
;
46 struct list_head list
;
53 unsigned char data
[SWITCHTEC_MRPC_PAYLOAD_SIZE
];
57 static struct switchtec_user
*stuser_create(struct switchtec_dev
*stdev
)
59 struct switchtec_user
*stuser
;
61 stuser
= kzalloc(sizeof(*stuser
), GFP_KERNEL
);
63 return ERR_PTR(-ENOMEM
);
65 get_device(&stdev
->dev
);
66 stuser
->stdev
= stdev
;
67 kref_init(&stuser
->kref
);
68 INIT_LIST_HEAD(&stuser
->list
);
69 init_completion(&stuser
->comp
);
70 stuser
->event_cnt
= atomic_read(&stdev
->event_cnt
);
72 dev_dbg(&stdev
->dev
, "%s: %p\n", __func__
, stuser
);
77 static void stuser_free(struct kref
*kref
)
79 struct switchtec_user
*stuser
;
81 stuser
= container_of(kref
, struct switchtec_user
, kref
);
83 dev_dbg(&stuser
->stdev
->dev
, "%s: %p\n", __func__
, stuser
);
85 put_device(&stuser
->stdev
->dev
);
89 static void stuser_put(struct switchtec_user
*stuser
)
91 kref_put(&stuser
->kref
, stuser_free
);
94 static void stuser_set_state(struct switchtec_user
*stuser
,
95 enum mrpc_state state
)
97 /* requires the mrpc_mutex to already be held when called */
99 const char * const state_names
[] = {
100 [MRPC_IDLE
] = "IDLE",
101 [MRPC_QUEUED
] = "QUEUED",
102 [MRPC_RUNNING
] = "RUNNING",
103 [MRPC_DONE
] = "DONE",
106 stuser
->state
= state
;
108 dev_dbg(&stuser
->stdev
->dev
, "stuser state %p -> %s",
109 stuser
, state_names
[state
]);
112 static void mrpc_complete_cmd(struct switchtec_dev
*stdev
);
114 static void mrpc_cmd_submit(struct switchtec_dev
*stdev
)
116 /* requires the mrpc_mutex to already be held when called */
118 struct switchtec_user
*stuser
;
120 if (stdev
->mrpc_busy
)
123 if (list_empty(&stdev
->mrpc_queue
))
126 stuser
= list_entry(stdev
->mrpc_queue
.next
, struct switchtec_user
,
129 stuser_set_state(stuser
, MRPC_RUNNING
);
130 stdev
->mrpc_busy
= 1;
131 memcpy_toio(&stdev
->mmio_mrpc
->input_data
,
132 stuser
->data
, stuser
->data_len
);
133 iowrite32(stuser
->cmd
, &stdev
->mmio_mrpc
->cmd
);
135 stuser
->status
= ioread32(&stdev
->mmio_mrpc
->status
);
136 if (stuser
->status
!= SWITCHTEC_MRPC_STATUS_INPROGRESS
)
137 mrpc_complete_cmd(stdev
);
139 schedule_delayed_work(&stdev
->mrpc_timeout
,
140 msecs_to_jiffies(500));
143 static int mrpc_queue_cmd(struct switchtec_user
*stuser
)
145 /* requires the mrpc_mutex to already be held when called */
147 struct switchtec_dev
*stdev
= stuser
->stdev
;
149 kref_get(&stuser
->kref
);
150 stuser
->read_len
= sizeof(stuser
->data
);
151 stuser_set_state(stuser
, MRPC_QUEUED
);
152 init_completion(&stuser
->comp
);
153 list_add_tail(&stuser
->list
, &stdev
->mrpc_queue
);
155 mrpc_cmd_submit(stdev
);
160 static void mrpc_complete_cmd(struct switchtec_dev
*stdev
)
162 /* requires the mrpc_mutex to already be held when called */
163 struct switchtec_user
*stuser
;
165 if (list_empty(&stdev
->mrpc_queue
))
168 stuser
= list_entry(stdev
->mrpc_queue
.next
, struct switchtec_user
,
171 stuser
->status
= ioread32(&stdev
->mmio_mrpc
->status
);
172 if (stuser
->status
== SWITCHTEC_MRPC_STATUS_INPROGRESS
)
175 stuser_set_state(stuser
, MRPC_DONE
);
176 stuser
->return_code
= 0;
178 if (stuser
->status
!= SWITCHTEC_MRPC_STATUS_DONE
)
181 stuser
->return_code
= ioread32(&stdev
->mmio_mrpc
->ret_value
);
182 if (stuser
->return_code
!= 0)
185 memcpy_fromio(stuser
->data
, &stdev
->mmio_mrpc
->output_data
,
189 complete_all(&stuser
->comp
);
190 list_del_init(&stuser
->list
);
192 stdev
->mrpc_busy
= 0;
194 mrpc_cmd_submit(stdev
);
197 static void mrpc_event_work(struct work_struct
*work
)
199 struct switchtec_dev
*stdev
;
201 stdev
= container_of(work
, struct switchtec_dev
, mrpc_work
);
203 dev_dbg(&stdev
->dev
, "%s\n", __func__
);
205 mutex_lock(&stdev
->mrpc_mutex
);
206 cancel_delayed_work(&stdev
->mrpc_timeout
);
207 mrpc_complete_cmd(stdev
);
208 mutex_unlock(&stdev
->mrpc_mutex
);
211 static void mrpc_timeout_work(struct work_struct
*work
)
213 struct switchtec_dev
*stdev
;
216 stdev
= container_of(work
, struct switchtec_dev
, mrpc_timeout
.work
);
218 dev_dbg(&stdev
->dev
, "%s\n", __func__
);
220 mutex_lock(&stdev
->mrpc_mutex
);
222 status
= ioread32(&stdev
->mmio_mrpc
->status
);
223 if (status
== SWITCHTEC_MRPC_STATUS_INPROGRESS
) {
224 schedule_delayed_work(&stdev
->mrpc_timeout
,
225 msecs_to_jiffies(500));
229 mrpc_complete_cmd(stdev
);
232 mutex_unlock(&stdev
->mrpc_mutex
);
235 static ssize_t
device_version_show(struct device
*dev
,
236 struct device_attribute
*attr
, char *buf
)
238 struct switchtec_dev
*stdev
= to_stdev(dev
);
241 ver
= ioread32(&stdev
->mmio_sys_info
->device_version
);
243 return sprintf(buf
, "%x\n", ver
);
245 static DEVICE_ATTR_RO(device_version
);
247 static ssize_t
fw_version_show(struct device
*dev
,
248 struct device_attribute
*attr
, char *buf
)
250 struct switchtec_dev
*stdev
= to_stdev(dev
);
253 ver
= ioread32(&stdev
->mmio_sys_info
->firmware_version
);
255 return sprintf(buf
, "%08x\n", ver
);
257 static DEVICE_ATTR_RO(fw_version
);
259 static ssize_t
io_string_show(char *buf
, void __iomem
*attr
, size_t len
)
263 memcpy_fromio(buf
, attr
, len
);
267 for (i
= len
- 1; i
> 0; i
--) {
277 #define DEVICE_ATTR_SYS_INFO_STR(field) \
278 static ssize_t field ## _show(struct device *dev, \
279 struct device_attribute *attr, char *buf) \
281 struct switchtec_dev *stdev = to_stdev(dev); \
282 return io_string_show(buf, &stdev->mmio_sys_info->field, \
283 sizeof(stdev->mmio_sys_info->field)); \
286 static DEVICE_ATTR_RO(field)
288 DEVICE_ATTR_SYS_INFO_STR(vendor_id
);
289 DEVICE_ATTR_SYS_INFO_STR(product_id
);
290 DEVICE_ATTR_SYS_INFO_STR(product_revision
);
291 DEVICE_ATTR_SYS_INFO_STR(component_vendor
);
293 static ssize_t
component_id_show(struct device
*dev
,
294 struct device_attribute
*attr
, char *buf
)
296 struct switchtec_dev
*stdev
= to_stdev(dev
);
297 int id
= ioread16(&stdev
->mmio_sys_info
->component_id
);
299 return sprintf(buf
, "PM%04X\n", id
);
301 static DEVICE_ATTR_RO(component_id
);
303 static ssize_t
component_revision_show(struct device
*dev
,
304 struct device_attribute
*attr
, char *buf
)
306 struct switchtec_dev
*stdev
= to_stdev(dev
);
307 int rev
= ioread8(&stdev
->mmio_sys_info
->component_revision
);
309 return sprintf(buf
, "%d\n", rev
);
311 static DEVICE_ATTR_RO(component_revision
);
313 static ssize_t
partition_show(struct device
*dev
,
314 struct device_attribute
*attr
, char *buf
)
316 struct switchtec_dev
*stdev
= to_stdev(dev
);
318 return sprintf(buf
, "%d\n", stdev
->partition
);
320 static DEVICE_ATTR_RO(partition
);
322 static ssize_t
partition_count_show(struct device
*dev
,
323 struct device_attribute
*attr
, char *buf
)
325 struct switchtec_dev
*stdev
= to_stdev(dev
);
327 return sprintf(buf
, "%d\n", stdev
->partition_count
);
329 static DEVICE_ATTR_RO(partition_count
);
331 static struct attribute
*switchtec_device_attrs
[] = {
332 &dev_attr_device_version
.attr
,
333 &dev_attr_fw_version
.attr
,
334 &dev_attr_vendor_id
.attr
,
335 &dev_attr_product_id
.attr
,
336 &dev_attr_product_revision
.attr
,
337 &dev_attr_component_vendor
.attr
,
338 &dev_attr_component_id
.attr
,
339 &dev_attr_component_revision
.attr
,
340 &dev_attr_partition
.attr
,
341 &dev_attr_partition_count
.attr
,
345 ATTRIBUTE_GROUPS(switchtec_device
);
347 static int switchtec_dev_open(struct inode
*inode
, struct file
*filp
)
349 struct switchtec_dev
*stdev
;
350 struct switchtec_user
*stuser
;
352 stdev
= container_of(inode
->i_cdev
, struct switchtec_dev
, cdev
);
354 stuser
= stuser_create(stdev
);
356 return PTR_ERR(stuser
);
358 filp
->private_data
= stuser
;
359 nonseekable_open(inode
, filp
);
361 dev_dbg(&stdev
->dev
, "%s: %p\n", __func__
, stuser
);
366 static int switchtec_dev_release(struct inode
*inode
, struct file
*filp
)
368 struct switchtec_user
*stuser
= filp
->private_data
;
375 static int lock_mutex_and_test_alive(struct switchtec_dev
*stdev
)
377 if (mutex_lock_interruptible(&stdev
->mrpc_mutex
))
381 mutex_unlock(&stdev
->mrpc_mutex
);
388 static ssize_t
switchtec_dev_write(struct file
*filp
, const char __user
*data
,
389 size_t size
, loff_t
*off
)
391 struct switchtec_user
*stuser
= filp
->private_data
;
392 struct switchtec_dev
*stdev
= stuser
->stdev
;
395 if (size
< sizeof(stuser
->cmd
) ||
396 size
> sizeof(stuser
->cmd
) + sizeof(stuser
->data
))
399 stuser
->data_len
= size
- sizeof(stuser
->cmd
);
401 rc
= lock_mutex_and_test_alive(stdev
);
405 if (stuser
->state
!= MRPC_IDLE
) {
410 rc
= copy_from_user(&stuser
->cmd
, data
, sizeof(stuser
->cmd
));
416 data
+= sizeof(stuser
->cmd
);
417 rc
= copy_from_user(&stuser
->data
, data
, size
- sizeof(stuser
->cmd
));
423 rc
= mrpc_queue_cmd(stuser
);
426 mutex_unlock(&stdev
->mrpc_mutex
);
434 static ssize_t
switchtec_dev_read(struct file
*filp
, char __user
*data
,
435 size_t size
, loff_t
*off
)
437 struct switchtec_user
*stuser
= filp
->private_data
;
438 struct switchtec_dev
*stdev
= stuser
->stdev
;
441 if (size
< sizeof(stuser
->cmd
) ||
442 size
> sizeof(stuser
->cmd
) + sizeof(stuser
->data
))
445 rc
= lock_mutex_and_test_alive(stdev
);
449 if (stuser
->state
== MRPC_IDLE
) {
450 mutex_unlock(&stdev
->mrpc_mutex
);
454 stuser
->read_len
= size
- sizeof(stuser
->return_code
);
456 mutex_unlock(&stdev
->mrpc_mutex
);
458 if (filp
->f_flags
& O_NONBLOCK
) {
459 if (!try_wait_for_completion(&stuser
->comp
))
462 rc
= wait_for_completion_interruptible(&stuser
->comp
);
467 rc
= lock_mutex_and_test_alive(stdev
);
471 if (stuser
->state
!= MRPC_DONE
) {
472 mutex_unlock(&stdev
->mrpc_mutex
);
476 rc
= copy_to_user(data
, &stuser
->return_code
,
477 sizeof(stuser
->return_code
));
483 data
+= sizeof(stuser
->return_code
);
484 rc
= copy_to_user(data
, &stuser
->data
,
485 size
- sizeof(stuser
->return_code
));
491 stuser_set_state(stuser
, MRPC_IDLE
);
494 mutex_unlock(&stdev
->mrpc_mutex
);
496 if (stuser
->status
== SWITCHTEC_MRPC_STATUS_DONE
)
498 else if (stuser
->status
== SWITCHTEC_MRPC_STATUS_INTERRUPTED
)
504 static __poll_t
switchtec_dev_poll(struct file
*filp
, poll_table
*wait
)
506 struct switchtec_user
*stuser
= filp
->private_data
;
507 struct switchtec_dev
*stdev
= stuser
->stdev
;
510 poll_wait(filp
, &stuser
->comp
.wait
, wait
);
511 poll_wait(filp
, &stdev
->event_wq
, wait
);
513 if (lock_mutex_and_test_alive(stdev
))
514 return EPOLLIN
| EPOLLRDHUP
| EPOLLOUT
| EPOLLERR
| EPOLLHUP
;
516 mutex_unlock(&stdev
->mrpc_mutex
);
518 if (try_wait_for_completion(&stuser
->comp
))
519 ret
|= EPOLLIN
| EPOLLRDNORM
;
521 if (stuser
->event_cnt
!= atomic_read(&stdev
->event_cnt
))
522 ret
|= EPOLLPRI
| EPOLLRDBAND
;
527 static int ioctl_flash_info(struct switchtec_dev
*stdev
,
528 struct switchtec_ioctl_flash_info __user
*uinfo
)
530 struct switchtec_ioctl_flash_info info
= {0};
531 struct flash_info_regs __iomem
*fi
= stdev
->mmio_flash_info
;
533 info
.flash_length
= ioread32(&fi
->flash_length
);
534 info
.num_partitions
= SWITCHTEC_IOCTL_NUM_PARTITIONS
;
536 if (copy_to_user(uinfo
, &info
, sizeof(info
)))
542 static void set_fw_info_part(struct switchtec_ioctl_flash_part_info
*info
,
543 struct partition_info __iomem
*pi
)
545 info
->address
= ioread32(&pi
->address
);
546 info
->length
= ioread32(&pi
->length
);
549 static int ioctl_flash_part_info(struct switchtec_dev
*stdev
,
550 struct switchtec_ioctl_flash_part_info __user
*uinfo
)
552 struct switchtec_ioctl_flash_part_info info
= {0};
553 struct flash_info_regs __iomem
*fi
= stdev
->mmio_flash_info
;
554 struct sys_info_regs __iomem
*si
= stdev
->mmio_sys_info
;
555 u32 active_addr
= -1;
557 if (copy_from_user(&info
, uinfo
, sizeof(info
)))
560 switch (info
.flash_partition
) {
561 case SWITCHTEC_IOCTL_PART_CFG0
:
562 active_addr
= ioread32(&fi
->active_cfg
);
563 set_fw_info_part(&info
, &fi
->cfg0
);
564 if (ioread16(&si
->cfg_running
) == SWITCHTEC_CFG0_RUNNING
)
565 info
.active
|= SWITCHTEC_IOCTL_PART_RUNNING
;
567 case SWITCHTEC_IOCTL_PART_CFG1
:
568 active_addr
= ioread32(&fi
->active_cfg
);
569 set_fw_info_part(&info
, &fi
->cfg1
);
570 if (ioread16(&si
->cfg_running
) == SWITCHTEC_CFG1_RUNNING
)
571 info
.active
|= SWITCHTEC_IOCTL_PART_RUNNING
;
573 case SWITCHTEC_IOCTL_PART_IMG0
:
574 active_addr
= ioread32(&fi
->active_img
);
575 set_fw_info_part(&info
, &fi
->img0
);
576 if (ioread16(&si
->img_running
) == SWITCHTEC_IMG0_RUNNING
)
577 info
.active
|= SWITCHTEC_IOCTL_PART_RUNNING
;
579 case SWITCHTEC_IOCTL_PART_IMG1
:
580 active_addr
= ioread32(&fi
->active_img
);
581 set_fw_info_part(&info
, &fi
->img1
);
582 if (ioread16(&si
->img_running
) == SWITCHTEC_IMG1_RUNNING
)
583 info
.active
|= SWITCHTEC_IOCTL_PART_RUNNING
;
585 case SWITCHTEC_IOCTL_PART_NVLOG
:
586 set_fw_info_part(&info
, &fi
->nvlog
);
588 case SWITCHTEC_IOCTL_PART_VENDOR0
:
589 set_fw_info_part(&info
, &fi
->vendor
[0]);
591 case SWITCHTEC_IOCTL_PART_VENDOR1
:
592 set_fw_info_part(&info
, &fi
->vendor
[1]);
594 case SWITCHTEC_IOCTL_PART_VENDOR2
:
595 set_fw_info_part(&info
, &fi
->vendor
[2]);
597 case SWITCHTEC_IOCTL_PART_VENDOR3
:
598 set_fw_info_part(&info
, &fi
->vendor
[3]);
600 case SWITCHTEC_IOCTL_PART_VENDOR4
:
601 set_fw_info_part(&info
, &fi
->vendor
[4]);
603 case SWITCHTEC_IOCTL_PART_VENDOR5
:
604 set_fw_info_part(&info
, &fi
->vendor
[5]);
606 case SWITCHTEC_IOCTL_PART_VENDOR6
:
607 set_fw_info_part(&info
, &fi
->vendor
[6]);
609 case SWITCHTEC_IOCTL_PART_VENDOR7
:
610 set_fw_info_part(&info
, &fi
->vendor
[7]);
616 if (info
.address
== active_addr
)
617 info
.active
|= SWITCHTEC_IOCTL_PART_ACTIVE
;
619 if (copy_to_user(uinfo
, &info
, sizeof(info
)))
625 static int ioctl_event_summary(struct switchtec_dev
*stdev
,
626 struct switchtec_user
*stuser
,
627 struct switchtec_ioctl_event_summary __user
*usum
)
629 struct switchtec_ioctl_event_summary s
= {0};
633 s
.global
= ioread32(&stdev
->mmio_sw_event
->global_summary
);
634 s
.part_bitmap
= ioread32(&stdev
->mmio_sw_event
->part_event_bitmap
);
635 s
.local_part
= ioread32(&stdev
->mmio_part_cfg
->part_event_summary
);
637 for (i
= 0; i
< stdev
->partition_count
; i
++) {
638 reg
= ioread32(&stdev
->mmio_part_cfg_all
[i
].part_event_summary
);
642 for (i
= 0; i
< SWITCHTEC_MAX_PFF_CSR
; i
++) {
643 reg
= ioread16(&stdev
->mmio_pff_csr
[i
].vendor_id
);
644 if (reg
!= MICROSEMI_VENDOR_ID
)
647 reg
= ioread32(&stdev
->mmio_pff_csr
[i
].pff_event_summary
);
651 if (copy_to_user(usum
, &s
, sizeof(s
)))
654 stuser
->event_cnt
= atomic_read(&stdev
->event_cnt
);
659 static u32 __iomem
*global_ev_reg(struct switchtec_dev
*stdev
,
660 size_t offset
, int index
)
662 return (void __iomem
*)stdev
->mmio_sw_event
+ offset
;
665 static u32 __iomem
*part_ev_reg(struct switchtec_dev
*stdev
,
666 size_t offset
, int index
)
668 return (void __iomem
*)&stdev
->mmio_part_cfg_all
[index
] + offset
;
671 static u32 __iomem
*pff_ev_reg(struct switchtec_dev
*stdev
,
672 size_t offset
, int index
)
674 return (void __iomem
*)&stdev
->mmio_pff_csr
[index
] + offset
;
677 #define EV_GLB(i, r)[i] = {offsetof(struct sw_event_regs, r), global_ev_reg}
678 #define EV_PAR(i, r)[i] = {offsetof(struct part_cfg_regs, r), part_ev_reg}
679 #define EV_PFF(i, r)[i] = {offsetof(struct pff_csr_regs, r), pff_ev_reg}
681 static const struct event_reg
{
683 u32 __iomem
*(*map_reg
)(struct switchtec_dev
*stdev
,
684 size_t offset
, int index
);
686 EV_GLB(SWITCHTEC_IOCTL_EVENT_STACK_ERROR
, stack_error_event_hdr
),
687 EV_GLB(SWITCHTEC_IOCTL_EVENT_PPU_ERROR
, ppu_error_event_hdr
),
688 EV_GLB(SWITCHTEC_IOCTL_EVENT_ISP_ERROR
, isp_error_event_hdr
),
689 EV_GLB(SWITCHTEC_IOCTL_EVENT_SYS_RESET
, sys_reset_event_hdr
),
690 EV_GLB(SWITCHTEC_IOCTL_EVENT_FW_EXC
, fw_exception_hdr
),
691 EV_GLB(SWITCHTEC_IOCTL_EVENT_FW_NMI
, fw_nmi_hdr
),
692 EV_GLB(SWITCHTEC_IOCTL_EVENT_FW_NON_FATAL
, fw_non_fatal_hdr
),
693 EV_GLB(SWITCHTEC_IOCTL_EVENT_FW_FATAL
, fw_fatal_hdr
),
694 EV_GLB(SWITCHTEC_IOCTL_EVENT_TWI_MRPC_COMP
, twi_mrpc_comp_hdr
),
695 EV_GLB(SWITCHTEC_IOCTL_EVENT_TWI_MRPC_COMP_ASYNC
,
696 twi_mrpc_comp_async_hdr
),
697 EV_GLB(SWITCHTEC_IOCTL_EVENT_CLI_MRPC_COMP
, cli_mrpc_comp_hdr
),
698 EV_GLB(SWITCHTEC_IOCTL_EVENT_CLI_MRPC_COMP_ASYNC
,
699 cli_mrpc_comp_async_hdr
),
700 EV_GLB(SWITCHTEC_IOCTL_EVENT_GPIO_INT
, gpio_interrupt_hdr
),
701 EV_GLB(SWITCHTEC_IOCTL_EVENT_GFMS
, gfms_event_hdr
),
702 EV_PAR(SWITCHTEC_IOCTL_EVENT_PART_RESET
, part_reset_hdr
),
703 EV_PAR(SWITCHTEC_IOCTL_EVENT_MRPC_COMP
, mrpc_comp_hdr
),
704 EV_PAR(SWITCHTEC_IOCTL_EVENT_MRPC_COMP_ASYNC
, mrpc_comp_async_hdr
),
705 EV_PAR(SWITCHTEC_IOCTL_EVENT_DYN_PART_BIND_COMP
, dyn_binding_hdr
),
706 EV_PFF(SWITCHTEC_IOCTL_EVENT_AER_IN_P2P
, aer_in_p2p_hdr
),
707 EV_PFF(SWITCHTEC_IOCTL_EVENT_AER_IN_VEP
, aer_in_vep_hdr
),
708 EV_PFF(SWITCHTEC_IOCTL_EVENT_DPC
, dpc_hdr
),
709 EV_PFF(SWITCHTEC_IOCTL_EVENT_CTS
, cts_hdr
),
710 EV_PFF(SWITCHTEC_IOCTL_EVENT_HOTPLUG
, hotplug_hdr
),
711 EV_PFF(SWITCHTEC_IOCTL_EVENT_IER
, ier_hdr
),
712 EV_PFF(SWITCHTEC_IOCTL_EVENT_THRESH
, threshold_hdr
),
713 EV_PFF(SWITCHTEC_IOCTL_EVENT_POWER_MGMT
, power_mgmt_hdr
),
714 EV_PFF(SWITCHTEC_IOCTL_EVENT_TLP_THROTTLING
, tlp_throttling_hdr
),
715 EV_PFF(SWITCHTEC_IOCTL_EVENT_FORCE_SPEED
, force_speed_hdr
),
716 EV_PFF(SWITCHTEC_IOCTL_EVENT_CREDIT_TIMEOUT
, credit_timeout_hdr
),
717 EV_PFF(SWITCHTEC_IOCTL_EVENT_LINK_STATE
, link_state_hdr
),
720 static u32 __iomem
*event_hdr_addr(struct switchtec_dev
*stdev
,
721 int event_id
, int index
)
725 if (event_id
< 0 || event_id
>= SWITCHTEC_IOCTL_MAX_EVENTS
)
726 return ERR_PTR(-EINVAL
);
728 off
= event_regs
[event_id
].offset
;
730 if (event_regs
[event_id
].map_reg
== part_ev_reg
) {
731 if (index
== SWITCHTEC_IOCTL_EVENT_LOCAL_PART_IDX
)
732 index
= stdev
->partition
;
733 else if (index
< 0 || index
>= stdev
->partition_count
)
734 return ERR_PTR(-EINVAL
);
735 } else if (event_regs
[event_id
].map_reg
== pff_ev_reg
) {
736 if (index
< 0 || index
>= stdev
->pff_csr_count
)
737 return ERR_PTR(-EINVAL
);
740 return event_regs
[event_id
].map_reg(stdev
, off
, index
);
743 static int event_ctl(struct switchtec_dev
*stdev
,
744 struct switchtec_ioctl_event_ctl
*ctl
)
750 reg
= event_hdr_addr(stdev
, ctl
->event_id
, ctl
->index
);
755 for (i
= 0; i
< ARRAY_SIZE(ctl
->data
); i
++)
756 ctl
->data
[i
] = ioread32(®
[i
+ 1]);
758 ctl
->occurred
= hdr
& SWITCHTEC_EVENT_OCCURRED
;
759 ctl
->count
= (hdr
>> 5) & 0xFF;
761 if (!(ctl
->flags
& SWITCHTEC_IOCTL_EVENT_FLAG_CLEAR
))
762 hdr
&= ~SWITCHTEC_EVENT_CLEAR
;
763 if (ctl
->flags
& SWITCHTEC_IOCTL_EVENT_FLAG_EN_POLL
)
764 hdr
|= SWITCHTEC_EVENT_EN_IRQ
;
765 if (ctl
->flags
& SWITCHTEC_IOCTL_EVENT_FLAG_DIS_POLL
)
766 hdr
&= ~SWITCHTEC_EVENT_EN_IRQ
;
767 if (ctl
->flags
& SWITCHTEC_IOCTL_EVENT_FLAG_EN_LOG
)
768 hdr
|= SWITCHTEC_EVENT_EN_LOG
;
769 if (ctl
->flags
& SWITCHTEC_IOCTL_EVENT_FLAG_DIS_LOG
)
770 hdr
&= ~SWITCHTEC_EVENT_EN_LOG
;
771 if (ctl
->flags
& SWITCHTEC_IOCTL_EVENT_FLAG_EN_CLI
)
772 hdr
|= SWITCHTEC_EVENT_EN_CLI
;
773 if (ctl
->flags
& SWITCHTEC_IOCTL_EVENT_FLAG_DIS_CLI
)
774 hdr
&= ~SWITCHTEC_EVENT_EN_CLI
;
775 if (ctl
->flags
& SWITCHTEC_IOCTL_EVENT_FLAG_EN_FATAL
)
776 hdr
|= SWITCHTEC_EVENT_FATAL
;
777 if (ctl
->flags
& SWITCHTEC_IOCTL_EVENT_FLAG_DIS_FATAL
)
778 hdr
&= ~SWITCHTEC_EVENT_FATAL
;
784 if (hdr
& SWITCHTEC_EVENT_EN_IRQ
)
785 ctl
->flags
|= SWITCHTEC_IOCTL_EVENT_FLAG_EN_POLL
;
786 if (hdr
& SWITCHTEC_EVENT_EN_LOG
)
787 ctl
->flags
|= SWITCHTEC_IOCTL_EVENT_FLAG_EN_LOG
;
788 if (hdr
& SWITCHTEC_EVENT_EN_CLI
)
789 ctl
->flags
|= SWITCHTEC_IOCTL_EVENT_FLAG_EN_CLI
;
790 if (hdr
& SWITCHTEC_EVENT_FATAL
)
791 ctl
->flags
|= SWITCHTEC_IOCTL_EVENT_FLAG_EN_FATAL
;
796 static int ioctl_event_ctl(struct switchtec_dev
*stdev
,
797 struct switchtec_ioctl_event_ctl __user
*uctl
)
801 struct switchtec_ioctl_event_ctl ctl
;
803 if (copy_from_user(&ctl
, uctl
, sizeof(ctl
)))
806 if (ctl
.event_id
>= SWITCHTEC_IOCTL_MAX_EVENTS
)
809 if (ctl
.flags
& SWITCHTEC_IOCTL_EVENT_FLAG_UNUSED
)
812 if (ctl
.index
== SWITCHTEC_IOCTL_EVENT_IDX_ALL
) {
813 if (event_regs
[ctl
.event_id
].map_reg
== global_ev_reg
)
815 else if (event_regs
[ctl
.event_id
].map_reg
== part_ev_reg
)
816 nr_idxs
= stdev
->partition_count
;
817 else if (event_regs
[ctl
.event_id
].map_reg
== pff_ev_reg
)
818 nr_idxs
= stdev
->pff_csr_count
;
822 for (ctl
.index
= 0; ctl
.index
< nr_idxs
; ctl
.index
++) {
823 ret
= event_ctl(stdev
, &ctl
);
828 ret
= event_ctl(stdev
, &ctl
);
833 if (copy_to_user(uctl
, &ctl
, sizeof(ctl
)))
839 static int ioctl_pff_to_port(struct switchtec_dev
*stdev
,
840 struct switchtec_ioctl_pff_port
*up
)
844 struct part_cfg_regs
*pcfg
;
845 struct switchtec_ioctl_pff_port p
;
847 if (copy_from_user(&p
, up
, sizeof(p
)))
851 for (part
= 0; part
< stdev
->partition_count
; part
++) {
852 pcfg
= &stdev
->mmio_part_cfg_all
[part
];
855 reg
= ioread32(&pcfg
->usp_pff_inst_id
);
861 reg
= ioread32(&pcfg
->vep_pff_inst_id
);
863 p
.port
= SWITCHTEC_IOCTL_PFF_VEP
;
867 for (i
= 0; i
< ARRAY_SIZE(pcfg
->dsp_pff_inst_id
); i
++) {
868 reg
= ioread32(&pcfg
->dsp_pff_inst_id
[i
]);
880 if (copy_to_user(up
, &p
, sizeof(p
)))
886 static int ioctl_port_to_pff(struct switchtec_dev
*stdev
,
887 struct switchtec_ioctl_pff_port
*up
)
889 struct switchtec_ioctl_pff_port p
;
890 struct part_cfg_regs
*pcfg
;
892 if (copy_from_user(&p
, up
, sizeof(p
)))
895 if (p
.partition
== SWITCHTEC_IOCTL_EVENT_LOCAL_PART_IDX
)
896 pcfg
= stdev
->mmio_part_cfg
;
897 else if (p
.partition
< stdev
->partition_count
)
898 pcfg
= &stdev
->mmio_part_cfg_all
[p
.partition
];
904 p
.pff
= ioread32(&pcfg
->usp_pff_inst_id
);
906 case SWITCHTEC_IOCTL_PFF_VEP
:
907 p
.pff
= ioread32(&pcfg
->vep_pff_inst_id
);
910 if (p
.port
> ARRAY_SIZE(pcfg
->dsp_pff_inst_id
))
912 p
.pff
= ioread32(&pcfg
->dsp_pff_inst_id
[p
.port
- 1]);
916 if (copy_to_user(up
, &p
, sizeof(p
)))
922 static long switchtec_dev_ioctl(struct file
*filp
, unsigned int cmd
,
925 struct switchtec_user
*stuser
= filp
->private_data
;
926 struct switchtec_dev
*stdev
= stuser
->stdev
;
928 void __user
*argp
= (void __user
*)arg
;
930 rc
= lock_mutex_and_test_alive(stdev
);
935 case SWITCHTEC_IOCTL_FLASH_INFO
:
936 rc
= ioctl_flash_info(stdev
, argp
);
938 case SWITCHTEC_IOCTL_FLASH_PART_INFO
:
939 rc
= ioctl_flash_part_info(stdev
, argp
);
941 case SWITCHTEC_IOCTL_EVENT_SUMMARY
:
942 rc
= ioctl_event_summary(stdev
, stuser
, argp
);
944 case SWITCHTEC_IOCTL_EVENT_CTL
:
945 rc
= ioctl_event_ctl(stdev
, argp
);
947 case SWITCHTEC_IOCTL_PFF_TO_PORT
:
948 rc
= ioctl_pff_to_port(stdev
, argp
);
950 case SWITCHTEC_IOCTL_PORT_TO_PFF
:
951 rc
= ioctl_port_to_pff(stdev
, argp
);
958 mutex_unlock(&stdev
->mrpc_mutex
);
962 static const struct file_operations switchtec_fops
= {
963 .owner
= THIS_MODULE
,
964 .open
= switchtec_dev_open
,
965 .release
= switchtec_dev_release
,
966 .write
= switchtec_dev_write
,
967 .read
= switchtec_dev_read
,
968 .poll
= switchtec_dev_poll
,
969 .unlocked_ioctl
= switchtec_dev_ioctl
,
970 .compat_ioctl
= switchtec_dev_ioctl
,
973 static void link_event_work(struct work_struct
*work
)
975 struct switchtec_dev
*stdev
;
977 stdev
= container_of(work
, struct switchtec_dev
, link_event_work
);
979 if (stdev
->link_notifier
)
980 stdev
->link_notifier(stdev
);
983 static void check_link_state_events(struct switchtec_dev
*stdev
)
990 for (idx
= 0; idx
< stdev
->pff_csr_count
; idx
++) {
991 reg
= ioread32(&stdev
->mmio_pff_csr
[idx
].link_state_hdr
);
992 dev_dbg(&stdev
->dev
, "link_state: %d->%08x\n", idx
, reg
);
993 count
= (reg
>> 5) & 0xFF;
995 if (count
!= stdev
->link_event_count
[idx
]) {
997 stdev
->link_event_count
[idx
] = count
;
1002 schedule_work(&stdev
->link_event_work
);
1005 static void enable_link_state_events(struct switchtec_dev
*stdev
)
1009 for (idx
= 0; idx
< stdev
->pff_csr_count
; idx
++) {
1010 iowrite32(SWITCHTEC_EVENT_CLEAR
|
1011 SWITCHTEC_EVENT_EN_IRQ
,
1012 &stdev
->mmio_pff_csr
[idx
].link_state_hdr
);
1016 static void stdev_release(struct device
*dev
)
1018 struct switchtec_dev
*stdev
= to_stdev(dev
);
1023 static void stdev_kill(struct switchtec_dev
*stdev
)
1025 struct switchtec_user
*stuser
, *tmpuser
;
1027 pci_clear_master(stdev
->pdev
);
1029 cancel_delayed_work_sync(&stdev
->mrpc_timeout
);
1031 /* Mark the hardware as unavailable and complete all completions */
1032 mutex_lock(&stdev
->mrpc_mutex
);
1033 stdev
->alive
= false;
1035 /* Wake up and kill any users waiting on an MRPC request */
1036 list_for_each_entry_safe(stuser
, tmpuser
, &stdev
->mrpc_queue
, list
) {
1037 complete_all(&stuser
->comp
);
1038 list_del_init(&stuser
->list
);
1042 mutex_unlock(&stdev
->mrpc_mutex
);
1044 /* Wake up any users waiting on event_wq */
1045 wake_up_interruptible(&stdev
->event_wq
);
1048 static struct switchtec_dev
*stdev_create(struct pci_dev
*pdev
)
1050 struct switchtec_dev
*stdev
;
1056 stdev
= kzalloc_node(sizeof(*stdev
), GFP_KERNEL
,
1057 dev_to_node(&pdev
->dev
));
1059 return ERR_PTR(-ENOMEM
);
1061 stdev
->alive
= true;
1063 INIT_LIST_HEAD(&stdev
->mrpc_queue
);
1064 mutex_init(&stdev
->mrpc_mutex
);
1065 stdev
->mrpc_busy
= 0;
1066 INIT_WORK(&stdev
->mrpc_work
, mrpc_event_work
);
1067 INIT_DELAYED_WORK(&stdev
->mrpc_timeout
, mrpc_timeout_work
);
1068 INIT_WORK(&stdev
->link_event_work
, link_event_work
);
1069 init_waitqueue_head(&stdev
->event_wq
);
1070 atomic_set(&stdev
->event_cnt
, 0);
1073 device_initialize(dev
);
1074 dev
->class = switchtec_class
;
1075 dev
->parent
= &pdev
->dev
;
1076 dev
->groups
= switchtec_device_groups
;
1077 dev
->release
= stdev_release
;
1079 minor
= ida_simple_get(&switchtec_minor_ida
, 0, 0,
1086 dev
->devt
= MKDEV(MAJOR(switchtec_devt
), minor
);
1087 dev_set_name(dev
, "switchtec%d", minor
);
1089 cdev
= &stdev
->cdev
;
1090 cdev_init(cdev
, &switchtec_fops
);
1091 cdev
->owner
= THIS_MODULE
;
1096 put_device(&stdev
->dev
);
1100 static int mask_event(struct switchtec_dev
*stdev
, int eid
, int idx
)
1102 size_t off
= event_regs
[eid
].offset
;
1103 u32 __iomem
*hdr_reg
;
1106 hdr_reg
= event_regs
[eid
].map_reg(stdev
, off
, idx
);
1107 hdr
= ioread32(hdr_reg
);
1109 if (!(hdr
& SWITCHTEC_EVENT_OCCURRED
&& hdr
& SWITCHTEC_EVENT_EN_IRQ
))
1112 if (eid
== SWITCHTEC_IOCTL_EVENT_LINK_STATE
)
1115 dev_dbg(&stdev
->dev
, "%s: %d %d %x\n", __func__
, eid
, idx
, hdr
);
1116 hdr
&= ~(SWITCHTEC_EVENT_EN_IRQ
| SWITCHTEC_EVENT_OCCURRED
);
1117 iowrite32(hdr
, hdr_reg
);
1122 static int mask_all_events(struct switchtec_dev
*stdev
, int eid
)
1127 if (event_regs
[eid
].map_reg
== part_ev_reg
) {
1128 for (idx
= 0; idx
< stdev
->partition_count
; idx
++)
1129 count
+= mask_event(stdev
, eid
, idx
);
1130 } else if (event_regs
[eid
].map_reg
== pff_ev_reg
) {
1131 for (idx
= 0; idx
< stdev
->pff_csr_count
; idx
++) {
1132 if (!stdev
->pff_local
[idx
])
1135 count
+= mask_event(stdev
, eid
, idx
);
1138 count
+= mask_event(stdev
, eid
, 0);
1144 static irqreturn_t
switchtec_event_isr(int irq
, void *dev
)
1146 struct switchtec_dev
*stdev
= dev
;
1148 irqreturn_t ret
= IRQ_NONE
;
1149 int eid
, event_count
= 0;
1151 reg
= ioread32(&stdev
->mmio_part_cfg
->mrpc_comp_hdr
);
1152 if (reg
& SWITCHTEC_EVENT_OCCURRED
) {
1153 dev_dbg(&stdev
->dev
, "%s: mrpc comp\n", __func__
);
1155 schedule_work(&stdev
->mrpc_work
);
1156 iowrite32(reg
, &stdev
->mmio_part_cfg
->mrpc_comp_hdr
);
1159 check_link_state_events(stdev
);
1161 for (eid
= 0; eid
< SWITCHTEC_IOCTL_MAX_EVENTS
; eid
++)
1162 event_count
+= mask_all_events(stdev
, eid
);
1165 atomic_inc(&stdev
->event_cnt
);
1166 wake_up_interruptible(&stdev
->event_wq
);
1167 dev_dbg(&stdev
->dev
, "%s: %d events\n", __func__
,
1175 static int switchtec_init_isr(struct switchtec_dev
*stdev
)
1180 nvecs
= pci_alloc_irq_vectors(stdev
->pdev
, 1, 4,
1181 PCI_IRQ_MSIX
| PCI_IRQ_MSI
);
1185 event_irq
= ioread32(&stdev
->mmio_part_cfg
->vep_vector_number
);
1186 if (event_irq
< 0 || event_irq
>= nvecs
)
1189 event_irq
= pci_irq_vector(stdev
->pdev
, event_irq
);
1193 return devm_request_irq(&stdev
->pdev
->dev
, event_irq
,
1194 switchtec_event_isr
, 0,
1195 KBUILD_MODNAME
, stdev
);
1198 static void init_pff(struct switchtec_dev
*stdev
)
1202 struct part_cfg_regs
*pcfg
= stdev
->mmio_part_cfg
;
1204 for (i
= 0; i
< SWITCHTEC_MAX_PFF_CSR
; i
++) {
1205 reg
= ioread16(&stdev
->mmio_pff_csr
[i
].vendor_id
);
1206 if (reg
!= MICROSEMI_VENDOR_ID
)
1210 stdev
->pff_csr_count
= i
;
1212 reg
= ioread32(&pcfg
->usp_pff_inst_id
);
1213 if (reg
< SWITCHTEC_MAX_PFF_CSR
)
1214 stdev
->pff_local
[reg
] = 1;
1216 reg
= ioread32(&pcfg
->vep_pff_inst_id
);
1217 if (reg
< SWITCHTEC_MAX_PFF_CSR
)
1218 stdev
->pff_local
[reg
] = 1;
1220 for (i
= 0; i
< ARRAY_SIZE(pcfg
->dsp_pff_inst_id
); i
++) {
1221 reg
= ioread32(&pcfg
->dsp_pff_inst_id
[i
]);
1222 if (reg
< SWITCHTEC_MAX_PFF_CSR
)
1223 stdev
->pff_local
[reg
] = 1;
1227 static int switchtec_init_pci(struct switchtec_dev
*stdev
,
1228 struct pci_dev
*pdev
)
1232 rc
= pcim_enable_device(pdev
);
1236 rc
= pcim_iomap_regions(pdev
, 0x1, KBUILD_MODNAME
);
1240 pci_set_master(pdev
);
1242 stdev
->mmio
= pcim_iomap_table(pdev
)[0];
1243 stdev
->mmio_mrpc
= stdev
->mmio
+ SWITCHTEC_GAS_MRPC_OFFSET
;
1244 stdev
->mmio_sw_event
= stdev
->mmio
+ SWITCHTEC_GAS_SW_EVENT_OFFSET
;
1245 stdev
->mmio_sys_info
= stdev
->mmio
+ SWITCHTEC_GAS_SYS_INFO_OFFSET
;
1246 stdev
->mmio_flash_info
= stdev
->mmio
+ SWITCHTEC_GAS_FLASH_INFO_OFFSET
;
1247 stdev
->mmio_ntb
= stdev
->mmio
+ SWITCHTEC_GAS_NTB_OFFSET
;
1248 stdev
->partition
= ioread8(&stdev
->mmio_sys_info
->partition_id
);
1249 stdev
->partition_count
= ioread8(&stdev
->mmio_ntb
->partition_count
);
1250 stdev
->mmio_part_cfg_all
= stdev
->mmio
+ SWITCHTEC_GAS_PART_CFG_OFFSET
;
1251 stdev
->mmio_part_cfg
= &stdev
->mmio_part_cfg_all
[stdev
->partition
];
1252 stdev
->mmio_pff_csr
= stdev
->mmio
+ SWITCHTEC_GAS_PFF_CSR_OFFSET
;
1254 if (stdev
->partition_count
< 1)
1255 stdev
->partition_count
= 1;
1259 pci_set_drvdata(pdev
, stdev
);
1264 static int switchtec_pci_probe(struct pci_dev
*pdev
,
1265 const struct pci_device_id
*id
)
1267 struct switchtec_dev
*stdev
;
1270 if (pdev
->class == MICROSEMI_NTB_CLASSCODE
)
1271 request_module_nowait("ntb_hw_switchtec");
1273 stdev
= stdev_create(pdev
);
1275 return PTR_ERR(stdev
);
1277 rc
= switchtec_init_pci(stdev
, pdev
);
1281 rc
= switchtec_init_isr(stdev
);
1283 dev_err(&stdev
->dev
, "failed to init isr.\n");
1287 iowrite32(SWITCHTEC_EVENT_CLEAR
|
1288 SWITCHTEC_EVENT_EN_IRQ
,
1289 &stdev
->mmio_part_cfg
->mrpc_comp_hdr
);
1290 enable_link_state_events(stdev
);
1292 rc
= cdev_device_add(&stdev
->cdev
, &stdev
->dev
);
1296 dev_info(&stdev
->dev
, "Management device registered.\n");
1303 ida_simple_remove(&switchtec_minor_ida
, MINOR(stdev
->dev
.devt
));
1304 put_device(&stdev
->dev
);
1308 static void switchtec_pci_remove(struct pci_dev
*pdev
)
1310 struct switchtec_dev
*stdev
= pci_get_drvdata(pdev
);
1312 pci_set_drvdata(pdev
, NULL
);
1314 cdev_device_del(&stdev
->cdev
, &stdev
->dev
);
1315 ida_simple_remove(&switchtec_minor_ida
, MINOR(stdev
->dev
.devt
));
1316 dev_info(&stdev
->dev
, "unregistered.\n");
1319 put_device(&stdev
->dev
);
1322 #define SWITCHTEC_PCI_DEVICE(device_id) \
1324 .vendor = MICROSEMI_VENDOR_ID, \
1325 .device = device_id, \
1326 .subvendor = PCI_ANY_ID, \
1327 .subdevice = PCI_ANY_ID, \
1328 .class = MICROSEMI_MGMT_CLASSCODE, \
1329 .class_mask = 0xFFFFFFFF, \
1332 .vendor = MICROSEMI_VENDOR_ID, \
1333 .device = device_id, \
1334 .subvendor = PCI_ANY_ID, \
1335 .subdevice = PCI_ANY_ID, \
1336 .class = MICROSEMI_NTB_CLASSCODE, \
1337 .class_mask = 0xFFFFFFFF, \
1340 static const struct pci_device_id switchtec_pci_tbl
[] = {
1341 SWITCHTEC_PCI_DEVICE(0x8531), //PFX 24xG3
1342 SWITCHTEC_PCI_DEVICE(0x8532), //PFX 32xG3
1343 SWITCHTEC_PCI_DEVICE(0x8533), //PFX 48xG3
1344 SWITCHTEC_PCI_DEVICE(0x8534), //PFX 64xG3
1345 SWITCHTEC_PCI_DEVICE(0x8535), //PFX 80xG3
1346 SWITCHTEC_PCI_DEVICE(0x8536), //PFX 96xG3
1347 SWITCHTEC_PCI_DEVICE(0x8541), //PSX 24xG3
1348 SWITCHTEC_PCI_DEVICE(0x8542), //PSX 32xG3
1349 SWITCHTEC_PCI_DEVICE(0x8543), //PSX 48xG3
1350 SWITCHTEC_PCI_DEVICE(0x8544), //PSX 64xG3
1351 SWITCHTEC_PCI_DEVICE(0x8545), //PSX 80xG3
1352 SWITCHTEC_PCI_DEVICE(0x8546), //PSX 96xG3
1353 SWITCHTEC_PCI_DEVICE(0x8551), //PAX 24XG3
1354 SWITCHTEC_PCI_DEVICE(0x8552), //PAX 32XG3
1355 SWITCHTEC_PCI_DEVICE(0x8553), //PAX 48XG3
1356 SWITCHTEC_PCI_DEVICE(0x8554), //PAX 64XG3
1357 SWITCHTEC_PCI_DEVICE(0x8555), //PAX 80XG3
1358 SWITCHTEC_PCI_DEVICE(0x8556), //PAX 96XG3
1359 SWITCHTEC_PCI_DEVICE(0x8561), //PFXL 24XG3
1360 SWITCHTEC_PCI_DEVICE(0x8562), //PFXL 32XG3
1361 SWITCHTEC_PCI_DEVICE(0x8563), //PFXL 48XG3
1362 SWITCHTEC_PCI_DEVICE(0x8564), //PFXL 64XG3
1363 SWITCHTEC_PCI_DEVICE(0x8565), //PFXL 80XG3
1364 SWITCHTEC_PCI_DEVICE(0x8566), //PFXL 96XG3
1365 SWITCHTEC_PCI_DEVICE(0x8571), //PFXI 24XG3
1366 SWITCHTEC_PCI_DEVICE(0x8572), //PFXI 32XG3
1367 SWITCHTEC_PCI_DEVICE(0x8573), //PFXI 48XG3
1368 SWITCHTEC_PCI_DEVICE(0x8574), //PFXI 64XG3
1369 SWITCHTEC_PCI_DEVICE(0x8575), //PFXI 80XG3
1370 SWITCHTEC_PCI_DEVICE(0x8576), //PFXI 96XG3
1373 MODULE_DEVICE_TABLE(pci
, switchtec_pci_tbl
);
1375 static struct pci_driver switchtec_pci_driver
= {
1376 .name
= KBUILD_MODNAME
,
1377 .id_table
= switchtec_pci_tbl
,
1378 .probe
= switchtec_pci_probe
,
1379 .remove
= switchtec_pci_remove
,
1382 static int __init
switchtec_init(void)
1386 rc
= alloc_chrdev_region(&switchtec_devt
, 0, max_devices
,
1391 switchtec_class
= class_create(THIS_MODULE
, "switchtec");
1392 if (IS_ERR(switchtec_class
)) {
1393 rc
= PTR_ERR(switchtec_class
);
1394 goto err_create_class
;
1397 rc
= pci_register_driver(&switchtec_pci_driver
);
1399 goto err_pci_register
;
1401 pr_info(KBUILD_MODNAME
": loaded.\n");
1406 class_destroy(switchtec_class
);
1409 unregister_chrdev_region(switchtec_devt
, max_devices
);
1413 module_init(switchtec_init
);
1415 static void __exit
switchtec_exit(void)
1417 pci_unregister_driver(&switchtec_pci_driver
);
1418 class_destroy(switchtec_class
);
1419 unregister_chrdev_region(switchtec_devt
, max_devices
);
1420 ida_destroy(&switchtec_minor_ida
);
1422 pr_info(KBUILD_MODNAME
": unloaded.\n");
1424 module_exit(switchtec_exit
);