2 * Marvell 88SE64xx/88SE94xx pci init
4 * Copyright 2007 Red Hat, Inc.
5 * Copyright 2008 Marvell. <kewei@marvell.com>
6 * Copyright 2009-2011 Marvell. <yuxiangl@marvell.com>
8 * This file is licensed under GPLv2.
10 * This program is free software; you can redistribute it and/or
11 * modify it under the terms of the GNU General Public License as
12 * published by the Free Software Foundation; version 2 of the
15 * This program is distributed in the hope that it will be useful,
16 * but WITHOUT ANY WARRANTY; without even the implied warranty of
17 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
18 * General Public License for more details.
20 * You should have received a copy of the GNU General Public License
21 * along with this program; if not, write to the Free Software
22 * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307
29 static int lldd_max_execute_num
= 1;
30 module_param_named(collector
, lldd_max_execute_num
, int, S_IRUGO
);
31 MODULE_PARM_DESC(collector
, "\n"
32 "\tIf greater than one, tells the SAS Layer to run in Task Collector\n"
33 "\tMode. If 1 or 0, tells the SAS Layer to run in Direct Mode.\n"
34 "\tThe mvsas SAS LLDD supports both modes.\n"
35 "\tDefault: 1 (Direct Mode).\n");
37 int interrupt_coalescing
= 0x80;
39 static struct scsi_transport_template
*mvs_stt
;
40 struct kmem_cache
*mvs_task_list_cache
;
41 static const struct mvs_chip_info mvs_chips
[] = {
42 [chip_6320
] = { 1, 2, 0x400, 17, 16, 6, 9, &mvs_64xx_dispatch
, },
43 [chip_6440
] = { 1, 4, 0x400, 17, 16, 6, 9, &mvs_64xx_dispatch
, },
44 [chip_6485
] = { 1, 8, 0x800, 33, 32, 6, 10, &mvs_64xx_dispatch
, },
45 [chip_9180
] = { 2, 4, 0x800, 17, 64, 8, 9, &mvs_94xx_dispatch
, },
46 [chip_9480
] = { 2, 4, 0x800, 17, 64, 8, 9, &mvs_94xx_dispatch
, },
47 [chip_9445
] = { 1, 4, 0x800, 17, 64, 8, 11, &mvs_94xx_dispatch
, },
48 [chip_9485
] = { 2, 4, 0x800, 17, 64, 8, 11, &mvs_94xx_dispatch
, },
49 [chip_1300
] = { 1, 4, 0x400, 17, 16, 6, 9, &mvs_64xx_dispatch
, },
50 [chip_1320
] = { 2, 4, 0x800, 17, 64, 8, 9, &mvs_94xx_dispatch
, },
53 struct device_attribute
*mvst_host_attrs
[];
57 static struct scsi_host_template mvs_sht
= {
58 .module
= THIS_MODULE
,
60 .queuecommand
= sas_queuecommand
,
61 .target_alloc
= sas_target_alloc
,
62 .slave_configure
= sas_slave_configure
,
63 .scan_finished
= mvs_scan_finished
,
64 .scan_start
= mvs_scan_start
,
65 .change_queue_depth
= sas_change_queue_depth
,
66 .change_queue_type
= sas_change_queue_type
,
67 .bios_param
= sas_bios_param
,
71 .sg_tablesize
= SG_ALL
,
72 .max_sectors
= SCSI_DEFAULT_MAX_SECTORS
,
73 .use_clustering
= ENABLE_CLUSTERING
,
74 .eh_device_reset_handler
= sas_eh_device_reset_handler
,
75 .eh_bus_reset_handler
= sas_eh_bus_reset_handler
,
76 .target_destroy
= sas_target_destroy
,
78 .shost_attrs
= mvst_host_attrs
,
81 static struct sas_domain_function_template mvs_transport_ops
= {
82 .lldd_dev_found
= mvs_dev_found
,
83 .lldd_dev_gone
= mvs_dev_gone
,
84 .lldd_execute_task
= mvs_queue_command
,
85 .lldd_control_phy
= mvs_phy_control
,
87 .lldd_abort_task
= mvs_abort_task
,
88 .lldd_abort_task_set
= mvs_abort_task_set
,
89 .lldd_clear_aca
= mvs_clear_aca
,
90 .lldd_clear_task_set
= mvs_clear_task_set
,
91 .lldd_I_T_nexus_reset
= mvs_I_T_nexus_reset
,
92 .lldd_lu_reset
= mvs_lu_reset
,
93 .lldd_query_task
= mvs_query_task
,
94 .lldd_port_formed
= mvs_port_formed
,
95 .lldd_port_deformed
= mvs_port_deformed
,
99 static void mvs_phy_init(struct mvs_info
*mvi
, int phy_id
)
101 struct mvs_phy
*phy
= &mvi
->phy
[phy_id
];
102 struct asd_sas_phy
*sas_phy
= &phy
->sas_phy
;
106 init_timer(&phy
->timer
);
107 sas_phy
->enabled
= (phy_id
< mvi
->chip
->n_phy
) ? 1 : 0;
108 sas_phy
->class = SAS
;
109 sas_phy
->iproto
= SAS_PROTOCOL_ALL
;
111 sas_phy
->type
= PHY_TYPE_PHYSICAL
;
112 sas_phy
->role
= PHY_ROLE_INITIATOR
;
113 sas_phy
->oob_mode
= OOB_NOT_CONNECTED
;
114 sas_phy
->linkrate
= SAS_LINK_RATE_UNKNOWN
;
116 sas_phy
->id
= phy_id
;
117 sas_phy
->sas_addr
= &mvi
->sas_addr
[0];
118 sas_phy
->frame_rcvd
= &phy
->frame_rcvd
[0];
119 sas_phy
->ha
= (struct sas_ha_struct
*)mvi
->shost
->hostdata
;
120 sas_phy
->lldd_phy
= phy
;
123 static void mvs_free(struct mvs_info
*mvi
)
131 if (mvi
->flags
& MVF_FLAG_SOC
)
132 slot_nr
= MVS_SOC_SLOTS
;
134 slot_nr
= MVS_CHIP_SLOT_SZ
;
137 pci_pool_destroy(mvi
->dma_pool
);
140 dma_free_coherent(mvi
->dev
,
141 sizeof(*mvi
->tx
) * MVS_CHIP_SLOT_SZ
,
142 mvi
->tx
, mvi
->tx_dma
);
144 dma_free_coherent(mvi
->dev
, MVS_RX_FISL_SZ
,
145 mvi
->rx_fis
, mvi
->rx_fis_dma
);
147 dma_free_coherent(mvi
->dev
,
148 sizeof(*mvi
->rx
) * (MVS_RX_RING_SZ
+ 1),
149 mvi
->rx
, mvi
->rx_dma
);
151 dma_free_coherent(mvi
->dev
,
152 sizeof(*mvi
->slot
) * slot_nr
,
153 mvi
->slot
, mvi
->slot_dma
);
155 if (mvi
->bulk_buffer
)
156 dma_free_coherent(mvi
->dev
, TRASH_BUCKET_SIZE
,
157 mvi
->bulk_buffer
, mvi
->bulk_buffer_dma
);
158 if (mvi
->bulk_buffer1
)
159 dma_free_coherent(mvi
->dev
, TRASH_BUCKET_SIZE
,
160 mvi
->bulk_buffer1
, mvi
->bulk_buffer_dma1
);
162 MVS_CHIP_DISP
->chip_iounmap(mvi
);
164 scsi_host_put(mvi
->shost
);
165 list_for_each_entry(mwq
, &mvi
->wq_list
, entry
)
166 cancel_delayed_work(&mwq
->work_q
);
171 #ifdef CONFIG_SCSI_MVSAS_TASKLET
172 static void mvs_tasklet(unsigned long opaque
)
177 struct mvs_info
*mvi
;
178 struct sas_ha_struct
*sha
= (struct sas_ha_struct
*)opaque
;
180 core_nr
= ((struct mvs_prv_info
*)sha
->lldd_ha
)->n_host
;
181 mvi
= ((struct mvs_prv_info
*)sha
->lldd_ha
)->mvi
[0];
186 stat
= MVS_CHIP_DISP
->isr_status(mvi
, mvi
->pdev
->irq
);
190 for (i
= 0; i
< core_nr
; i
++) {
191 mvi
= ((struct mvs_prv_info
*)sha
->lldd_ha
)->mvi
[i
];
192 MVS_CHIP_DISP
->isr(mvi
, mvi
->pdev
->irq
, stat
);
195 MVS_CHIP_DISP
->interrupt_enable(mvi
);
200 static irqreturn_t
mvs_interrupt(int irq
, void *opaque
)
204 struct mvs_info
*mvi
;
205 struct sas_ha_struct
*sha
= opaque
;
206 #ifndef CONFIG_SCSI_MVSAS_TASKLET
210 core_nr
= ((struct mvs_prv_info
*)sha
->lldd_ha
)->n_host
;
211 mvi
= ((struct mvs_prv_info
*)sha
->lldd_ha
)->mvi
[0];
215 #ifdef CONFIG_SCSI_MVSAS_TASKLET
216 MVS_CHIP_DISP
->interrupt_disable(mvi
);
219 stat
= MVS_CHIP_DISP
->isr_status(mvi
, irq
);
221 #ifdef CONFIG_SCSI_MVSAS_TASKLET
222 MVS_CHIP_DISP
->interrupt_enable(mvi
);
227 #ifdef CONFIG_SCSI_MVSAS_TASKLET
228 tasklet_schedule(&((struct mvs_prv_info
*)sha
->lldd_ha
)->mv_tasklet
);
230 for (i
= 0; i
< core_nr
; i
++) {
231 mvi
= ((struct mvs_prv_info
*)sha
->lldd_ha
)->mvi
[i
];
232 MVS_CHIP_DISP
->isr(mvi
, irq
, stat
);
238 static int mvs_alloc(struct mvs_info
*mvi
, struct Scsi_Host
*shost
)
243 if (mvi
->flags
& MVF_FLAG_SOC
)
244 slot_nr
= MVS_SOC_SLOTS
;
246 slot_nr
= MVS_CHIP_SLOT_SZ
;
248 spin_lock_init(&mvi
->lock
);
249 for (i
= 0; i
< mvi
->chip
->n_phy
; i
++) {
250 mvs_phy_init(mvi
, i
);
251 mvi
->port
[i
].wide_port_phymap
= 0;
252 mvi
->port
[i
].port_attached
= 0;
253 INIT_LIST_HEAD(&mvi
->port
[i
].list
);
255 for (i
= 0; i
< MVS_MAX_DEVICES
; i
++) {
256 mvi
->devices
[i
].taskfileset
= MVS_ID_NOT_MAPPED
;
257 mvi
->devices
[i
].dev_type
= SAS_PHY_UNUSED
;
258 mvi
->devices
[i
].device_id
= i
;
259 mvi
->devices
[i
].dev_status
= MVS_DEV_NORMAL
;
260 init_timer(&mvi
->devices
[i
].timer
);
264 * alloc and init our DMA areas
266 mvi
->tx
= dma_alloc_coherent(mvi
->dev
,
267 sizeof(*mvi
->tx
) * MVS_CHIP_SLOT_SZ
,
268 &mvi
->tx_dma
, GFP_KERNEL
);
271 memset(mvi
->tx
, 0, sizeof(*mvi
->tx
) * MVS_CHIP_SLOT_SZ
);
272 mvi
->rx_fis
= dma_alloc_coherent(mvi
->dev
, MVS_RX_FISL_SZ
,
273 &mvi
->rx_fis_dma
, GFP_KERNEL
);
276 memset(mvi
->rx_fis
, 0, MVS_RX_FISL_SZ
);
278 mvi
->rx
= dma_alloc_coherent(mvi
->dev
,
279 sizeof(*mvi
->rx
) * (MVS_RX_RING_SZ
+ 1),
280 &mvi
->rx_dma
, GFP_KERNEL
);
283 memset(mvi
->rx
, 0, sizeof(*mvi
->rx
) * (MVS_RX_RING_SZ
+ 1));
284 mvi
->rx
[0] = cpu_to_le32(0xfff);
285 mvi
->rx_cons
= 0xfff;
287 mvi
->slot
= dma_alloc_coherent(mvi
->dev
,
288 sizeof(*mvi
->slot
) * slot_nr
,
289 &mvi
->slot_dma
, GFP_KERNEL
);
292 memset(mvi
->slot
, 0, sizeof(*mvi
->slot
) * slot_nr
);
294 mvi
->bulk_buffer
= dma_alloc_coherent(mvi
->dev
,
296 &mvi
->bulk_buffer_dma
, GFP_KERNEL
);
297 if (!mvi
->bulk_buffer
)
300 mvi
->bulk_buffer1
= dma_alloc_coherent(mvi
->dev
,
302 &mvi
->bulk_buffer_dma1
, GFP_KERNEL
);
303 if (!mvi
->bulk_buffer1
)
306 sprintf(pool_name
, "%s%d", "mvs_dma_pool", mvi
->id
);
307 mvi
->dma_pool
= pci_pool_create(pool_name
, mvi
->pdev
, MVS_SLOT_BUF_SZ
, 16, 0);
308 if (!mvi
->dma_pool
) {
309 printk(KERN_DEBUG
"failed to create dma pool %s.\n", pool_name
);
312 mvi
->tags_num
= slot_nr
;
314 /* Initialize tags */
322 int mvs_ioremap(struct mvs_info
*mvi
, int bar
, int bar_ex
)
324 unsigned long res_start
, res_len
, res_flag
, res_flag_ex
= 0;
325 struct pci_dev
*pdev
= mvi
->pdev
;
328 * ioremap main and peripheral registers
330 res_start
= pci_resource_start(pdev
, bar_ex
);
331 res_len
= pci_resource_len(pdev
, bar_ex
);
332 if (!res_start
|| !res_len
)
335 res_flag_ex
= pci_resource_flags(pdev
, bar_ex
);
336 if (res_flag_ex
& IORESOURCE_MEM
) {
337 if (res_flag_ex
& IORESOURCE_CACHEABLE
)
338 mvi
->regs_ex
= ioremap(res_start
, res_len
);
340 mvi
->regs_ex
= ioremap_nocache(res_start
,
343 mvi
->regs_ex
= (void *)res_start
;
348 res_start
= pci_resource_start(pdev
, bar
);
349 res_len
= pci_resource_len(pdev
, bar
);
350 if (!res_start
|| !res_len
)
353 res_flag
= pci_resource_flags(pdev
, bar
);
354 if (res_flag
& IORESOURCE_CACHEABLE
)
355 mvi
->regs
= ioremap(res_start
, res_len
);
357 mvi
->regs
= ioremap_nocache(res_start
, res_len
);
360 if (mvi
->regs_ex
&& (res_flag_ex
& IORESOURCE_MEM
))
361 iounmap(mvi
->regs_ex
);
371 void mvs_iounmap(void __iomem
*regs
)
376 static struct mvs_info
*mvs_pci_alloc(struct pci_dev
*pdev
,
377 const struct pci_device_id
*ent
,
378 struct Scsi_Host
*shost
, unsigned int id
)
380 struct mvs_info
*mvi
= NULL
;
381 struct sas_ha_struct
*sha
= SHOST_TO_SAS_HA(shost
);
383 mvi
= kzalloc(sizeof(*mvi
) +
384 (1L << mvs_chips
[ent
->driver_data
].slot_width
) *
385 sizeof(struct mvs_slot_info
), GFP_KERNEL
);
390 mvi
->dev
= &pdev
->dev
;
391 mvi
->chip_id
= ent
->driver_data
;
392 mvi
->chip
= &mvs_chips
[mvi
->chip_id
];
393 INIT_LIST_HEAD(&mvi
->wq_list
);
395 ((struct mvs_prv_info
*)sha
->lldd_ha
)->mvi
[id
] = mvi
;
396 ((struct mvs_prv_info
*)sha
->lldd_ha
)->n_phy
= mvi
->chip
->n_phy
;
402 mvi
->tags
= kzalloc(MVS_CHIP_SLOT_SZ
>>3, GFP_KERNEL
);
406 if (MVS_CHIP_DISP
->chip_ioremap(mvi
))
408 if (!mvs_alloc(mvi
, shost
))
415 static int pci_go_64(struct pci_dev
*pdev
)
419 if (!pci_set_dma_mask(pdev
, DMA_BIT_MASK(64))) {
420 rc
= pci_set_consistent_dma_mask(pdev
, DMA_BIT_MASK(64));
422 rc
= pci_set_consistent_dma_mask(pdev
, DMA_BIT_MASK(32));
424 dev_printk(KERN_ERR
, &pdev
->dev
,
425 "64-bit DMA enable failed\n");
430 rc
= pci_set_dma_mask(pdev
, DMA_BIT_MASK(32));
432 dev_printk(KERN_ERR
, &pdev
->dev
,
433 "32-bit DMA enable failed\n");
436 rc
= pci_set_consistent_dma_mask(pdev
, DMA_BIT_MASK(32));
438 dev_printk(KERN_ERR
, &pdev
->dev
,
439 "32-bit consistent DMA enable failed\n");
447 static int mvs_prep_sas_ha_init(struct Scsi_Host
*shost
,
448 const struct mvs_chip_info
*chip_info
)
450 int phy_nr
, port_nr
; unsigned short core_nr
;
451 struct asd_sas_phy
**arr_phy
;
452 struct asd_sas_port
**arr_port
;
453 struct sas_ha_struct
*sha
= SHOST_TO_SAS_HA(shost
);
455 core_nr
= chip_info
->n_host
;
456 phy_nr
= core_nr
* chip_info
->n_phy
;
459 memset(sha
, 0x00, sizeof(struct sas_ha_struct
));
460 arr_phy
= kcalloc(phy_nr
, sizeof(void *), GFP_KERNEL
);
461 arr_port
= kcalloc(port_nr
, sizeof(void *), GFP_KERNEL
);
462 if (!arr_phy
|| !arr_port
)
465 sha
->sas_phy
= arr_phy
;
466 sha
->sas_port
= arr_port
;
467 sha
->core
.shost
= shost
;
469 sha
->lldd_ha
= kzalloc(sizeof(struct mvs_prv_info
), GFP_KERNEL
);
473 ((struct mvs_prv_info
*)sha
->lldd_ha
)->n_host
= core_nr
;
475 shost
->transportt
= mvs_stt
;
476 shost
->max_id
= MVS_MAX_DEVICES
;
478 shost
->max_channel
= 1;
479 shost
->max_cmd_len
= 16;
489 static void mvs_post_sas_ha_init(struct Scsi_Host
*shost
,
490 const struct mvs_chip_info
*chip_info
)
492 int can_queue
, i
= 0, j
= 0;
493 struct mvs_info
*mvi
= NULL
;
494 struct sas_ha_struct
*sha
= SHOST_TO_SAS_HA(shost
);
495 unsigned short nr_core
= ((struct mvs_prv_info
*)sha
->lldd_ha
)->n_host
;
497 for (j
= 0; j
< nr_core
; j
++) {
498 mvi
= ((struct mvs_prv_info
*)sha
->lldd_ha
)->mvi
[j
];
499 for (i
= 0; i
< chip_info
->n_phy
; i
++) {
500 sha
->sas_phy
[j
* chip_info
->n_phy
+ i
] =
501 &mvi
->phy
[i
].sas_phy
;
502 sha
->sas_port
[j
* chip_info
->n_phy
+ i
] =
503 &mvi
->port
[i
].sas_port
;
507 sha
->sas_ha_name
= DRV_NAME
;
509 sha
->lldd_module
= THIS_MODULE
;
510 sha
->sas_addr
= &mvi
->sas_addr
[0];
512 sha
->num_phys
= nr_core
* chip_info
->n_phy
;
514 sha
->lldd_max_execute_num
= lldd_max_execute_num
;
516 if (mvi
->flags
& MVF_FLAG_SOC
)
517 can_queue
= MVS_SOC_CAN_QUEUE
;
519 can_queue
= MVS_CHIP_SLOT_SZ
;
521 sha
->lldd_queue_size
= can_queue
;
522 shost
->sg_tablesize
= min_t(u16
, SG_ALL
, MVS_MAX_SG
);
523 shost
->can_queue
= can_queue
;
524 mvi
->shost
->cmd_per_lun
= MVS_QUEUE_SIZE
;
525 sha
->core
.shost
= mvi
->shost
;
528 static void mvs_init_sas_add(struct mvs_info
*mvi
)
531 for (i
= 0; i
< mvi
->chip
->n_phy
; i
++) {
532 mvi
->phy
[i
].dev_sas_addr
= 0x5005043011ab0000ULL
;
533 mvi
->phy
[i
].dev_sas_addr
=
534 cpu_to_be64((u64
)(*(u64
*)&mvi
->phy
[i
].dev_sas_addr
));
537 memcpy(mvi
->sas_addr
, &mvi
->phy
[0].dev_sas_addr
, SAS_ADDR_SIZE
);
540 static int mvs_pci_init(struct pci_dev
*pdev
, const struct pci_device_id
*ent
)
542 unsigned int rc
, nhost
= 0;
543 struct mvs_info
*mvi
;
544 struct mvs_prv_info
*mpi
;
545 irq_handler_t irq_handler
= mvs_interrupt
;
546 struct Scsi_Host
*shost
= NULL
;
547 const struct mvs_chip_info
*chip
;
549 dev_printk(KERN_INFO
, &pdev
->dev
,
550 "mvsas: driver version %s\n", DRV_VERSION
);
551 rc
= pci_enable_device(pdev
);
555 pci_set_master(pdev
);
557 rc
= pci_request_regions(pdev
, DRV_NAME
);
559 goto err_out_disable
;
561 rc
= pci_go_64(pdev
);
563 goto err_out_regions
;
565 shost
= scsi_host_alloc(&mvs_sht
, sizeof(void *));
568 goto err_out_regions
;
571 chip
= &mvs_chips
[ent
->driver_data
];
572 SHOST_TO_SAS_HA(shost
) =
573 kcalloc(1, sizeof(struct sas_ha_struct
), GFP_KERNEL
);
574 if (!SHOST_TO_SAS_HA(shost
)) {
577 goto err_out_regions
;
580 rc
= mvs_prep_sas_ha_init(shost
, chip
);
584 goto err_out_regions
;
587 pci_set_drvdata(pdev
, SHOST_TO_SAS_HA(shost
));
590 mvi
= mvs_pci_alloc(pdev
, ent
, shost
, nhost
);
593 goto err_out_regions
;
596 memset(&mvi
->hba_info_param
, 0xFF,
597 sizeof(struct hba_info_page
));
599 mvs_init_sas_add(mvi
);
601 mvi
->instance
= nhost
;
602 rc
= MVS_CHIP_DISP
->chip_init(mvi
);
605 goto err_out_regions
;
608 } while (nhost
< chip
->n_host
);
609 mpi
= (struct mvs_prv_info
*)(SHOST_TO_SAS_HA(shost
)->lldd_ha
);
610 #ifdef CONFIG_SCSI_MVSAS_TASKLET
611 tasklet_init(&(mpi
->mv_tasklet
), mvs_tasklet
,
612 (unsigned long)SHOST_TO_SAS_HA(shost
));
615 mvs_post_sas_ha_init(shost
, chip
);
617 rc
= scsi_add_host(shost
, &pdev
->dev
);
621 rc
= sas_register_ha(SHOST_TO_SAS_HA(shost
));
624 rc
= request_irq(pdev
->irq
, irq_handler
, IRQF_SHARED
,
625 DRV_NAME
, SHOST_TO_SAS_HA(shost
));
629 MVS_CHIP_DISP
->interrupt_enable(mvi
);
631 scsi_scan_host(mvi
->shost
);
636 sas_unregister_ha(SHOST_TO_SAS_HA(shost
));
638 scsi_remove_host(mvi
->shost
);
640 pci_release_regions(pdev
);
642 pci_disable_device(pdev
);
647 static void mvs_pci_remove(struct pci_dev
*pdev
)
649 unsigned short core_nr
, i
= 0;
650 struct sas_ha_struct
*sha
= pci_get_drvdata(pdev
);
651 struct mvs_info
*mvi
= NULL
;
653 core_nr
= ((struct mvs_prv_info
*)sha
->lldd_ha
)->n_host
;
654 mvi
= ((struct mvs_prv_info
*)sha
->lldd_ha
)->mvi
[0];
656 #ifdef CONFIG_SCSI_MVSAS_TASKLET
657 tasklet_kill(&((struct mvs_prv_info
*)sha
->lldd_ha
)->mv_tasklet
);
660 sas_unregister_ha(sha
);
661 sas_remove_host(mvi
->shost
);
662 scsi_remove_host(mvi
->shost
);
664 MVS_CHIP_DISP
->interrupt_disable(mvi
);
665 free_irq(mvi
->pdev
->irq
, sha
);
666 for (i
= 0; i
< core_nr
; i
++) {
667 mvi
= ((struct mvs_prv_info
*)sha
->lldd_ha
)->mvi
[i
];
671 kfree(sha
->sas_port
);
673 pci_release_regions(pdev
);
674 pci_disable_device(pdev
);
678 static struct pci_device_id mvs_pci_table
[] = {
679 { PCI_VDEVICE(MARVELL
, 0x6320), chip_6320
},
680 { PCI_VDEVICE(MARVELL
, 0x6340), chip_6440
},
682 .vendor
= PCI_VENDOR_ID_MARVELL
,
684 .subvendor
= PCI_ANY_ID
,
688 .driver_data
= chip_6485
,
690 { PCI_VDEVICE(MARVELL
, 0x6440), chip_6440
},
691 { PCI_VDEVICE(MARVELL
, 0x6485), chip_6485
},
692 { PCI_VDEVICE(MARVELL
, 0x9480), chip_9480
},
693 { PCI_VDEVICE(MARVELL
, 0x9180), chip_9180
},
694 { PCI_VDEVICE(ARECA
, PCI_DEVICE_ID_ARECA_1300
), chip_1300
},
695 { PCI_VDEVICE(ARECA
, PCI_DEVICE_ID_ARECA_1320
), chip_1320
},
696 { PCI_VDEVICE(ADAPTEC2
, 0x0450), chip_6440
},
697 { PCI_VDEVICE(TTI
, 0x2710), chip_9480
},
698 { PCI_VDEVICE(TTI
, 0x2720), chip_9480
},
699 { PCI_VDEVICE(TTI
, 0x2721), chip_9480
},
700 { PCI_VDEVICE(TTI
, 0x2722), chip_9480
},
701 { PCI_VDEVICE(TTI
, 0x2740), chip_9480
},
702 { PCI_VDEVICE(TTI
, 0x2744), chip_9480
},
703 { PCI_VDEVICE(TTI
, 0x2760), chip_9480
},
705 .vendor
= PCI_VENDOR_ID_MARVELL_EXT
,
707 .subvendor
= PCI_ANY_ID
,
711 .driver_data
= chip_9480
,
714 .vendor
= PCI_VENDOR_ID_MARVELL_EXT
,
716 .subvendor
= PCI_ANY_ID
,
720 .driver_data
= chip_9445
,
723 .vendor
= PCI_VENDOR_ID_MARVELL_EXT
,
725 .subvendor
= PCI_ANY_ID
,
729 .driver_data
= chip_9485
,
731 { PCI_VDEVICE(OCZ
, 0x1021), chip_9485
}, /* OCZ RevoDrive3 */
732 { PCI_VDEVICE(OCZ
, 0x1022), chip_9485
}, /* OCZ RevoDrive3/zDriveR4 (exact model unknown) */
733 { PCI_VDEVICE(OCZ
, 0x1040), chip_9485
}, /* OCZ RevoDrive3/zDriveR4 (exact model unknown) */
734 { PCI_VDEVICE(OCZ
, 0x1041), chip_9485
}, /* OCZ RevoDrive3/zDriveR4 (exact model unknown) */
735 { PCI_VDEVICE(OCZ
, 0x1042), chip_9485
}, /* OCZ RevoDrive3/zDriveR4 (exact model unknown) */
736 { PCI_VDEVICE(OCZ
, 0x1043), chip_9485
}, /* OCZ RevoDrive3/zDriveR4 (exact model unknown) */
737 { PCI_VDEVICE(OCZ
, 0x1044), chip_9485
}, /* OCZ RevoDrive3/zDriveR4 (exact model unknown) */
738 { PCI_VDEVICE(OCZ
, 0x1080), chip_9485
}, /* OCZ RevoDrive3/zDriveR4 (exact model unknown) */
739 { PCI_VDEVICE(OCZ
, 0x1083), chip_9485
}, /* OCZ RevoDrive3/zDriveR4 (exact model unknown) */
740 { PCI_VDEVICE(OCZ
, 0x1084), chip_9485
}, /* OCZ RevoDrive3/zDriveR4 (exact model unknown) */
742 { } /* terminate list */
745 static struct pci_driver mvs_pci_driver
= {
747 .id_table
= mvs_pci_table
,
748 .probe
= mvs_pci_init
,
749 .remove
= mvs_pci_remove
,
753 mvs_show_driver_version(struct device
*cdev
,
754 struct device_attribute
*attr
, char *buffer
)
756 return snprintf(buffer
, PAGE_SIZE
, "%s\n", DRV_VERSION
);
759 static DEVICE_ATTR(driver_version
,
761 mvs_show_driver_version
,
765 mvs_store_interrupt_coalescing(struct device
*cdev
,
766 struct device_attribute
*attr
,
767 const char *buffer
, size_t size
)
770 struct mvs_info
*mvi
= NULL
;
771 struct Scsi_Host
*shost
= class_to_shost(cdev
);
772 struct sas_ha_struct
*sha
= SHOST_TO_SAS_HA(shost
);
777 if (sscanf(buffer
, "%d", &val
) != 1)
780 if (val
>= 0x10000) {
781 mv_dprintk("interrupt coalescing timer %d us is"
783 return strlen(buffer
);
786 interrupt_coalescing
= val
;
788 core_nr
= ((struct mvs_prv_info
*)sha
->lldd_ha
)->n_host
;
789 mvi
= ((struct mvs_prv_info
*)sha
->lldd_ha
)->mvi
[0];
794 for (i
= 0; i
< core_nr
; i
++) {
795 mvi
= ((struct mvs_prv_info
*)sha
->lldd_ha
)->mvi
[i
];
796 if (MVS_CHIP_DISP
->tune_interrupt
)
797 MVS_CHIP_DISP
->tune_interrupt(mvi
,
798 interrupt_coalescing
);
800 mv_dprintk("set interrupt coalescing time to %d us\n",
801 interrupt_coalescing
);
802 return strlen(buffer
);
805 static ssize_t
mvs_show_interrupt_coalescing(struct device
*cdev
,
806 struct device_attribute
*attr
, char *buffer
)
808 return snprintf(buffer
, PAGE_SIZE
, "%d\n", interrupt_coalescing
);
811 static DEVICE_ATTR(interrupt_coalescing
,
813 mvs_show_interrupt_coalescing
,
814 mvs_store_interrupt_coalescing
);
817 struct task_struct
*mvs_th
;
818 static int __init
mvs_init(void)
821 mvs_stt
= sas_domain_attach_transport(&mvs_transport_ops
);
825 mvs_task_list_cache
= kmem_cache_create("mvs_task_list", sizeof(struct mvs_task_list
),
826 0, SLAB_HWCACHE_ALIGN
, NULL
);
827 if (!mvs_task_list_cache
) {
829 mv_printk("%s: mvs_task_list_cache alloc failed! \n", __func__
);
833 rc
= pci_register_driver(&mvs_pci_driver
);
841 sas_release_transport(mvs_stt
);
845 static void __exit
mvs_exit(void)
847 pci_unregister_driver(&mvs_pci_driver
);
848 sas_release_transport(mvs_stt
);
849 kmem_cache_destroy(mvs_task_list_cache
);
852 struct device_attribute
*mvst_host_attrs
[] = {
853 &dev_attr_driver_version
,
854 &dev_attr_interrupt_coalescing
,
858 module_init(mvs_init
);
859 module_exit(mvs_exit
);
861 MODULE_AUTHOR("Jeff Garzik <jgarzik@pobox.com>");
862 MODULE_DESCRIPTION("Marvell 88SE6440 SAS/SATA controller driver");
863 MODULE_VERSION(DRV_VERSION
);
864 MODULE_LICENSE("GPL");
866 MODULE_DEVICE_TABLE(pci
, mvs_pci_table
);