1 // SPDX-License-Identifier: GPL-2.0
3 * Linux Driver for Mylex DAC960/AcceleRAID/eXtremeRAID PCI RAID Controllers
5 * This driver supports the newer, SCSI-based firmware interface only.
7 * Copyright 2017 Hannes Reinecke, SUSE Linux GmbH <hare@suse.com>
9 * Based on the original DAC960 driver, which has
10 * Copyright 1998-2001 by Leonard N. Zubkoff <lnz@dandelion.com>
11 * Portions Copyright 2002 by Mylex (An IBM Business Unit)
14 #include <linux/module.h>
15 #include <linux/types.h>
16 #include <linux/delay.h>
17 #include <linux/interrupt.h>
18 #include <linux/pci.h>
19 #include <linux/raid_class.h>
20 #include <asm/unaligned.h>
21 #include <scsi/scsi.h>
22 #include <scsi/scsi_host.h>
23 #include <scsi/scsi_device.h>
24 #include <scsi/scsi_cmnd.h>
25 #include <scsi/scsi_tcq.h>
28 static struct raid_template
*myrs_raid_template
;
30 static struct myrs_devstate_name_entry
{
31 enum myrs_devstate state
;
33 } myrs_devstate_name_list
[] = {
34 { MYRS_DEVICE_UNCONFIGURED
, "Unconfigured" },
35 { MYRS_DEVICE_ONLINE
, "Online" },
36 { MYRS_DEVICE_REBUILD
, "Rebuild" },
37 { MYRS_DEVICE_MISSING
, "Missing" },
38 { MYRS_DEVICE_SUSPECTED_CRITICAL
, "SuspectedCritical" },
39 { MYRS_DEVICE_OFFLINE
, "Offline" },
40 { MYRS_DEVICE_CRITICAL
, "Critical" },
41 { MYRS_DEVICE_SUSPECTED_DEAD
, "SuspectedDead" },
42 { MYRS_DEVICE_COMMANDED_OFFLINE
, "CommandedOffline" },
43 { MYRS_DEVICE_STANDBY
, "Standby" },
44 { MYRS_DEVICE_INVALID_STATE
, "Invalid" },
47 static char *myrs_devstate_name(enum myrs_devstate state
)
49 struct myrs_devstate_name_entry
*entry
= myrs_devstate_name_list
;
52 for (i
= 0; i
< ARRAY_SIZE(myrs_devstate_name_list
); i
++) {
53 if (entry
[i
].state
== state
)
59 static struct myrs_raid_level_name_entry
{
60 enum myrs_raid_level level
;
62 } myrs_raid_level_name_list
[] = {
63 { MYRS_RAID_LEVEL0
, "RAID0" },
64 { MYRS_RAID_LEVEL1
, "RAID1" },
65 { MYRS_RAID_LEVEL3
, "RAID3 right asymmetric parity" },
66 { MYRS_RAID_LEVEL5
, "RAID5 right asymmetric parity" },
67 { MYRS_RAID_LEVEL6
, "RAID6" },
68 { MYRS_RAID_JBOD
, "JBOD" },
69 { MYRS_RAID_NEWSPAN
, "New Mylex SPAN" },
70 { MYRS_RAID_LEVEL3F
, "RAID3 fixed parity" },
71 { MYRS_RAID_LEVEL3L
, "RAID3 left symmetric parity" },
72 { MYRS_RAID_SPAN
, "Mylex SPAN" },
73 { MYRS_RAID_LEVEL5L
, "RAID5 left symmetric parity" },
74 { MYRS_RAID_LEVELE
, "RAIDE (concatenation)" },
75 { MYRS_RAID_PHYSICAL
, "Physical device" },
78 static char *myrs_raid_level_name(enum myrs_raid_level level
)
80 struct myrs_raid_level_name_entry
*entry
= myrs_raid_level_name_list
;
83 for (i
= 0; i
< ARRAY_SIZE(myrs_raid_level_name_list
); i
++) {
84 if (entry
[i
].level
== level
)
91 * myrs_reset_cmd - clears critical fields in struct myrs_cmdblk
93 static inline void myrs_reset_cmd(struct myrs_cmdblk
*cmd_blk
)
95 union myrs_cmd_mbox
*mbox
= &cmd_blk
->mbox
;
97 memset(mbox
, 0, sizeof(union myrs_cmd_mbox
));
102 * myrs_qcmd - queues Command for DAC960 V2 Series Controllers.
104 static void myrs_qcmd(struct myrs_hba
*cs
, struct myrs_cmdblk
*cmd_blk
)
106 void __iomem
*base
= cs
->io_base
;
107 union myrs_cmd_mbox
*mbox
= &cmd_blk
->mbox
;
108 union myrs_cmd_mbox
*next_mbox
= cs
->next_cmd_mbox
;
110 cs
->write_cmd_mbox(next_mbox
, mbox
);
112 if (cs
->prev_cmd_mbox1
->words
[0] == 0 ||
113 cs
->prev_cmd_mbox2
->words
[0] == 0)
114 cs
->get_cmd_mbox(base
);
116 cs
->prev_cmd_mbox2
= cs
->prev_cmd_mbox1
;
117 cs
->prev_cmd_mbox1
= next_mbox
;
119 if (++next_mbox
> cs
->last_cmd_mbox
)
120 next_mbox
= cs
->first_cmd_mbox
;
122 cs
->next_cmd_mbox
= next_mbox
;
126 * myrs_exec_cmd - executes V2 Command and waits for completion.
128 static void myrs_exec_cmd(struct myrs_hba
*cs
,
129 struct myrs_cmdblk
*cmd_blk
)
131 DECLARE_COMPLETION_ONSTACK(complete
);
134 cmd_blk
->complete
= &complete
;
135 spin_lock_irqsave(&cs
->queue_lock
, flags
);
136 myrs_qcmd(cs
, cmd_blk
);
137 spin_unlock_irqrestore(&cs
->queue_lock
, flags
);
139 WARN_ON(in_interrupt());
140 wait_for_completion(&complete
);
144 * myrs_report_progress - prints progress message
146 static void myrs_report_progress(struct myrs_hba
*cs
, unsigned short ldev_num
,
147 unsigned char *msg
, unsigned long blocks
,
150 shost_printk(KERN_INFO
, cs
->host
,
151 "Logical Drive %d: %s in Progress: %d%% completed\n",
153 (100 * (int)(blocks
>> 7)) / (int)(size
>> 7));
157 * myrs_get_ctlr_info - executes a Controller Information IOCTL Command
159 static unsigned char myrs_get_ctlr_info(struct myrs_hba
*cs
)
161 struct myrs_cmdblk
*cmd_blk
= &cs
->dcmd_blk
;
162 union myrs_cmd_mbox
*mbox
= &cmd_blk
->mbox
;
163 dma_addr_t ctlr_info_addr
;
165 unsigned char status
;
166 unsigned short ldev_present
, ldev_critical
, ldev_offline
;
168 ldev_present
= cs
->ctlr_info
->ldev_present
;
169 ldev_critical
= cs
->ctlr_info
->ldev_critical
;
170 ldev_offline
= cs
->ctlr_info
->ldev_offline
;
172 ctlr_info_addr
= dma_map_single(&cs
->pdev
->dev
, cs
->ctlr_info
,
173 sizeof(struct myrs_ctlr_info
),
175 if (dma_mapping_error(&cs
->pdev
->dev
, ctlr_info_addr
))
176 return MYRS_STATUS_FAILED
;
178 mutex_lock(&cs
->dcmd_mutex
);
179 myrs_reset_cmd(cmd_blk
);
180 mbox
->ctlr_info
.id
= MYRS_DCMD_TAG
;
181 mbox
->ctlr_info
.opcode
= MYRS_CMD_OP_IOCTL
;
182 mbox
->ctlr_info
.control
.dma_ctrl_to_host
= true;
183 mbox
->ctlr_info
.control
.no_autosense
= true;
184 mbox
->ctlr_info
.dma_size
= sizeof(struct myrs_ctlr_info
);
185 mbox
->ctlr_info
.ctlr_num
= 0;
186 mbox
->ctlr_info
.ioctl_opcode
= MYRS_IOCTL_GET_CTLR_INFO
;
187 sgl
= &mbox
->ctlr_info
.dma_addr
;
188 sgl
->sge
[0].sge_addr
= ctlr_info_addr
;
189 sgl
->sge
[0].sge_count
= mbox
->ctlr_info
.dma_size
;
190 dev_dbg(&cs
->host
->shost_gendev
, "Sending GetControllerInfo\n");
191 myrs_exec_cmd(cs
, cmd_blk
);
192 status
= cmd_blk
->status
;
193 mutex_unlock(&cs
->dcmd_mutex
);
194 dma_unmap_single(&cs
->pdev
->dev
, ctlr_info_addr
,
195 sizeof(struct myrs_ctlr_info
), DMA_FROM_DEVICE
);
196 if (status
== MYRS_STATUS_SUCCESS
) {
197 if (cs
->ctlr_info
->bg_init_active
+
198 cs
->ctlr_info
->ldev_init_active
+
199 cs
->ctlr_info
->pdev_init_active
+
200 cs
->ctlr_info
->cc_active
+
201 cs
->ctlr_info
->rbld_active
+
202 cs
->ctlr_info
->exp_active
!= 0)
203 cs
->needs_update
= true;
204 if (cs
->ctlr_info
->ldev_present
!= ldev_present
||
205 cs
->ctlr_info
->ldev_critical
!= ldev_critical
||
206 cs
->ctlr_info
->ldev_offline
!= ldev_offline
)
207 shost_printk(KERN_INFO
, cs
->host
,
208 "Logical drive count changes (%d/%d/%d)\n",
209 cs
->ctlr_info
->ldev_critical
,
210 cs
->ctlr_info
->ldev_offline
,
211 cs
->ctlr_info
->ldev_present
);
218 * myrs_get_ldev_info - executes a Logical Device Information IOCTL Command
220 static unsigned char myrs_get_ldev_info(struct myrs_hba
*cs
,
221 unsigned short ldev_num
, struct myrs_ldev_info
*ldev_info
)
223 struct myrs_cmdblk
*cmd_blk
= &cs
->dcmd_blk
;
224 union myrs_cmd_mbox
*mbox
= &cmd_blk
->mbox
;
225 dma_addr_t ldev_info_addr
;
226 struct myrs_ldev_info ldev_info_orig
;
228 unsigned char status
;
230 memcpy(&ldev_info_orig
, ldev_info
, sizeof(struct myrs_ldev_info
));
231 ldev_info_addr
= dma_map_single(&cs
->pdev
->dev
, ldev_info
,
232 sizeof(struct myrs_ldev_info
),
234 if (dma_mapping_error(&cs
->pdev
->dev
, ldev_info_addr
))
235 return MYRS_STATUS_FAILED
;
237 mutex_lock(&cs
->dcmd_mutex
);
238 myrs_reset_cmd(cmd_blk
);
239 mbox
->ldev_info
.id
= MYRS_DCMD_TAG
;
240 mbox
->ldev_info
.opcode
= MYRS_CMD_OP_IOCTL
;
241 mbox
->ldev_info
.control
.dma_ctrl_to_host
= true;
242 mbox
->ldev_info
.control
.no_autosense
= true;
243 mbox
->ldev_info
.dma_size
= sizeof(struct myrs_ldev_info
);
244 mbox
->ldev_info
.ldev
.ldev_num
= ldev_num
;
245 mbox
->ldev_info
.ioctl_opcode
= MYRS_IOCTL_GET_LDEV_INFO_VALID
;
246 sgl
= &mbox
->ldev_info
.dma_addr
;
247 sgl
->sge
[0].sge_addr
= ldev_info_addr
;
248 sgl
->sge
[0].sge_count
= mbox
->ldev_info
.dma_size
;
249 dev_dbg(&cs
->host
->shost_gendev
,
250 "Sending GetLogicalDeviceInfoValid for ldev %d\n", ldev_num
);
251 myrs_exec_cmd(cs
, cmd_blk
);
252 status
= cmd_blk
->status
;
253 mutex_unlock(&cs
->dcmd_mutex
);
254 dma_unmap_single(&cs
->pdev
->dev
, ldev_info_addr
,
255 sizeof(struct myrs_ldev_info
), DMA_FROM_DEVICE
);
256 if (status
== MYRS_STATUS_SUCCESS
) {
257 unsigned short ldev_num
= ldev_info
->ldev_num
;
258 struct myrs_ldev_info
*new = ldev_info
;
259 struct myrs_ldev_info
*old
= &ldev_info_orig
;
260 unsigned long ldev_size
= new->cfg_devsize
;
262 if (new->dev_state
!= old
->dev_state
) {
265 name
= myrs_devstate_name(new->dev_state
);
266 shost_printk(KERN_INFO
, cs
->host
,
267 "Logical Drive %d is now %s\n",
268 ldev_num
, name
? name
: "Invalid");
270 if ((new->soft_errs
!= old
->soft_errs
) ||
271 (new->cmds_failed
!= old
->cmds_failed
) ||
272 (new->deferred_write_errs
!= old
->deferred_write_errs
))
273 shost_printk(KERN_INFO
, cs
->host
,
274 "Logical Drive %d Errors: Soft = %d, Failed = %d, Deferred Write = %d\n",
275 ldev_num
, new->soft_errs
,
277 new->deferred_write_errs
);
278 if (new->bg_init_active
)
279 myrs_report_progress(cs
, ldev_num
,
280 "Background Initialization",
281 new->bg_init_lba
, ldev_size
);
282 else if (new->fg_init_active
)
283 myrs_report_progress(cs
, ldev_num
,
284 "Foreground Initialization",
285 new->fg_init_lba
, ldev_size
);
286 else if (new->migration_active
)
287 myrs_report_progress(cs
, ldev_num
,
289 new->migration_lba
, ldev_size
);
290 else if (new->patrol_active
)
291 myrs_report_progress(cs
, ldev_num
,
293 new->patrol_lba
, ldev_size
);
294 if (old
->bg_init_active
&& !new->bg_init_active
)
295 shost_printk(KERN_INFO
, cs
->host
,
296 "Logical Drive %d: Background Initialization %s\n",
298 (new->ldev_control
.ldev_init_done
?
299 "Completed" : "Failed"));
305 * myrs_get_pdev_info - executes a "Read Physical Device Information" Command
307 static unsigned char myrs_get_pdev_info(struct myrs_hba
*cs
,
308 unsigned char channel
, unsigned char target
, unsigned char lun
,
309 struct myrs_pdev_info
*pdev_info
)
311 struct myrs_cmdblk
*cmd_blk
= &cs
->dcmd_blk
;
312 union myrs_cmd_mbox
*mbox
= &cmd_blk
->mbox
;
313 dma_addr_t pdev_info_addr
;
315 unsigned char status
;
317 pdev_info_addr
= dma_map_single(&cs
->pdev
->dev
, pdev_info
,
318 sizeof(struct myrs_pdev_info
),
320 if (dma_mapping_error(&cs
->pdev
->dev
, pdev_info_addr
))
321 return MYRS_STATUS_FAILED
;
323 mutex_lock(&cs
->dcmd_mutex
);
324 myrs_reset_cmd(cmd_blk
);
325 mbox
->pdev_info
.opcode
= MYRS_CMD_OP_IOCTL
;
326 mbox
->pdev_info
.id
= MYRS_DCMD_TAG
;
327 mbox
->pdev_info
.control
.dma_ctrl_to_host
= true;
328 mbox
->pdev_info
.control
.no_autosense
= true;
329 mbox
->pdev_info
.dma_size
= sizeof(struct myrs_pdev_info
);
330 mbox
->pdev_info
.pdev
.lun
= lun
;
331 mbox
->pdev_info
.pdev
.target
= target
;
332 mbox
->pdev_info
.pdev
.channel
= channel
;
333 mbox
->pdev_info
.ioctl_opcode
= MYRS_IOCTL_GET_PDEV_INFO_VALID
;
334 sgl
= &mbox
->pdev_info
.dma_addr
;
335 sgl
->sge
[0].sge_addr
= pdev_info_addr
;
336 sgl
->sge
[0].sge_count
= mbox
->pdev_info
.dma_size
;
337 dev_dbg(&cs
->host
->shost_gendev
,
338 "Sending GetPhysicalDeviceInfoValid for pdev %d:%d:%d\n",
339 channel
, target
, lun
);
340 myrs_exec_cmd(cs
, cmd_blk
);
341 status
= cmd_blk
->status
;
342 mutex_unlock(&cs
->dcmd_mutex
);
343 dma_unmap_single(&cs
->pdev
->dev
, pdev_info_addr
,
344 sizeof(struct myrs_pdev_info
), DMA_FROM_DEVICE
);
349 * myrs_dev_op - executes a "Device Operation" Command
351 static unsigned char myrs_dev_op(struct myrs_hba
*cs
,
352 enum myrs_ioctl_opcode opcode
, enum myrs_opdev opdev
)
354 struct myrs_cmdblk
*cmd_blk
= &cs
->dcmd_blk
;
355 union myrs_cmd_mbox
*mbox
= &cmd_blk
->mbox
;
356 unsigned char status
;
358 mutex_lock(&cs
->dcmd_mutex
);
359 myrs_reset_cmd(cmd_blk
);
360 mbox
->dev_op
.opcode
= MYRS_CMD_OP_IOCTL
;
361 mbox
->dev_op
.id
= MYRS_DCMD_TAG
;
362 mbox
->dev_op
.control
.dma_ctrl_to_host
= true;
363 mbox
->dev_op
.control
.no_autosense
= true;
364 mbox
->dev_op
.ioctl_opcode
= opcode
;
365 mbox
->dev_op
.opdev
= opdev
;
366 myrs_exec_cmd(cs
, cmd_blk
);
367 status
= cmd_blk
->status
;
368 mutex_unlock(&cs
->dcmd_mutex
);
373 * myrs_translate_pdev - translates a Physical Device Channel and
374 * TargetID into a Logical Device.
376 static unsigned char myrs_translate_pdev(struct myrs_hba
*cs
,
377 unsigned char channel
, unsigned char target
, unsigned char lun
,
378 struct myrs_devmap
*devmap
)
380 struct pci_dev
*pdev
= cs
->pdev
;
381 dma_addr_t devmap_addr
;
382 struct myrs_cmdblk
*cmd_blk
;
383 union myrs_cmd_mbox
*mbox
;
385 unsigned char status
;
387 memset(devmap
, 0x0, sizeof(struct myrs_devmap
));
388 devmap_addr
= dma_map_single(&pdev
->dev
, devmap
,
389 sizeof(struct myrs_devmap
),
391 if (dma_mapping_error(&pdev
->dev
, devmap_addr
))
392 return MYRS_STATUS_FAILED
;
394 mutex_lock(&cs
->dcmd_mutex
);
395 cmd_blk
= &cs
->dcmd_blk
;
396 mbox
= &cmd_blk
->mbox
;
397 mbox
->pdev_info
.opcode
= MYRS_CMD_OP_IOCTL
;
398 mbox
->pdev_info
.control
.dma_ctrl_to_host
= true;
399 mbox
->pdev_info
.control
.no_autosense
= true;
400 mbox
->pdev_info
.dma_size
= sizeof(struct myrs_devmap
);
401 mbox
->pdev_info
.pdev
.target
= target
;
402 mbox
->pdev_info
.pdev
.channel
= channel
;
403 mbox
->pdev_info
.pdev
.lun
= lun
;
404 mbox
->pdev_info
.ioctl_opcode
= MYRS_IOCTL_XLATE_PDEV_TO_LDEV
;
405 sgl
= &mbox
->pdev_info
.dma_addr
;
406 sgl
->sge
[0].sge_addr
= devmap_addr
;
407 sgl
->sge
[0].sge_count
= mbox
->pdev_info
.dma_size
;
409 myrs_exec_cmd(cs
, cmd_blk
);
410 status
= cmd_blk
->status
;
411 mutex_unlock(&cs
->dcmd_mutex
);
412 dma_unmap_single(&pdev
->dev
, devmap_addr
,
413 sizeof(struct myrs_devmap
), DMA_FROM_DEVICE
);
418 * myrs_get_event - executes a Get Event Command
420 static unsigned char myrs_get_event(struct myrs_hba
*cs
,
421 unsigned int event_num
, struct myrs_event
*event_buf
)
423 struct pci_dev
*pdev
= cs
->pdev
;
424 dma_addr_t event_addr
;
425 struct myrs_cmdblk
*cmd_blk
= &cs
->mcmd_blk
;
426 union myrs_cmd_mbox
*mbox
= &cmd_blk
->mbox
;
428 unsigned char status
;
430 event_addr
= dma_map_single(&pdev
->dev
, event_buf
,
431 sizeof(struct myrs_event
), DMA_FROM_DEVICE
);
432 if (dma_mapping_error(&pdev
->dev
, event_addr
))
433 return MYRS_STATUS_FAILED
;
435 mbox
->get_event
.opcode
= MYRS_CMD_OP_IOCTL
;
436 mbox
->get_event
.dma_size
= sizeof(struct myrs_event
);
437 mbox
->get_event
.evnum_upper
= event_num
>> 16;
438 mbox
->get_event
.ctlr_num
= 0;
439 mbox
->get_event
.ioctl_opcode
= MYRS_IOCTL_GET_EVENT
;
440 mbox
->get_event
.evnum_lower
= event_num
& 0xFFFF;
441 sgl
= &mbox
->get_event
.dma_addr
;
442 sgl
->sge
[0].sge_addr
= event_addr
;
443 sgl
->sge
[0].sge_count
= mbox
->get_event
.dma_size
;
444 myrs_exec_cmd(cs
, cmd_blk
);
445 status
= cmd_blk
->status
;
446 dma_unmap_single(&pdev
->dev
, event_addr
,
447 sizeof(struct myrs_event
), DMA_FROM_DEVICE
);
453 * myrs_get_fwstatus - executes a Get Health Status Command
455 static unsigned char myrs_get_fwstatus(struct myrs_hba
*cs
)
457 struct myrs_cmdblk
*cmd_blk
= &cs
->mcmd_blk
;
458 union myrs_cmd_mbox
*mbox
= &cmd_blk
->mbox
;
460 unsigned char status
= cmd_blk
->status
;
462 myrs_reset_cmd(cmd_blk
);
463 mbox
->common
.opcode
= MYRS_CMD_OP_IOCTL
;
464 mbox
->common
.id
= MYRS_MCMD_TAG
;
465 mbox
->common
.control
.dma_ctrl_to_host
= true;
466 mbox
->common
.control
.no_autosense
= true;
467 mbox
->common
.dma_size
= sizeof(struct myrs_fwstat
);
468 mbox
->common
.ioctl_opcode
= MYRS_IOCTL_GET_HEALTH_STATUS
;
469 sgl
= &mbox
->common
.dma_addr
;
470 sgl
->sge
[0].sge_addr
= cs
->fwstat_addr
;
471 sgl
->sge
[0].sge_count
= mbox
->ctlr_info
.dma_size
;
472 dev_dbg(&cs
->host
->shost_gendev
, "Sending GetHealthStatus\n");
473 myrs_exec_cmd(cs
, cmd_blk
);
474 status
= cmd_blk
->status
;
480 * myrs_enable_mmio_mbox - enables the Memory Mailbox Interface
482 static bool myrs_enable_mmio_mbox(struct myrs_hba
*cs
,
483 enable_mbox_t enable_mbox_fn
)
485 void __iomem
*base
= cs
->io_base
;
486 struct pci_dev
*pdev
= cs
->pdev
;
487 union myrs_cmd_mbox
*cmd_mbox
;
488 struct myrs_stat_mbox
*stat_mbox
;
489 union myrs_cmd_mbox
*mbox
;
490 dma_addr_t mbox_addr
;
491 unsigned char status
= MYRS_STATUS_FAILED
;
493 if (dma_set_mask(&pdev
->dev
, DMA_BIT_MASK(64)))
494 if (dma_set_mask(&pdev
->dev
, DMA_BIT_MASK(32))) {
495 dev_err(&pdev
->dev
, "DMA mask out of range\n");
499 /* Temporary dma mapping, used only in the scope of this function */
500 mbox
= dma_alloc_coherent(&pdev
->dev
, sizeof(union myrs_cmd_mbox
),
501 &mbox_addr
, GFP_KERNEL
);
502 if (dma_mapping_error(&pdev
->dev
, mbox_addr
))
505 /* These are the base addresses for the command memory mailbox array */
506 cs
->cmd_mbox_size
= MYRS_MAX_CMD_MBOX
* sizeof(union myrs_cmd_mbox
);
507 cmd_mbox
= dma_alloc_coherent(&pdev
->dev
, cs
->cmd_mbox_size
,
508 &cs
->cmd_mbox_addr
, GFP_KERNEL
);
509 if (dma_mapping_error(&pdev
->dev
, cs
->cmd_mbox_addr
)) {
510 dev_err(&pdev
->dev
, "Failed to map command mailbox\n");
513 cs
->first_cmd_mbox
= cmd_mbox
;
514 cmd_mbox
+= MYRS_MAX_CMD_MBOX
- 1;
515 cs
->last_cmd_mbox
= cmd_mbox
;
516 cs
->next_cmd_mbox
= cs
->first_cmd_mbox
;
517 cs
->prev_cmd_mbox1
= cs
->last_cmd_mbox
;
518 cs
->prev_cmd_mbox2
= cs
->last_cmd_mbox
- 1;
520 /* These are the base addresses for the status memory mailbox array */
521 cs
->stat_mbox_size
= MYRS_MAX_STAT_MBOX
* sizeof(struct myrs_stat_mbox
);
522 stat_mbox
= dma_alloc_coherent(&pdev
->dev
, cs
->stat_mbox_size
,
523 &cs
->stat_mbox_addr
, GFP_KERNEL
);
524 if (dma_mapping_error(&pdev
->dev
, cs
->stat_mbox_addr
)) {
525 dev_err(&pdev
->dev
, "Failed to map status mailbox\n");
529 cs
->first_stat_mbox
= stat_mbox
;
530 stat_mbox
+= MYRS_MAX_STAT_MBOX
- 1;
531 cs
->last_stat_mbox
= stat_mbox
;
532 cs
->next_stat_mbox
= cs
->first_stat_mbox
;
534 cs
->fwstat_buf
= dma_alloc_coherent(&pdev
->dev
,
535 sizeof(struct myrs_fwstat
),
536 &cs
->fwstat_addr
, GFP_KERNEL
);
537 if (dma_mapping_error(&pdev
->dev
, cs
->fwstat_addr
)) {
538 dev_err(&pdev
->dev
, "Failed to map firmware health buffer\n");
539 cs
->fwstat_buf
= NULL
;
542 cs
->ctlr_info
= kzalloc(sizeof(struct myrs_ctlr_info
),
543 GFP_KERNEL
| GFP_DMA
);
547 cs
->event_buf
= kzalloc(sizeof(struct myrs_event
),
548 GFP_KERNEL
| GFP_DMA
);
552 /* Enable the Memory Mailbox Interface. */
553 memset(mbox
, 0, sizeof(union myrs_cmd_mbox
));
554 mbox
->set_mbox
.id
= 1;
555 mbox
->set_mbox
.opcode
= MYRS_CMD_OP_IOCTL
;
556 mbox
->set_mbox
.control
.no_autosense
= true;
557 mbox
->set_mbox
.first_cmd_mbox_size_kb
=
558 (MYRS_MAX_CMD_MBOX
* sizeof(union myrs_cmd_mbox
)) >> 10;
559 mbox
->set_mbox
.first_stat_mbox_size_kb
=
560 (MYRS_MAX_STAT_MBOX
* sizeof(struct myrs_stat_mbox
)) >> 10;
561 mbox
->set_mbox
.second_cmd_mbox_size_kb
= 0;
562 mbox
->set_mbox
.second_stat_mbox_size_kb
= 0;
563 mbox
->set_mbox
.sense_len
= 0;
564 mbox
->set_mbox
.ioctl_opcode
= MYRS_IOCTL_SET_MEM_MBOX
;
565 mbox
->set_mbox
.fwstat_buf_size_kb
= 1;
566 mbox
->set_mbox
.fwstat_buf_addr
= cs
->fwstat_addr
;
567 mbox
->set_mbox
.first_cmd_mbox_addr
= cs
->cmd_mbox_addr
;
568 mbox
->set_mbox
.first_stat_mbox_addr
= cs
->stat_mbox_addr
;
569 status
= enable_mbox_fn(base
, mbox_addr
);
572 dma_free_coherent(&pdev
->dev
, sizeof(union myrs_cmd_mbox
),
574 if (status
!= MYRS_STATUS_SUCCESS
)
575 dev_err(&pdev
->dev
, "Failed to enable mailbox, status %X\n",
577 return (status
== MYRS_STATUS_SUCCESS
);
581 * myrs_get_config - reads the Configuration Information
583 static int myrs_get_config(struct myrs_hba
*cs
)
585 struct myrs_ctlr_info
*info
= cs
->ctlr_info
;
586 struct Scsi_Host
*shost
= cs
->host
;
587 unsigned char status
;
588 unsigned char model
[20];
589 unsigned char fw_version
[12];
592 /* Get data into dma-able area, then copy into permanent location */
593 mutex_lock(&cs
->cinfo_mutex
);
594 status
= myrs_get_ctlr_info(cs
);
595 mutex_unlock(&cs
->cinfo_mutex
);
596 if (status
!= MYRS_STATUS_SUCCESS
) {
597 shost_printk(KERN_ERR
, shost
,
598 "Failed to get controller information\n");
602 /* Initialize the Controller Model Name and Full Model Name fields. */
603 model_len
= sizeof(info
->ctlr_name
);
604 if (model_len
> sizeof(model
)-1)
605 model_len
= sizeof(model
)-1;
606 memcpy(model
, info
->ctlr_name
, model_len
);
608 while (model
[model_len
] == ' ' || model
[model_len
] == '\0')
610 model
[++model_len
] = '\0';
611 strcpy(cs
->model_name
, "DAC960 ");
612 strcat(cs
->model_name
, model
);
613 /* Initialize the Controller Firmware Version field. */
614 sprintf(fw_version
, "%d.%02d-%02d",
615 info
->fw_major_version
, info
->fw_minor_version
,
616 info
->fw_turn_number
);
617 if (info
->fw_major_version
== 6 &&
618 info
->fw_minor_version
== 0 &&
619 info
->fw_turn_number
< 1) {
620 shost_printk(KERN_WARNING
, shost
,
621 "FIRMWARE VERSION %s DOES NOT PROVIDE THE CONTROLLER\n"
622 "STATUS MONITORING FUNCTIONALITY NEEDED BY THIS DRIVER.\n"
623 "PLEASE UPGRADE TO VERSION 6.00-01 OR ABOVE.\n",
627 /* Initialize the Controller Channels and Targets. */
628 shost
->max_channel
= info
->physchan_present
+ info
->virtchan_present
;
629 shost
->max_id
= info
->max_targets
[0];
630 for (i
= 1; i
< 16; i
++) {
631 if (!info
->max_targets
[i
])
633 if (shost
->max_id
< info
->max_targets
[i
])
634 shost
->max_id
= info
->max_targets
[i
];
638 * Initialize the Controller Queue Depth, Driver Queue Depth,
639 * Logical Drive Count, Maximum Blocks per Command, Controller
640 * Scatter/Gather Limit, and Driver Scatter/Gather Limit.
641 * The Driver Queue Depth must be at most three less than
642 * the Controller Queue Depth; tag '1' is reserved for
643 * direct commands, and tag '2' for monitoring commands.
645 shost
->can_queue
= info
->max_tcq
- 3;
646 if (shost
->can_queue
> MYRS_MAX_CMD_MBOX
- 3)
647 shost
->can_queue
= MYRS_MAX_CMD_MBOX
- 3;
648 shost
->max_sectors
= info
->max_transfer_size
;
649 shost
->sg_tablesize
= info
->max_sge
;
650 if (shost
->sg_tablesize
> MYRS_SG_LIMIT
)
651 shost
->sg_tablesize
= MYRS_SG_LIMIT
;
653 shost_printk(KERN_INFO
, shost
,
654 "Configuring %s PCI RAID Controller\n", model
);
655 shost_printk(KERN_INFO
, shost
,
656 " Firmware Version: %s, Channels: %d, Memory Size: %dMB\n",
657 fw_version
, info
->physchan_present
, info
->mem_size_mb
);
659 shost_printk(KERN_INFO
, shost
,
660 " Controller Queue Depth: %d, Maximum Blocks per Command: %d\n",
661 shost
->can_queue
, shost
->max_sectors
);
663 shost_printk(KERN_INFO
, shost
,
664 " Driver Queue Depth: %d, Scatter/Gather Limit: %d of %d Segments\n",
665 shost
->can_queue
, shost
->sg_tablesize
, MYRS_SG_LIMIT
);
666 for (i
= 0; i
< info
->physchan_max
; i
++) {
667 if (!info
->max_targets
[i
])
669 shost_printk(KERN_INFO
, shost
,
670 " Device Channel %d: max %d devices\n",
671 i
, info
->max_targets
[i
]);
673 shost_printk(KERN_INFO
, shost
,
674 " Physical: %d/%d channels, %d disks, %d devices\n",
675 info
->physchan_present
, info
->physchan_max
,
676 info
->pdisk_present
, info
->pdev_present
);
678 shost_printk(KERN_INFO
, shost
,
679 " Logical: %d/%d channels, %d disks\n",
680 info
->virtchan_present
, info
->virtchan_max
,
686 * myrs_log_event - prints a Controller Event message
690 unsigned char *ev_msg
;
692 /* Physical Device Events (0x0000 - 0x007F) */
693 { 0x0001, "P Online" },
694 { 0x0002, "P Standby" },
695 { 0x0005, "P Automatic Rebuild Started" },
696 { 0x0006, "P Manual Rebuild Started" },
697 { 0x0007, "P Rebuild Completed" },
698 { 0x0008, "P Rebuild Cancelled" },
699 { 0x0009, "P Rebuild Failed for Unknown Reasons" },
700 { 0x000A, "P Rebuild Failed due to New Physical Device" },
701 { 0x000B, "P Rebuild Failed due to Logical Drive Failure" },
702 { 0x000C, "S Offline" },
703 { 0x000D, "P Found" },
704 { 0x000E, "P Removed" },
705 { 0x000F, "P Unconfigured" },
706 { 0x0010, "P Expand Capacity Started" },
707 { 0x0011, "P Expand Capacity Completed" },
708 { 0x0012, "P Expand Capacity Failed" },
709 { 0x0013, "P Command Timed Out" },
710 { 0x0014, "P Command Aborted" },
711 { 0x0015, "P Command Retried" },
712 { 0x0016, "P Parity Error" },
713 { 0x0017, "P Soft Error" },
714 { 0x0018, "P Miscellaneous Error" },
715 { 0x0019, "P Reset" },
716 { 0x001A, "P Active Spare Found" },
717 { 0x001B, "P Warm Spare Found" },
718 { 0x001C, "S Sense Data Received" },
719 { 0x001D, "P Initialization Started" },
720 { 0x001E, "P Initialization Completed" },
721 { 0x001F, "P Initialization Failed" },
722 { 0x0020, "P Initialization Cancelled" },
723 { 0x0021, "P Failed because Write Recovery Failed" },
724 { 0x0022, "P Failed because SCSI Bus Reset Failed" },
725 { 0x0023, "P Failed because of Double Check Condition" },
726 { 0x0024, "P Failed because Device Cannot Be Accessed" },
727 { 0x0025, "P Failed because of Gross Error on SCSI Processor" },
728 { 0x0026, "P Failed because of Bad Tag from Device" },
729 { 0x0027, "P Failed because of Command Timeout" },
730 { 0x0028, "P Failed because of System Reset" },
731 { 0x0029, "P Failed because of Busy Status or Parity Error" },
732 { 0x002A, "P Failed because Host Set Device to Failed State" },
733 { 0x002B, "P Failed because of Selection Timeout" },
734 { 0x002C, "P Failed because of SCSI Bus Phase Error" },
735 { 0x002D, "P Failed because Device Returned Unknown Status" },
736 { 0x002E, "P Failed because Device Not Ready" },
737 { 0x002F, "P Failed because Device Not Found at Startup" },
738 { 0x0030, "P Failed because COD Write Operation Failed" },
739 { 0x0031, "P Failed because BDT Write Operation Failed" },
740 { 0x0039, "P Missing at Startup" },
741 { 0x003A, "P Start Rebuild Failed due to Physical Drive Too Small" },
742 { 0x003C, "P Temporarily Offline Device Automatically Made Online" },
743 { 0x003D, "P Standby Rebuild Started" },
744 /* Logical Device Events (0x0080 - 0x00FF) */
745 { 0x0080, "M Consistency Check Started" },
746 { 0x0081, "M Consistency Check Completed" },
747 { 0x0082, "M Consistency Check Cancelled" },
748 { 0x0083, "M Consistency Check Completed With Errors" },
749 { 0x0084, "M Consistency Check Failed due to Logical Drive Failure" },
750 { 0x0085, "M Consistency Check Failed due to Physical Device Failure" },
751 { 0x0086, "L Offline" },
752 { 0x0087, "L Critical" },
753 { 0x0088, "L Online" },
754 { 0x0089, "M Automatic Rebuild Started" },
755 { 0x008A, "M Manual Rebuild Started" },
756 { 0x008B, "M Rebuild Completed" },
757 { 0x008C, "M Rebuild Cancelled" },
758 { 0x008D, "M Rebuild Failed for Unknown Reasons" },
759 { 0x008E, "M Rebuild Failed due to New Physical Device" },
760 { 0x008F, "M Rebuild Failed due to Logical Drive Failure" },
761 { 0x0090, "M Initialization Started" },
762 { 0x0091, "M Initialization Completed" },
763 { 0x0092, "M Initialization Cancelled" },
764 { 0x0093, "M Initialization Failed" },
765 { 0x0094, "L Found" },
766 { 0x0095, "L Deleted" },
767 { 0x0096, "M Expand Capacity Started" },
768 { 0x0097, "M Expand Capacity Completed" },
769 { 0x0098, "M Expand Capacity Failed" },
770 { 0x0099, "L Bad Block Found" },
771 { 0x009A, "L Size Changed" },
772 { 0x009B, "L Type Changed" },
773 { 0x009C, "L Bad Data Block Found" },
774 { 0x009E, "L Read of Data Block in BDT" },
775 { 0x009F, "L Write Back Data for Disk Block Lost" },
776 { 0x00A0, "L Temporarily Offline RAID-5/3 Drive Made Online" },
777 { 0x00A1, "L Temporarily Offline RAID-6/1/0/7 Drive Made Online" },
778 { 0x00A2, "L Standby Rebuild Started" },
779 /* Fault Management Events (0x0100 - 0x017F) */
780 { 0x0140, "E Fan %d Failed" },
781 { 0x0141, "E Fan %d OK" },
782 { 0x0142, "E Fan %d Not Present" },
783 { 0x0143, "E Power Supply %d Failed" },
784 { 0x0144, "E Power Supply %d OK" },
785 { 0x0145, "E Power Supply %d Not Present" },
786 { 0x0146, "E Temperature Sensor %d Temperature Exceeds Safe Limit" },
787 { 0x0147, "E Temperature Sensor %d Temperature Exceeds Working Limit" },
788 { 0x0148, "E Temperature Sensor %d Temperature Normal" },
789 { 0x0149, "E Temperature Sensor %d Not Present" },
790 { 0x014A, "E Enclosure Management Unit %d Access Critical" },
791 { 0x014B, "E Enclosure Management Unit %d Access OK" },
792 { 0x014C, "E Enclosure Management Unit %d Access Offline" },
793 /* Controller Events (0x0180 - 0x01FF) */
794 { 0x0181, "C Cache Write Back Error" },
795 { 0x0188, "C Battery Backup Unit Found" },
796 { 0x0189, "C Battery Backup Unit Charge Level Low" },
797 { 0x018A, "C Battery Backup Unit Charge Level OK" },
798 { 0x0193, "C Installation Aborted" },
799 { 0x0195, "C Battery Backup Unit Physically Removed" },
800 { 0x0196, "C Memory Error During Warm Boot" },
801 { 0x019E, "C Memory Soft ECC Error Corrected" },
802 { 0x019F, "C Memory Hard ECC Error Corrected" },
803 { 0x01A2, "C Battery Backup Unit Failed" },
804 { 0x01AB, "C Mirror Race Recovery Failed" },
805 { 0x01AC, "C Mirror Race on Critical Drive" },
806 /* Controller Internal Processor Events */
807 { 0x0380, "C Internal Controller Hung" },
808 { 0x0381, "C Internal Controller Firmware Breakpoint" },
809 { 0x0390, "C Internal Controller i960 Processor Specific Error" },
810 { 0x03A0, "C Internal Controller StrongARM Processor Specific Error" },
814 static void myrs_log_event(struct myrs_hba
*cs
, struct myrs_event
*ev
)
816 unsigned char msg_buf
[MYRS_LINE_BUFFER_SIZE
];
817 int ev_idx
= 0, ev_code
;
818 unsigned char ev_type
, *ev_msg
;
819 struct Scsi_Host
*shost
= cs
->host
;
820 struct scsi_device
*sdev
;
821 struct scsi_sense_hdr sshdr
= {0};
822 unsigned char sense_info
[4];
823 unsigned char cmd_specific
[4];
825 if (ev
->ev_code
== 0x1C) {
826 if (!scsi_normalize_sense(ev
->sense_data
, 40, &sshdr
)) {
827 memset(&sshdr
, 0x0, sizeof(sshdr
));
828 memset(sense_info
, 0x0, sizeof(sense_info
));
829 memset(cmd_specific
, 0x0, sizeof(cmd_specific
));
831 memcpy(sense_info
, &ev
->sense_data
[3], 4);
832 memcpy(cmd_specific
, &ev
->sense_data
[7], 4);
835 if (sshdr
.sense_key
== VENDOR_SPECIFIC
&&
836 (sshdr
.asc
== 0x80 || sshdr
.asc
== 0x81))
837 ev
->ev_code
= ((sshdr
.asc
- 0x80) << 8 | sshdr
.ascq
);
839 ev_code
= myrs_ev_list
[ev_idx
].ev_code
;
840 if (ev_code
== ev
->ev_code
|| ev_code
== 0)
844 ev_type
= myrs_ev_list
[ev_idx
].ev_msg
[0];
845 ev_msg
= &myrs_ev_list
[ev_idx
].ev_msg
[2];
847 shost_printk(KERN_WARNING
, shost
,
848 "Unknown Controller Event Code %04X\n",
854 sdev
= scsi_device_lookup(shost
, ev
->channel
,
856 sdev_printk(KERN_INFO
, sdev
, "event %d: Physical Device %s\n",
858 if (sdev
&& sdev
->hostdata
&&
859 sdev
->channel
< cs
->ctlr_info
->physchan_present
) {
860 struct myrs_pdev_info
*pdev_info
= sdev
->hostdata
;
862 switch (ev
->ev_code
) {
865 pdev_info
->dev_state
= MYRS_DEVICE_ONLINE
;
868 pdev_info
->dev_state
= MYRS_DEVICE_STANDBY
;
871 pdev_info
->dev_state
= MYRS_DEVICE_OFFLINE
;
874 pdev_info
->dev_state
= MYRS_DEVICE_MISSING
;
877 pdev_info
->dev_state
= MYRS_DEVICE_UNCONFIGURED
;
883 shost_printk(KERN_INFO
, shost
,
884 "event %d: Logical Drive %d %s\n",
885 ev
->ev_seq
, ev
->lun
, ev_msg
);
886 cs
->needs_update
= true;
889 shost_printk(KERN_INFO
, shost
,
890 "event %d: Logical Drive %d %s\n",
891 ev
->ev_seq
, ev
->lun
, ev_msg
);
892 cs
->needs_update
= true;
895 if (sshdr
.sense_key
== NO_SENSE
||
896 (sshdr
.sense_key
== NOT_READY
&&
897 sshdr
.asc
== 0x04 && (sshdr
.ascq
== 0x01 ||
898 sshdr
.ascq
== 0x02)))
900 shost_printk(KERN_INFO
, shost
,
901 "event %d: Physical Device %d:%d %s\n",
902 ev
->ev_seq
, ev
->channel
, ev
->target
, ev_msg
);
903 shost_printk(KERN_INFO
, shost
,
904 "Physical Device %d:%d Sense Key = %X, ASC = %02X, ASCQ = %02X\n",
905 ev
->channel
, ev
->target
,
906 sshdr
.sense_key
, sshdr
.asc
, sshdr
.ascq
);
907 shost_printk(KERN_INFO
, shost
,
908 "Physical Device %d:%d Sense Information = %02X%02X%02X%02X %02X%02X%02X%02X\n",
909 ev
->channel
, ev
->target
,
910 sense_info
[0], sense_info
[1],
911 sense_info
[2], sense_info
[3],
912 cmd_specific
[0], cmd_specific
[1],
913 cmd_specific
[2], cmd_specific
[3]);
916 if (cs
->disable_enc_msg
)
918 sprintf(msg_buf
, ev_msg
, ev
->lun
);
919 shost_printk(KERN_INFO
, shost
, "event %d: Enclosure %d %s\n",
920 ev
->ev_seq
, ev
->target
, msg_buf
);
923 shost_printk(KERN_INFO
, shost
, "event %d: Controller %s\n",
927 shost_printk(KERN_INFO
, shost
,
928 "event %d: Unknown Event Code %04X\n",
929 ev
->ev_seq
, ev
->ev_code
);
935 * SCSI sysfs interface functions
937 static ssize_t
raid_state_show(struct device
*dev
,
938 struct device_attribute
*attr
, char *buf
)
940 struct scsi_device
*sdev
= to_scsi_device(dev
);
941 struct myrs_hba
*cs
= shost_priv(sdev
->host
);
945 return snprintf(buf
, 16, "Unknown\n");
947 if (sdev
->channel
>= cs
->ctlr_info
->physchan_present
) {
948 struct myrs_ldev_info
*ldev_info
= sdev
->hostdata
;
951 name
= myrs_devstate_name(ldev_info
->dev_state
);
953 ret
= snprintf(buf
, 32, "%s\n", name
);
955 ret
= snprintf(buf
, 32, "Invalid (%02X)\n",
956 ldev_info
->dev_state
);
958 struct myrs_pdev_info
*pdev_info
;
961 pdev_info
= sdev
->hostdata
;
962 name
= myrs_devstate_name(pdev_info
->dev_state
);
964 ret
= snprintf(buf
, 32, "%s\n", name
);
966 ret
= snprintf(buf
, 32, "Invalid (%02X)\n",
967 pdev_info
->dev_state
);
972 static ssize_t
raid_state_store(struct device
*dev
,
973 struct device_attribute
*attr
, const char *buf
, size_t count
)
975 struct scsi_device
*sdev
= to_scsi_device(dev
);
976 struct myrs_hba
*cs
= shost_priv(sdev
->host
);
977 struct myrs_cmdblk
*cmd_blk
;
978 union myrs_cmd_mbox
*mbox
;
979 enum myrs_devstate new_state
;
980 unsigned short ldev_num
;
981 unsigned char status
;
983 if (!strncmp(buf
, "offline", 7) ||
984 !strncmp(buf
, "kill", 4))
985 new_state
= MYRS_DEVICE_OFFLINE
;
986 else if (!strncmp(buf
, "online", 6))
987 new_state
= MYRS_DEVICE_ONLINE
;
988 else if (!strncmp(buf
, "standby", 7))
989 new_state
= MYRS_DEVICE_STANDBY
;
993 if (sdev
->channel
< cs
->ctlr_info
->physchan_present
) {
994 struct myrs_pdev_info
*pdev_info
= sdev
->hostdata
;
995 struct myrs_devmap
*pdev_devmap
=
996 (struct myrs_devmap
*)&pdev_info
->rsvd13
;
998 if (pdev_info
->dev_state
== new_state
) {
999 sdev_printk(KERN_INFO
, sdev
,
1000 "Device already in %s\n",
1001 myrs_devstate_name(new_state
));
1004 status
= myrs_translate_pdev(cs
, sdev
->channel
, sdev
->id
,
1005 sdev
->lun
, pdev_devmap
);
1006 if (status
!= MYRS_STATUS_SUCCESS
)
1008 ldev_num
= pdev_devmap
->ldev_num
;
1010 struct myrs_ldev_info
*ldev_info
= sdev
->hostdata
;
1012 if (ldev_info
->dev_state
== new_state
) {
1013 sdev_printk(KERN_INFO
, sdev
,
1014 "Device already in %s\n",
1015 myrs_devstate_name(new_state
));
1018 ldev_num
= ldev_info
->ldev_num
;
1020 mutex_lock(&cs
->dcmd_mutex
);
1021 cmd_blk
= &cs
->dcmd_blk
;
1022 myrs_reset_cmd(cmd_blk
);
1023 mbox
= &cmd_blk
->mbox
;
1024 mbox
->common
.opcode
= MYRS_CMD_OP_IOCTL
;
1025 mbox
->common
.id
= MYRS_DCMD_TAG
;
1026 mbox
->common
.control
.dma_ctrl_to_host
= true;
1027 mbox
->common
.control
.no_autosense
= true;
1028 mbox
->set_devstate
.ioctl_opcode
= MYRS_IOCTL_SET_DEVICE_STATE
;
1029 mbox
->set_devstate
.state
= new_state
;
1030 mbox
->set_devstate
.ldev
.ldev_num
= ldev_num
;
1031 myrs_exec_cmd(cs
, cmd_blk
);
1032 status
= cmd_blk
->status
;
1033 mutex_unlock(&cs
->dcmd_mutex
);
1034 if (status
== MYRS_STATUS_SUCCESS
) {
1035 if (sdev
->channel
< cs
->ctlr_info
->physchan_present
) {
1036 struct myrs_pdev_info
*pdev_info
= sdev
->hostdata
;
1038 pdev_info
->dev_state
= new_state
;
1040 struct myrs_ldev_info
*ldev_info
= sdev
->hostdata
;
1042 ldev_info
->dev_state
= new_state
;
1044 sdev_printk(KERN_INFO
, sdev
,
1045 "Set device state to %s\n",
1046 myrs_devstate_name(new_state
));
1049 sdev_printk(KERN_INFO
, sdev
,
1050 "Failed to set device state to %s, status 0x%02x\n",
1051 myrs_devstate_name(new_state
), status
);
1054 static DEVICE_ATTR_RW(raid_state
);
1056 static ssize_t
raid_level_show(struct device
*dev
,
1057 struct device_attribute
*attr
, char *buf
)
1059 struct scsi_device
*sdev
= to_scsi_device(dev
);
1060 struct myrs_hba
*cs
= shost_priv(sdev
->host
);
1061 const char *name
= NULL
;
1063 if (!sdev
->hostdata
)
1064 return snprintf(buf
, 16, "Unknown\n");
1066 if (sdev
->channel
>= cs
->ctlr_info
->physchan_present
) {
1067 struct myrs_ldev_info
*ldev_info
;
1069 ldev_info
= sdev
->hostdata
;
1070 name
= myrs_raid_level_name(ldev_info
->raid_level
);
1072 return snprintf(buf
, 32, "Invalid (%02X)\n",
1073 ldev_info
->dev_state
);
1076 name
= myrs_raid_level_name(MYRS_RAID_PHYSICAL
);
1078 return snprintf(buf
, 32, "%s\n", name
);
1080 static DEVICE_ATTR_RO(raid_level
);
1082 static ssize_t
rebuild_show(struct device
*dev
,
1083 struct device_attribute
*attr
, char *buf
)
1085 struct scsi_device
*sdev
= to_scsi_device(dev
);
1086 struct myrs_hba
*cs
= shost_priv(sdev
->host
);
1087 struct myrs_ldev_info
*ldev_info
;
1088 unsigned short ldev_num
;
1089 unsigned char status
;
1091 if (sdev
->channel
< cs
->ctlr_info
->physchan_present
)
1092 return snprintf(buf
, 32, "physical device - not rebuilding\n");
1094 ldev_info
= sdev
->hostdata
;
1095 ldev_num
= ldev_info
->ldev_num
;
1096 status
= myrs_get_ldev_info(cs
, ldev_num
, ldev_info
);
1097 if (status
!= MYRS_STATUS_SUCCESS
) {
1098 sdev_printk(KERN_INFO
, sdev
,
1099 "Failed to get device information, status 0x%02x\n",
1103 if (ldev_info
->rbld_active
) {
1104 return snprintf(buf
, 32, "rebuilding block %zu of %zu\n",
1105 (size_t)ldev_info
->rbld_lba
,
1106 (size_t)ldev_info
->cfg_devsize
);
1108 return snprintf(buf
, 32, "not rebuilding\n");
1111 static ssize_t
rebuild_store(struct device
*dev
,
1112 struct device_attribute
*attr
, const char *buf
, size_t count
)
1114 struct scsi_device
*sdev
= to_scsi_device(dev
);
1115 struct myrs_hba
*cs
= shost_priv(sdev
->host
);
1116 struct myrs_ldev_info
*ldev_info
;
1117 struct myrs_cmdblk
*cmd_blk
;
1118 union myrs_cmd_mbox
*mbox
;
1119 unsigned short ldev_num
;
1120 unsigned char status
;
1123 if (sdev
->channel
< cs
->ctlr_info
->physchan_present
)
1126 ldev_info
= sdev
->hostdata
;
1129 ldev_num
= ldev_info
->ldev_num
;
1131 ret
= kstrtoint(buf
, 0, &rebuild
);
1135 status
= myrs_get_ldev_info(cs
, ldev_num
, ldev_info
);
1136 if (status
!= MYRS_STATUS_SUCCESS
) {
1137 sdev_printk(KERN_INFO
, sdev
,
1138 "Failed to get device information, status 0x%02x\n",
1143 if (rebuild
&& ldev_info
->rbld_active
) {
1144 sdev_printk(KERN_INFO
, sdev
,
1145 "Rebuild Not Initiated; already in progress\n");
1148 if (!rebuild
&& !ldev_info
->rbld_active
) {
1149 sdev_printk(KERN_INFO
, sdev
,
1150 "Rebuild Not Cancelled; no rebuild in progress\n");
1154 mutex_lock(&cs
->dcmd_mutex
);
1155 cmd_blk
= &cs
->dcmd_blk
;
1156 myrs_reset_cmd(cmd_blk
);
1157 mbox
= &cmd_blk
->mbox
;
1158 mbox
->common
.opcode
= MYRS_CMD_OP_IOCTL
;
1159 mbox
->common
.id
= MYRS_DCMD_TAG
;
1160 mbox
->common
.control
.dma_ctrl_to_host
= true;
1161 mbox
->common
.control
.no_autosense
= true;
1163 mbox
->ldev_info
.ldev
.ldev_num
= ldev_num
;
1164 mbox
->ldev_info
.ioctl_opcode
= MYRS_IOCTL_RBLD_DEVICE_START
;
1166 mbox
->ldev_info
.ldev
.ldev_num
= ldev_num
;
1167 mbox
->ldev_info
.ioctl_opcode
= MYRS_IOCTL_RBLD_DEVICE_STOP
;
1169 myrs_exec_cmd(cs
, cmd_blk
);
1170 status
= cmd_blk
->status
;
1171 mutex_unlock(&cs
->dcmd_mutex
);
1173 sdev_printk(KERN_INFO
, sdev
,
1174 "Rebuild Not %s, status 0x%02x\n",
1175 rebuild
? "Initiated" : "Cancelled", status
);
1178 sdev_printk(KERN_INFO
, sdev
, "Rebuild %s\n",
1179 rebuild
? "Initiated" : "Cancelled");
1185 static DEVICE_ATTR_RW(rebuild
);
1187 static ssize_t
consistency_check_show(struct device
*dev
,
1188 struct device_attribute
*attr
, char *buf
)
1190 struct scsi_device
*sdev
= to_scsi_device(dev
);
1191 struct myrs_hba
*cs
= shost_priv(sdev
->host
);
1192 struct myrs_ldev_info
*ldev_info
;
1193 unsigned short ldev_num
;
1194 unsigned char status
;
1196 if (sdev
->channel
< cs
->ctlr_info
->physchan_present
)
1197 return snprintf(buf
, 32, "physical device - not checking\n");
1199 ldev_info
= sdev
->hostdata
;
1202 ldev_num
= ldev_info
->ldev_num
;
1203 status
= myrs_get_ldev_info(cs
, ldev_num
, ldev_info
);
1204 if (ldev_info
->cc_active
)
1205 return snprintf(buf
, 32, "checking block %zu of %zu\n",
1206 (size_t)ldev_info
->cc_lba
,
1207 (size_t)ldev_info
->cfg_devsize
);
1209 return snprintf(buf
, 32, "not checking\n");
1212 static ssize_t
consistency_check_store(struct device
*dev
,
1213 struct device_attribute
*attr
, const char *buf
, size_t count
)
1215 struct scsi_device
*sdev
= to_scsi_device(dev
);
1216 struct myrs_hba
*cs
= shost_priv(sdev
->host
);
1217 struct myrs_ldev_info
*ldev_info
;
1218 struct myrs_cmdblk
*cmd_blk
;
1219 union myrs_cmd_mbox
*mbox
;
1220 unsigned short ldev_num
;
1221 unsigned char status
;
1224 if (sdev
->channel
< cs
->ctlr_info
->physchan_present
)
1227 ldev_info
= sdev
->hostdata
;
1230 ldev_num
= ldev_info
->ldev_num
;
1232 ret
= kstrtoint(buf
, 0, &check
);
1236 status
= myrs_get_ldev_info(cs
, ldev_num
, ldev_info
);
1237 if (status
!= MYRS_STATUS_SUCCESS
) {
1238 sdev_printk(KERN_INFO
, sdev
,
1239 "Failed to get device information, status 0x%02x\n",
1243 if (check
&& ldev_info
->cc_active
) {
1244 sdev_printk(KERN_INFO
, sdev
,
1245 "Consistency Check Not Initiated; "
1246 "already in progress\n");
1249 if (!check
&& !ldev_info
->cc_active
) {
1250 sdev_printk(KERN_INFO
, sdev
,
1251 "Consistency Check Not Cancelled; "
1252 "check not in progress\n");
1256 mutex_lock(&cs
->dcmd_mutex
);
1257 cmd_blk
= &cs
->dcmd_blk
;
1258 myrs_reset_cmd(cmd_blk
);
1259 mbox
= &cmd_blk
->mbox
;
1260 mbox
->common
.opcode
= MYRS_CMD_OP_IOCTL
;
1261 mbox
->common
.id
= MYRS_DCMD_TAG
;
1262 mbox
->common
.control
.dma_ctrl_to_host
= true;
1263 mbox
->common
.control
.no_autosense
= true;
1265 mbox
->cc
.ldev
.ldev_num
= ldev_num
;
1266 mbox
->cc
.ioctl_opcode
= MYRS_IOCTL_CC_START
;
1267 mbox
->cc
.restore_consistency
= true;
1268 mbox
->cc
.initialized_area_only
= false;
1270 mbox
->cc
.ldev
.ldev_num
= ldev_num
;
1271 mbox
->cc
.ioctl_opcode
= MYRS_IOCTL_CC_STOP
;
1273 myrs_exec_cmd(cs
, cmd_blk
);
1274 status
= cmd_blk
->status
;
1275 mutex_unlock(&cs
->dcmd_mutex
);
1276 if (status
!= MYRS_STATUS_SUCCESS
) {
1277 sdev_printk(KERN_INFO
, sdev
,
1278 "Consistency Check Not %s, status 0x%02x\n",
1279 check
? "Initiated" : "Cancelled", status
);
1282 sdev_printk(KERN_INFO
, sdev
, "Consistency Check %s\n",
1283 check
? "Initiated" : "Cancelled");
1289 static DEVICE_ATTR_RW(consistency_check
);
1291 static struct device_attribute
*myrs_sdev_attrs
[] = {
1292 &dev_attr_consistency_check
,
1294 &dev_attr_raid_state
,
1295 &dev_attr_raid_level
,
1299 static ssize_t
serial_show(struct device
*dev
,
1300 struct device_attribute
*attr
, char *buf
)
1302 struct Scsi_Host
*shost
= class_to_shost(dev
);
1303 struct myrs_hba
*cs
= shost_priv(shost
);
1306 memcpy(serial
, cs
->ctlr_info
->serial_number
, 16);
1308 return snprintf(buf
, 16, "%s\n", serial
);
1310 static DEVICE_ATTR_RO(serial
);
1312 static ssize_t
ctlr_num_show(struct device
*dev
,
1313 struct device_attribute
*attr
, char *buf
)
1315 struct Scsi_Host
*shost
= class_to_shost(dev
);
1316 struct myrs_hba
*cs
= shost_priv(shost
);
1318 return snprintf(buf
, 20, "%d\n", cs
->host
->host_no
);
1320 static DEVICE_ATTR_RO(ctlr_num
);
1322 static struct myrs_cpu_type_tbl
{
1323 enum myrs_cpu_type type
;
1325 } myrs_cpu_type_names
[] = {
1326 { MYRS_CPUTYPE_i960CA
, "i960CA" },
1327 { MYRS_CPUTYPE_i960RD
, "i960RD" },
1328 { MYRS_CPUTYPE_i960RN
, "i960RN" },
1329 { MYRS_CPUTYPE_i960RP
, "i960RP" },
1330 { MYRS_CPUTYPE_NorthBay
, "NorthBay" },
1331 { MYRS_CPUTYPE_StrongArm
, "StrongARM" },
1332 { MYRS_CPUTYPE_i960RM
, "i960RM" },
1335 static ssize_t
processor_show(struct device
*dev
,
1336 struct device_attribute
*attr
, char *buf
)
1338 struct Scsi_Host
*shost
= class_to_shost(dev
);
1339 struct myrs_hba
*cs
= shost_priv(shost
);
1340 struct myrs_cpu_type_tbl
*tbl
;
1341 const char *first_processor
= NULL
;
1342 const char *second_processor
= NULL
;
1343 struct myrs_ctlr_info
*info
= cs
->ctlr_info
;
1347 if (info
->cpu
[0].cpu_count
) {
1348 tbl
= myrs_cpu_type_names
;
1349 for (i
= 0; i
< ARRAY_SIZE(myrs_cpu_type_names
); i
++) {
1350 if (tbl
[i
].type
== info
->cpu
[0].cpu_type
) {
1351 first_processor
= tbl
[i
].name
;
1356 if (info
->cpu
[1].cpu_count
) {
1357 tbl
= myrs_cpu_type_names
;
1358 for (i
= 0; i
< ARRAY_SIZE(myrs_cpu_type_names
); i
++) {
1359 if (tbl
[i
].type
== info
->cpu
[1].cpu_type
) {
1360 second_processor
= tbl
[i
].name
;
1365 if (first_processor
&& second_processor
)
1366 ret
= snprintf(buf
, 64, "1: %s (%s, %d cpus)\n"
1367 "2: %s (%s, %d cpus)\n",
1368 info
->cpu
[0].cpu_name
,
1369 first_processor
, info
->cpu
[0].cpu_count
,
1370 info
->cpu
[1].cpu_name
,
1371 second_processor
, info
->cpu
[1].cpu_count
);
1372 else if (first_processor
&& !second_processor
)
1373 ret
= snprintf(buf
, 64, "1: %s (%s, %d cpus)\n2: absent\n",
1374 info
->cpu
[0].cpu_name
,
1375 first_processor
, info
->cpu
[0].cpu_count
);
1376 else if (!first_processor
&& second_processor
)
1377 ret
= snprintf(buf
, 64, "1: absent\n2: %s (%s, %d cpus)\n",
1378 info
->cpu
[1].cpu_name
,
1379 second_processor
, info
->cpu
[1].cpu_count
);
1381 ret
= snprintf(buf
, 64, "1: absent\n2: absent\n");
1385 static DEVICE_ATTR_RO(processor
);
1387 static ssize_t
model_show(struct device
*dev
,
1388 struct device_attribute
*attr
, char *buf
)
1390 struct Scsi_Host
*shost
= class_to_shost(dev
);
1391 struct myrs_hba
*cs
= shost_priv(shost
);
1393 return snprintf(buf
, 28, "%s\n", cs
->model_name
);
1395 static DEVICE_ATTR_RO(model
);
1397 static ssize_t
ctlr_type_show(struct device
*dev
,
1398 struct device_attribute
*attr
, char *buf
)
1400 struct Scsi_Host
*shost
= class_to_shost(dev
);
1401 struct myrs_hba
*cs
= shost_priv(shost
);
1403 return snprintf(buf
, 4, "%d\n", cs
->ctlr_info
->ctlr_type
);
1405 static DEVICE_ATTR_RO(ctlr_type
);
1407 static ssize_t
cache_size_show(struct device
*dev
,
1408 struct device_attribute
*attr
, char *buf
)
1410 struct Scsi_Host
*shost
= class_to_shost(dev
);
1411 struct myrs_hba
*cs
= shost_priv(shost
);
1413 return snprintf(buf
, 8, "%d MB\n", cs
->ctlr_info
->cache_size_mb
);
1415 static DEVICE_ATTR_RO(cache_size
);
1417 static ssize_t
firmware_show(struct device
*dev
,
1418 struct device_attribute
*attr
, char *buf
)
1420 struct Scsi_Host
*shost
= class_to_shost(dev
);
1421 struct myrs_hba
*cs
= shost_priv(shost
);
1423 return snprintf(buf
, 16, "%d.%02d-%02d\n",
1424 cs
->ctlr_info
->fw_major_version
,
1425 cs
->ctlr_info
->fw_minor_version
,
1426 cs
->ctlr_info
->fw_turn_number
);
1428 static DEVICE_ATTR_RO(firmware
);
1430 static ssize_t
discovery_store(struct device
*dev
,
1431 struct device_attribute
*attr
, const char *buf
, size_t count
)
1433 struct Scsi_Host
*shost
= class_to_shost(dev
);
1434 struct myrs_hba
*cs
= shost_priv(shost
);
1435 struct myrs_cmdblk
*cmd_blk
;
1436 union myrs_cmd_mbox
*mbox
;
1437 unsigned char status
;
1439 mutex_lock(&cs
->dcmd_mutex
);
1440 cmd_blk
= &cs
->dcmd_blk
;
1441 myrs_reset_cmd(cmd_blk
);
1442 mbox
= &cmd_blk
->mbox
;
1443 mbox
->common
.opcode
= MYRS_CMD_OP_IOCTL
;
1444 mbox
->common
.id
= MYRS_DCMD_TAG
;
1445 mbox
->common
.control
.dma_ctrl_to_host
= true;
1446 mbox
->common
.control
.no_autosense
= true;
1447 mbox
->common
.ioctl_opcode
= MYRS_IOCTL_START_DISCOVERY
;
1448 myrs_exec_cmd(cs
, cmd_blk
);
1449 status
= cmd_blk
->status
;
1450 mutex_unlock(&cs
->dcmd_mutex
);
1451 if (status
!= MYRS_STATUS_SUCCESS
) {
1452 shost_printk(KERN_INFO
, shost
,
1453 "Discovery Not Initiated, status %02X\n",
1457 shost_printk(KERN_INFO
, shost
, "Discovery Initiated\n");
1459 cs
->needs_update
= true;
1460 queue_delayed_work(cs
->work_q
, &cs
->monitor_work
, 1);
1461 flush_delayed_work(&cs
->monitor_work
);
1462 shost_printk(KERN_INFO
, shost
, "Discovery Completed\n");
1466 static DEVICE_ATTR_WO(discovery
);
1468 static ssize_t
flush_cache_store(struct device
*dev
,
1469 struct device_attribute
*attr
, const char *buf
, size_t count
)
1471 struct Scsi_Host
*shost
= class_to_shost(dev
);
1472 struct myrs_hba
*cs
= shost_priv(shost
);
1473 unsigned char status
;
1475 status
= myrs_dev_op(cs
, MYRS_IOCTL_FLUSH_DEVICE_DATA
,
1476 MYRS_RAID_CONTROLLER
);
1477 if (status
== MYRS_STATUS_SUCCESS
) {
1478 shost_printk(KERN_INFO
, shost
, "Cache Flush Completed\n");
1481 shost_printk(KERN_INFO
, shost
,
1482 "Cache Flush failed, status 0x%02x\n", status
);
1485 static DEVICE_ATTR_WO(flush_cache
);
1487 static ssize_t
disable_enclosure_messages_show(struct device
*dev
,
1488 struct device_attribute
*attr
, char *buf
)
1490 struct Scsi_Host
*shost
= class_to_shost(dev
);
1491 struct myrs_hba
*cs
= shost_priv(shost
);
1493 return snprintf(buf
, 3, "%d\n", cs
->disable_enc_msg
);
1496 static ssize_t
disable_enclosure_messages_store(struct device
*dev
,
1497 struct device_attribute
*attr
, const char *buf
, size_t count
)
1499 struct scsi_device
*sdev
= to_scsi_device(dev
);
1500 struct myrs_hba
*cs
= shost_priv(sdev
->host
);
1503 ret
= kstrtoint(buf
, 0, &value
);
1510 cs
->disable_enc_msg
= value
;
1513 static DEVICE_ATTR_RW(disable_enclosure_messages
);
1515 static struct device_attribute
*myrs_shost_attrs
[] = {
1518 &dev_attr_processor
,
1520 &dev_attr_ctlr_type
,
1521 &dev_attr_cache_size
,
1523 &dev_attr_discovery
,
1524 &dev_attr_flush_cache
,
1525 &dev_attr_disable_enclosure_messages
,
1530 * SCSI midlayer interface
1532 int myrs_host_reset(struct scsi_cmnd
*scmd
)
1534 struct Scsi_Host
*shost
= scmd
->device
->host
;
1535 struct myrs_hba
*cs
= shost_priv(shost
);
1537 cs
->reset(cs
->io_base
);
1541 static void myrs_mode_sense(struct myrs_hba
*cs
, struct scsi_cmnd
*scmd
,
1542 struct myrs_ldev_info
*ldev_info
)
1544 unsigned char modes
[32], *mode_pg
;
1548 dbd
= (scmd
->cmnd
[1] & 0x08) == 0x08;
1551 mode_pg
= &modes
[4];
1554 mode_pg
= &modes
[12];
1556 memset(modes
, 0, sizeof(modes
));
1557 modes
[0] = mode_len
- 1;
1558 modes
[2] = 0x10; /* Enable FUA */
1559 if (ldev_info
->ldev_control
.wce
== MYRS_LOGICALDEVICE_RO
)
1562 unsigned char *block_desc
= &modes
[4];
1565 put_unaligned_be32(ldev_info
->cfg_devsize
, &block_desc
[0]);
1566 put_unaligned_be32(ldev_info
->devsize_bytes
, &block_desc
[5]);
1570 if (ldev_info
->ldev_control
.rce
== MYRS_READCACHE_DISABLED
)
1572 if (ldev_info
->ldev_control
.wce
== MYRS_WRITECACHE_ENABLED
||
1573 ldev_info
->ldev_control
.wce
== MYRS_INTELLIGENT_WRITECACHE_ENABLED
)
1575 if (ldev_info
->cacheline_size
) {
1577 put_unaligned_be16(1 << ldev_info
->cacheline_size
,
1581 scsi_sg_copy_from_buffer(scmd
, modes
, mode_len
);
1584 static int myrs_queuecommand(struct Scsi_Host
*shost
,
1585 struct scsi_cmnd
*scmd
)
1587 struct myrs_hba
*cs
= shost_priv(shost
);
1588 struct myrs_cmdblk
*cmd_blk
= scsi_cmd_priv(scmd
);
1589 union myrs_cmd_mbox
*mbox
= &cmd_blk
->mbox
;
1590 struct scsi_device
*sdev
= scmd
->device
;
1591 union myrs_sgl
*hw_sge
;
1592 dma_addr_t sense_addr
;
1593 struct scatterlist
*sgl
;
1594 unsigned long flags
, timeout
;
1597 if (!scmd
->device
->hostdata
) {
1598 scmd
->result
= (DID_NO_CONNECT
<< 16);
1599 scmd
->scsi_done(scmd
);
1603 switch (scmd
->cmnd
[0]) {
1605 scsi_build_sense_buffer(0, scmd
->sense_buffer
, ILLEGAL_REQUEST
,
1607 scmd
->result
= (DRIVER_SENSE
<< 24) | SAM_STAT_CHECK_CONDITION
;
1608 scmd
->scsi_done(scmd
);
1611 if (scmd
->device
->channel
>= cs
->ctlr_info
->physchan_present
) {
1612 struct myrs_ldev_info
*ldev_info
= sdev
->hostdata
;
1614 if ((scmd
->cmnd
[2] & 0x3F) != 0x3F &&
1615 (scmd
->cmnd
[2] & 0x3F) != 0x08) {
1616 /* Illegal request, invalid field in CDB */
1617 scsi_build_sense_buffer(0, scmd
->sense_buffer
,
1618 ILLEGAL_REQUEST
, 0x24, 0);
1619 scmd
->result
= (DRIVER_SENSE
<< 24) |
1620 SAM_STAT_CHECK_CONDITION
;
1622 myrs_mode_sense(cs
, scmd
, ldev_info
);
1623 scmd
->result
= (DID_OK
<< 16);
1625 scmd
->scsi_done(scmd
);
1631 myrs_reset_cmd(cmd_blk
);
1632 cmd_blk
->sense
= dma_pool_alloc(cs
->sense_pool
, GFP_ATOMIC
,
1634 if (!cmd_blk
->sense
)
1635 return SCSI_MLQUEUE_HOST_BUSY
;
1636 cmd_blk
->sense_addr
= sense_addr
;
1638 timeout
= scmd
->request
->timeout
;
1639 if (scmd
->cmd_len
<= 10) {
1640 if (scmd
->device
->channel
>= cs
->ctlr_info
->physchan_present
) {
1641 struct myrs_ldev_info
*ldev_info
= sdev
->hostdata
;
1643 mbox
->SCSI_10
.opcode
= MYRS_CMD_OP_SCSI_10
;
1644 mbox
->SCSI_10
.pdev
.lun
= ldev_info
->lun
;
1645 mbox
->SCSI_10
.pdev
.target
= ldev_info
->target
;
1646 mbox
->SCSI_10
.pdev
.channel
= ldev_info
->channel
;
1647 mbox
->SCSI_10
.pdev
.ctlr
= 0;
1649 mbox
->SCSI_10
.opcode
= MYRS_CMD_OP_SCSI_10_PASSTHRU
;
1650 mbox
->SCSI_10
.pdev
.lun
= sdev
->lun
;
1651 mbox
->SCSI_10
.pdev
.target
= sdev
->id
;
1652 mbox
->SCSI_10
.pdev
.channel
= sdev
->channel
;
1654 mbox
->SCSI_10
.id
= scmd
->request
->tag
+ 3;
1655 mbox
->SCSI_10
.control
.dma_ctrl_to_host
=
1656 (scmd
->sc_data_direction
== DMA_FROM_DEVICE
);
1657 if (scmd
->request
->cmd_flags
& REQ_FUA
)
1658 mbox
->SCSI_10
.control
.fua
= true;
1659 mbox
->SCSI_10
.dma_size
= scsi_bufflen(scmd
);
1660 mbox
->SCSI_10
.sense_addr
= cmd_blk
->sense_addr
;
1661 mbox
->SCSI_10
.sense_len
= MYRS_SENSE_SIZE
;
1662 mbox
->SCSI_10
.cdb_len
= scmd
->cmd_len
;
1664 mbox
->SCSI_10
.tmo
.tmo_scale
= MYRS_TMO_SCALE_MINUTES
;
1665 mbox
->SCSI_10
.tmo
.tmo_val
= timeout
/ 60;
1667 mbox
->SCSI_10
.tmo
.tmo_scale
= MYRS_TMO_SCALE_SECONDS
;
1668 mbox
->SCSI_10
.tmo
.tmo_val
= timeout
;
1670 memcpy(&mbox
->SCSI_10
.cdb
, scmd
->cmnd
, scmd
->cmd_len
);
1671 hw_sge
= &mbox
->SCSI_10
.dma_addr
;
1672 cmd_blk
->dcdb
= NULL
;
1674 dma_addr_t dcdb_dma
;
1676 cmd_blk
->dcdb
= dma_pool_alloc(cs
->dcdb_pool
, GFP_ATOMIC
,
1678 if (!cmd_blk
->dcdb
) {
1679 dma_pool_free(cs
->sense_pool
, cmd_blk
->sense
,
1680 cmd_blk
->sense_addr
);
1681 cmd_blk
->sense
= NULL
;
1682 cmd_blk
->sense_addr
= 0;
1683 return SCSI_MLQUEUE_HOST_BUSY
;
1685 cmd_blk
->dcdb_dma
= dcdb_dma
;
1686 if (scmd
->device
->channel
>= cs
->ctlr_info
->physchan_present
) {
1687 struct myrs_ldev_info
*ldev_info
= sdev
->hostdata
;
1689 mbox
->SCSI_255
.opcode
= MYRS_CMD_OP_SCSI_256
;
1690 mbox
->SCSI_255
.pdev
.lun
= ldev_info
->lun
;
1691 mbox
->SCSI_255
.pdev
.target
= ldev_info
->target
;
1692 mbox
->SCSI_255
.pdev
.channel
= ldev_info
->channel
;
1693 mbox
->SCSI_255
.pdev
.ctlr
= 0;
1695 mbox
->SCSI_255
.opcode
= MYRS_CMD_OP_SCSI_255_PASSTHRU
;
1696 mbox
->SCSI_255
.pdev
.lun
= sdev
->lun
;
1697 mbox
->SCSI_255
.pdev
.target
= sdev
->id
;
1698 mbox
->SCSI_255
.pdev
.channel
= sdev
->channel
;
1700 mbox
->SCSI_255
.id
= scmd
->request
->tag
+ 3;
1701 mbox
->SCSI_255
.control
.dma_ctrl_to_host
=
1702 (scmd
->sc_data_direction
== DMA_FROM_DEVICE
);
1703 if (scmd
->request
->cmd_flags
& REQ_FUA
)
1704 mbox
->SCSI_255
.control
.fua
= true;
1705 mbox
->SCSI_255
.dma_size
= scsi_bufflen(scmd
);
1706 mbox
->SCSI_255
.sense_addr
= cmd_blk
->sense_addr
;
1707 mbox
->SCSI_255
.sense_len
= MYRS_SENSE_SIZE
;
1708 mbox
->SCSI_255
.cdb_len
= scmd
->cmd_len
;
1709 mbox
->SCSI_255
.cdb_addr
= cmd_blk
->dcdb_dma
;
1711 mbox
->SCSI_255
.tmo
.tmo_scale
= MYRS_TMO_SCALE_MINUTES
;
1712 mbox
->SCSI_255
.tmo
.tmo_val
= timeout
/ 60;
1714 mbox
->SCSI_255
.tmo
.tmo_scale
= MYRS_TMO_SCALE_SECONDS
;
1715 mbox
->SCSI_255
.tmo
.tmo_val
= timeout
;
1717 memcpy(cmd_blk
->dcdb
, scmd
->cmnd
, scmd
->cmd_len
);
1718 hw_sge
= &mbox
->SCSI_255
.dma_addr
;
1720 if (scmd
->sc_data_direction
== DMA_NONE
)
1722 nsge
= scsi_dma_map(scmd
);
1724 sgl
= scsi_sglist(scmd
);
1725 hw_sge
->sge
[0].sge_addr
= (u64
)sg_dma_address(sgl
);
1726 hw_sge
->sge
[0].sge_count
= (u64
)sg_dma_len(sgl
);
1728 struct myrs_sge
*hw_sgl
;
1729 dma_addr_t hw_sgl_addr
;
1733 hw_sgl
= dma_pool_alloc(cs
->sg_pool
, GFP_ATOMIC
,
1735 if (WARN_ON(!hw_sgl
)) {
1736 if (cmd_blk
->dcdb
) {
1737 dma_pool_free(cs
->dcdb_pool
,
1740 cmd_blk
->dcdb
= NULL
;
1741 cmd_blk
->dcdb_dma
= 0;
1743 dma_pool_free(cs
->sense_pool
,
1745 cmd_blk
->sense_addr
);
1746 cmd_blk
->sense
= NULL
;
1747 cmd_blk
->sense_addr
= 0;
1748 return SCSI_MLQUEUE_HOST_BUSY
;
1750 cmd_blk
->sgl
= hw_sgl
;
1751 cmd_blk
->sgl_addr
= hw_sgl_addr
;
1752 if (scmd
->cmd_len
<= 10)
1753 mbox
->SCSI_10
.control
.add_sge_mem
= true;
1755 mbox
->SCSI_255
.control
.add_sge_mem
= true;
1756 hw_sge
->ext
.sge0_len
= nsge
;
1757 hw_sge
->ext
.sge0_addr
= cmd_blk
->sgl_addr
;
1759 hw_sgl
= hw_sge
->sge
;
1761 scsi_for_each_sg(scmd
, sgl
, nsge
, i
) {
1762 if (WARN_ON(!hw_sgl
)) {
1763 scsi_dma_unmap(scmd
);
1764 scmd
->result
= (DID_ERROR
<< 16);
1765 scmd
->scsi_done(scmd
);
1768 hw_sgl
->sge_addr
= (u64
)sg_dma_address(sgl
);
1769 hw_sgl
->sge_count
= (u64
)sg_dma_len(sgl
);
1774 spin_lock_irqsave(&cs
->queue_lock
, flags
);
1775 myrs_qcmd(cs
, cmd_blk
);
1776 spin_unlock_irqrestore(&cs
->queue_lock
, flags
);
1781 static unsigned short myrs_translate_ldev(struct myrs_hba
*cs
,
1782 struct scsi_device
*sdev
)
1784 unsigned short ldev_num
;
1785 unsigned int chan_offset
=
1786 sdev
->channel
- cs
->ctlr_info
->physchan_present
;
1788 ldev_num
= sdev
->id
+ chan_offset
* sdev
->host
->max_id
;
1793 static int myrs_slave_alloc(struct scsi_device
*sdev
)
1795 struct myrs_hba
*cs
= shost_priv(sdev
->host
);
1796 unsigned char status
;
1798 if (sdev
->channel
> sdev
->host
->max_channel
)
1801 if (sdev
->channel
>= cs
->ctlr_info
->physchan_present
) {
1802 struct myrs_ldev_info
*ldev_info
;
1803 unsigned short ldev_num
;
1808 ldev_num
= myrs_translate_ldev(cs
, sdev
);
1810 ldev_info
= kzalloc(sizeof(*ldev_info
), GFP_KERNEL
|GFP_DMA
);
1814 status
= myrs_get_ldev_info(cs
, ldev_num
, ldev_info
);
1815 if (status
!= MYRS_STATUS_SUCCESS
) {
1816 sdev
->hostdata
= NULL
;
1819 enum raid_level level
;
1821 dev_dbg(&sdev
->sdev_gendev
,
1822 "Logical device mapping %d:%d:%d -> %d\n",
1823 ldev_info
->channel
, ldev_info
->target
,
1824 ldev_info
->lun
, ldev_info
->ldev_num
);
1826 sdev
->hostdata
= ldev_info
;
1827 switch (ldev_info
->raid_level
) {
1828 case MYRS_RAID_LEVEL0
:
1829 level
= RAID_LEVEL_LINEAR
;
1831 case MYRS_RAID_LEVEL1
:
1832 level
= RAID_LEVEL_1
;
1834 case MYRS_RAID_LEVEL3
:
1835 case MYRS_RAID_LEVEL3F
:
1836 case MYRS_RAID_LEVEL3L
:
1837 level
= RAID_LEVEL_3
;
1839 case MYRS_RAID_LEVEL5
:
1840 case MYRS_RAID_LEVEL5L
:
1841 level
= RAID_LEVEL_5
;
1843 case MYRS_RAID_LEVEL6
:
1844 level
= RAID_LEVEL_6
;
1846 case MYRS_RAID_LEVELE
:
1847 case MYRS_RAID_NEWSPAN
:
1848 case MYRS_RAID_SPAN
:
1849 level
= RAID_LEVEL_LINEAR
;
1851 case MYRS_RAID_JBOD
:
1852 level
= RAID_LEVEL_JBOD
;
1855 level
= RAID_LEVEL_UNKNOWN
;
1858 raid_set_level(myrs_raid_template
,
1859 &sdev
->sdev_gendev
, level
);
1860 if (ldev_info
->dev_state
!= MYRS_DEVICE_ONLINE
) {
1863 name
= myrs_devstate_name(ldev_info
->dev_state
);
1864 sdev_printk(KERN_DEBUG
, sdev
,
1865 "logical device in state %s\n",
1866 name
? name
: "Invalid");
1870 struct myrs_pdev_info
*pdev_info
;
1872 pdev_info
= kzalloc(sizeof(*pdev_info
), GFP_KERNEL
|GFP_DMA
);
1876 status
= myrs_get_pdev_info(cs
, sdev
->channel
,
1877 sdev
->id
, sdev
->lun
,
1879 if (status
!= MYRS_STATUS_SUCCESS
) {
1880 sdev
->hostdata
= NULL
;
1884 sdev
->hostdata
= pdev_info
;
1889 static int myrs_slave_configure(struct scsi_device
*sdev
)
1891 struct myrs_hba
*cs
= shost_priv(sdev
->host
);
1892 struct myrs_ldev_info
*ldev_info
;
1894 if (sdev
->channel
> sdev
->host
->max_channel
)
1897 if (sdev
->channel
< cs
->ctlr_info
->physchan_present
) {
1898 /* Skip HBA device */
1899 if (sdev
->type
== TYPE_RAID
)
1901 sdev
->no_uld_attach
= 1;
1907 ldev_info
= sdev
->hostdata
;
1910 if (ldev_info
->ldev_control
.wce
== MYRS_WRITECACHE_ENABLED
||
1911 ldev_info
->ldev_control
.wce
== MYRS_INTELLIGENT_WRITECACHE_ENABLED
)
1912 sdev
->wce_default_on
= 1;
1913 sdev
->tagged_supported
= 1;
1917 static void myrs_slave_destroy(struct scsi_device
*sdev
)
1919 kfree(sdev
->hostdata
);
1922 struct scsi_host_template myrs_template
= {
1923 .module
= THIS_MODULE
,
1925 .proc_name
= "myrs",
1926 .queuecommand
= myrs_queuecommand
,
1927 .eh_host_reset_handler
= myrs_host_reset
,
1928 .slave_alloc
= myrs_slave_alloc
,
1929 .slave_configure
= myrs_slave_configure
,
1930 .slave_destroy
= myrs_slave_destroy
,
1931 .cmd_size
= sizeof(struct myrs_cmdblk
),
1932 .shost_attrs
= myrs_shost_attrs
,
1933 .sdev_attrs
= myrs_sdev_attrs
,
1937 static struct myrs_hba
*myrs_alloc_host(struct pci_dev
*pdev
,
1938 const struct pci_device_id
*entry
)
1940 struct Scsi_Host
*shost
;
1941 struct myrs_hba
*cs
;
1943 shost
= scsi_host_alloc(&myrs_template
, sizeof(struct myrs_hba
));
1947 shost
->max_cmd_len
= 16;
1948 shost
->max_lun
= 256;
1949 cs
= shost_priv(shost
);
1950 mutex_init(&cs
->dcmd_mutex
);
1951 mutex_init(&cs
->cinfo_mutex
);
1958 * RAID template functions
1962 * myrs_is_raid - return boolean indicating device is raid volume
1963 * @dev the device struct object
1966 myrs_is_raid(struct device
*dev
)
1968 struct scsi_device
*sdev
= to_scsi_device(dev
);
1969 struct myrs_hba
*cs
= shost_priv(sdev
->host
);
1971 return (sdev
->channel
>= cs
->ctlr_info
->physchan_present
) ? 1 : 0;
1975 * myrs_get_resync - get raid volume resync percent complete
1976 * @dev the device struct object
1979 myrs_get_resync(struct device
*dev
)
1981 struct scsi_device
*sdev
= to_scsi_device(dev
);
1982 struct myrs_hba
*cs
= shost_priv(sdev
->host
);
1983 struct myrs_ldev_info
*ldev_info
= sdev
->hostdata
;
1984 u64 percent_complete
= 0;
1987 if (sdev
->channel
< cs
->ctlr_info
->physchan_present
|| !ldev_info
)
1989 if (ldev_info
->rbld_active
) {
1990 unsigned short ldev_num
= ldev_info
->ldev_num
;
1992 status
= myrs_get_ldev_info(cs
, ldev_num
, ldev_info
);
1993 percent_complete
= ldev_info
->rbld_lba
* 100;
1994 do_div(percent_complete
, ldev_info
->cfg_devsize
);
1996 raid_set_resync(myrs_raid_template
, dev
, percent_complete
);
2000 * myrs_get_state - get raid volume status
2001 * @dev the device struct object
2004 myrs_get_state(struct device
*dev
)
2006 struct scsi_device
*sdev
= to_scsi_device(dev
);
2007 struct myrs_hba
*cs
= shost_priv(sdev
->host
);
2008 struct myrs_ldev_info
*ldev_info
= sdev
->hostdata
;
2009 enum raid_state state
= RAID_STATE_UNKNOWN
;
2011 if (sdev
->channel
< cs
->ctlr_info
->physchan_present
|| !ldev_info
)
2012 state
= RAID_STATE_UNKNOWN
;
2014 switch (ldev_info
->dev_state
) {
2015 case MYRS_DEVICE_ONLINE
:
2016 state
= RAID_STATE_ACTIVE
;
2018 case MYRS_DEVICE_SUSPECTED_CRITICAL
:
2019 case MYRS_DEVICE_CRITICAL
:
2020 state
= RAID_STATE_DEGRADED
;
2022 case MYRS_DEVICE_REBUILD
:
2023 state
= RAID_STATE_RESYNCING
;
2025 case MYRS_DEVICE_UNCONFIGURED
:
2026 case MYRS_DEVICE_INVALID_STATE
:
2027 state
= RAID_STATE_UNKNOWN
;
2030 state
= RAID_STATE_OFFLINE
;
2033 raid_set_state(myrs_raid_template
, dev
, state
);
2036 struct raid_function_template myrs_raid_functions
= {
2037 .cookie
= &myrs_template
,
2038 .is_raid
= myrs_is_raid
,
2039 .get_resync
= myrs_get_resync
,
2040 .get_state
= myrs_get_state
,
2044 * PCI interface functions
2046 void myrs_flush_cache(struct myrs_hba
*cs
)
2048 myrs_dev_op(cs
, MYRS_IOCTL_FLUSH_DEVICE_DATA
, MYRS_RAID_CONTROLLER
);
2051 static void myrs_handle_scsi(struct myrs_hba
*cs
, struct myrs_cmdblk
*cmd_blk
,
2052 struct scsi_cmnd
*scmd
)
2054 unsigned char status
;
2059 scsi_dma_unmap(scmd
);
2060 status
= cmd_blk
->status
;
2061 if (cmd_blk
->sense
) {
2062 if (status
== MYRS_STATUS_FAILED
&& cmd_blk
->sense_len
) {
2063 unsigned int sense_len
= SCSI_SENSE_BUFFERSIZE
;
2065 if (sense_len
> cmd_blk
->sense_len
)
2066 sense_len
= cmd_blk
->sense_len
;
2067 memcpy(scmd
->sense_buffer
, cmd_blk
->sense
, sense_len
);
2069 dma_pool_free(cs
->sense_pool
, cmd_blk
->sense
,
2070 cmd_blk
->sense_addr
);
2071 cmd_blk
->sense
= NULL
;
2072 cmd_blk
->sense_addr
= 0;
2074 if (cmd_blk
->dcdb
) {
2075 dma_pool_free(cs
->dcdb_pool
, cmd_blk
->dcdb
,
2077 cmd_blk
->dcdb
= NULL
;
2078 cmd_blk
->dcdb_dma
= 0;
2081 dma_pool_free(cs
->sg_pool
, cmd_blk
->sgl
,
2083 cmd_blk
->sgl
= NULL
;
2084 cmd_blk
->sgl_addr
= 0;
2086 if (cmd_blk
->residual
)
2087 scsi_set_resid(scmd
, cmd_blk
->residual
);
2088 if (status
== MYRS_STATUS_DEVICE_NON_RESPONSIVE
||
2089 status
== MYRS_STATUS_DEVICE_NON_RESPONSIVE2
)
2090 scmd
->result
= (DID_BAD_TARGET
<< 16);
2092 scmd
->result
= (DID_OK
<< 16) | status
;
2093 scmd
->scsi_done(scmd
);
2096 static void myrs_handle_cmdblk(struct myrs_hba
*cs
, struct myrs_cmdblk
*cmd_blk
)
2101 if (cmd_blk
->complete
) {
2102 complete(cmd_blk
->complete
);
2103 cmd_blk
->complete
= NULL
;
2107 static void myrs_monitor(struct work_struct
*work
)
2109 struct myrs_hba
*cs
= container_of(work
, struct myrs_hba
,
2111 struct Scsi_Host
*shost
= cs
->host
;
2112 struct myrs_ctlr_info
*info
= cs
->ctlr_info
;
2113 unsigned int epoch
= cs
->fwstat_buf
->epoch
;
2114 unsigned long interval
= MYRS_PRIMARY_MONITOR_INTERVAL
;
2115 unsigned char status
;
2117 dev_dbg(&shost
->shost_gendev
, "monitor tick\n");
2119 status
= myrs_get_fwstatus(cs
);
2121 if (cs
->needs_update
) {
2122 cs
->needs_update
= false;
2123 mutex_lock(&cs
->cinfo_mutex
);
2124 status
= myrs_get_ctlr_info(cs
);
2125 mutex_unlock(&cs
->cinfo_mutex
);
2127 if (cs
->fwstat_buf
->next_evseq
- cs
->next_evseq
> 0) {
2128 status
= myrs_get_event(cs
, cs
->next_evseq
,
2130 if (status
== MYRS_STATUS_SUCCESS
) {
2131 myrs_log_event(cs
, cs
->event_buf
);
2137 if (time_after(jiffies
, cs
->secondary_monitor_time
2138 + MYRS_SECONDARY_MONITOR_INTERVAL
))
2139 cs
->secondary_monitor_time
= jiffies
;
2141 if (info
->bg_init_active
+
2142 info
->ldev_init_active
+
2143 info
->pdev_init_active
+
2146 info
->exp_active
!= 0) {
2147 struct scsi_device
*sdev
;
2149 shost_for_each_device(sdev
, shost
) {
2150 struct myrs_ldev_info
*ldev_info
;
2153 if (sdev
->channel
< info
->physchan_present
)
2155 ldev_info
= sdev
->hostdata
;
2158 ldev_num
= ldev_info
->ldev_num
;
2159 myrs_get_ldev_info(cs
, ldev_num
, ldev_info
);
2161 cs
->needs_update
= true;
2163 if (epoch
== cs
->epoch
&&
2164 cs
->fwstat_buf
->next_evseq
== cs
->next_evseq
&&
2165 (cs
->needs_update
== false ||
2166 time_before(jiffies
, cs
->primary_monitor_time
2167 + MYRS_PRIMARY_MONITOR_INTERVAL
))) {
2168 interval
= MYRS_SECONDARY_MONITOR_INTERVAL
;
2172 cs
->primary_monitor_time
= jiffies
;
2173 queue_delayed_work(cs
->work_q
, &cs
->monitor_work
, interval
);
2176 static bool myrs_create_mempools(struct pci_dev
*pdev
, struct myrs_hba
*cs
)
2178 struct Scsi_Host
*shost
= cs
->host
;
2179 size_t elem_size
, elem_align
;
2181 elem_align
= sizeof(struct myrs_sge
);
2182 elem_size
= shost
->sg_tablesize
* elem_align
;
2183 cs
->sg_pool
= dma_pool_create("myrs_sg", &pdev
->dev
,
2184 elem_size
, elem_align
, 0);
2185 if (cs
->sg_pool
== NULL
) {
2186 shost_printk(KERN_ERR
, shost
,
2187 "Failed to allocate SG pool\n");
2191 cs
->sense_pool
= dma_pool_create("myrs_sense", &pdev
->dev
,
2192 MYRS_SENSE_SIZE
, sizeof(int), 0);
2193 if (cs
->sense_pool
== NULL
) {
2194 dma_pool_destroy(cs
->sg_pool
);
2196 shost_printk(KERN_ERR
, shost
,
2197 "Failed to allocate sense data pool\n");
2201 cs
->dcdb_pool
= dma_pool_create("myrs_dcdb", &pdev
->dev
,
2203 sizeof(unsigned char), 0);
2204 if (!cs
->dcdb_pool
) {
2205 dma_pool_destroy(cs
->sg_pool
);
2207 dma_pool_destroy(cs
->sense_pool
);
2208 cs
->sense_pool
= NULL
;
2209 shost_printk(KERN_ERR
, shost
,
2210 "Failed to allocate DCDB pool\n");
2214 snprintf(cs
->work_q_name
, sizeof(cs
->work_q_name
),
2215 "myrs_wq_%d", shost
->host_no
);
2216 cs
->work_q
= create_singlethread_workqueue(cs
->work_q_name
);
2218 dma_pool_destroy(cs
->dcdb_pool
);
2219 cs
->dcdb_pool
= NULL
;
2220 dma_pool_destroy(cs
->sg_pool
);
2222 dma_pool_destroy(cs
->sense_pool
);
2223 cs
->sense_pool
= NULL
;
2224 shost_printk(KERN_ERR
, shost
,
2225 "Failed to create workqueue\n");
2229 /* Initialize the Monitoring Timer. */
2230 INIT_DELAYED_WORK(&cs
->monitor_work
, myrs_monitor
);
2231 queue_delayed_work(cs
->work_q
, &cs
->monitor_work
, 1);
2236 static void myrs_destroy_mempools(struct myrs_hba
*cs
)
2238 cancel_delayed_work_sync(&cs
->monitor_work
);
2239 destroy_workqueue(cs
->work_q
);
2241 dma_pool_destroy(cs
->sg_pool
);
2242 dma_pool_destroy(cs
->dcdb_pool
);
2243 dma_pool_destroy(cs
->sense_pool
);
2246 static void myrs_unmap(struct myrs_hba
*cs
)
2248 kfree(cs
->event_buf
);
2249 kfree(cs
->ctlr_info
);
2250 if (cs
->fwstat_buf
) {
2251 dma_free_coherent(&cs
->pdev
->dev
, sizeof(struct myrs_fwstat
),
2252 cs
->fwstat_buf
, cs
->fwstat_addr
);
2253 cs
->fwstat_buf
= NULL
;
2255 if (cs
->first_stat_mbox
) {
2256 dma_free_coherent(&cs
->pdev
->dev
, cs
->stat_mbox_size
,
2257 cs
->first_stat_mbox
, cs
->stat_mbox_addr
);
2258 cs
->first_stat_mbox
= NULL
;
2260 if (cs
->first_cmd_mbox
) {
2261 dma_free_coherent(&cs
->pdev
->dev
, cs
->cmd_mbox_size
,
2262 cs
->first_cmd_mbox
, cs
->cmd_mbox_addr
);
2263 cs
->first_cmd_mbox
= NULL
;
2267 static void myrs_cleanup(struct myrs_hba
*cs
)
2269 struct pci_dev
*pdev
= cs
->pdev
;
2271 /* Free the memory mailbox, status, and related structures */
2274 if (cs
->mmio_base
) {
2275 cs
->disable_intr(cs
);
2276 iounmap(cs
->mmio_base
);
2279 free_irq(cs
->irq
, cs
);
2281 release_region(cs
->io_addr
, 0x80);
2282 iounmap(cs
->mmio_base
);
2283 pci_set_drvdata(pdev
, NULL
);
2284 pci_disable_device(pdev
);
2285 scsi_host_put(cs
->host
);
2288 static struct myrs_hba
*myrs_detect(struct pci_dev
*pdev
,
2289 const struct pci_device_id
*entry
)
2291 struct myrs_privdata
*privdata
=
2292 (struct myrs_privdata
*)entry
->driver_data
;
2293 irq_handler_t irq_handler
= privdata
->irq_handler
;
2294 unsigned int mmio_size
= privdata
->mmio_size
;
2295 struct myrs_hba
*cs
= NULL
;
2297 cs
= myrs_alloc_host(pdev
, entry
);
2299 dev_err(&pdev
->dev
, "Unable to allocate Controller\n");
2304 if (pci_enable_device(pdev
))
2307 cs
->pci_addr
= pci_resource_start(pdev
, 0);
2309 pci_set_drvdata(pdev
, cs
);
2310 spin_lock_init(&cs
->queue_lock
);
2311 /* Map the Controller Register Window. */
2312 if (mmio_size
< PAGE_SIZE
)
2313 mmio_size
= PAGE_SIZE
;
2314 cs
->mmio_base
= ioremap_nocache(cs
->pci_addr
& PAGE_MASK
, mmio_size
);
2315 if (cs
->mmio_base
== NULL
) {
2317 "Unable to map Controller Register Window\n");
2321 cs
->io_base
= cs
->mmio_base
+ (cs
->pci_addr
& ~PAGE_MASK
);
2322 if (privdata
->hw_init(pdev
, cs
, cs
->io_base
))
2325 /* Acquire shared access to the IRQ Channel. */
2326 if (request_irq(pdev
->irq
, irq_handler
, IRQF_SHARED
, "myrs", cs
) < 0) {
2328 "Unable to acquire IRQ Channel %d\n", pdev
->irq
);
2331 cs
->irq
= pdev
->irq
;
2336 "Failed to initialize Controller\n");
2342 * myrs_err_status reports Controller BIOS Messages passed through
2343 the Error Status Register when the driver performs the BIOS handshaking.
2344 It returns true for fatal errors and false otherwise.
2347 static bool myrs_err_status(struct myrs_hba
*cs
, unsigned char status
,
2348 unsigned char parm0
, unsigned char parm1
)
2350 struct pci_dev
*pdev
= cs
->pdev
;
2354 dev_info(&pdev
->dev
,
2355 "Physical Device %d:%d Not Responding\n",
2359 dev_notice(&pdev
->dev
, "Spinning Up Drives\n");
2362 dev_notice(&pdev
->dev
, "Configuration Checksum Error\n");
2365 dev_notice(&pdev
->dev
, "Mirror Race Recovery Failed\n");
2368 dev_notice(&pdev
->dev
, "Mirror Race Recovery In Progress\n");
2371 dev_notice(&pdev
->dev
, "Physical Device %d:%d COD Mismatch\n",
2375 dev_notice(&pdev
->dev
, "Logical Drive Installation Aborted\n");
2378 dev_notice(&pdev
->dev
, "Mirror Race On A Critical Logical Drive\n");
2381 dev_notice(&pdev
->dev
, "New Controller Configuration Found\n");
2384 dev_err(&pdev
->dev
, "Fatal Memory Parity Error\n");
2387 dev_err(&pdev
->dev
, "Unknown Initialization Error %02X\n",
2395 * Hardware-specific functions
2399 * DAC960 GEM Series Controllers.
2402 static inline void DAC960_GEM_hw_mbox_new_cmd(void __iomem
*base
)
2404 __le32 val
= cpu_to_le32(DAC960_GEM_IDB_HWMBOX_NEW_CMD
<< 24);
2406 writel(val
, base
+ DAC960_GEM_IDB_READ_OFFSET
);
2409 static inline void DAC960_GEM_ack_hw_mbox_status(void __iomem
*base
)
2411 __le32 val
= cpu_to_le32(DAC960_GEM_IDB_HWMBOX_ACK_STS
<< 24);
2413 writel(val
, base
+ DAC960_GEM_IDB_CLEAR_OFFSET
);
2416 static inline void DAC960_GEM_gen_intr(void __iomem
*base
)
2418 __le32 val
= cpu_to_le32(DAC960_GEM_IDB_GEN_IRQ
<< 24);
2420 writel(val
, base
+ DAC960_GEM_IDB_READ_OFFSET
);
2423 static inline void DAC960_GEM_reset_ctrl(void __iomem
*base
)
2425 __le32 val
= cpu_to_le32(DAC960_GEM_IDB_CTRL_RESET
<< 24);
2427 writel(val
, base
+ DAC960_GEM_IDB_READ_OFFSET
);
2430 static inline void DAC960_GEM_mem_mbox_new_cmd(void __iomem
*base
)
2432 __le32 val
= cpu_to_le32(DAC960_GEM_IDB_HWMBOX_NEW_CMD
<< 24);
2434 writel(val
, base
+ DAC960_GEM_IDB_READ_OFFSET
);
2437 static inline bool DAC960_GEM_hw_mbox_is_full(void __iomem
*base
)
2441 val
= readl(base
+ DAC960_GEM_IDB_READ_OFFSET
);
2442 return (le32_to_cpu(val
) >> 24) & DAC960_GEM_IDB_HWMBOX_FULL
;
2445 static inline bool DAC960_GEM_init_in_progress(void __iomem
*base
)
2449 val
= readl(base
+ DAC960_GEM_IDB_READ_OFFSET
);
2450 return (le32_to_cpu(val
) >> 24) & DAC960_GEM_IDB_INIT_IN_PROGRESS
;
2453 static inline void DAC960_GEM_ack_hw_mbox_intr(void __iomem
*base
)
2455 __le32 val
= cpu_to_le32(DAC960_GEM_ODB_HWMBOX_ACK_IRQ
<< 24);
2457 writel(val
, base
+ DAC960_GEM_ODB_CLEAR_OFFSET
);
2460 static inline void DAC960_GEM_ack_mem_mbox_intr(void __iomem
*base
)
2462 __le32 val
= cpu_to_le32(DAC960_GEM_ODB_MMBOX_ACK_IRQ
<< 24);
2464 writel(val
, base
+ DAC960_GEM_ODB_CLEAR_OFFSET
);
2467 static inline void DAC960_GEM_ack_intr(void __iomem
*base
)
2469 __le32 val
= cpu_to_le32((DAC960_GEM_ODB_HWMBOX_ACK_IRQ
|
2470 DAC960_GEM_ODB_MMBOX_ACK_IRQ
) << 24);
2472 writel(val
, base
+ DAC960_GEM_ODB_CLEAR_OFFSET
);
2475 static inline bool DAC960_GEM_hw_mbox_status_available(void __iomem
*base
)
2479 val
= readl(base
+ DAC960_GEM_ODB_READ_OFFSET
);
2480 return (le32_to_cpu(val
) >> 24) & DAC960_GEM_ODB_HWMBOX_STS_AVAIL
;
2483 static inline bool DAC960_GEM_mem_mbox_status_available(void __iomem
*base
)
2487 val
= readl(base
+ DAC960_GEM_ODB_READ_OFFSET
);
2488 return (le32_to_cpu(val
) >> 24) & DAC960_GEM_ODB_MMBOX_STS_AVAIL
;
2491 static inline void DAC960_GEM_enable_intr(void __iomem
*base
)
2493 __le32 val
= cpu_to_le32((DAC960_GEM_IRQMASK_HWMBOX_IRQ
|
2494 DAC960_GEM_IRQMASK_MMBOX_IRQ
) << 24);
2495 writel(val
, base
+ DAC960_GEM_IRQMASK_CLEAR_OFFSET
);
2498 static inline void DAC960_GEM_disable_intr(void __iomem
*base
)
2502 writel(val
, base
+ DAC960_GEM_IRQMASK_READ_OFFSET
);
2505 static inline bool DAC960_GEM_intr_enabled(void __iomem
*base
)
2509 val
= readl(base
+ DAC960_GEM_IRQMASK_READ_OFFSET
);
2510 return !((le32_to_cpu(val
) >> 24) &
2511 (DAC960_GEM_IRQMASK_HWMBOX_IRQ
|
2512 DAC960_GEM_IRQMASK_MMBOX_IRQ
));
2515 static inline void DAC960_GEM_write_cmd_mbox(union myrs_cmd_mbox
*mem_mbox
,
2516 union myrs_cmd_mbox
*mbox
)
2518 memcpy(&mem_mbox
->words
[1], &mbox
->words
[1],
2519 sizeof(union myrs_cmd_mbox
) - sizeof(unsigned int));
2520 /* Barrier to avoid reordering */
2522 mem_mbox
->words
[0] = mbox
->words
[0];
2523 /* Barrier to force PCI access */
2527 static inline void DAC960_GEM_write_hw_mbox(void __iomem
*base
,
2528 dma_addr_t cmd_mbox_addr
)
2530 dma_addr_writeql(cmd_mbox_addr
, base
+ DAC960_GEM_CMDMBX_OFFSET
);
2533 static inline unsigned short DAC960_GEM_read_cmd_ident(void __iomem
*base
)
2535 return readw(base
+ DAC960_GEM_CMDSTS_OFFSET
);
2538 static inline unsigned char DAC960_GEM_read_cmd_status(void __iomem
*base
)
2540 return readw(base
+ DAC960_GEM_CMDSTS_OFFSET
+ 2);
2544 DAC960_GEM_read_error_status(void __iomem
*base
, unsigned char *error
,
2545 unsigned char *param0
, unsigned char *param1
)
2549 val
= readl(base
+ DAC960_GEM_ERRSTS_READ_OFFSET
);
2550 if (!((le32_to_cpu(val
) >> 24) & DAC960_GEM_ERRSTS_PENDING
))
2552 *error
= val
& ~(DAC960_GEM_ERRSTS_PENDING
<< 24);
2553 *param0
= readb(base
+ DAC960_GEM_CMDMBX_OFFSET
+ 0);
2554 *param1
= readb(base
+ DAC960_GEM_CMDMBX_OFFSET
+ 1);
2555 writel(0x03000000, base
+ DAC960_GEM_ERRSTS_CLEAR_OFFSET
);
2559 static inline unsigned char
2560 DAC960_GEM_mbox_init(void __iomem
*base
, dma_addr_t mbox_addr
)
2562 unsigned char status
;
2564 while (DAC960_GEM_hw_mbox_is_full(base
))
2566 DAC960_GEM_write_hw_mbox(base
, mbox_addr
);
2567 DAC960_GEM_hw_mbox_new_cmd(base
);
2568 while (!DAC960_GEM_hw_mbox_status_available(base
))
2570 status
= DAC960_GEM_read_cmd_status(base
);
2571 DAC960_GEM_ack_hw_mbox_intr(base
);
2572 DAC960_GEM_ack_hw_mbox_status(base
);
2577 static int DAC960_GEM_hw_init(struct pci_dev
*pdev
,
2578 struct myrs_hba
*cs
, void __iomem
*base
)
2581 unsigned char status
, parm0
, parm1
;
2583 DAC960_GEM_disable_intr(base
);
2584 DAC960_GEM_ack_hw_mbox_status(base
);
2586 while (DAC960_GEM_init_in_progress(base
) &&
2587 timeout
< MYRS_MAILBOX_TIMEOUT
) {
2588 if (DAC960_GEM_read_error_status(base
, &status
,
2590 myrs_err_status(cs
, status
, parm0
, parm1
))
2595 if (timeout
== MYRS_MAILBOX_TIMEOUT
) {
2597 "Timeout waiting for Controller Initialisation\n");
2600 if (!myrs_enable_mmio_mbox(cs
, DAC960_GEM_mbox_init
)) {
2602 "Unable to Enable Memory Mailbox Interface\n");
2603 DAC960_GEM_reset_ctrl(base
);
2606 DAC960_GEM_enable_intr(base
);
2607 cs
->write_cmd_mbox
= DAC960_GEM_write_cmd_mbox
;
2608 cs
->get_cmd_mbox
= DAC960_GEM_mem_mbox_new_cmd
;
2609 cs
->disable_intr
= DAC960_GEM_disable_intr
;
2610 cs
->reset
= DAC960_GEM_reset_ctrl
;
2614 static irqreturn_t
DAC960_GEM_intr_handler(int irq
, void *arg
)
2616 struct myrs_hba
*cs
= arg
;
2617 void __iomem
*base
= cs
->io_base
;
2618 struct myrs_stat_mbox
*next_stat_mbox
;
2619 unsigned long flags
;
2621 spin_lock_irqsave(&cs
->queue_lock
, flags
);
2622 DAC960_GEM_ack_intr(base
);
2623 next_stat_mbox
= cs
->next_stat_mbox
;
2624 while (next_stat_mbox
->id
> 0) {
2625 unsigned short id
= next_stat_mbox
->id
;
2626 struct scsi_cmnd
*scmd
= NULL
;
2627 struct myrs_cmdblk
*cmd_blk
= NULL
;
2629 if (id
== MYRS_DCMD_TAG
)
2630 cmd_blk
= &cs
->dcmd_blk
;
2631 else if (id
== MYRS_MCMD_TAG
)
2632 cmd_blk
= &cs
->mcmd_blk
;
2634 scmd
= scsi_host_find_tag(cs
->host
, id
- 3);
2636 cmd_blk
= scsi_cmd_priv(scmd
);
2639 cmd_blk
->status
= next_stat_mbox
->status
;
2640 cmd_blk
->sense_len
= next_stat_mbox
->sense_len
;
2641 cmd_blk
->residual
= next_stat_mbox
->residual
;
2643 dev_err(&cs
->pdev
->dev
,
2644 "Unhandled command completion %d\n", id
);
2646 memset(next_stat_mbox
, 0, sizeof(struct myrs_stat_mbox
));
2647 if (++next_stat_mbox
> cs
->last_stat_mbox
)
2648 next_stat_mbox
= cs
->first_stat_mbox
;
2652 myrs_handle_cmdblk(cs
, cmd_blk
);
2654 myrs_handle_scsi(cs
, cmd_blk
, scmd
);
2657 cs
->next_stat_mbox
= next_stat_mbox
;
2658 spin_unlock_irqrestore(&cs
->queue_lock
, flags
);
2662 struct myrs_privdata DAC960_GEM_privdata
= {
2663 .hw_init
= DAC960_GEM_hw_init
,
2664 .irq_handler
= DAC960_GEM_intr_handler
,
2665 .mmio_size
= DAC960_GEM_mmio_size
,
2669 * DAC960 BA Series Controllers.
2672 static inline void DAC960_BA_hw_mbox_new_cmd(void __iomem
*base
)
2674 writeb(DAC960_BA_IDB_HWMBOX_NEW_CMD
, base
+ DAC960_BA_IDB_OFFSET
);
2677 static inline void DAC960_BA_ack_hw_mbox_status(void __iomem
*base
)
2679 writeb(DAC960_BA_IDB_HWMBOX_ACK_STS
, base
+ DAC960_BA_IDB_OFFSET
);
2682 static inline void DAC960_BA_gen_intr(void __iomem
*base
)
2684 writeb(DAC960_BA_IDB_GEN_IRQ
, base
+ DAC960_BA_IDB_OFFSET
);
2687 static inline void DAC960_BA_reset_ctrl(void __iomem
*base
)
2689 writeb(DAC960_BA_IDB_CTRL_RESET
, base
+ DAC960_BA_IDB_OFFSET
);
2692 static inline void DAC960_BA_mem_mbox_new_cmd(void __iomem
*base
)
2694 writeb(DAC960_BA_IDB_MMBOX_NEW_CMD
, base
+ DAC960_BA_IDB_OFFSET
);
2697 static inline bool DAC960_BA_hw_mbox_is_full(void __iomem
*base
)
2701 val
= readb(base
+ DAC960_BA_IDB_OFFSET
);
2702 return !(val
& DAC960_BA_IDB_HWMBOX_EMPTY
);
2705 static inline bool DAC960_BA_init_in_progress(void __iomem
*base
)
2709 val
= readb(base
+ DAC960_BA_IDB_OFFSET
);
2710 return !(val
& DAC960_BA_IDB_INIT_DONE
);
2713 static inline void DAC960_BA_ack_hw_mbox_intr(void __iomem
*base
)
2715 writeb(DAC960_BA_ODB_HWMBOX_ACK_IRQ
, base
+ DAC960_BA_ODB_OFFSET
);
2718 static inline void DAC960_BA_ack_mem_mbox_intr(void __iomem
*base
)
2720 writeb(DAC960_BA_ODB_MMBOX_ACK_IRQ
, base
+ DAC960_BA_ODB_OFFSET
);
2723 static inline void DAC960_BA_ack_intr(void __iomem
*base
)
2725 writeb(DAC960_BA_ODB_HWMBOX_ACK_IRQ
| DAC960_BA_ODB_MMBOX_ACK_IRQ
,
2726 base
+ DAC960_BA_ODB_OFFSET
);
2729 static inline bool DAC960_BA_hw_mbox_status_available(void __iomem
*base
)
2733 val
= readb(base
+ DAC960_BA_ODB_OFFSET
);
2734 return val
& DAC960_BA_ODB_HWMBOX_STS_AVAIL
;
2737 static inline bool DAC960_BA_mem_mbox_status_available(void __iomem
*base
)
2741 val
= readb(base
+ DAC960_BA_ODB_OFFSET
);
2742 return val
& DAC960_BA_ODB_MMBOX_STS_AVAIL
;
2745 static inline void DAC960_BA_enable_intr(void __iomem
*base
)
2747 writeb(~DAC960_BA_IRQMASK_DISABLE_IRQ
, base
+ DAC960_BA_IRQMASK_OFFSET
);
2750 static inline void DAC960_BA_disable_intr(void __iomem
*base
)
2752 writeb(0xFF, base
+ DAC960_BA_IRQMASK_OFFSET
);
2755 static inline bool DAC960_BA_intr_enabled(void __iomem
*base
)
2759 val
= readb(base
+ DAC960_BA_IRQMASK_OFFSET
);
2760 return !(val
& DAC960_BA_IRQMASK_DISABLE_IRQ
);
2763 static inline void DAC960_BA_write_cmd_mbox(union myrs_cmd_mbox
*mem_mbox
,
2764 union myrs_cmd_mbox
*mbox
)
2766 memcpy(&mem_mbox
->words
[1], &mbox
->words
[1],
2767 sizeof(union myrs_cmd_mbox
) - sizeof(unsigned int));
2768 /* Barrier to avoid reordering */
2770 mem_mbox
->words
[0] = mbox
->words
[0];
2771 /* Barrier to force PCI access */
2776 static inline void DAC960_BA_write_hw_mbox(void __iomem
*base
,
2777 dma_addr_t cmd_mbox_addr
)
2779 dma_addr_writeql(cmd_mbox_addr
, base
+ DAC960_BA_CMDMBX_OFFSET
);
2782 static inline unsigned short DAC960_BA_read_cmd_ident(void __iomem
*base
)
2784 return readw(base
+ DAC960_BA_CMDSTS_OFFSET
);
2787 static inline unsigned char DAC960_BA_read_cmd_status(void __iomem
*base
)
2789 return readw(base
+ DAC960_BA_CMDSTS_OFFSET
+ 2);
2793 DAC960_BA_read_error_status(void __iomem
*base
, unsigned char *error
,
2794 unsigned char *param0
, unsigned char *param1
)
2798 val
= readb(base
+ DAC960_BA_ERRSTS_OFFSET
);
2799 if (!(val
& DAC960_BA_ERRSTS_PENDING
))
2801 val
&= ~DAC960_BA_ERRSTS_PENDING
;
2803 *param0
= readb(base
+ DAC960_BA_CMDMBX_OFFSET
+ 0);
2804 *param1
= readb(base
+ DAC960_BA_CMDMBX_OFFSET
+ 1);
2805 writeb(0xFF, base
+ DAC960_BA_ERRSTS_OFFSET
);
2809 static inline unsigned char
2810 DAC960_BA_mbox_init(void __iomem
*base
, dma_addr_t mbox_addr
)
2812 unsigned char status
;
2814 while (DAC960_BA_hw_mbox_is_full(base
))
2816 DAC960_BA_write_hw_mbox(base
, mbox_addr
);
2817 DAC960_BA_hw_mbox_new_cmd(base
);
2818 while (!DAC960_BA_hw_mbox_status_available(base
))
2820 status
= DAC960_BA_read_cmd_status(base
);
2821 DAC960_BA_ack_hw_mbox_intr(base
);
2822 DAC960_BA_ack_hw_mbox_status(base
);
2827 static int DAC960_BA_hw_init(struct pci_dev
*pdev
,
2828 struct myrs_hba
*cs
, void __iomem
*base
)
2831 unsigned char status
, parm0
, parm1
;
2833 DAC960_BA_disable_intr(base
);
2834 DAC960_BA_ack_hw_mbox_status(base
);
2836 while (DAC960_BA_init_in_progress(base
) &&
2837 timeout
< MYRS_MAILBOX_TIMEOUT
) {
2838 if (DAC960_BA_read_error_status(base
, &status
,
2840 myrs_err_status(cs
, status
, parm0
, parm1
))
2845 if (timeout
== MYRS_MAILBOX_TIMEOUT
) {
2847 "Timeout waiting for Controller Initialisation\n");
2850 if (!myrs_enable_mmio_mbox(cs
, DAC960_BA_mbox_init
)) {
2852 "Unable to Enable Memory Mailbox Interface\n");
2853 DAC960_BA_reset_ctrl(base
);
2856 DAC960_BA_enable_intr(base
);
2857 cs
->write_cmd_mbox
= DAC960_BA_write_cmd_mbox
;
2858 cs
->get_cmd_mbox
= DAC960_BA_mem_mbox_new_cmd
;
2859 cs
->disable_intr
= DAC960_BA_disable_intr
;
2860 cs
->reset
= DAC960_BA_reset_ctrl
;
2864 static irqreturn_t
DAC960_BA_intr_handler(int irq
, void *arg
)
2866 struct myrs_hba
*cs
= arg
;
2867 void __iomem
*base
= cs
->io_base
;
2868 struct myrs_stat_mbox
*next_stat_mbox
;
2869 unsigned long flags
;
2871 spin_lock_irqsave(&cs
->queue_lock
, flags
);
2872 DAC960_BA_ack_intr(base
);
2873 next_stat_mbox
= cs
->next_stat_mbox
;
2874 while (next_stat_mbox
->id
> 0) {
2875 unsigned short id
= next_stat_mbox
->id
;
2876 struct scsi_cmnd
*scmd
= NULL
;
2877 struct myrs_cmdblk
*cmd_blk
= NULL
;
2879 if (id
== MYRS_DCMD_TAG
)
2880 cmd_blk
= &cs
->dcmd_blk
;
2881 else if (id
== MYRS_MCMD_TAG
)
2882 cmd_blk
= &cs
->mcmd_blk
;
2884 scmd
= scsi_host_find_tag(cs
->host
, id
- 3);
2886 cmd_blk
= scsi_cmd_priv(scmd
);
2889 cmd_blk
->status
= next_stat_mbox
->status
;
2890 cmd_blk
->sense_len
= next_stat_mbox
->sense_len
;
2891 cmd_blk
->residual
= next_stat_mbox
->residual
;
2893 dev_err(&cs
->pdev
->dev
,
2894 "Unhandled command completion %d\n", id
);
2896 memset(next_stat_mbox
, 0, sizeof(struct myrs_stat_mbox
));
2897 if (++next_stat_mbox
> cs
->last_stat_mbox
)
2898 next_stat_mbox
= cs
->first_stat_mbox
;
2902 myrs_handle_cmdblk(cs
, cmd_blk
);
2904 myrs_handle_scsi(cs
, cmd_blk
, scmd
);
2907 cs
->next_stat_mbox
= next_stat_mbox
;
2908 spin_unlock_irqrestore(&cs
->queue_lock
, flags
);
2912 struct myrs_privdata DAC960_BA_privdata
= {
2913 .hw_init
= DAC960_BA_hw_init
,
2914 .irq_handler
= DAC960_BA_intr_handler
,
2915 .mmio_size
= DAC960_BA_mmio_size
,
2919 * DAC960 LP Series Controllers.
2922 static inline void DAC960_LP_hw_mbox_new_cmd(void __iomem
*base
)
2924 writeb(DAC960_LP_IDB_HWMBOX_NEW_CMD
, base
+ DAC960_LP_IDB_OFFSET
);
2927 static inline void DAC960_LP_ack_hw_mbox_status(void __iomem
*base
)
2929 writeb(DAC960_LP_IDB_HWMBOX_ACK_STS
, base
+ DAC960_LP_IDB_OFFSET
);
2932 static inline void DAC960_LP_gen_intr(void __iomem
*base
)
2934 writeb(DAC960_LP_IDB_GEN_IRQ
, base
+ DAC960_LP_IDB_OFFSET
);
2937 static inline void DAC960_LP_reset_ctrl(void __iomem
*base
)
2939 writeb(DAC960_LP_IDB_CTRL_RESET
, base
+ DAC960_LP_IDB_OFFSET
);
2942 static inline void DAC960_LP_mem_mbox_new_cmd(void __iomem
*base
)
2944 writeb(DAC960_LP_IDB_MMBOX_NEW_CMD
, base
+ DAC960_LP_IDB_OFFSET
);
2947 static inline bool DAC960_LP_hw_mbox_is_full(void __iomem
*base
)
2951 val
= readb(base
+ DAC960_LP_IDB_OFFSET
);
2952 return val
& DAC960_LP_IDB_HWMBOX_FULL
;
2955 static inline bool DAC960_LP_init_in_progress(void __iomem
*base
)
2959 val
= readb(base
+ DAC960_LP_IDB_OFFSET
);
2960 return val
& DAC960_LP_IDB_INIT_IN_PROGRESS
;
2963 static inline void DAC960_LP_ack_hw_mbox_intr(void __iomem
*base
)
2965 writeb(DAC960_LP_ODB_HWMBOX_ACK_IRQ
, base
+ DAC960_LP_ODB_OFFSET
);
2968 static inline void DAC960_LP_ack_mem_mbox_intr(void __iomem
*base
)
2970 writeb(DAC960_LP_ODB_MMBOX_ACK_IRQ
, base
+ DAC960_LP_ODB_OFFSET
);
2973 static inline void DAC960_LP_ack_intr(void __iomem
*base
)
2975 writeb(DAC960_LP_ODB_HWMBOX_ACK_IRQ
| DAC960_LP_ODB_MMBOX_ACK_IRQ
,
2976 base
+ DAC960_LP_ODB_OFFSET
);
2979 static inline bool DAC960_LP_hw_mbox_status_available(void __iomem
*base
)
2983 val
= readb(base
+ DAC960_LP_ODB_OFFSET
);
2984 return val
& DAC960_LP_ODB_HWMBOX_STS_AVAIL
;
2987 static inline bool DAC960_LP_mem_mbox_status_available(void __iomem
*base
)
2991 val
= readb(base
+ DAC960_LP_ODB_OFFSET
);
2992 return val
& DAC960_LP_ODB_MMBOX_STS_AVAIL
;
2995 static inline void DAC960_LP_enable_intr(void __iomem
*base
)
2997 writeb(~DAC960_LP_IRQMASK_DISABLE_IRQ
, base
+ DAC960_LP_IRQMASK_OFFSET
);
3000 static inline void DAC960_LP_disable_intr(void __iomem
*base
)
3002 writeb(0xFF, base
+ DAC960_LP_IRQMASK_OFFSET
);
3005 static inline bool DAC960_LP_intr_enabled(void __iomem
*base
)
3009 val
= readb(base
+ DAC960_LP_IRQMASK_OFFSET
);
3010 return !(val
& DAC960_LP_IRQMASK_DISABLE_IRQ
);
3013 static inline void DAC960_LP_write_cmd_mbox(union myrs_cmd_mbox
*mem_mbox
,
3014 union myrs_cmd_mbox
*mbox
)
3016 memcpy(&mem_mbox
->words
[1], &mbox
->words
[1],
3017 sizeof(union myrs_cmd_mbox
) - sizeof(unsigned int));
3018 /* Barrier to avoid reordering */
3020 mem_mbox
->words
[0] = mbox
->words
[0];
3021 /* Barrier to force PCI access */
3025 static inline void DAC960_LP_write_hw_mbox(void __iomem
*base
,
3026 dma_addr_t cmd_mbox_addr
)
3028 dma_addr_writeql(cmd_mbox_addr
, base
+ DAC960_LP_CMDMBX_OFFSET
);
3031 static inline unsigned short DAC960_LP_read_cmd_ident(void __iomem
*base
)
3033 return readw(base
+ DAC960_LP_CMDSTS_OFFSET
);
3036 static inline unsigned char DAC960_LP_read_cmd_status(void __iomem
*base
)
3038 return readw(base
+ DAC960_LP_CMDSTS_OFFSET
+ 2);
3042 DAC960_LP_read_error_status(void __iomem
*base
, unsigned char *error
,
3043 unsigned char *param0
, unsigned char *param1
)
3047 val
= readb(base
+ DAC960_LP_ERRSTS_OFFSET
);
3048 if (!(val
& DAC960_LP_ERRSTS_PENDING
))
3050 val
&= ~DAC960_LP_ERRSTS_PENDING
;
3052 *param0
= readb(base
+ DAC960_LP_CMDMBX_OFFSET
+ 0);
3053 *param1
= readb(base
+ DAC960_LP_CMDMBX_OFFSET
+ 1);
3054 writeb(0xFF, base
+ DAC960_LP_ERRSTS_OFFSET
);
3058 static inline unsigned char
3059 DAC960_LP_mbox_init(void __iomem
*base
, dma_addr_t mbox_addr
)
3061 unsigned char status
;
3063 while (DAC960_LP_hw_mbox_is_full(base
))
3065 DAC960_LP_write_hw_mbox(base
, mbox_addr
);
3066 DAC960_LP_hw_mbox_new_cmd(base
);
3067 while (!DAC960_LP_hw_mbox_status_available(base
))
3069 status
= DAC960_LP_read_cmd_status(base
);
3070 DAC960_LP_ack_hw_mbox_intr(base
);
3071 DAC960_LP_ack_hw_mbox_status(base
);
3076 static int DAC960_LP_hw_init(struct pci_dev
*pdev
,
3077 struct myrs_hba
*cs
, void __iomem
*base
)
3080 unsigned char status
, parm0
, parm1
;
3082 DAC960_LP_disable_intr(base
);
3083 DAC960_LP_ack_hw_mbox_status(base
);
3085 while (DAC960_LP_init_in_progress(base
) &&
3086 timeout
< MYRS_MAILBOX_TIMEOUT
) {
3087 if (DAC960_LP_read_error_status(base
, &status
,
3089 myrs_err_status(cs
, status
, parm0
, parm1
))
3094 if (timeout
== MYRS_MAILBOX_TIMEOUT
) {
3096 "Timeout waiting for Controller Initialisation\n");
3099 if (!myrs_enable_mmio_mbox(cs
, DAC960_LP_mbox_init
)) {
3101 "Unable to Enable Memory Mailbox Interface\n");
3102 DAC960_LP_reset_ctrl(base
);
3105 DAC960_LP_enable_intr(base
);
3106 cs
->write_cmd_mbox
= DAC960_LP_write_cmd_mbox
;
3107 cs
->get_cmd_mbox
= DAC960_LP_mem_mbox_new_cmd
;
3108 cs
->disable_intr
= DAC960_LP_disable_intr
;
3109 cs
->reset
= DAC960_LP_reset_ctrl
;
3114 static irqreturn_t
DAC960_LP_intr_handler(int irq
, void *arg
)
3116 struct myrs_hba
*cs
= arg
;
3117 void __iomem
*base
= cs
->io_base
;
3118 struct myrs_stat_mbox
*next_stat_mbox
;
3119 unsigned long flags
;
3121 spin_lock_irqsave(&cs
->queue_lock
, flags
);
3122 DAC960_LP_ack_intr(base
);
3123 next_stat_mbox
= cs
->next_stat_mbox
;
3124 while (next_stat_mbox
->id
> 0) {
3125 unsigned short id
= next_stat_mbox
->id
;
3126 struct scsi_cmnd
*scmd
= NULL
;
3127 struct myrs_cmdblk
*cmd_blk
= NULL
;
3129 if (id
== MYRS_DCMD_TAG
)
3130 cmd_blk
= &cs
->dcmd_blk
;
3131 else if (id
== MYRS_MCMD_TAG
)
3132 cmd_blk
= &cs
->mcmd_blk
;
3134 scmd
= scsi_host_find_tag(cs
->host
, id
- 3);
3136 cmd_blk
= scsi_cmd_priv(scmd
);
3139 cmd_blk
->status
= next_stat_mbox
->status
;
3140 cmd_blk
->sense_len
= next_stat_mbox
->sense_len
;
3141 cmd_blk
->residual
= next_stat_mbox
->residual
;
3143 dev_err(&cs
->pdev
->dev
,
3144 "Unhandled command completion %d\n", id
);
3146 memset(next_stat_mbox
, 0, sizeof(struct myrs_stat_mbox
));
3147 if (++next_stat_mbox
> cs
->last_stat_mbox
)
3148 next_stat_mbox
= cs
->first_stat_mbox
;
3152 myrs_handle_cmdblk(cs
, cmd_blk
);
3154 myrs_handle_scsi(cs
, cmd_blk
, scmd
);
3157 cs
->next_stat_mbox
= next_stat_mbox
;
3158 spin_unlock_irqrestore(&cs
->queue_lock
, flags
);
3162 struct myrs_privdata DAC960_LP_privdata
= {
3163 .hw_init
= DAC960_LP_hw_init
,
3164 .irq_handler
= DAC960_LP_intr_handler
,
3165 .mmio_size
= DAC960_LP_mmio_size
,
3172 myrs_probe(struct pci_dev
*dev
, const struct pci_device_id
*entry
)
3174 struct myrs_hba
*cs
;
3177 cs
= myrs_detect(dev
, entry
);
3181 ret
= myrs_get_config(cs
);
3187 if (!myrs_create_mempools(dev
, cs
)) {
3192 ret
= scsi_add_host(cs
->host
, &dev
->dev
);
3194 dev_err(&dev
->dev
, "scsi_add_host failed with %d\n", ret
);
3195 myrs_destroy_mempools(cs
);
3198 scsi_scan_host(cs
->host
);
3206 static void myrs_remove(struct pci_dev
*pdev
)
3208 struct myrs_hba
*cs
= pci_get_drvdata(pdev
);
3213 shost_printk(KERN_NOTICE
, cs
->host
, "Flushing Cache...");
3214 myrs_flush_cache(cs
);
3215 myrs_destroy_mempools(cs
);
3220 static const struct pci_device_id myrs_id_table
[] = {
3222 PCI_DEVICE_SUB(PCI_VENDOR_ID_MYLEX
,
3223 PCI_DEVICE_ID_MYLEX_DAC960_GEM
,
3224 PCI_VENDOR_ID_MYLEX
, PCI_ANY_ID
),
3225 .driver_data
= (unsigned long) &DAC960_GEM_privdata
,
3228 PCI_DEVICE_DATA(MYLEX
, DAC960_BA
, &DAC960_BA_privdata
),
3231 PCI_DEVICE_DATA(MYLEX
, DAC960_LP
, &DAC960_LP_privdata
),
3236 MODULE_DEVICE_TABLE(pci
, myrs_id_table
);
3238 static struct pci_driver myrs_pci_driver
= {
3240 .id_table
= myrs_id_table
,
3241 .probe
= myrs_probe
,
3242 .remove
= myrs_remove
,
3245 static int __init
myrs_init_module(void)
3249 myrs_raid_template
= raid_class_attach(&myrs_raid_functions
);
3250 if (!myrs_raid_template
)
3253 ret
= pci_register_driver(&myrs_pci_driver
);
3255 raid_class_release(myrs_raid_template
);
3260 static void __exit
myrs_cleanup_module(void)
3262 pci_unregister_driver(&myrs_pci_driver
);
3263 raid_class_release(myrs_raid_template
);
3266 module_init(myrs_init_module
);
3267 module_exit(myrs_cleanup_module
);
3269 MODULE_DESCRIPTION("Mylex DAC960/AcceleRAID/eXtremeRAID driver (SCSI Interface)");
3270 MODULE_AUTHOR("Hannes Reinecke <hare@suse.com>");
3271 MODULE_LICENSE("GPL");