2 * Adaptec AAC series RAID controller driver
3 * (c) Copyright 2001 Red Hat Inc.
5 * based on the old aacraid driver that is..
6 * Adaptec aacraid device driver for Linux.
8 * Copyright (c) 2000-2010 Adaptec, Inc.
9 * 2010-2015 PMC-Sierra, Inc. (aacraid@pmc-sierra.com)
10 * 2016-2017 Microsemi Corp. (aacraid@microsemi.com)
12 * This program is free software; you can redistribute it and/or modify
13 * it under the terms of the GNU General Public License as published by
14 * the Free Software Foundation; either version 2, or (at your option)
17 * This program is distributed in the hope that it will be useful,
18 * but WITHOUT ANY WARRANTY; without even the implied warranty of
19 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
20 * GNU General Public License for more details.
22 * You should have received a copy of the GNU General Public License
23 * along with this program; see the file COPYING. If not, write to
24 * the Free Software Foundation, 675 Mass Ave, Cambridge, MA 02139, USA.
29 * Abstract: Contains all routines for control of the AFA comm layer
33 #include <linux/kernel.h>
34 #include <linux/init.h>
35 #include <linux/types.h>
36 #include <linux/pci.h>
37 #include <linux/spinlock.h>
38 #include <linux/slab.h>
39 #include <linux/completion.h>
40 #include <linux/dma-mapping.h>
41 #include <linux/blkdev.h>
42 #include <linux/delay.h> /* ssleep prototype */
43 #include <linux/kthread.h>
44 #include <linux/semaphore.h>
45 #include <linux/uaccess.h>
46 #include <scsi/scsi_host.h>
51 * ioctl_send_fib - send a FIB from userspace
52 * @dev: adapter is being processed
53 * @arg: arguments to the ioctl call
55 * This routine sends a fib to the adapter on behalf of a user level
58 # define AAC_DEBUG_PREAMBLE KERN_INFO
59 # define AAC_DEBUG_POSTAMBLE
61 static int ioctl_send_fib(struct aac_dev
* dev
, void __user
*arg
)
65 struct hw_fib
* hw_fib
= (struct hw_fib
*)0;
66 dma_addr_t hw_fib_pa
= (dma_addr_t
)0LL;
67 unsigned int size
, osize
;
73 fibptr
= aac_fib_alloc(dev
);
78 kfib
= fibptr
->hw_fib_va
;
80 * First copy in the header so that we can check the size field.
82 if (copy_from_user((void *)kfib
, arg
, sizeof(struct aac_fibhdr
))) {
87 * Since we copy based on the fib header size, make sure that we
88 * will not overrun the buffer when we copy the memory. Return
89 * an error if we would.
91 osize
= size
= le16_to_cpu(kfib
->header
.Size
) +
92 sizeof(struct aac_fibhdr
);
93 if (size
< le16_to_cpu(kfib
->header
.SenderSize
))
94 size
= le16_to_cpu(kfib
->header
.SenderSize
);
95 if (size
> dev
->max_fib_size
) {
103 kfib
= dma_alloc_coherent(&dev
->pdev
->dev
, size
, &daddr
,
110 /* Highjack the hw_fib */
111 hw_fib
= fibptr
->hw_fib_va
;
112 hw_fib_pa
= fibptr
->hw_fib_pa
;
113 fibptr
->hw_fib_va
= kfib
;
114 fibptr
->hw_fib_pa
= daddr
;
115 memset(((char *)kfib
) + dev
->max_fib_size
, 0, size
- dev
->max_fib_size
);
116 memcpy(kfib
, hw_fib
, dev
->max_fib_size
);
119 if (copy_from_user(kfib
, arg
, size
)) {
124 /* Sanity check the second copy */
125 if ((osize
!= le16_to_cpu(kfib
->header
.Size
) +
126 sizeof(struct aac_fibhdr
))
127 || (size
< le16_to_cpu(kfib
->header
.SenderSize
))) {
132 if (kfib
->header
.Command
== cpu_to_le16(TakeABreakPt
)) {
133 aac_adapter_interrupt(dev
);
135 * Since we didn't really send a fib, zero out the state to allow
136 * cleanup code not to assert.
138 kfib
->header
.XferState
= 0;
140 retval
= aac_fib_send(le16_to_cpu(kfib
->header
.Command
), fibptr
,
141 le16_to_cpu(kfib
->header
.Size
) , FsaNormal
,
146 if (aac_fib_complete(fibptr
) != 0) {
152 * Make sure that the size returned by the adapter (which includes
153 * the header) is less than or equal to the size of a fib, so we
154 * don't corrupt application data. Then copy that size to the user
155 * buffer. (Don't try to add the header information again, since it
156 * was already included by the adapter.)
160 if (copy_to_user(arg
, (void *)kfib
, size
))
164 dma_free_coherent(&dev
->pdev
->dev
, size
, kfib
,
166 fibptr
->hw_fib_pa
= hw_fib_pa
;
167 fibptr
->hw_fib_va
= hw_fib
;
169 if (retval
!= -ERESTARTSYS
)
170 aac_fib_free(fibptr
);
175 * open_getadapter_fib - Get the next fib
177 * This routine will get the next Fib, if available, from the AdapterFibContext
178 * passed in from the user.
181 static int open_getadapter_fib(struct aac_dev
* dev
, void __user
*arg
)
183 struct aac_fib_context
* fibctx
;
186 fibctx
= kmalloc(sizeof(struct aac_fib_context
), GFP_KERNEL
);
187 if (fibctx
== NULL
) {
191 struct list_head
* entry
;
192 struct aac_fib_context
* context
;
194 fibctx
->type
= FSAFS_NTC_GET_ADAPTER_FIB_CONTEXT
;
195 fibctx
->size
= sizeof(struct aac_fib_context
);
197 * Yes yes, I know this could be an index, but we have a
198 * better guarantee of uniqueness for the locked loop below.
199 * Without the aid of a persistent history, this also helps
200 * reduce the chance that the opaque context would be reused.
202 fibctx
->unique
= (u32
)((ulong
)fibctx
& 0xFFFFFFFF);
204 * Initialize the mutex used to wait for the next AIF.
206 sema_init(&fibctx
->wait_sem
, 0);
209 * Initialize the fibs and set the count of fibs on
213 INIT_LIST_HEAD(&fibctx
->fib_list
);
214 fibctx
->jiffies
= jiffies
/HZ
;
216 * Now add this context onto the adapter's
217 * AdapterFibContext list.
219 spin_lock_irqsave(&dev
->fib_lock
, flags
);
220 /* Ensure that we have a unique identifier */
221 entry
= dev
->fib_list
.next
;
222 while (entry
!= &dev
->fib_list
) {
223 context
= list_entry(entry
, struct aac_fib_context
, next
);
224 if (context
->unique
== fibctx
->unique
) {
225 /* Not unique (32 bits) */
227 entry
= dev
->fib_list
.next
;
232 list_add_tail(&fibctx
->next
, &dev
->fib_list
);
233 spin_unlock_irqrestore(&dev
->fib_lock
, flags
);
234 if (copy_to_user(arg
, &fibctx
->unique
,
235 sizeof(fibctx
->unique
))) {
245 * next_getadapter_fib - get the next fib
246 * @dev: adapter to use
247 * @arg: ioctl argument
249 * This routine will get the next Fib, if available, from the AdapterFibContext
250 * passed in from the user.
253 static int next_getadapter_fib(struct aac_dev
* dev
, void __user
*arg
)
257 struct aac_fib_context
*fibctx
;
259 struct list_head
* entry
;
262 if(copy_from_user((void *)&f
, arg
, sizeof(struct fib_ioctl
)))
265 * Verify that the HANDLE passed in was a valid AdapterFibContext
267 * Search the list of AdapterFibContext addresses on the adapter
268 * to be sure this is a valid address
270 spin_lock_irqsave(&dev
->fib_lock
, flags
);
271 entry
= dev
->fib_list
.next
;
274 while (entry
!= &dev
->fib_list
) {
275 fibctx
= list_entry(entry
, struct aac_fib_context
, next
);
277 * Extract the AdapterFibContext from the Input parameters.
279 if (fibctx
->unique
== f
.fibctx
) { /* We found a winner */
286 spin_unlock_irqrestore(&dev
->fib_lock
, flags
);
287 dprintk ((KERN_INFO
"Fib Context not found\n"));
291 if((fibctx
->type
!= FSAFS_NTC_GET_ADAPTER_FIB_CONTEXT
) ||
292 (fibctx
->size
!= sizeof(struct aac_fib_context
))) {
293 spin_unlock_irqrestore(&dev
->fib_lock
, flags
);
294 dprintk ((KERN_INFO
"Fib Context corrupt?\n"));
299 * If there are no fibs to send back, then either wait or return
303 if (!list_empty(&fibctx
->fib_list
)) {
305 * Pull the next fib from the fibs
307 entry
= fibctx
->fib_list
.next
;
310 fib
= list_entry(entry
, struct fib
, fiblink
);
312 spin_unlock_irqrestore(&dev
->fib_lock
, flags
);
313 if (copy_to_user(f
.fib
, fib
->hw_fib_va
, sizeof(struct hw_fib
))) {
314 kfree(fib
->hw_fib_va
);
319 * Free the space occupied by this copy of the fib.
321 kfree(fib
->hw_fib_va
);
325 spin_unlock_irqrestore(&dev
->fib_lock
, flags
);
326 /* If someone killed the AIF aacraid thread, restart it */
327 status
= !dev
->aif_thread
;
328 if (status
&& !dev
->in_reset
&& dev
->queues
&& dev
->fsa_dev
) {
329 /* Be paranoid, be very paranoid! */
330 kthread_stop(dev
->thread
);
333 dev
->thread
= kthread_run(aac_command_thread
, dev
,
338 if(down_interruptible(&fibctx
->wait_sem
) < 0) {
339 status
= -ERESTARTSYS
;
341 /* Lock again and retry */
342 spin_lock_irqsave(&dev
->fib_lock
, flags
);
349 fibctx
->jiffies
= jiffies
/HZ
;
353 int aac_close_fib_context(struct aac_dev
* dev
, struct aac_fib_context
* fibctx
)
358 * First free any FIBs that have not been consumed.
360 while (!list_empty(&fibctx
->fib_list
)) {
361 struct list_head
* entry
;
363 * Pull the next fib from the fibs
365 entry
= fibctx
->fib_list
.next
;
367 fib
= list_entry(entry
, struct fib
, fiblink
);
370 * Free the space occupied by this copy of the fib.
372 kfree(fib
->hw_fib_va
);
376 * Remove the Context from the AdapterFibContext List
378 list_del(&fibctx
->next
);
384 * Free the space occupied by the Context
391 * close_getadapter_fib - close down user fib context
393 * @arg: ioctl arguments
395 * This routine will close down the fibctx passed in from the user.
398 static int close_getadapter_fib(struct aac_dev
* dev
, void __user
*arg
)
400 struct aac_fib_context
*fibctx
;
403 struct list_head
* entry
;
406 * Verify that the HANDLE passed in was a valid AdapterFibContext
408 * Search the list of AdapterFibContext addresses on the adapter
409 * to be sure this is a valid address
412 entry
= dev
->fib_list
.next
;
415 while(entry
!= &dev
->fib_list
) {
416 fibctx
= list_entry(entry
, struct aac_fib_context
, next
);
418 * Extract the fibctx from the input parameters
420 if (fibctx
->unique
== (u32
)(uintptr_t)arg
) /* We found a winner */
427 return 0; /* Already gone */
429 if((fibctx
->type
!= FSAFS_NTC_GET_ADAPTER_FIB_CONTEXT
) ||
430 (fibctx
->size
!= sizeof(struct aac_fib_context
)))
432 spin_lock_irqsave(&dev
->fib_lock
, flags
);
433 status
= aac_close_fib_context(dev
, fibctx
);
434 spin_unlock_irqrestore(&dev
->fib_lock
, flags
);
439 * check_revision - close down user fib context
441 * @arg: ioctl arguments
443 * This routine returns the driver version.
444 * Under Linux, there have been no version incompatibilities, so this is
448 static int check_revision(struct aac_dev
*dev
, void __user
*arg
)
450 struct revision response
;
451 char *driver_version
= aac_driver_version
;
455 version
= (simple_strtol(driver_version
,
456 &driver_version
, 10) << 24) | 0x00000400;
457 version
+= simple_strtol(driver_version
+ 1, &driver_version
, 10) << 16;
458 version
+= simple_strtol(driver_version
+ 1, NULL
, 10);
459 response
.version
= cpu_to_le32(version
);
460 # ifdef AAC_DRIVER_BUILD
461 response
.build
= cpu_to_le32(AAC_DRIVER_BUILD
);
463 response
.build
= cpu_to_le32(9999);
466 if (copy_to_user(arg
, &response
, sizeof(response
)))
478 static int aac_send_raw_srb(struct aac_dev
* dev
, void __user
* arg
)
482 struct aac_srb
*srbcmd
= NULL
;
483 struct aac_hba_cmd_req
*hbacmd
= NULL
;
484 struct user_aac_srb
*user_srbcmd
= NULL
;
485 struct user_aac_srb __user
*user_srb
= arg
;
486 struct aac_srb_reply __user
*user_reply
;
492 void __user
*sg_user
[HBA_MAX_SG_EMBEDDED
];
493 void *sg_list
[HBA_MAX_SG_EMBEDDED
];
494 u32 sg_count
[HBA_MAX_SG_EMBEDDED
];
497 u32 actual_fibsize64
, actual_fibsize
= 0;
499 int is_native_device
;
504 dprintk((KERN_DEBUG
"aacraid: send raw srb -EBUSY\n"));
507 if (!capable(CAP_SYS_ADMIN
)){
508 dprintk((KERN_DEBUG
"aacraid: No permission to send raw srb\n"));
512 * Allocate and initialize a Fib then setup a SRB command
514 if (!(srbfib
= aac_fib_alloc(dev
))) {
518 memset(sg_list
, 0, sizeof(sg_list
)); /* cleanup may take issue */
519 if(copy_from_user(&fibsize
, &user_srb
->count
,sizeof(u32
))){
520 dprintk((KERN_DEBUG
"aacraid: Could not copy data size from user\n"));
525 if ((fibsize
< (sizeof(struct user_aac_srb
) - sizeof(struct user_sgentry
))) ||
526 (fibsize
> (dev
->max_fib_size
- sizeof(struct aac_fibhdr
)))) {
531 user_srbcmd
= kmalloc(fibsize
, GFP_KERNEL
);
533 dprintk((KERN_DEBUG
"aacraid: Could not make a copy of the srb\n"));
537 if(copy_from_user(user_srbcmd
, user_srb
,fibsize
)){
538 dprintk((KERN_DEBUG
"aacraid: Could not copy srb from user\n"));
543 flags
= user_srbcmd
->flags
; /* from user in cpu order */
544 switch (flags
& (SRB_DataIn
| SRB_DataOut
)) {
546 data_dir
= DMA_TO_DEVICE
;
548 case (SRB_DataIn
| SRB_DataOut
):
549 data_dir
= DMA_BIDIRECTIONAL
;
552 data_dir
= DMA_FROM_DEVICE
;
557 if (user_srbcmd
->sg
.count
> ARRAY_SIZE(sg_list
)) {
558 dprintk((KERN_DEBUG
"aacraid: too many sg entries %d\n",
559 user_srbcmd
->sg
.count
));
563 if ((data_dir
== DMA_NONE
) && user_srbcmd
->sg
.count
) {
564 dprintk((KERN_DEBUG
"aacraid:SG with no direction specified\n"));
568 actual_fibsize
= sizeof(struct aac_srb
) - sizeof(struct sgentry
) +
569 ((user_srbcmd
->sg
.count
& 0xff) * sizeof(struct sgentry
));
570 actual_fibsize64
= actual_fibsize
+ (user_srbcmd
->sg
.count
& 0xff) *
571 (sizeof(struct sgentry64
) - sizeof(struct sgentry
));
572 /* User made a mistake - should not continue */
573 if ((actual_fibsize
!= fibsize
) && (actual_fibsize64
!= fibsize
)) {
574 dprintk((KERN_DEBUG
"aacraid: Bad Size specified in "
575 "Raw SRB command calculated fibsize=%lu;%lu "
576 "user_srbcmd->sg.count=%d aac_srb=%lu sgentry=%lu;%lu "
577 "issued fibsize=%d\n",
578 actual_fibsize
, actual_fibsize64
, user_srbcmd
->sg
.count
,
579 sizeof(struct aac_srb
), sizeof(struct sgentry
),
580 sizeof(struct sgentry64
), fibsize
));
585 chn
= user_srbcmd
->channel
;
586 if (chn
< AAC_MAX_BUSES
&& user_srbcmd
->id
< AAC_MAX_TARGETS
&&
587 dev
->hba_map
[chn
][user_srbcmd
->id
].devtype
==
588 AAC_DEVTYPE_NATIVE_RAW
) {
589 is_native_device
= 1;
590 hbacmd
= (struct aac_hba_cmd_req
*)srbfib
->hw_fib_va
;
591 memset(hbacmd
, 0, 96); /* sizeof(*hbacmd) is not necessary */
593 /* iu_type is a parameter of aac_hba_send */
598 case DMA_FROM_DEVICE
:
599 case DMA_BIDIRECTIONAL
:
606 hbacmd
->lun
[1] = cpu_to_le32(user_srbcmd
->lun
);
607 hbacmd
->it_nexus
= dev
->hba_map
[chn
][user_srbcmd
->id
].rmw_nexus
;
610 * we fill in reply_qid later in aac_src_deliver_message
611 * we fill in iu_type, request_id later in aac_hba_send
612 * we fill in emb_data_desc_count, data_length later
616 memcpy(hbacmd
->cdb
, user_srbcmd
->cdb
, sizeof(hbacmd
->cdb
));
618 address
= (u64
)srbfib
->hw_error_pa
;
619 hbacmd
->error_ptr_hi
= cpu_to_le32((u32
)(address
>> 32));
620 hbacmd
->error_ptr_lo
= cpu_to_le32((u32
)(address
& 0xffffffff));
621 hbacmd
->error_length
= cpu_to_le32(FW_ERROR_BUFFER_SIZE
);
622 hbacmd
->emb_data_desc_count
=
623 cpu_to_le32(user_srbcmd
->sg
.count
);
624 srbfib
->hbacmd_size
= 64 +
625 user_srbcmd
->sg
.count
* sizeof(struct aac_hba_sgl
);
628 is_native_device
= 0;
629 aac_fib_init(srbfib
);
631 /* raw_srb FIB is not FastResponseCapable */
632 srbfib
->hw_fib_va
->header
.XferState
&=
633 ~cpu_to_le32(FastResponseCapable
);
635 srbcmd
= (struct aac_srb
*) fib_data(srbfib
);
637 // Fix up srb for endian and force some values
639 srbcmd
->function
= cpu_to_le32(SRBF_ExecuteScsi
); // Force this
640 srbcmd
->channel
= cpu_to_le32(user_srbcmd
->channel
);
641 srbcmd
->id
= cpu_to_le32(user_srbcmd
->id
);
642 srbcmd
->lun
= cpu_to_le32(user_srbcmd
->lun
);
643 srbcmd
->timeout
= cpu_to_le32(user_srbcmd
->timeout
);
644 srbcmd
->flags
= cpu_to_le32(flags
);
645 srbcmd
->retry_limit
= 0; // Obsolete parameter
646 srbcmd
->cdb_size
= cpu_to_le32(user_srbcmd
->cdb_size
);
647 memcpy(srbcmd
->cdb
, user_srbcmd
->cdb
, sizeof(srbcmd
->cdb
));
651 if (is_native_device
) {
652 struct user_sgmap
*usg32
= &user_srbcmd
->sg
;
653 struct user_sgmap64
*usg64
=
654 (struct user_sgmap64
*)&user_srbcmd
->sg
;
656 for (i
= 0; i
< usg32
->count
; i
++) {
660 sg_count
[i
] = (actual_fibsize64
== fibsize
) ?
661 usg64
->sg
[i
].count
: usg32
->sg
[i
].count
;
663 (dev
->scsi_host_ptr
->max_sectors
<< 9)) {
664 pr_err("aacraid: upsg->sg[%d].count=%u>%u\n",
666 dev
->scsi_host_ptr
->max_sectors
<< 9);
671 p
= kmalloc(sg_count
[i
], GFP_KERNEL
);
677 if (actual_fibsize64
== fibsize
) {
678 addr
= (u64
)usg64
->sg
[i
].addr
[0];
679 addr
+= ((u64
)usg64
->sg
[i
].addr
[1]) << 32;
681 addr
= (u64
)usg32
->sg
[i
].addr
;
684 sg_user
[i
] = (void __user
*)(uintptr_t)addr
;
685 sg_list
[i
] = p
; // save so we can clean up later
688 if (flags
& SRB_DataOut
) {
689 if (copy_from_user(p
, sg_user
[i
],
695 addr
= pci_map_single(dev
->pdev
, p
, sg_count
[i
],
697 hbacmd
->sge
[i
].addr_hi
= cpu_to_le32((u32
)(addr
>>32));
698 hbacmd
->sge
[i
].addr_lo
= cpu_to_le32(
699 (u32
)(addr
& 0xffffffff));
700 hbacmd
->sge
[i
].len
= cpu_to_le32(sg_count
[i
]);
701 hbacmd
->sge
[i
].flags
= 0;
702 byte_count
+= sg_count
[i
];
705 if (usg32
->count
> 0) /* embedded sglist */
706 hbacmd
->sge
[usg32
->count
-1].flags
=
707 cpu_to_le32(0x40000000);
708 hbacmd
->data_length
= cpu_to_le32(byte_count
);
710 status
= aac_hba_send(HBA_IU_TYPE_SCSI_CMD_REQ
, srbfib
,
713 } else if (dev
->adapter_info
.options
& AAC_OPT_SGMAP_HOST64
) {
714 struct user_sgmap64
* upsg
= (struct user_sgmap64
*)&user_srbcmd
->sg
;
715 struct sgmap64
* psg
= (struct sgmap64
*)&srbcmd
->sg
;
718 * This should also catch if user used the 32 bit sgmap
720 if (actual_fibsize64
== fibsize
) {
721 actual_fibsize
= actual_fibsize64
;
722 for (i
= 0; i
< upsg
->count
; i
++) {
726 sg_count
[i
] = upsg
->sg
[i
].count
;
728 ((dev
->adapter_info
.options
&
730 (dev
->scsi_host_ptr
->max_sectors
<< 9) :
736 p
= kmalloc(sg_count
[i
], GFP_KERNEL
);
738 dprintk((KERN_DEBUG
"aacraid: Could not allocate SG buffer - size = %d buffer number %d of %d\n",
739 sg_count
[i
], i
, upsg
->count
));
743 addr
= (u64
)upsg
->sg
[i
].addr
[0];
744 addr
+= ((u64
)upsg
->sg
[i
].addr
[1]) << 32;
745 sg_user
[i
] = (void __user
*)(uintptr_t)addr
;
746 sg_list
[i
] = p
; // save so we can clean up later
749 if (flags
& SRB_DataOut
) {
750 if (copy_from_user(p
, sg_user
[i
],
752 dprintk((KERN_DEBUG
"aacraid: Could not copy sg data from user\n"));
757 addr
= pci_map_single(dev
->pdev
, p
,
758 sg_count
[i
], data_dir
);
760 psg
->sg
[i
].addr
[0] = cpu_to_le32(addr
& 0xffffffff);
761 psg
->sg
[i
].addr
[1] = cpu_to_le32(addr
>>32);
762 byte_count
+= sg_count
[i
];
763 psg
->sg
[i
].count
= cpu_to_le32(sg_count
[i
]);
766 struct user_sgmap
* usg
;
768 actual_fibsize
- sizeof(struct aac_srb
)
769 + sizeof(struct sgmap
), GFP_KERNEL
);
771 dprintk((KERN_DEBUG
"aacraid: Allocation error in Raw SRB command\n"));
775 actual_fibsize
= actual_fibsize64
;
777 for (i
= 0; i
< usg
->count
; i
++) {
781 sg_count
[i
] = usg
->sg
[i
].count
;
783 ((dev
->adapter_info
.options
&
785 (dev
->scsi_host_ptr
->max_sectors
<< 9) :
792 p
= kmalloc(sg_count
[i
], GFP_KERNEL
);
794 dprintk((KERN_DEBUG
"aacraid: Could not allocate SG buffer - size = %d buffer number %d of %d\n",
795 sg_count
[i
], i
, usg
->count
));
800 sg_user
[i
] = (void __user
*)(uintptr_t)usg
->sg
[i
].addr
;
801 sg_list
[i
] = p
; // save so we can clean up later
804 if (flags
& SRB_DataOut
) {
805 if (copy_from_user(p
, sg_user
[i
],
808 dprintk((KERN_DEBUG
"aacraid: Could not copy sg data from user\n"));
813 addr
= pci_map_single(dev
->pdev
, p
,
814 sg_count
[i
], data_dir
);
816 psg
->sg
[i
].addr
[0] = cpu_to_le32(addr
& 0xffffffff);
817 psg
->sg
[i
].addr
[1] = cpu_to_le32(addr
>>32);
818 byte_count
+= sg_count
[i
];
819 psg
->sg
[i
].count
= cpu_to_le32(sg_count
[i
]);
823 srbcmd
->count
= cpu_to_le32(byte_count
);
824 if (user_srbcmd
->sg
.count
)
825 psg
->count
= cpu_to_le32(sg_indx
+1);
828 status
= aac_fib_send(ScsiPortCommand64
, srbfib
, actual_fibsize
, FsaNormal
, 1, 1,NULL
,NULL
);
830 struct user_sgmap
* upsg
= &user_srbcmd
->sg
;
831 struct sgmap
* psg
= &srbcmd
->sg
;
833 if (actual_fibsize64
== fibsize
) {
834 struct user_sgmap64
* usg
= (struct user_sgmap64
*)upsg
;
835 for (i
= 0; i
< upsg
->count
; i
++) {
839 sg_count
[i
] = usg
->sg
[i
].count
;
841 ((dev
->adapter_info
.options
&
843 (dev
->scsi_host_ptr
->max_sectors
<< 9) :
848 p
= kmalloc(sg_count
[i
], GFP_KERNEL
);
850 dprintk((KERN_DEBUG
"aacraid: Could not allocate SG buffer - size = %d buffer number %d of %d\n",
851 sg_count
[i
], i
, usg
->count
));
855 addr
= (u64
)usg
->sg
[i
].addr
[0];
856 addr
+= ((u64
)usg
->sg
[i
].addr
[1]) << 32;
857 sg_user
[i
] = (void __user
*)addr
;
858 sg_list
[i
] = p
; // save so we can clean up later
861 if (flags
& SRB_DataOut
) {
862 if (copy_from_user(p
, sg_user
[i
],
864 dprintk((KERN_DEBUG
"aacraid: Could not copy sg data from user\n"));
869 addr
= pci_map_single(dev
->pdev
, p
, usg
->sg
[i
].count
, data_dir
);
871 psg
->sg
[i
].addr
= cpu_to_le32(addr
& 0xffffffff);
872 byte_count
+= usg
->sg
[i
].count
;
873 psg
->sg
[i
].count
= cpu_to_le32(sg_count
[i
]);
876 for (i
= 0; i
< upsg
->count
; i
++) {
880 sg_count
[i
] = upsg
->sg
[i
].count
;
882 ((dev
->adapter_info
.options
&
884 (dev
->scsi_host_ptr
->max_sectors
<< 9) :
889 p
= kmalloc(sg_count
[i
], GFP_KERNEL
);
891 dprintk((KERN_DEBUG
"aacraid: Could not allocate SG buffer - size = %d buffer number %d of %d\n",
892 sg_count
[i
], i
, upsg
->count
));
896 sg_user
[i
] = (void __user
*)(uintptr_t)upsg
->sg
[i
].addr
;
897 sg_list
[i
] = p
; // save so we can clean up later
900 if (flags
& SRB_DataOut
) {
901 if (copy_from_user(p
, sg_user
[i
],
903 dprintk((KERN_DEBUG
"aacraid: Could not copy sg data from user\n"));
908 addr
= pci_map_single(dev
->pdev
, p
,
909 sg_count
[i
], data_dir
);
911 psg
->sg
[i
].addr
= cpu_to_le32(addr
);
912 byte_count
+= sg_count
[i
];
913 psg
->sg
[i
].count
= cpu_to_le32(sg_count
[i
]);
916 srbcmd
->count
= cpu_to_le32(byte_count
);
917 if (user_srbcmd
->sg
.count
)
918 psg
->count
= cpu_to_le32(sg_indx
+1);
921 status
= aac_fib_send(ScsiPortCommand
, srbfib
, actual_fibsize
, FsaNormal
, 1, 1, NULL
, NULL
);
924 if (status
== -ERESTARTSYS
) {
925 rcode
= -ERESTARTSYS
;
930 dprintk((KERN_DEBUG
"aacraid: Could not send raw srb fib to hba\n"));
935 if (flags
& SRB_DataIn
) {
936 for(i
= 0 ; i
<= sg_indx
; i
++){
937 if (copy_to_user(sg_user
[i
], sg_list
[i
], sg_count
[i
])) {
938 dprintk((KERN_DEBUG
"aacraid: Could not copy sg data to user\n"));
946 user_reply
= arg
+ fibsize
;
947 if (is_native_device
) {
948 struct aac_hba_resp
*err
=
949 &((struct aac_native_hba
*)srbfib
->hw_fib_va
)->resp
.err
;
950 struct aac_srb_reply reply
;
952 memset(&reply
, 0, sizeof(reply
));
953 reply
.status
= ST_OK
;
954 if (srbfib
->flags
& FIB_CONTEXT_FLAG_FASTRESP
) {
956 reply
.srb_status
= SRB_STATUS_SUCCESS
;
957 reply
.scsi_status
= 0;
958 reply
.data_xfer_length
= byte_count
;
959 reply
.sense_data_size
= 0;
960 memset(reply
.sense_data
, 0, AAC_SENSE_BUFFERSIZE
);
962 reply
.srb_status
= err
->service_response
;
963 reply
.scsi_status
= err
->status
;
964 reply
.data_xfer_length
= byte_count
-
965 le32_to_cpu(err
->residual_count
);
966 reply
.sense_data_size
= err
->sense_response_data_len
;
967 memcpy(reply
.sense_data
, err
->sense_response_buf
,
968 AAC_SENSE_BUFFERSIZE
);
970 if (copy_to_user(user_reply
, &reply
,
971 sizeof(struct aac_srb_reply
))) {
972 dprintk((KERN_DEBUG
"aacraid: Copy to user failed\n"));
977 struct aac_srb_reply
*reply
;
979 reply
= (struct aac_srb_reply
*) fib_data(srbfib
);
980 if (copy_to_user(user_reply
, reply
,
981 sizeof(struct aac_srb_reply
))) {
982 dprintk((KERN_DEBUG
"aacraid: Copy to user failed\n"));
990 if (rcode
!= -ERESTARTSYS
) {
991 for (i
= 0; i
<= sg_indx
; i
++)
993 aac_fib_complete(srbfib
);
994 aac_fib_free(srbfib
);
1000 struct aac_pci_info
{
1006 static int aac_get_pci_info(struct aac_dev
* dev
, void __user
*arg
)
1008 struct aac_pci_info pci_info
;
1010 pci_info
.bus
= dev
->pdev
->bus
->number
;
1011 pci_info
.slot
= PCI_SLOT(dev
->pdev
->devfn
);
1013 if (copy_to_user(arg
, &pci_info
, sizeof(struct aac_pci_info
))) {
1014 dprintk((KERN_DEBUG
"aacraid: Could not copy pci info\n"));
1020 static int aac_get_hba_info(struct aac_dev
*dev
, void __user
*arg
)
1022 struct aac_hba_info hbainfo
;
1024 memset(&hbainfo
, 0, sizeof(hbainfo
));
1025 hbainfo
.adapter_number
= (u8
) dev
->id
;
1026 hbainfo
.system_io_bus_number
= dev
->pdev
->bus
->number
;
1027 hbainfo
.device_number
= (dev
->pdev
->devfn
>> 3);
1028 hbainfo
.function_number
= (dev
->pdev
->devfn
& 0x0007);
1030 hbainfo
.vendor_id
= dev
->pdev
->vendor
;
1031 hbainfo
.device_id
= dev
->pdev
->device
;
1032 hbainfo
.sub_vendor_id
= dev
->pdev
->subsystem_vendor
;
1033 hbainfo
.sub_system_id
= dev
->pdev
->subsystem_device
;
1035 if (copy_to_user(arg
, &hbainfo
, sizeof(struct aac_hba_info
))) {
1036 dprintk((KERN_DEBUG
"aacraid: Could not copy hba info\n"));
1043 struct aac_reset_iop
{
1047 static int aac_send_reset_adapter(struct aac_dev
*dev
, void __user
*arg
)
1049 struct aac_reset_iop reset
;
1052 if (copy_from_user((void *)&reset
, arg
, sizeof(struct aac_reset_iop
)))
1055 dev
->adapter_shutdown
= 1;
1057 mutex_unlock(&dev
->ioctl_mutex
);
1058 retval
= aac_reset_adapter(dev
, 0, reset
.reset_type
);
1059 mutex_lock(&dev
->ioctl_mutex
);
1064 int aac_do_ioctl(struct aac_dev
* dev
, int cmd
, void __user
*arg
)
1068 mutex_lock(&dev
->ioctl_mutex
);
1070 if (dev
->adapter_shutdown
) {
1076 * HBA gets first crack
1079 status
= aac_dev_ioctl(dev
, cmd
, arg
);
1080 if (status
!= -ENOTTY
)
1084 case FSACTL_MINIPORT_REV_CHECK
:
1085 status
= check_revision(dev
, arg
);
1087 case FSACTL_SEND_LARGE_FIB
:
1088 case FSACTL_SENDFIB
:
1089 status
= ioctl_send_fib(dev
, arg
);
1091 case FSACTL_OPEN_GET_ADAPTER_FIB
:
1092 status
= open_getadapter_fib(dev
, arg
);
1094 case FSACTL_GET_NEXT_ADAPTER_FIB
:
1095 status
= next_getadapter_fib(dev
, arg
);
1097 case FSACTL_CLOSE_GET_ADAPTER_FIB
:
1098 status
= close_getadapter_fib(dev
, arg
);
1100 case FSACTL_SEND_RAW_SRB
:
1101 status
= aac_send_raw_srb(dev
,arg
);
1103 case FSACTL_GET_PCI_INFO
:
1104 status
= aac_get_pci_info(dev
,arg
);
1106 case FSACTL_GET_HBA_INFO
:
1107 status
= aac_get_hba_info(dev
, arg
);
1109 case FSACTL_RESET_IOP
:
1110 status
= aac_send_reset_adapter(dev
, arg
);
1119 mutex_unlock(&dev
->ioctl_mutex
);