| /* |
| * Adaptec AAC series RAID controller driver |
| * (c) Copyright 2001 Red Hat Inc. |
| * |
| * based on the old aacraid driver that is.. |
| * Adaptec aacraid device driver for Linux. |
| * |
| * Copyright (c) 2000-2010 Adaptec, Inc. |
| * 2010-2015 PMC-Sierra, Inc. (aacraid@pmc-sierra.com) |
| * 2016-2017 Microsemi Corp. (aacraid@microsemi.com) |
| * |
| * This program is free software; you can redistribute it and/or modify |
| * it under the terms of the GNU General Public License as published by |
| * the Free Software Foundation; either version 2, or (at your option) |
| * any later version. |
| * |
| * This program is distributed in the hope that it will be useful, |
| * but WITHOUT ANY WARRANTY; without even the implied warranty of |
| * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the |
| * GNU General Public License for more details. |
| * |
| * You should have received a copy of the GNU General Public License |
| * along with this program; see the file COPYING. If not, write to |
| * the Free Software Foundation, 675 Mass Ave, Cambridge, MA 02139, USA. |
| * |
| * Module Name: |
| * commctrl.c |
| * |
| * Abstract: Contains all routines for control of the AFA comm layer |
| * |
| */ |
| |
| #include <linux/kernel.h> |
| #include <linux/init.h> |
| #include <linux/types.h> |
| #include <linux/pci.h> |
| #include <linux/spinlock.h> |
| #include <linux/slab.h> |
| #include <linux/completion.h> |
| #include <linux/dma-mapping.h> |
| #include <linux/blkdev.h> |
| #include <linux/delay.h> /* ssleep prototype */ |
| #include <linux/kthread.h> |
| #include <linux/semaphore.h> |
| #include <linux/uaccess.h> |
| #include <scsi/scsi_host.h> |
| |
| #include "aacraid.h" |
| |
| /** |
| * ioctl_send_fib - send a FIB from userspace |
| * @dev: adapter is being processed |
| * @arg: arguments to the ioctl call |
| * |
| * This routine sends a fib to the adapter on behalf of a user level |
| * program. |
| */ |
| # define AAC_DEBUG_PREAMBLE KERN_INFO |
| # define AAC_DEBUG_POSTAMBLE |
| |
| static int ioctl_send_fib(struct aac_dev * dev, void __user *arg) |
| { |
| struct hw_fib * kfib; |
| struct fib *fibptr; |
| struct hw_fib * hw_fib = (struct hw_fib *)0; |
| dma_addr_t hw_fib_pa = (dma_addr_t)0LL; |
| unsigned int size, osize; |
| int retval; |
| |
| if (dev->in_reset) { |
| return -EBUSY; |
| } |
| fibptr = aac_fib_alloc(dev); |
| if(fibptr == NULL) { |
| return -ENOMEM; |
| } |
| |
| kfib = fibptr->hw_fib_va; |
| /* |
| * First copy in the header so that we can check the size field. |
| */ |
| if (copy_from_user((void *)kfib, arg, sizeof(struct aac_fibhdr))) { |
| aac_fib_free(fibptr); |
| return -EFAULT; |
| } |
| /* |
| * Since we copy based on the fib header size, make sure that we |
| * will not overrun the buffer when we copy the memory. Return |
| * an error if we would. |
| */ |
| osize = size = le16_to_cpu(kfib->header.Size) + |
| sizeof(struct aac_fibhdr); |
| if (size < le16_to_cpu(kfib->header.SenderSize)) |
| size = le16_to_cpu(kfib->header.SenderSize); |
| if (size > dev->max_fib_size) { |
| dma_addr_t daddr; |
| |
| if (size > 2048) { |
| retval = -EINVAL; |
| goto cleanup; |
| } |
| |
| kfib = dma_alloc_coherent(&dev->pdev->dev, size, &daddr, |
| GFP_KERNEL); |
| if (!kfib) { |
| retval = -ENOMEM; |
| goto cleanup; |
| } |
| |
| /* Highjack the hw_fib */ |
| hw_fib = fibptr->hw_fib_va; |
| hw_fib_pa = fibptr->hw_fib_pa; |
| fibptr->hw_fib_va = kfib; |
| fibptr->hw_fib_pa = daddr; |
| memset(((char *)kfib) + dev->max_fib_size, 0, size - dev->max_fib_size); |
| memcpy(kfib, hw_fib, dev->max_fib_size); |
| } |
| |
| if (copy_from_user(kfib, arg, size)) { |
| retval = -EFAULT; |
| goto cleanup; |
| } |
| |
| /* Sanity check the second copy */ |
| if ((osize != le16_to_cpu(kfib->header.Size) + |
| sizeof(struct aac_fibhdr)) |
| || (size < le16_to_cpu(kfib->header.SenderSize))) { |
| retval = -EINVAL; |
| goto cleanup; |
| } |
| |
| if (kfib->header.Command == cpu_to_le16(TakeABreakPt)) { |
| aac_adapter_interrupt(dev); |
| /* |
| * Since we didn't really send a fib, zero out the state to allow |
| * cleanup code not to assert. |
| */ |
| kfib->header.XferState = 0; |
| } else { |
| retval = aac_fib_send(le16_to_cpu(kfib->header.Command), fibptr, |
| le16_to_cpu(kfib->header.Size) , FsaNormal, |
| 1, 1, NULL, NULL); |
| if (retval) { |
| goto cleanup; |
| } |
| if (aac_fib_complete(fibptr) != 0) { |
| retval = -EINVAL; |
| goto cleanup; |
| } |
| } |
| /* |
| * Make sure that the size returned by the adapter (which includes |
| * the header) is less than or equal to the size of a fib, so we |
| * don't corrupt application data. Then copy that size to the user |
| * buffer. (Don't try to add the header information again, since it |
| * was already included by the adapter.) |
| */ |
| |
| retval = 0; |
| if (copy_to_user(arg, (void *)kfib, size)) |
| retval = -EFAULT; |
| cleanup: |
| if (hw_fib) { |
| dma_free_coherent(&dev->pdev->dev, size, kfib, |
| fibptr->hw_fib_pa); |
| fibptr->hw_fib_pa = hw_fib_pa; |
| fibptr->hw_fib_va = hw_fib; |
| } |
| if (retval != -ERESTARTSYS) |
| aac_fib_free(fibptr); |
| return retval; |
| } |
| |
| /** |
| * open_getadapter_fib - Get the next fib |
| * |
| * This routine will get the next Fib, if available, from the AdapterFibContext |
| * passed in from the user. |
| */ |
| |
| static int open_getadapter_fib(struct aac_dev * dev, void __user *arg) |
| { |
| struct aac_fib_context * fibctx; |
| int status; |
| |
| fibctx = kmalloc(sizeof(struct aac_fib_context), GFP_KERNEL); |
| if (fibctx == NULL) { |
| status = -ENOMEM; |
| } else { |
| unsigned long flags; |
| struct list_head * entry; |
| struct aac_fib_context * context; |
| |
| fibctx->type = FSAFS_NTC_GET_ADAPTER_FIB_CONTEXT; |
| fibctx->size = sizeof(struct aac_fib_context); |
| /* |
| * Yes yes, I know this could be an index, but we have a |
| * better guarantee of uniqueness for the locked loop below. |
| * Without the aid of a persistent history, this also helps |
| * reduce the chance that the opaque context would be reused. |
| */ |
| fibctx->unique = (u32)((ulong)fibctx & 0xFFFFFFFF); |
| /* |
| * Initialize the mutex used to wait for the next AIF. |
| */ |
| sema_init(&fibctx->wait_sem, 0); |
| fibctx->wait = 0; |
| /* |
| * Initialize the fibs and set the count of fibs on |
| * the list to 0. |
| */ |
| fibctx->count = 0; |
| INIT_LIST_HEAD(&fibctx->fib_list); |
| fibctx->jiffies = jiffies/HZ; |
| /* |
| * Now add this context onto the adapter's |
| * AdapterFibContext list. |
| */ |
| spin_lock_irqsave(&dev->fib_lock, flags); |
| /* Ensure that we have a unique identifier */ |
| entry = dev->fib_list.next; |
| while (entry != &dev->fib_list) { |
| context = list_entry(entry, struct aac_fib_context, next); |
| if (context->unique == fibctx->unique) { |
| /* Not unique (32 bits) */ |
| fibctx->unique++; |
| entry = dev->fib_list.next; |
| } else { |
| entry = entry->next; |
| } |
| } |
| list_add_tail(&fibctx->next, &dev->fib_list); |
| spin_unlock_irqrestore(&dev->fib_lock, flags); |
| if (copy_to_user(arg, &fibctx->unique, |
| sizeof(fibctx->unique))) { |
| status = -EFAULT; |
| } else { |
| status = 0; |
| } |
| } |
| return status; |
| } |
| |
| /** |
| * next_getadapter_fib - get the next fib |
| * @dev: adapter to use |
| * @arg: ioctl argument |
| * |
| * This routine will get the next Fib, if available, from the AdapterFibContext |
| * passed in from the user. |
| */ |
| |
| static int next_getadapter_fib(struct aac_dev * dev, void __user *arg) |
| { |
| struct fib_ioctl f; |
| struct fib *fib; |
| struct aac_fib_context *fibctx; |
| int status; |
| struct list_head * entry; |
| unsigned long flags; |
| |
| if(copy_from_user((void *)&f, arg, sizeof(struct fib_ioctl))) |
| return -EFAULT; |
| /* |
| * Verify that the HANDLE passed in was a valid AdapterFibContext |
| * |
| * Search the list of AdapterFibContext addresses on the adapter |
| * to be sure this is a valid address |
| */ |
| spin_lock_irqsave(&dev->fib_lock, flags); |
| entry = dev->fib_list.next; |
| fibctx = NULL; |
| |
| while (entry != &dev->fib_list) { |
| fibctx = list_entry(entry, struct aac_fib_context, next); |
| /* |
| * Extract the AdapterFibContext from the Input parameters. |
| */ |
| if (fibctx->unique == f.fibctx) { /* We found a winner */ |
| break; |
| } |
| entry = entry->next; |
| fibctx = NULL; |
| } |
| if (!fibctx) { |
| spin_unlock_irqrestore(&dev->fib_lock, flags); |
| dprintk ((KERN_INFO "Fib Context not found\n")); |
| return -EINVAL; |
| } |
| |
| if((fibctx->type != FSAFS_NTC_GET_ADAPTER_FIB_CONTEXT) || |
| (fibctx->size != sizeof(struct aac_fib_context))) { |
| spin_unlock_irqrestore(&dev->fib_lock, flags); |
| dprintk ((KERN_INFO "Fib Context corrupt?\n")); |
| return -EINVAL; |
| } |
| status = 0; |
| /* |
| * If there are no fibs to send back, then either wait or return |
| * -EAGAIN |
| */ |
| return_fib: |
| if (!list_empty(&fibctx->fib_list)) { |
| /* |
| * Pull the next fib from the fibs |
| */ |
| entry = fibctx->fib_list.next; |
| list_del(entry); |
| |
| fib = list_entry(entry, struct fib, fiblink); |
| fibctx->count--; |
| spin_unlock_irqrestore(&dev->fib_lock, flags); |
| if (copy_to_user(f.fib, fib->hw_fib_va, sizeof(struct hw_fib))) { |
| kfree(fib->hw_fib_va); |
| kfree(fib); |
| return -EFAULT; |
| } |
| /* |
| * Free the space occupied by this copy of the fib. |
| */ |
| kfree(fib->hw_fib_va); |
| kfree(fib); |
| status = 0; |
| } else { |
| spin_unlock_irqrestore(&dev->fib_lock, flags); |
| /* If someone killed the AIF aacraid thread, restart it */ |
| status = !dev->aif_thread; |
| if (status && !dev->in_reset && dev->queues && dev->fsa_dev) { |
| /* Be paranoid, be very paranoid! */ |
| kthread_stop(dev->thread); |
| ssleep(1); |
| dev->aif_thread = 0; |
| dev->thread = kthread_run(aac_command_thread, dev, |
| "%s", dev->name); |
| ssleep(1); |
| } |
| if (f.wait) { |
| if(down_interruptible(&fibctx->wait_sem) < 0) { |
| status = -ERESTARTSYS; |
| } else { |
| /* Lock again and retry */ |
| spin_lock_irqsave(&dev->fib_lock, flags); |
| goto return_fib; |
| } |
| } else { |
| status = -EAGAIN; |
| } |
| } |
| fibctx->jiffies = jiffies/HZ; |
| return status; |
| } |
| |
| int aac_close_fib_context(struct aac_dev * dev, struct aac_fib_context * fibctx) |
| { |
| struct fib *fib; |
| |
| /* |
| * First free any FIBs that have not been consumed. |
| */ |
| while (!list_empty(&fibctx->fib_list)) { |
| struct list_head * entry; |
| /* |
| * Pull the next fib from the fibs |
| */ |
| entry = fibctx->fib_list.next; |
| list_del(entry); |
| fib = list_entry(entry, struct fib, fiblink); |
| fibctx->count--; |
| /* |
| * Free the space occupied by this copy of the fib. |
| */ |
| kfree(fib->hw_fib_va); |
| kfree(fib); |
| } |
| /* |
| * Remove the Context from the AdapterFibContext List |
| */ |
| list_del(&fibctx->next); |
| /* |
| * Invalidate context |
| */ |
| fibctx->type = 0; |
| /* |
| * Free the space occupied by the Context |
| */ |
| kfree(fibctx); |
| return 0; |
| } |
| |
| /** |
| * close_getadapter_fib - close down user fib context |
| * @dev: adapter |
| * @arg: ioctl arguments |
| * |
| * This routine will close down the fibctx passed in from the user. |
| */ |
| |
| static int close_getadapter_fib(struct aac_dev * dev, void __user *arg) |
| { |
| struct aac_fib_context *fibctx; |
| int status; |
| unsigned long flags; |
| struct list_head * entry; |
| |
| /* |
| * Verify that the HANDLE passed in was a valid AdapterFibContext |
| * |
| * Search the list of AdapterFibContext addresses on the adapter |
| * to be sure this is a valid address |
| */ |
| |
| entry = dev->fib_list.next; |
| fibctx = NULL; |
| |
| while(entry != &dev->fib_list) { |
| fibctx = list_entry(entry, struct aac_fib_context, next); |
| /* |
| * Extract the fibctx from the input parameters |
| */ |
| if (fibctx->unique == (u32)(uintptr_t)arg) /* We found a winner */ |
| break; |
| entry = entry->next; |
| fibctx = NULL; |
| } |
| |
| if (!fibctx) |
| return 0; /* Already gone */ |
| |
| if((fibctx->type != FSAFS_NTC_GET_ADAPTER_FIB_CONTEXT) || |
| (fibctx->size != sizeof(struct aac_fib_context))) |
| return -EINVAL; |
| spin_lock_irqsave(&dev->fib_lock, flags); |
| status = aac_close_fib_context(dev, fibctx); |
| spin_unlock_irqrestore(&dev->fib_lock, flags); |
| return status; |
| } |
| |
| /** |
| * check_revision - close down user fib context |
| * @dev: adapter |
| * @arg: ioctl arguments |
| * |
| * This routine returns the driver version. |
| * Under Linux, there have been no version incompatibilities, so this is |
| * simple! |
| */ |
| |
| static int check_revision(struct aac_dev *dev, void __user *arg) |
| { |
| struct revision response; |
| char *driver_version = aac_driver_version; |
| u32 version; |
| |
| response.compat = 1; |
| version = (simple_strtol(driver_version, |
| &driver_version, 10) << 24) | 0x00000400; |
| version += simple_strtol(driver_version + 1, &driver_version, 10) << 16; |
| version += simple_strtol(driver_version + 1, NULL, 10); |
| response.version = cpu_to_le32(version); |
| # ifdef AAC_DRIVER_BUILD |
| response.build = cpu_to_le32(AAC_DRIVER_BUILD); |
| # else |
| response.build = cpu_to_le32(9999); |
| # endif |
| |
| if (copy_to_user(arg, &response, sizeof(response))) |
| return -EFAULT; |
| return 0; |
| } |
| |
| |
| /** |
| * |
| * aac_send_raw_scb |
| * |
| */ |
| |
| static int aac_send_raw_srb(struct aac_dev* dev, void __user * arg) |
| { |
| struct fib* srbfib; |
| int status; |
| struct aac_srb *srbcmd = NULL; |
| struct aac_hba_cmd_req *hbacmd = NULL; |
| struct user_aac_srb *user_srbcmd = NULL; |
| struct user_aac_srb __user *user_srb = arg; |
| struct aac_srb_reply __user *user_reply; |
| u32 chn; |
| u32 fibsize = 0; |
| u32 flags = 0; |
| s32 rcode = 0; |
| u32 data_dir; |
| void __user *sg_user[HBA_MAX_SG_EMBEDDED]; |
| void *sg_list[HBA_MAX_SG_EMBEDDED]; |
| u32 sg_count[HBA_MAX_SG_EMBEDDED]; |
| u32 sg_indx = 0; |
| u32 byte_count = 0; |
| u32 actual_fibsize64, actual_fibsize = 0; |
| int i; |
| int is_native_device; |
| u64 address; |
| |
| |
| if (dev->in_reset) { |
| dprintk((KERN_DEBUG"aacraid: send raw srb -EBUSY\n")); |
| return -EBUSY; |
| } |
| if (!capable(CAP_SYS_ADMIN)){ |
| dprintk((KERN_DEBUG"aacraid: No permission to send raw srb\n")); |
| return -EPERM; |
| } |
| /* |
| * Allocate and initialize a Fib then setup a SRB command |
| */ |
| if (!(srbfib = aac_fib_alloc(dev))) { |
| return -ENOMEM; |
| } |
| |
| memset(sg_list, 0, sizeof(sg_list)); /* cleanup may take issue */ |
| if(copy_from_user(&fibsize, &user_srb->count,sizeof(u32))){ |
| dprintk((KERN_DEBUG"aacraid: Could not copy data size from user\n")); |
| rcode = -EFAULT; |
| goto cleanup; |
| } |
| |
| if ((fibsize < (sizeof(struct user_aac_srb) - sizeof(struct user_sgentry))) || |
| (fibsize > (dev->max_fib_size - sizeof(struct aac_fibhdr)))) { |
| rcode = -EINVAL; |
| goto cleanup; |
| } |
| |
| user_srbcmd = kmalloc(fibsize, GFP_KERNEL); |
| if (!user_srbcmd) { |
| dprintk((KERN_DEBUG"aacraid: Could not make a copy of the srb\n")); |
| rcode = -ENOMEM; |
| goto cleanup; |
| } |
| if(copy_from_user(user_srbcmd, user_srb,fibsize)){ |
| dprintk((KERN_DEBUG"aacraid: Could not copy srb from user\n")); |
| rcode = -EFAULT; |
| goto cleanup; |
| } |
| |
| flags = user_srbcmd->flags; /* from user in cpu order */ |
| switch (flags & (SRB_DataIn | SRB_DataOut)) { |
| case SRB_DataOut: |
| data_dir = DMA_TO_DEVICE; |
| break; |
| case (SRB_DataIn | SRB_DataOut): |
| data_dir = DMA_BIDIRECTIONAL; |
| break; |
| case SRB_DataIn: |
| data_dir = DMA_FROM_DEVICE; |
| break; |
| default: |
| data_dir = DMA_NONE; |
| } |
| if (user_srbcmd->sg.count > ARRAY_SIZE(sg_list)) { |
| dprintk((KERN_DEBUG"aacraid: too many sg entries %d\n", |
| user_srbcmd->sg.count)); |
| rcode = -EINVAL; |
| goto cleanup; |
| } |
| if ((data_dir == DMA_NONE) && user_srbcmd->sg.count) { |
| dprintk((KERN_DEBUG"aacraid:SG with no direction specified\n")); |
| rcode = -EINVAL; |
| goto cleanup; |
| } |
| actual_fibsize = sizeof(struct aac_srb) - sizeof(struct sgentry) + |
| ((user_srbcmd->sg.count & 0xff) * sizeof(struct sgentry)); |
| actual_fibsize64 = actual_fibsize + (user_srbcmd->sg.count & 0xff) * |
| (sizeof(struct sgentry64) - sizeof(struct sgentry)); |
| /* User made a mistake - should not continue */ |
| if ((actual_fibsize != fibsize) && (actual_fibsize64 != fibsize)) { |
| dprintk((KERN_DEBUG"aacraid: Bad Size specified in " |
| "Raw SRB command calculated fibsize=%lu;%lu " |
| "user_srbcmd->sg.count=%d aac_srb=%lu sgentry=%lu;%lu " |
| "issued fibsize=%d\n", |
| actual_fibsize, actual_fibsize64, user_srbcmd->sg.count, |
| sizeof(struct aac_srb), sizeof(struct sgentry), |
| sizeof(struct sgentry64), fibsize)); |
| rcode = -EINVAL; |
| goto cleanup; |
| } |
| |
| chn = user_srbcmd->channel; |
| if (chn < AAC_MAX_BUSES && user_srbcmd->id < AAC_MAX_TARGETS && |
| dev->hba_map[chn][user_srbcmd->id].devtype == |
| AAC_DEVTYPE_NATIVE_RAW) { |
| is_native_device = 1; |
| hbacmd = (struct aac_hba_cmd_req *)srbfib->hw_fib_va; |
| memset(hbacmd, 0, 96); /* sizeof(*hbacmd) is not necessary */ |
| |
| /* iu_type is a parameter of aac_hba_send */ |
| switch (data_dir) { |
| case DMA_TO_DEVICE: |
| hbacmd->byte1 = 2; |
| break; |
| case DMA_FROM_DEVICE: |
| case DMA_BIDIRECTIONAL: |
| hbacmd->byte1 = 1; |
| break; |
| case DMA_NONE: |
| default: |
| break; |
| } |
| hbacmd->lun[1] = cpu_to_le32(user_srbcmd->lun); |
| hbacmd->it_nexus = dev->hba_map[chn][user_srbcmd->id].rmw_nexus; |
| |
| /* |
| * we fill in reply_qid later in aac_src_deliver_message |
| * we fill in iu_type, request_id later in aac_hba_send |
| * we fill in emb_data_desc_count, data_length later |
| * in sg list build |
| */ |
| |
| memcpy(hbacmd->cdb, user_srbcmd->cdb, sizeof(hbacmd->cdb)); |
| |
| address = (u64)srbfib->hw_error_pa; |
| hbacmd->error_ptr_hi = cpu_to_le32((u32)(address >> 32)); |
| hbacmd->error_ptr_lo = cpu_to_le32((u32)(address & 0xffffffff)); |
| hbacmd->error_length = cpu_to_le32(FW_ERROR_BUFFER_SIZE); |
| hbacmd->emb_data_desc_count = |
| cpu_to_le32(user_srbcmd->sg.count); |
| srbfib->hbacmd_size = 64 + |
| user_srbcmd->sg.count * sizeof(struct aac_hba_sgl); |
| |
| } else { |
| is_native_device = 0; |
| aac_fib_init(srbfib); |
| |
| /* raw_srb FIB is not FastResponseCapable */ |
| srbfib->hw_fib_va->header.XferState &= |
| ~cpu_to_le32(FastResponseCapable); |
| |
| srbcmd = (struct aac_srb *) fib_data(srbfib); |
| |
| // Fix up srb for endian and force some values |
| |
| srbcmd->function = cpu_to_le32(SRBF_ExecuteScsi); // Force this |
| srbcmd->channel = cpu_to_le32(user_srbcmd->channel); |
| srbcmd->id = cpu_to_le32(user_srbcmd->id); |
| srbcmd->lun = cpu_to_le32(user_srbcmd->lun); |
| srbcmd->timeout = cpu_to_le32(user_srbcmd->timeout); |
| srbcmd->flags = cpu_to_le32(flags); |
| srbcmd->retry_limit = 0; // Obsolete parameter |
| srbcmd->cdb_size = cpu_to_le32(user_srbcmd->cdb_size); |
| memcpy(srbcmd->cdb, user_srbcmd->cdb, sizeof(srbcmd->cdb)); |
| } |
| |
| byte_count = 0; |
| if (is_native_device) { |
| struct user_sgmap *usg32 = &user_srbcmd->sg; |
| struct user_sgmap64 *usg64 = |
| (struct user_sgmap64 *)&user_srbcmd->sg; |
| |
| for (i = 0; i < usg32->count; i++) { |
| void *p; |
| u64 addr; |
| |
| sg_count[i] = (actual_fibsize64 == fibsize) ? |
| usg64->sg[i].count : usg32->sg[i].count; |
| if (sg_count[i] > |
| (dev->scsi_host_ptr->max_sectors << 9)) { |
| pr_err("aacraid: upsg->sg[%d].count=%u>%u\n", |
| i, sg_count[i], |
| dev->scsi_host_ptr->max_sectors << 9); |
| rcode = -EINVAL; |
| goto cleanup; |
| } |
| |
| p = kmalloc(sg_count[i], GFP_KERNEL); |
| if (!p) { |
| rcode = -ENOMEM; |
| goto cleanup; |
| } |
| |
| if (actual_fibsize64 == fibsize) { |
| addr = (u64)usg64->sg[i].addr[0]; |
| addr += ((u64)usg64->sg[i].addr[1]) << 32; |
| } else { |
| addr = (u64)usg32->sg[i].addr; |
| } |
| |
| sg_user[i] = (void __user *)(uintptr_t)addr; |
| sg_list[i] = p; // save so we can clean up later |
| sg_indx = i; |
| |
| if (flags & SRB_DataOut) { |
| if (copy_from_user(p, sg_user[i], |
| sg_count[i])) { |
| rcode = -EFAULT; |
| goto cleanup; |
| } |
| } |
| addr = pci_map_single(dev->pdev, p, sg_count[i], |
| data_dir); |
| hbacmd->sge[i].addr_hi = cpu_to_le32((u32)(addr>>32)); |
| hbacmd->sge[i].addr_lo = cpu_to_le32( |
| (u32)(addr & 0xffffffff)); |
| hbacmd->sge[i].len = cpu_to_le32(sg_count[i]); |
| hbacmd->sge[i].flags = 0; |
| byte_count += sg_count[i]; |
| } |
| |
| if (usg32->count > 0) /* embedded sglist */ |
| hbacmd->sge[usg32->count-1].flags = |
| cpu_to_le32(0x40000000); |
| hbacmd->data_length = cpu_to_le32(byte_count); |
| |
| status = aac_hba_send(HBA_IU_TYPE_SCSI_CMD_REQ, srbfib, |
| NULL, NULL); |
| |
| } else if (dev->adapter_info.options & AAC_OPT_SGMAP_HOST64) { |
| struct user_sgmap64* upsg = (struct user_sgmap64*)&user_srbcmd->sg; |
| struct sgmap64* psg = (struct sgmap64*)&srbcmd->sg; |
| |
| /* |
| * This should also catch if user used the 32 bit sgmap |
| */ |
| if (actual_fibsize64 == fibsize) { |
| actual_fibsize = actual_fibsize64; |
| for (i = 0; i < upsg->count; i++) { |
| u64 addr; |
| void* p; |
| |
| sg_count[i] = upsg->sg[i].count; |
| if (sg_count[i] > |
| ((dev->adapter_info.options & |
| AAC_OPT_NEW_COMM) ? |
| (dev->scsi_host_ptr->max_sectors << 9) : |
| 65536)) { |
| rcode = -EINVAL; |
| goto cleanup; |
| } |
| |
| p = kmalloc(sg_count[i], GFP_KERNEL); |
| if(!p) { |
| dprintk((KERN_DEBUG"aacraid: Could not allocate SG buffer - size = %d buffer number %d of %d\n", |
| sg_count[i], i, upsg->count)); |
| rcode = -ENOMEM; |
| goto cleanup; |
| } |
| addr = (u64)upsg->sg[i].addr[0]; |
| addr += ((u64)upsg->sg[i].addr[1]) << 32; |
| sg_user[i] = (void __user *)(uintptr_t)addr; |
| sg_list[i] = p; // save so we can clean up later |
| sg_indx = i; |
| |
| if (flags & SRB_DataOut) { |
| if (copy_from_user(p, sg_user[i], |
| sg_count[i])){ |
| dprintk((KERN_DEBUG"aacraid: Could not copy sg data from user\n")); |
| rcode = -EFAULT; |
| goto cleanup; |
| } |
| } |
| addr = pci_map_single(dev->pdev, p, |
| sg_count[i], data_dir); |
| |
| psg->sg[i].addr[0] = cpu_to_le32(addr & 0xffffffff); |
| psg->sg[i].addr[1] = cpu_to_le32(addr>>32); |
| byte_count += sg_count[i]; |
| psg->sg[i].count = cpu_to_le32(sg_count[i]); |
| } |
| } else { |
| struct user_sgmap* usg; |
| usg = kmemdup(upsg, |
| actual_fibsize - sizeof(struct aac_srb) |
| + sizeof(struct sgmap), GFP_KERNEL); |
| if (!usg) { |
| dprintk((KERN_DEBUG"aacraid: Allocation error in Raw SRB command\n")); |
| rcode = -ENOMEM; |
| goto cleanup; |
| } |
| actual_fibsize = actual_fibsize64; |
| |
| for (i = 0; i < usg->count; i++) { |
| u64 addr; |
| void* p; |
| |
| sg_count[i] = usg->sg[i].count; |
| if (sg_count[i] > |
| ((dev->adapter_info.options & |
| AAC_OPT_NEW_COMM) ? |
| (dev->scsi_host_ptr->max_sectors << 9) : |
| 65536)) { |
| kfree(usg); |
| rcode = -EINVAL; |
| goto cleanup; |
| } |
| |
| p = kmalloc(sg_count[i], GFP_KERNEL); |
| if(!p) { |
| dprintk((KERN_DEBUG "aacraid: Could not allocate SG buffer - size = %d buffer number %d of %d\n", |
| sg_count[i], i, usg->count)); |
| kfree(usg); |
| rcode = -ENOMEM; |
| goto cleanup; |
| } |
| sg_user[i] = (void __user *)(uintptr_t)usg->sg[i].addr; |
| sg_list[i] = p; // save so we can clean up later |
| sg_indx = i; |
| |
| if (flags & SRB_DataOut) { |
| if (copy_from_user(p, sg_user[i], |
| sg_count[i])) { |
| kfree (usg); |
| dprintk((KERN_DEBUG"aacraid: Could not copy sg data from user\n")); |
| rcode = -EFAULT; |
| goto cleanup; |
| } |
| } |
| addr = pci_map_single(dev->pdev, p, |
| sg_count[i], data_dir); |
| |
| psg->sg[i].addr[0] = cpu_to_le32(addr & 0xffffffff); |
| psg->sg[i].addr[1] = cpu_to_le32(addr>>32); |
| byte_count += sg_count[i]; |
| psg->sg[i].count = cpu_to_le32(sg_count[i]); |
| } |
| kfree (usg); |
| } |
| srbcmd->count = cpu_to_le32(byte_count); |
| if (user_srbcmd->sg.count) |
| psg->count = cpu_to_le32(sg_indx+1); |
| else |
| psg->count = 0; |
| status = aac_fib_send(ScsiPortCommand64, srbfib, actual_fibsize, FsaNormal, 1, 1,NULL,NULL); |
| } else { |
| struct user_sgmap* upsg = &user_srbcmd->sg; |
| struct sgmap* psg = &srbcmd->sg; |
| |
| if (actual_fibsize64 == fibsize) { |
| struct user_sgmap64* usg = (struct user_sgmap64 *)upsg; |
| for (i = 0; i < upsg->count; i++) { |
| uintptr_t addr; |
| void* p; |
| |
| sg_count[i] = usg->sg[i].count; |
| if (sg_count[i] > |
| ((dev->adapter_info.options & |
| AAC_OPT_NEW_COMM) ? |
| (dev->scsi_host_ptr->max_sectors << 9) : |
| 65536)) { |
| rcode = -EINVAL; |
| goto cleanup; |
| } |
| p = kmalloc(sg_count[i], GFP_KERNEL|GFP_DMA32); |
| if (!p) { |
| dprintk((KERN_DEBUG"aacraid: Could not allocate SG buffer - size = %d buffer number %d of %d\n", |
| sg_count[i], i, usg->count)); |
| rcode = -ENOMEM; |
| goto cleanup; |
| } |
| addr = (u64)usg->sg[i].addr[0]; |
| addr += ((u64)usg->sg[i].addr[1]) << 32; |
| sg_user[i] = (void __user *)addr; |
| sg_list[i] = p; // save so we can clean up later |
| sg_indx = i; |
| |
| if (flags & SRB_DataOut) { |
| if (copy_from_user(p, sg_user[i], |
| sg_count[i])){ |
| dprintk((KERN_DEBUG"aacraid: Could not copy sg data from user\n")); |
| rcode = -EFAULT; |
| goto cleanup; |
| } |
| } |
| addr = pci_map_single(dev->pdev, p, usg->sg[i].count, data_dir); |
| |
| psg->sg[i].addr = cpu_to_le32(addr & 0xffffffff); |
| byte_count += usg->sg[i].count; |
| psg->sg[i].count = cpu_to_le32(sg_count[i]); |
| } |
| } else { |
| for (i = 0; i < upsg->count; i++) { |
| dma_addr_t addr; |
| void* p; |
| |
| sg_count[i] = upsg->sg[i].count; |
| if (sg_count[i] > |
| ((dev->adapter_info.options & |
| AAC_OPT_NEW_COMM) ? |
| (dev->scsi_host_ptr->max_sectors << 9) : |
| 65536)) { |
| rcode = -EINVAL; |
| goto cleanup; |
| } |
| p = kmalloc(sg_count[i], GFP_KERNEL|GFP_DMA32); |
| if (!p) { |
| dprintk((KERN_DEBUG"aacraid: Could not allocate SG buffer - size = %d buffer number %d of %d\n", |
| sg_count[i], i, upsg->count)); |
| rcode = -ENOMEM; |
| goto cleanup; |
| } |
| sg_user[i] = (void __user *)(uintptr_t)upsg->sg[i].addr; |
| sg_list[i] = p; // save so we can clean up later |
| sg_indx = i; |
| |
| if (flags & SRB_DataOut) { |
| if (copy_from_user(p, sg_user[i], |
| sg_count[i])) { |
| dprintk((KERN_DEBUG"aacraid: Could not copy sg data from user\n")); |
| rcode = -EFAULT; |
| goto cleanup; |
| } |
| } |
| addr = pci_map_single(dev->pdev, p, |
| sg_count[i], data_dir); |
| |
| psg->sg[i].addr = cpu_to_le32(addr); |
| byte_count += sg_count[i]; |
| psg->sg[i].count = cpu_to_le32(sg_count[i]); |
| } |
| } |
| srbcmd->count = cpu_to_le32(byte_count); |
| if (user_srbcmd->sg.count) |
| psg->count = cpu_to_le32(sg_indx+1); |
| else |
| psg->count = 0; |
| status = aac_fib_send(ScsiPortCommand, srbfib, actual_fibsize, FsaNormal, 1, 1, NULL, NULL); |
| } |
| |
| if (status == -ERESTARTSYS) { |
| rcode = -ERESTARTSYS; |
| goto cleanup; |
| } |
| |
| if (status != 0) { |
| dprintk((KERN_DEBUG"aacraid: Could not send raw srb fib to hba\n")); |
| rcode = -ENXIO; |
| goto cleanup; |
| } |
| |
| if (flags & SRB_DataIn) { |
| for(i = 0 ; i <= sg_indx; i++){ |
| if (copy_to_user(sg_user[i], sg_list[i], sg_count[i])) { |
| dprintk((KERN_DEBUG"aacraid: Could not copy sg data to user\n")); |
| rcode = -EFAULT; |
| goto cleanup; |
| |
| } |
| } |
| } |
| |
| user_reply = arg + fibsize; |
| if (is_native_device) { |
| struct aac_hba_resp *err = |
| &((struct aac_native_hba *)srbfib->hw_fib_va)->resp.err; |
| struct aac_srb_reply reply; |
| |
| memset(&reply, 0, sizeof(reply)); |
| reply.status = ST_OK; |
| if (srbfib->flags & FIB_CONTEXT_FLAG_FASTRESP) { |
| /* fast response */ |
| reply.srb_status = SRB_STATUS_SUCCESS; |
| reply.scsi_status = 0; |
| reply.data_xfer_length = byte_count; |
| reply.sense_data_size = 0; |
| memset(reply.sense_data, 0, AAC_SENSE_BUFFERSIZE); |
| } else { |
| reply.srb_status = err->service_response; |
| reply.scsi_status = err->status; |
| reply.data_xfer_length = byte_count - |
| le32_to_cpu(err->residual_count); |
| reply.sense_data_size = err->sense_response_data_len; |
| memcpy(reply.sense_data, err->sense_response_buf, |
| AAC_SENSE_BUFFERSIZE); |
| } |
| if (copy_to_user(user_reply, &reply, |
| sizeof(struct aac_srb_reply))) { |
| dprintk((KERN_DEBUG"aacraid: Copy to user failed\n")); |
| rcode = -EFAULT; |
| goto cleanup; |
| } |
| } else { |
| struct aac_srb_reply *reply; |
| |
| reply = (struct aac_srb_reply *) fib_data(srbfib); |
| if (copy_to_user(user_reply, reply, |
| sizeof(struct aac_srb_reply))) { |
| dprintk((KERN_DEBUG"aacraid: Copy to user failed\n")); |
| rcode = -EFAULT; |
| goto cleanup; |
| } |
| } |
| |
| cleanup: |
| kfree(user_srbcmd); |
| if (rcode != -ERESTARTSYS) { |
| for (i = 0; i <= sg_indx; i++) |
| kfree(sg_list[i]); |
| aac_fib_complete(srbfib); |
| aac_fib_free(srbfib); |
| } |
| |
| return rcode; |
| } |
| |
| struct aac_pci_info { |
| u32 bus; |
| u32 slot; |
| }; |
| |
| |
| static int aac_get_pci_info(struct aac_dev* dev, void __user *arg) |
| { |
| struct aac_pci_info pci_info; |
| |
| pci_info.bus = dev->pdev->bus->number; |
| pci_info.slot = PCI_SLOT(dev->pdev->devfn); |
| |
| if (copy_to_user(arg, &pci_info, sizeof(struct aac_pci_info))) { |
| dprintk((KERN_DEBUG "aacraid: Could not copy pci info\n")); |
| return -EFAULT; |
| } |
| return 0; |
| } |
| |
| static int aac_get_hba_info(struct aac_dev *dev, void __user *arg) |
| { |
| struct aac_hba_info hbainfo; |
| |
| memset(&hbainfo, 0, sizeof(hbainfo)); |
| hbainfo.adapter_number = (u8) dev->id; |
| hbainfo.system_io_bus_number = dev->pdev->bus->number; |
| hbainfo.device_number = (dev->pdev->devfn >> 3); |
| hbainfo.function_number = (dev->pdev->devfn & 0x0007); |
| |
| hbainfo.vendor_id = dev->pdev->vendor; |
| hbainfo.device_id = dev->pdev->device; |
| hbainfo.sub_vendor_id = dev->pdev->subsystem_vendor; |
| hbainfo.sub_system_id = dev->pdev->subsystem_device; |
| |
| if (copy_to_user(arg, &hbainfo, sizeof(struct aac_hba_info))) { |
| dprintk((KERN_DEBUG "aacraid: Could not copy hba info\n")); |
| return -EFAULT; |
| } |
| |
| return 0; |
| } |
| |
| struct aac_reset_iop { |
| u8 reset_type; |
| }; |
| |
| static int aac_send_reset_adapter(struct aac_dev *dev, void __user *arg) |
| { |
| struct aac_reset_iop reset; |
| int retval; |
| |
| if (copy_from_user((void *)&reset, arg, sizeof(struct aac_reset_iop))) |
| return -EFAULT; |
| |
| retval = aac_reset_adapter(dev, 0, reset.reset_type); |
| return retval; |
| |
| } |
| |
| int aac_do_ioctl(struct aac_dev * dev, int cmd, void __user *arg) |
| { |
| int status; |
| |
| mutex_lock(&dev->ioctl_mutex); |
| |
| if (dev->adapter_shutdown) { |
| status = -EACCES; |
| goto cleanup; |
| } |
| |
| /* |
| * HBA gets first crack |
| */ |
| |
| status = aac_dev_ioctl(dev, cmd, arg); |
| if (status != -ENOTTY) |
| goto cleanup; |
| |
| switch (cmd) { |
| case FSACTL_MINIPORT_REV_CHECK: |
| status = check_revision(dev, arg); |
| break; |
| case FSACTL_SEND_LARGE_FIB: |
| case FSACTL_SENDFIB: |
| status = ioctl_send_fib(dev, arg); |
| break; |
| case FSACTL_OPEN_GET_ADAPTER_FIB: |
| status = open_getadapter_fib(dev, arg); |
| break; |
| case FSACTL_GET_NEXT_ADAPTER_FIB: |
| status = next_getadapter_fib(dev, arg); |
| break; |
| case FSACTL_CLOSE_GET_ADAPTER_FIB: |
| status = close_getadapter_fib(dev, arg); |
| break; |
| case FSACTL_SEND_RAW_SRB: |
| status = aac_send_raw_srb(dev,arg); |
| break; |
| case FSACTL_GET_PCI_INFO: |
| status = aac_get_pci_info(dev,arg); |
| break; |
| case FSACTL_GET_HBA_INFO: |
| status = aac_get_hba_info(dev, arg); |
| break; |
| case FSACTL_RESET_IOP: |
| status = aac_send_reset_adapter(dev, arg); |
| break; |
| |
| default: |
| status = -ENOTTY; |
| break; |
| } |
| |
| cleanup: |
| mutex_unlock(&dev->ioctl_mutex); |
| |
| return status; |
| } |
| |