| /* |
| * qdio queue initialization |
| * |
| * Copyright IBM Corp. 2008 |
| * Author(s): Jan Glauber <jang@linux.vnet.ibm.com> |
| */ |
| #include <linux/kernel.h> |
| #include <linux/slab.h> |
| #include <linux/export.h> |
| #include <asm/qdio.h> |
| |
| #include "cio.h" |
| #include "css.h" |
| #include "device.h" |
| #include "ioasm.h" |
| #include "chsc.h" |
| #include "qdio.h" |
| #include "qdio_debug.h" |
| |
| static struct kmem_cache *qdio_q_cache; |
| static struct kmem_cache *qdio_aob_cache; |
| |
| struct qaob *qdio_allocate_aob(void) |
| { |
| return kmem_cache_zalloc(qdio_aob_cache, GFP_ATOMIC); |
| } |
| EXPORT_SYMBOL_GPL(qdio_allocate_aob); |
| |
| void qdio_release_aob(struct qaob *aob) |
| { |
| kmem_cache_free(qdio_aob_cache, aob); |
| } |
| EXPORT_SYMBOL_GPL(qdio_release_aob); |
| |
| /* |
| * qebsm is only available under 64bit but the adapter sets the feature |
| * flag anyway, so we manually override it. |
| */ |
| static inline int qebsm_possible(void) |
| { |
| #ifdef CONFIG_64BIT |
| return css_general_characteristics.qebsm; |
| #endif |
| return 0; |
| } |
| |
| /* |
| * qib_param_field: pointer to 128 bytes or NULL, if no param field |
| * nr_input_qs: pointer to nr_queues*128 words of data or NULL |
| */ |
| static void set_impl_params(struct qdio_irq *irq_ptr, |
| unsigned int qib_param_field_format, |
| unsigned char *qib_param_field, |
| unsigned long *input_slib_elements, |
| unsigned long *output_slib_elements) |
| { |
| struct qdio_q *q; |
| int i, j; |
| |
| if (!irq_ptr) |
| return; |
| |
| irq_ptr->qib.pfmt = qib_param_field_format; |
| if (qib_param_field) |
| memcpy(irq_ptr->qib.parm, qib_param_field, |
| QDIO_MAX_BUFFERS_PER_Q); |
| |
| if (!input_slib_elements) |
| goto output; |
| |
| for_each_input_queue(irq_ptr, q, i) { |
| for (j = 0; j < QDIO_MAX_BUFFERS_PER_Q; j++) |
| q->slib->slibe[j].parms = |
| input_slib_elements[i * QDIO_MAX_BUFFERS_PER_Q + j]; |
| } |
| output: |
| if (!output_slib_elements) |
| return; |
| |
| for_each_output_queue(irq_ptr, q, i) { |
| for (j = 0; j < QDIO_MAX_BUFFERS_PER_Q; j++) |
| q->slib->slibe[j].parms = |
| output_slib_elements[i * QDIO_MAX_BUFFERS_PER_Q + j]; |
| } |
| } |
| |
| static int __qdio_allocate_qs(struct qdio_q **irq_ptr_qs, int nr_queues) |
| { |
| struct qdio_q *q; |
| int i; |
| |
| for (i = 0; i < nr_queues; i++) { |
| q = kmem_cache_alloc(qdio_q_cache, GFP_KERNEL); |
| if (!q) |
| return -ENOMEM; |
| |
| q->slib = (struct slib *) __get_free_page(GFP_KERNEL); |
| if (!q->slib) { |
| kmem_cache_free(qdio_q_cache, q); |
| return -ENOMEM; |
| } |
| irq_ptr_qs[i] = q; |
| } |
| return 0; |
| } |
| |
| int qdio_allocate_qs(struct qdio_irq *irq_ptr, int nr_input_qs, int nr_output_qs) |
| { |
| int rc; |
| |
| rc = __qdio_allocate_qs(irq_ptr->input_qs, nr_input_qs); |
| if (rc) |
| return rc; |
| rc = __qdio_allocate_qs(irq_ptr->output_qs, nr_output_qs); |
| return rc; |
| } |
| |
| static void setup_queues_misc(struct qdio_q *q, struct qdio_irq *irq_ptr, |
| qdio_handler_t *handler, int i) |
| { |
| struct slib *slib = q->slib; |
| |
| /* queue must be cleared for qdio_establish */ |
| memset(q, 0, sizeof(*q)); |
| memset(slib, 0, PAGE_SIZE); |
| q->slib = slib; |
| q->irq_ptr = irq_ptr; |
| q->mask = 1 << (31 - i); |
| q->nr = i; |
| q->handler = handler; |
| } |
| |
| static void setup_storage_lists(struct qdio_q *q, struct qdio_irq *irq_ptr, |
| void **sbals_array, int i) |
| { |
| struct qdio_q *prev; |
| int j; |
| |
| DBF_HEX(&q, sizeof(void *)); |
| q->sl = (struct sl *)((char *)q->slib + PAGE_SIZE / 2); |
| |
| /* fill in sbal */ |
| for (j = 0; j < QDIO_MAX_BUFFERS_PER_Q; j++) |
| q->sbal[j] = *sbals_array++; |
| |
| /* fill in slib */ |
| if (i > 0) { |
| prev = (q->is_input_q) ? irq_ptr->input_qs[i - 1] |
| : irq_ptr->output_qs[i - 1]; |
| prev->slib->nsliba = (unsigned long)q->slib; |
| } |
| |
| q->slib->sla = (unsigned long)q->sl; |
| q->slib->slsba = (unsigned long)&q->slsb.val[0]; |
| |
| /* fill in sl */ |
| for (j = 0; j < QDIO_MAX_BUFFERS_PER_Q; j++) |
| q->sl->element[j].sbal = (unsigned long)q->sbal[j]; |
| } |
| |
| static void setup_queues(struct qdio_irq *irq_ptr, |
| struct qdio_initialize *qdio_init) |
| { |
| struct qdio_q *q; |
| void **input_sbal_array = qdio_init->input_sbal_addr_array; |
| void **output_sbal_array = qdio_init->output_sbal_addr_array; |
| struct qdio_outbuf_state *output_sbal_state_array = |
| qdio_init->output_sbal_state_array; |
| int i; |
| |
| for_each_input_queue(irq_ptr, q, i) { |
| DBF_EVENT("inq:%1d", i); |
| setup_queues_misc(q, irq_ptr, qdio_init->input_handler, i); |
| |
| q->is_input_q = 1; |
| q->u.in.queue_start_poll = qdio_init->queue_start_poll_array ? |
| qdio_init->queue_start_poll_array[i] : NULL; |
| |
| setup_storage_lists(q, irq_ptr, input_sbal_array, i); |
| input_sbal_array += QDIO_MAX_BUFFERS_PER_Q; |
| |
| if (is_thinint_irq(irq_ptr)) { |
| tasklet_init(&q->tasklet, tiqdio_inbound_processing, |
| (unsigned long) q); |
| } else { |
| tasklet_init(&q->tasklet, qdio_inbound_processing, |
| (unsigned long) q); |
| } |
| } |
| |
| for_each_output_queue(irq_ptr, q, i) { |
| DBF_EVENT("outq:%1d", i); |
| setup_queues_misc(q, irq_ptr, qdio_init->output_handler, i); |
| |
| q->u.out.sbal_state = output_sbal_state_array; |
| output_sbal_state_array += QDIO_MAX_BUFFERS_PER_Q; |
| |
| q->is_input_q = 0; |
| q->u.out.scan_threshold = qdio_init->scan_threshold; |
| setup_storage_lists(q, irq_ptr, output_sbal_array, i); |
| output_sbal_array += QDIO_MAX_BUFFERS_PER_Q; |
| |
| tasklet_init(&q->tasklet, qdio_outbound_processing, |
| (unsigned long) q); |
| setup_timer(&q->u.out.timer, (void(*)(unsigned long)) |
| &qdio_outbound_timer, (unsigned long)q); |
| } |
| } |
| |
| static void process_ac_flags(struct qdio_irq *irq_ptr, unsigned char qdioac) |
| { |
| if (qdioac & AC1_SIGA_INPUT_NEEDED) |
| irq_ptr->siga_flag.input = 1; |
| if (qdioac & AC1_SIGA_OUTPUT_NEEDED) |
| irq_ptr->siga_flag.output = 1; |
| if (qdioac & AC1_SIGA_SYNC_NEEDED) |
| irq_ptr->siga_flag.sync = 1; |
| if (!(qdioac & AC1_AUTOMATIC_SYNC_ON_THININT)) |
| irq_ptr->siga_flag.sync_after_ai = 1; |
| if (!(qdioac & AC1_AUTOMATIC_SYNC_ON_OUT_PCI)) |
| irq_ptr->siga_flag.sync_out_after_pci = 1; |
| } |
| |
| static void check_and_setup_qebsm(struct qdio_irq *irq_ptr, |
| unsigned char qdioac, unsigned long token) |
| { |
| if (!(irq_ptr->qib.rflags & QIB_RFLAGS_ENABLE_QEBSM)) |
| goto no_qebsm; |
| if (!(qdioac & AC1_SC_QEBSM_AVAILABLE) || |
| (!(qdioac & AC1_SC_QEBSM_ENABLED))) |
| goto no_qebsm; |
| |
| irq_ptr->sch_token = token; |
| |
| DBF_EVENT("V=V:1"); |
| DBF_EVENT("%8lx", irq_ptr->sch_token); |
| return; |
| |
| no_qebsm: |
| irq_ptr->sch_token = 0; |
| irq_ptr->qib.rflags &= ~QIB_RFLAGS_ENABLE_QEBSM; |
| DBF_EVENT("noV=V"); |
| } |
| |
| /* |
| * If there is a qdio_irq we use the chsc_page and store the information |
| * in the qdio_irq, otherwise we copy it to the specified structure. |
| */ |
| int qdio_setup_get_ssqd(struct qdio_irq *irq_ptr, |
| struct subchannel_id *schid, |
| struct qdio_ssqd_desc *data) |
| { |
| struct chsc_ssqd_area *ssqd; |
| int rc; |
| |
| DBF_EVENT("getssqd:%4x", schid->sch_no); |
| if (irq_ptr != NULL) |
| ssqd = (struct chsc_ssqd_area *)irq_ptr->chsc_page; |
| else |
| ssqd = (struct chsc_ssqd_area *)__get_free_page(GFP_KERNEL); |
| memset(ssqd, 0, PAGE_SIZE); |
| |
| ssqd->request = (struct chsc_header) { |
| .length = 0x0010, |
| .code = 0x0024, |
| }; |
| ssqd->first_sch = schid->sch_no; |
| ssqd->last_sch = schid->sch_no; |
| ssqd->ssid = schid->ssid; |
| |
| if (chsc(ssqd)) |
| return -EIO; |
| rc = chsc_error_from_response(ssqd->response.code); |
| if (rc) |
| return rc; |
| |
| if (!(ssqd->qdio_ssqd.flags & CHSC_FLAG_QDIO_CAPABILITY) || |
| !(ssqd->qdio_ssqd.flags & CHSC_FLAG_VALIDITY) || |
| (ssqd->qdio_ssqd.sch != schid->sch_no)) |
| return -EINVAL; |
| |
| if (irq_ptr != NULL) |
| memcpy(&irq_ptr->ssqd_desc, &ssqd->qdio_ssqd, |
| sizeof(struct qdio_ssqd_desc)); |
| else { |
| memcpy(data, &ssqd->qdio_ssqd, |
| sizeof(struct qdio_ssqd_desc)); |
| free_page((unsigned long)ssqd); |
| } |
| return 0; |
| } |
| |
| void qdio_setup_ssqd_info(struct qdio_irq *irq_ptr) |
| { |
| unsigned char qdioac; |
| int rc; |
| |
| rc = qdio_setup_get_ssqd(irq_ptr, &irq_ptr->schid, NULL); |
| if (rc) { |
| DBF_ERROR("%4x ssqd ERR", irq_ptr->schid.sch_no); |
| DBF_ERROR("rc:%x", rc); |
| /* all flags set, worst case */ |
| qdioac = AC1_SIGA_INPUT_NEEDED | AC1_SIGA_OUTPUT_NEEDED | |
| AC1_SIGA_SYNC_NEEDED; |
| } else |
| qdioac = irq_ptr->ssqd_desc.qdioac1; |
| |
| check_and_setup_qebsm(irq_ptr, qdioac, irq_ptr->ssqd_desc.sch_token); |
| process_ac_flags(irq_ptr, qdioac); |
| DBF_EVENT("ac 1:%2x 2:%4x", qdioac, irq_ptr->ssqd_desc.qdioac2); |
| DBF_EVENT("3:%4x qib:%4x", irq_ptr->ssqd_desc.qdioac3, irq_ptr->qib.ac); |
| } |
| |
| void qdio_release_memory(struct qdio_irq *irq_ptr) |
| { |
| struct qdio_q *q; |
| int i; |
| |
| /* |
| * Must check queue array manually since irq_ptr->nr_input_queues / |
| * irq_ptr->nr_input_queues may not yet be set. |
| */ |
| for (i = 0; i < QDIO_MAX_QUEUES_PER_IRQ; i++) { |
| q = irq_ptr->input_qs[i]; |
| if (q) { |
| free_page((unsigned long) q->slib); |
| kmem_cache_free(qdio_q_cache, q); |
| } |
| } |
| for (i = 0; i < QDIO_MAX_QUEUES_PER_IRQ; i++) { |
| q = irq_ptr->output_qs[i]; |
| if (q) { |
| if (q->u.out.use_cq) { |
| int n; |
| |
| for (n = 0; n < QDIO_MAX_BUFFERS_PER_Q; ++n) { |
| struct qaob *aob = q->u.out.aobs[n]; |
| if (aob) { |
| qdio_release_aob(aob); |
| q->u.out.aobs[n] = NULL; |
| } |
| } |
| |
| qdio_disable_async_operation(&q->u.out); |
| } |
| free_page((unsigned long) q->slib); |
| kmem_cache_free(qdio_q_cache, q); |
| } |
| } |
| free_page((unsigned long) irq_ptr->qdr); |
| free_page(irq_ptr->chsc_page); |
| free_page((unsigned long) irq_ptr); |
| } |
| |
| static void __qdio_allocate_fill_qdr(struct qdio_irq *irq_ptr, |
| struct qdio_q **irq_ptr_qs, |
| int i, int nr) |
| { |
| irq_ptr->qdr->qdf0[i + nr].sliba = |
| (unsigned long)irq_ptr_qs[i]->slib; |
| |
| irq_ptr->qdr->qdf0[i + nr].sla = |
| (unsigned long)irq_ptr_qs[i]->sl; |
| |
| irq_ptr->qdr->qdf0[i + nr].slsba = |
| (unsigned long)&irq_ptr_qs[i]->slsb.val[0]; |
| |
| irq_ptr->qdr->qdf0[i + nr].akey = PAGE_DEFAULT_KEY >> 4; |
| irq_ptr->qdr->qdf0[i + nr].bkey = PAGE_DEFAULT_KEY >> 4; |
| irq_ptr->qdr->qdf0[i + nr].ckey = PAGE_DEFAULT_KEY >> 4; |
| irq_ptr->qdr->qdf0[i + nr].dkey = PAGE_DEFAULT_KEY >> 4; |
| } |
| |
| static void setup_qdr(struct qdio_irq *irq_ptr, |
| struct qdio_initialize *qdio_init) |
| { |
| int i; |
| |
| irq_ptr->qdr->qfmt = qdio_init->q_format; |
| irq_ptr->qdr->ac = qdio_init->qdr_ac; |
| irq_ptr->qdr->iqdcnt = qdio_init->no_input_qs; |
| irq_ptr->qdr->oqdcnt = qdio_init->no_output_qs; |
| irq_ptr->qdr->iqdsz = sizeof(struct qdesfmt0) / 4; /* size in words */ |
| irq_ptr->qdr->oqdsz = sizeof(struct qdesfmt0) / 4; |
| irq_ptr->qdr->qiba = (unsigned long)&irq_ptr->qib; |
| irq_ptr->qdr->qkey = PAGE_DEFAULT_KEY >> 4; |
| |
| for (i = 0; i < qdio_init->no_input_qs; i++) |
| __qdio_allocate_fill_qdr(irq_ptr, irq_ptr->input_qs, i, 0); |
| |
| for (i = 0; i < qdio_init->no_output_qs; i++) |
| __qdio_allocate_fill_qdr(irq_ptr, irq_ptr->output_qs, i, |
| qdio_init->no_input_qs); |
| } |
| |
| static void setup_qib(struct qdio_irq *irq_ptr, |
| struct qdio_initialize *init_data) |
| { |
| if (qebsm_possible()) |
| irq_ptr->qib.rflags |= QIB_RFLAGS_ENABLE_QEBSM; |
| |
| irq_ptr->qib.rflags |= init_data->qib_rflags; |
| |
| irq_ptr->qib.qfmt = init_data->q_format; |
| if (init_data->no_input_qs) |
| irq_ptr->qib.isliba = |
| (unsigned long)(irq_ptr->input_qs[0]->slib); |
| if (init_data->no_output_qs) |
| irq_ptr->qib.osliba = |
| (unsigned long)(irq_ptr->output_qs[0]->slib); |
| memcpy(irq_ptr->qib.ebcnam, init_data->adapter_name, 8); |
| } |
| |
| int qdio_setup_irq(struct qdio_initialize *init_data) |
| { |
| struct ciw *ciw; |
| struct qdio_irq *irq_ptr = init_data->cdev->private->qdio_data; |
| int rc; |
| |
| memset(&irq_ptr->qib, 0, sizeof(irq_ptr->qib)); |
| memset(&irq_ptr->siga_flag, 0, sizeof(irq_ptr->siga_flag)); |
| memset(&irq_ptr->ccw, 0, sizeof(irq_ptr->ccw)); |
| memset(&irq_ptr->ssqd_desc, 0, sizeof(irq_ptr->ssqd_desc)); |
| memset(&irq_ptr->perf_stat, 0, sizeof(irq_ptr->perf_stat)); |
| |
| irq_ptr->debugfs_dev = irq_ptr->debugfs_perf = NULL; |
| irq_ptr->sch_token = irq_ptr->state = irq_ptr->perf_stat_enabled = 0; |
| |
| /* wipes qib.ac, required by ar7063 */ |
| memset(irq_ptr->qdr, 0, sizeof(struct qdr)); |
| |
| irq_ptr->int_parm = init_data->int_parm; |
| irq_ptr->nr_input_qs = init_data->no_input_qs; |
| irq_ptr->nr_output_qs = init_data->no_output_qs; |
| irq_ptr->cdev = init_data->cdev; |
| ccw_device_get_schid(irq_ptr->cdev, &irq_ptr->schid); |
| setup_queues(irq_ptr, init_data); |
| |
| setup_qib(irq_ptr, init_data); |
| qdio_setup_thinint(irq_ptr); |
| set_impl_params(irq_ptr, init_data->qib_param_field_format, |
| init_data->qib_param_field, |
| init_data->input_slib_elements, |
| init_data->output_slib_elements); |
| |
| /* fill input and output descriptors */ |
| setup_qdr(irq_ptr, init_data); |
| |
| /* qdr, qib, sls, slsbs, slibs, sbales are filled now */ |
| |
| /* get qdio commands */ |
| ciw = ccw_device_get_ciw(init_data->cdev, CIW_TYPE_EQUEUE); |
| if (!ciw) { |
| DBF_ERROR("%4x NO EQ", irq_ptr->schid.sch_no); |
| rc = -EINVAL; |
| goto out_err; |
| } |
| irq_ptr->equeue = *ciw; |
| |
| ciw = ccw_device_get_ciw(init_data->cdev, CIW_TYPE_AQUEUE); |
| if (!ciw) { |
| DBF_ERROR("%4x NO AQ", irq_ptr->schid.sch_no); |
| rc = -EINVAL; |
| goto out_err; |
| } |
| irq_ptr->aqueue = *ciw; |
| |
| /* set new interrupt handler */ |
| irq_ptr->orig_handler = init_data->cdev->handler; |
| init_data->cdev->handler = qdio_int_handler; |
| return 0; |
| out_err: |
| qdio_release_memory(irq_ptr); |
| return rc; |
| } |
| |
| void qdio_print_subchannel_info(struct qdio_irq *irq_ptr, |
| struct ccw_device *cdev) |
| { |
| char s[80]; |
| |
| snprintf(s, 80, "qdio: %s %s on SC %x using " |
| "AI:%d QEBSM:%d PCI:%d TDD:%d SIGA:%s%s%s%s%s\n", |
| dev_name(&cdev->dev), |
| (irq_ptr->qib.qfmt == QDIO_QETH_QFMT) ? "OSA" : |
| ((irq_ptr->qib.qfmt == QDIO_ZFCP_QFMT) ? "ZFCP" : "HS"), |
| irq_ptr->schid.sch_no, |
| is_thinint_irq(irq_ptr), |
| (irq_ptr->sch_token) ? 1 : 0, |
| (irq_ptr->qib.ac & QIB_AC_OUTBOUND_PCI_SUPPORTED) ? 1 : 0, |
| css_general_characteristics.aif_tdd, |
| (irq_ptr->siga_flag.input) ? "R" : " ", |
| (irq_ptr->siga_flag.output) ? "W" : " ", |
| (irq_ptr->siga_flag.sync) ? "S" : " ", |
| (irq_ptr->siga_flag.sync_after_ai) ? "A" : " ", |
| (irq_ptr->siga_flag.sync_out_after_pci) ? "P" : " "); |
| printk(KERN_INFO "%s", s); |
| } |
| |
| int qdio_enable_async_operation(struct qdio_output_q *outq) |
| { |
| outq->aobs = kzalloc(sizeof(struct qaob *) * QDIO_MAX_BUFFERS_PER_Q, |
| GFP_ATOMIC); |
| if (!outq->aobs) { |
| outq->use_cq = 0; |
| return -ENOMEM; |
| } |
| outq->use_cq = 1; |
| return 0; |
| } |
| |
| void qdio_disable_async_operation(struct qdio_output_q *q) |
| { |
| kfree(q->aobs); |
| q->aobs = NULL; |
| q->use_cq = 0; |
| } |
| |
| int __init qdio_setup_init(void) |
| { |
| int rc; |
| |
| qdio_q_cache = kmem_cache_create("qdio_q", sizeof(struct qdio_q), |
| 256, 0, NULL); |
| if (!qdio_q_cache) |
| return -ENOMEM; |
| |
| qdio_aob_cache = kmem_cache_create("qdio_aob", |
| sizeof(struct qaob), |
| sizeof(struct qaob), |
| 0, |
| NULL); |
| if (!qdio_aob_cache) { |
| rc = -ENOMEM; |
| goto free_qdio_q_cache; |
| } |
| |
| /* Check for OSA/FCP thin interrupts (bit 67). */ |
| DBF_EVENT("thinint:%1d", |
| (css_general_characteristics.aif_osa) ? 1 : 0); |
| |
| /* Check for QEBSM support in general (bit 58). */ |
| DBF_EVENT("cssQEBSM:%1d", (qebsm_possible()) ? 1 : 0); |
| rc = 0; |
| out: |
| return rc; |
| free_qdio_q_cache: |
| kmem_cache_destroy(qdio_q_cache); |
| goto out; |
| } |
| |
| void qdio_setup_exit(void) |
| { |
| kmem_cache_destroy(qdio_aob_cache); |
| kmem_cache_destroy(qdio_q_cache); |
| } |