| /* |
| * drivers/s390/cio/cio.c |
| * S/390 common I/O routines -- low level i/o calls |
| * |
| * Copyright (C) IBM Corp. 1999,2006 |
| * Author(s): Ingo Adlung (adlung@de.ibm.com) |
| * Cornelia Huck (cornelia.huck@de.ibm.com) |
| * Arnd Bergmann (arndb@de.ibm.com) |
| * Martin Schwidefsky (schwidefsky@de.ibm.com) |
| */ |
| |
| #include <linux/module.h> |
| #include <linux/init.h> |
| #include <linux/slab.h> |
| #include <linux/device.h> |
| #include <linux/kernel_stat.h> |
| #include <linux/interrupt.h> |
| #include <asm/cio.h> |
| #include <asm/delay.h> |
| #include <asm/irq.h> |
| #include <asm/irq_regs.h> |
| #include <asm/setup.h> |
| #include <asm/reset.h> |
| #include <asm/ipl.h> |
| #include <asm/chpid.h> |
| #include <asm/airq.h> |
| #include <asm/cpu.h> |
| #include "cio.h" |
| #include "css.h" |
| #include "chsc.h" |
| #include "ioasm.h" |
| #include "io_sch.h" |
| #include "blacklist.h" |
| #include "cio_debug.h" |
| #include "chp.h" |
| #include "../s390mach.h" |
| |
| debug_info_t *cio_debug_msg_id; |
| debug_info_t *cio_debug_trace_id; |
| debug_info_t *cio_debug_crw_id; |
| |
| int cio_show_msg; |
| |
| static int __init |
| cio_setup (char *parm) |
| { |
| if (!strcmp (parm, "yes")) |
| cio_show_msg = 1; |
| else if (!strcmp (parm, "no")) |
| cio_show_msg = 0; |
| else |
| printk(KERN_ERR "cio: cio_setup: " |
| "invalid cio_msg parameter '%s'", parm); |
| return 1; |
| } |
| |
| __setup ("cio_msg=", cio_setup); |
| |
| /* |
| * Function: cio_debug_init |
| * Initializes three debug logs for common I/O: |
| * - cio_msg logs generic cio messages |
| * - cio_trace logs the calling of different functions |
| * - cio_crw logs machine check related cio messages |
| */ |
| static int __init cio_debug_init(void) |
| { |
| cio_debug_msg_id = debug_register("cio_msg", 16, 1, 16 * sizeof(long)); |
| if (!cio_debug_msg_id) |
| goto out_unregister; |
| debug_register_view(cio_debug_msg_id, &debug_sprintf_view); |
| debug_set_level(cio_debug_msg_id, 2); |
| cio_debug_trace_id = debug_register("cio_trace", 16, 1, 16); |
| if (!cio_debug_trace_id) |
| goto out_unregister; |
| debug_register_view(cio_debug_trace_id, &debug_hex_ascii_view); |
| debug_set_level(cio_debug_trace_id, 2); |
| cio_debug_crw_id = debug_register("cio_crw", 16, 1, 16 * sizeof(long)); |
| if (!cio_debug_crw_id) |
| goto out_unregister; |
| debug_register_view(cio_debug_crw_id, &debug_sprintf_view); |
| debug_set_level(cio_debug_crw_id, 4); |
| return 0; |
| |
| out_unregister: |
| if (cio_debug_msg_id) |
| debug_unregister(cio_debug_msg_id); |
| if (cio_debug_trace_id) |
| debug_unregister(cio_debug_trace_id); |
| if (cio_debug_crw_id) |
| debug_unregister(cio_debug_crw_id); |
| printk(KERN_WARNING"cio: could not initialize debugging\n"); |
| return -1; |
| } |
| |
| arch_initcall (cio_debug_init); |
| |
| int |
| cio_set_options (struct subchannel *sch, int flags) |
| { |
| sch->options.suspend = (flags & DOIO_ALLOW_SUSPEND) != 0; |
| sch->options.prefetch = (flags & DOIO_DENY_PREFETCH) != 0; |
| sch->options.inter = (flags & DOIO_SUPPRESS_INTER) != 0; |
| return 0; |
| } |
| |
| /* FIXME: who wants to use this? */ |
| int |
| cio_get_options (struct subchannel *sch) |
| { |
| int flags; |
| |
| flags = 0; |
| if (sch->options.suspend) |
| flags |= DOIO_ALLOW_SUSPEND; |
| if (sch->options.prefetch) |
| flags |= DOIO_DENY_PREFETCH; |
| if (sch->options.inter) |
| flags |= DOIO_SUPPRESS_INTER; |
| return flags; |
| } |
| |
| /* |
| * Use tpi to get a pending interrupt, call the interrupt handler and |
| * return a pointer to the subchannel structure. |
| */ |
| static int |
| cio_tpi(void) |
| { |
| struct tpi_info *tpi_info; |
| struct subchannel *sch; |
| struct irb *irb; |
| |
| tpi_info = (struct tpi_info *) __LC_SUBCHANNEL_ID; |
| if (tpi (NULL) != 1) |
| return 0; |
| irb = (struct irb *) __LC_IRB; |
| /* Store interrupt response block to lowcore. */ |
| if (tsch (tpi_info->schid, irb) != 0) |
| /* Not status pending or not operational. */ |
| return 1; |
| sch = (struct subchannel *)(unsigned long)tpi_info->intparm; |
| if (!sch) |
| return 1; |
| local_bh_disable(); |
| irq_enter (); |
| spin_lock(sch->lock); |
| memcpy (&sch->schib.scsw, &irb->scsw, sizeof (struct scsw)); |
| if (sch->driver && sch->driver->irq) |
| sch->driver->irq(sch); |
| spin_unlock(sch->lock); |
| irq_exit (); |
| _local_bh_enable(); |
| return 1; |
| } |
| |
| static int |
| cio_start_handle_notoper(struct subchannel *sch, __u8 lpm) |
| { |
| char dbf_text[15]; |
| |
| if (lpm != 0) |
| sch->lpm &= ~lpm; |
| else |
| sch->lpm = 0; |
| |
| stsch (sch->schid, &sch->schib); |
| |
| CIO_MSG_EVENT(0, "cio_start: 'not oper' status for " |
| "subchannel 0.%x.%04x!\n", sch->schid.ssid, |
| sch->schid.sch_no); |
| sprintf(dbf_text, "no%s", sch->dev.bus_id); |
| CIO_TRACE_EVENT(0, dbf_text); |
| CIO_HEX_EVENT(0, &sch->schib, sizeof (struct schib)); |
| |
| return (sch->lpm ? -EACCES : -ENODEV); |
| } |
| |
| int |
| cio_start_key (struct subchannel *sch, /* subchannel structure */ |
| struct ccw1 * cpa, /* logical channel prog addr */ |
| __u8 lpm, /* logical path mask */ |
| __u8 key) /* storage key */ |
| { |
| char dbf_txt[15]; |
| int ccode; |
| struct orb *orb; |
| |
| CIO_TRACE_EVENT(4, "stIO"); |
| CIO_TRACE_EVENT(4, sch->dev.bus_id); |
| |
| orb = &to_io_private(sch)->orb; |
| /* sch is always under 2G. */ |
| orb->intparm = (u32)(addr_t)sch; |
| orb->fmt = 1; |
| |
| orb->pfch = sch->options.prefetch == 0; |
| orb->spnd = sch->options.suspend; |
| orb->ssic = sch->options.suspend && sch->options.inter; |
| orb->lpm = (lpm != 0) ? lpm : sch->lpm; |
| #ifdef CONFIG_64BIT |
| /* |
| * for 64 bit we always support 64 bit IDAWs with 4k page size only |
| */ |
| orb->c64 = 1; |
| orb->i2k = 0; |
| #endif |
| orb->key = key >> 4; |
| /* issue "Start Subchannel" */ |
| orb->cpa = (__u32) __pa(cpa); |
| ccode = ssch(sch->schid, orb); |
| |
| /* process condition code */ |
| sprintf(dbf_txt, "ccode:%d", ccode); |
| CIO_TRACE_EVENT(4, dbf_txt); |
| |
| switch (ccode) { |
| case 0: |
| /* |
| * initialize device status information |
| */ |
| sch->schib.scsw.actl |= SCSW_ACTL_START_PEND; |
| return 0; |
| case 1: /* status pending */ |
| case 2: /* busy */ |
| return -EBUSY; |
| default: /* device/path not operational */ |
| return cio_start_handle_notoper(sch, lpm); |
| } |
| } |
| |
| int |
| cio_start (struct subchannel *sch, struct ccw1 *cpa, __u8 lpm) |
| { |
| return cio_start_key(sch, cpa, lpm, PAGE_DEFAULT_KEY); |
| } |
| |
| /* |
| * resume suspended I/O operation |
| */ |
| int |
| cio_resume (struct subchannel *sch) |
| { |
| char dbf_txt[15]; |
| int ccode; |
| |
| CIO_TRACE_EVENT (4, "resIO"); |
| CIO_TRACE_EVENT (4, sch->dev.bus_id); |
| |
| ccode = rsch (sch->schid); |
| |
| sprintf (dbf_txt, "ccode:%d", ccode); |
| CIO_TRACE_EVENT (4, dbf_txt); |
| |
| switch (ccode) { |
| case 0: |
| sch->schib.scsw.actl |= SCSW_ACTL_RESUME_PEND; |
| return 0; |
| case 1: |
| return -EBUSY; |
| case 2: |
| return -EINVAL; |
| default: |
| /* |
| * useless to wait for request completion |
| * as device is no longer operational ! |
| */ |
| return -ENODEV; |
| } |
| } |
| |
| /* |
| * halt I/O operation |
| */ |
| int |
| cio_halt(struct subchannel *sch) |
| { |
| char dbf_txt[15]; |
| int ccode; |
| |
| if (!sch) |
| return -ENODEV; |
| |
| CIO_TRACE_EVENT (2, "haltIO"); |
| CIO_TRACE_EVENT (2, sch->dev.bus_id); |
| |
| /* |
| * Issue "Halt subchannel" and process condition code |
| */ |
| ccode = hsch (sch->schid); |
| |
| sprintf (dbf_txt, "ccode:%d", ccode); |
| CIO_TRACE_EVENT (2, dbf_txt); |
| |
| switch (ccode) { |
| case 0: |
| sch->schib.scsw.actl |= SCSW_ACTL_HALT_PEND; |
| return 0; |
| case 1: /* status pending */ |
| case 2: /* busy */ |
| return -EBUSY; |
| default: /* device not operational */ |
| return -ENODEV; |
| } |
| } |
| |
| /* |
| * Clear I/O operation |
| */ |
| int |
| cio_clear(struct subchannel *sch) |
| { |
| char dbf_txt[15]; |
| int ccode; |
| |
| if (!sch) |
| return -ENODEV; |
| |
| CIO_TRACE_EVENT (2, "clearIO"); |
| CIO_TRACE_EVENT (2, sch->dev.bus_id); |
| |
| /* |
| * Issue "Clear subchannel" and process condition code |
| */ |
| ccode = csch (sch->schid); |
| |
| sprintf (dbf_txt, "ccode:%d", ccode); |
| CIO_TRACE_EVENT (2, dbf_txt); |
| |
| switch (ccode) { |
| case 0: |
| sch->schib.scsw.actl |= SCSW_ACTL_CLEAR_PEND; |
| return 0; |
| default: /* device not operational */ |
| return -ENODEV; |
| } |
| } |
| |
| /* |
| * Function: cio_cancel |
| * Issues a "Cancel Subchannel" on the specified subchannel |
| * Note: We don't need any fancy intparms and flags here |
| * since xsch is executed synchronously. |
| * Only for common I/O internal use as for now. |
| */ |
| int |
| cio_cancel (struct subchannel *sch) |
| { |
| char dbf_txt[15]; |
| int ccode; |
| |
| if (!sch) |
| return -ENODEV; |
| |
| CIO_TRACE_EVENT (2, "cancelIO"); |
| CIO_TRACE_EVENT (2, sch->dev.bus_id); |
| |
| ccode = xsch (sch->schid); |
| |
| sprintf (dbf_txt, "ccode:%d", ccode); |
| CIO_TRACE_EVENT (2, dbf_txt); |
| |
| switch (ccode) { |
| case 0: /* success */ |
| /* Update information in scsw. */ |
| stsch (sch->schid, &sch->schib); |
| return 0; |
| case 1: /* status pending */ |
| return -EBUSY; |
| case 2: /* not applicable */ |
| return -EINVAL; |
| default: /* not oper */ |
| return -ENODEV; |
| } |
| } |
| |
| /* |
| * Function: cio_modify |
| * Issues a "Modify Subchannel" on the specified subchannel |
| */ |
| int |
| cio_modify (struct subchannel *sch) |
| { |
| int ccode, retry, ret; |
| |
| ret = 0; |
| for (retry = 0; retry < 5; retry++) { |
| ccode = msch_err (sch->schid, &sch->schib); |
| if (ccode < 0) /* -EIO if msch gets a program check. */ |
| return ccode; |
| switch (ccode) { |
| case 0: /* successfull */ |
| return 0; |
| case 1: /* status pending */ |
| return -EBUSY; |
| case 2: /* busy */ |
| udelay (100); /* allow for recovery */ |
| ret = -EBUSY; |
| break; |
| case 3: /* not operational */ |
| return -ENODEV; |
| } |
| } |
| return ret; |
| } |
| |
| /* |
| * Enable subchannel. |
| */ |
| int cio_enable_subchannel(struct subchannel *sch, unsigned int isc, |
| u32 intparm) |
| { |
| char dbf_txt[15]; |
| int ccode; |
| int retry; |
| int ret; |
| |
| CIO_TRACE_EVENT (2, "ensch"); |
| CIO_TRACE_EVENT (2, sch->dev.bus_id); |
| |
| if (sch_is_pseudo_sch(sch)) |
| return -EINVAL; |
| ccode = stsch (sch->schid, &sch->schib); |
| if (ccode) |
| return -ENODEV; |
| |
| for (retry = 5, ret = 0; retry > 0; retry--) { |
| sch->schib.pmcw.ena = 1; |
| sch->schib.pmcw.isc = isc; |
| sch->schib.pmcw.intparm = intparm; |
| ret = cio_modify(sch); |
| if (ret == -ENODEV) |
| break; |
| if (ret == -EIO) |
| /* |
| * Got a program check in cio_modify. Try without |
| * the concurrent sense bit the next time. |
| */ |
| sch->schib.pmcw.csense = 0; |
| if (ret == 0) { |
| stsch (sch->schid, &sch->schib); |
| if (sch->schib.pmcw.ena) |
| break; |
| } |
| if (ret == -EBUSY) { |
| struct irb irb; |
| if (tsch(sch->schid, &irb) != 0) |
| break; |
| } |
| } |
| sprintf (dbf_txt, "ret:%d", ret); |
| CIO_TRACE_EVENT (2, dbf_txt); |
| return ret; |
| } |
| |
| /* |
| * Disable subchannel. |
| */ |
| int |
| cio_disable_subchannel (struct subchannel *sch) |
| { |
| char dbf_txt[15]; |
| int ccode; |
| int retry; |
| int ret; |
| |
| CIO_TRACE_EVENT (2, "dissch"); |
| CIO_TRACE_EVENT (2, sch->dev.bus_id); |
| |
| if (sch_is_pseudo_sch(sch)) |
| return 0; |
| ccode = stsch (sch->schid, &sch->schib); |
| if (ccode == 3) /* Not operational. */ |
| return -ENODEV; |
| |
| if (sch->schib.scsw.actl != 0) |
| /* |
| * the disable function must not be called while there are |
| * requests pending for completion ! |
| */ |
| return -EBUSY; |
| |
| for (retry = 5, ret = 0; retry > 0; retry--) { |
| sch->schib.pmcw.ena = 0; |
| ret = cio_modify(sch); |
| if (ret == -ENODEV) |
| break; |
| if (ret == -EBUSY) |
| /* |
| * The subchannel is busy or status pending. |
| * We'll disable when the next interrupt was delivered |
| * via the state machine. |
| */ |
| break; |
| if (ret == 0) { |
| stsch (sch->schid, &sch->schib); |
| if (!sch->schib.pmcw.ena) |
| break; |
| } |
| } |
| sprintf (dbf_txt, "ret:%d", ret); |
| CIO_TRACE_EVENT (2, dbf_txt); |
| return ret; |
| } |
| |
| int cio_create_sch_lock(struct subchannel *sch) |
| { |
| sch->lock = kmalloc(sizeof(spinlock_t), GFP_KERNEL); |
| if (!sch->lock) |
| return -ENOMEM; |
| spin_lock_init(sch->lock); |
| return 0; |
| } |
| |
| /* |
| * cio_validate_subchannel() |
| * |
| * Find out subchannel type and initialize struct subchannel. |
| * Return codes: |
| * SUBCHANNEL_TYPE_IO for a normal io subchannel |
| * SUBCHANNEL_TYPE_CHSC for a chsc subchannel |
| * SUBCHANNEL_TYPE_MESSAGE for a messaging subchannel |
| * SUBCHANNEL_TYPE_ADM for a adm(?) subchannel |
| * -ENXIO for non-defined subchannels |
| * -ENODEV for subchannels with invalid device number or blacklisted devices |
| */ |
| int |
| cio_validate_subchannel (struct subchannel *sch, struct subchannel_id schid) |
| { |
| char dbf_txt[15]; |
| int ccode; |
| int err; |
| |
| sprintf (dbf_txt, "valsch%x", schid.sch_no); |
| CIO_TRACE_EVENT (4, dbf_txt); |
| |
| /* Nuke all fields. */ |
| memset(sch, 0, sizeof(struct subchannel)); |
| |
| sch->schid = schid; |
| if (cio_is_console(schid)) { |
| sch->lock = cio_get_console_lock(); |
| } else { |
| err = cio_create_sch_lock(sch); |
| if (err) |
| goto out; |
| } |
| mutex_init(&sch->reg_mutex); |
| /* Set a name for the subchannel */ |
| snprintf (sch->dev.bus_id, BUS_ID_SIZE, "0.%x.%04x", schid.ssid, |
| schid.sch_no); |
| |
| /* |
| * The first subchannel that is not-operational (ccode==3) |
| * indicates that there aren't any more devices available. |
| * If stsch gets an exception, it means the current subchannel set |
| * is not valid. |
| */ |
| ccode = stsch_err (schid, &sch->schib); |
| if (ccode) { |
| err = (ccode == 3) ? -ENXIO : ccode; |
| goto out; |
| } |
| /* Copy subchannel type from path management control word. */ |
| sch->st = sch->schib.pmcw.st; |
| |
| /* |
| * ... just being curious we check for non I/O subchannels |
| */ |
| if (sch->st != 0) { |
| CIO_DEBUG(KERN_INFO, 0, |
| "Subchannel 0.%x.%04x reports " |
| "non-I/O subchannel type %04X\n", |
| sch->schid.ssid, sch->schid.sch_no, sch->st); |
| /* We stop here for non-io subchannels. */ |
| err = sch->st; |
| goto out; |
| } |
| |
| /* Initialization for io subchannels. */ |
| if (!css_sch_is_valid(&sch->schib)) { |
| err = -ENODEV; |
| goto out; |
| } |
| |
| /* Devno is valid. */ |
| if (is_blacklisted (sch->schid.ssid, sch->schib.pmcw.dev)) { |
| /* |
| * This device must not be known to Linux. So we simply |
| * say that there is no device and return ENODEV. |
| */ |
| CIO_MSG_EVENT(4, "Blacklisted device detected " |
| "at devno %04X, subchannel set %x\n", |
| sch->schib.pmcw.dev, sch->schid.ssid); |
| err = -ENODEV; |
| goto out; |
| } |
| if (cio_is_console(sch->schid)) |
| sch->opm = 0xff; |
| else |
| sch->opm = chp_get_sch_opm(sch); |
| sch->lpm = sch->schib.pmcw.pam & sch->opm; |
| |
| CIO_DEBUG(KERN_INFO, 0, |
| "Detected device %04x on subchannel 0.%x.%04X" |
| " - PIM = %02X, PAM = %02X, POM = %02X\n", |
| sch->schib.pmcw.dev, sch->schid.ssid, |
| sch->schid.sch_no, sch->schib.pmcw.pim, |
| sch->schib.pmcw.pam, sch->schib.pmcw.pom); |
| |
| /* |
| * We now have to initially ... |
| * ... set "interruption subclass" |
| * ... enable "concurrent sense" |
| * ... enable "multipath mode" if more than one |
| * CHPID is available. This is done regardless |
| * whether multiple paths are available for us. |
| */ |
| sch->schib.pmcw.isc = 3; /* could be smth. else */ |
| sch->schib.pmcw.csense = 1; /* concurrent sense */ |
| sch->schib.pmcw.ena = 0; |
| if ((sch->lpm & (sch->lpm - 1)) != 0) |
| sch->schib.pmcw.mp = 1; /* multipath mode */ |
| /* clean up possible residual cmf stuff */ |
| sch->schib.pmcw.mme = 0; |
| sch->schib.pmcw.mbfc = 0; |
| sch->schib.pmcw.mbi = 0; |
| sch->schib.mba = 0; |
| return 0; |
| out: |
| if (!cio_is_console(schid)) |
| kfree(sch->lock); |
| sch->lock = NULL; |
| return err; |
| } |
| |
| /* |
| * do_IRQ() handles all normal I/O device IRQ's (the special |
| * SMP cross-CPU interrupts have their own specific |
| * handlers). |
| * |
| */ |
| void |
| do_IRQ (struct pt_regs *regs) |
| { |
| struct tpi_info *tpi_info; |
| struct subchannel *sch; |
| struct irb *irb; |
| struct pt_regs *old_regs; |
| |
| old_regs = set_irq_regs(regs); |
| irq_enter(); |
| s390_idle_check(); |
| if (S390_lowcore.int_clock >= S390_lowcore.clock_comparator) |
| /* Serve timer interrupts first. */ |
| clock_comparator_work(); |
| /* |
| * Get interrupt information from lowcore |
| */ |
| tpi_info = (struct tpi_info *) __LC_SUBCHANNEL_ID; |
| irb = (struct irb *) __LC_IRB; |
| do { |
| kstat_cpu(smp_processor_id()).irqs[IO_INTERRUPT]++; |
| /* |
| * Non I/O-subchannel thin interrupts are processed differently |
| */ |
| if (tpi_info->adapter_IO == 1 && |
| tpi_info->int_type == IO_INTERRUPT_TYPE) { |
| do_adapter_IO(); |
| continue; |
| } |
| sch = (struct subchannel *)(unsigned long)tpi_info->intparm; |
| if (!sch) { |
| /* Clear pending interrupt condition. */ |
| tsch(tpi_info->schid, irb); |
| continue; |
| } |
| spin_lock(sch->lock); |
| /* Store interrupt response block to lowcore. */ |
| if (tsch(tpi_info->schid, irb) == 0) { |
| /* Keep subchannel information word up to date. */ |
| memcpy (&sch->schib.scsw, &irb->scsw, |
| sizeof (irb->scsw)); |
| /* Call interrupt handler if there is one. */ |
| if (sch->driver && sch->driver->irq) |
| sch->driver->irq(sch); |
| } |
| spin_unlock(sch->lock); |
| /* |
| * Are more interrupts pending? |
| * If so, the tpi instruction will update the lowcore |
| * to hold the info for the next interrupt. |
| * We don't do this for VM because a tpi drops the cpu |
| * out of the sie which costs more cycles than it saves. |
| */ |
| } while (!MACHINE_IS_VM && tpi (NULL) != 0); |
| irq_exit(); |
| set_irq_regs(old_regs); |
| } |
| |
| #ifdef CONFIG_CCW_CONSOLE |
| static struct subchannel console_subchannel; |
| static struct io_subchannel_private console_priv; |
| static int console_subchannel_in_use; |
| |
| void *cio_get_console_priv(void) |
| { |
| return &console_priv; |
| } |
| |
| /* |
| * busy wait for the next interrupt on the console |
| */ |
| void wait_cons_dev(void) |
| __releases(console_subchannel.lock) |
| __acquires(console_subchannel.lock) |
| { |
| unsigned long cr6 __attribute__ ((aligned (8))); |
| unsigned long save_cr6 __attribute__ ((aligned (8))); |
| |
| /* |
| * before entering the spinlock we may already have |
| * processed the interrupt on a different CPU... |
| */ |
| if (!console_subchannel_in_use) |
| return; |
| |
| /* disable all but isc 7 (console device) */ |
| __ctl_store (save_cr6, 6, 6); |
| cr6 = 0x01000000; |
| __ctl_load (cr6, 6, 6); |
| |
| do { |
| spin_unlock(console_subchannel.lock); |
| if (!cio_tpi()) |
| cpu_relax(); |
| spin_lock(console_subchannel.lock); |
| } while (console_subchannel.schib.scsw.actl != 0); |
| /* |
| * restore previous isc value |
| */ |
| __ctl_load (save_cr6, 6, 6); |
| } |
| |
| static int |
| cio_test_for_console(struct subchannel_id schid, void *data) |
| { |
| if (stsch_err(schid, &console_subchannel.schib) != 0) |
| return -ENXIO; |
| if ((console_subchannel.schib.pmcw.st == SUBCHANNEL_TYPE_IO) && |
| console_subchannel.schib.pmcw.dnv && |
| (console_subchannel.schib.pmcw.dev == console_devno)) { |
| console_irq = schid.sch_no; |
| return 1; /* found */ |
| } |
| return 0; |
| } |
| |
| |
| static int |
| cio_get_console_sch_no(void) |
| { |
| struct subchannel_id schid; |
| |
| init_subchannel_id(&schid); |
| if (console_irq != -1) { |
| /* VM provided us with the irq number of the console. */ |
| schid.sch_no = console_irq; |
| if (stsch(schid, &console_subchannel.schib) != 0 || |
| (console_subchannel.schib.pmcw.st != SUBCHANNEL_TYPE_IO) || |
| !console_subchannel.schib.pmcw.dnv) |
| return -1; |
| console_devno = console_subchannel.schib.pmcw.dev; |
| } else if (console_devno != -1) { |
| /* At least the console device number is known. */ |
| for_each_subchannel(cio_test_for_console, NULL); |
| if (console_irq == -1) |
| return -1; |
| } else { |
| /* unlike in 2.4, we cannot autoprobe here, since |
| * the channel subsystem is not fully initialized. |
| * With some luck, the HWC console can take over */ |
| printk(KERN_WARNING "cio: No ccw console found!\n"); |
| return -1; |
| } |
| return console_irq; |
| } |
| |
| struct subchannel * |
| cio_probe_console(void) |
| { |
| int sch_no, ret; |
| struct subchannel_id schid; |
| |
| if (xchg(&console_subchannel_in_use, 1) != 0) |
| return ERR_PTR(-EBUSY); |
| sch_no = cio_get_console_sch_no(); |
| if (sch_no == -1) { |
| console_subchannel_in_use = 0; |
| return ERR_PTR(-ENODEV); |
| } |
| memset(&console_subchannel, 0, sizeof(struct subchannel)); |
| init_subchannel_id(&schid); |
| schid.sch_no = sch_no; |
| ret = cio_validate_subchannel(&console_subchannel, schid); |
| if (ret) { |
| console_subchannel_in_use = 0; |
| return ERR_PTR(-ENODEV); |
| } |
| |
| /* |
| * enable console I/O-interrupt subclass 7 |
| */ |
| ctl_set_bit(6, 24); |
| console_subchannel.schib.pmcw.isc = 7; |
| console_subchannel.schib.pmcw.intparm = |
| (u32)(addr_t)&console_subchannel; |
| ret = cio_modify(&console_subchannel); |
| if (ret) { |
| console_subchannel_in_use = 0; |
| return ERR_PTR(ret); |
| } |
| return &console_subchannel; |
| } |
| |
| void |
| cio_release_console(void) |
| { |
| console_subchannel.schib.pmcw.intparm = 0; |
| cio_modify(&console_subchannel); |
| ctl_clear_bit(6, 24); |
| console_subchannel_in_use = 0; |
| } |
| |
| /* Bah... hack to catch console special sausages. */ |
| int |
| cio_is_console(struct subchannel_id schid) |
| { |
| if (!console_subchannel_in_use) |
| return 0; |
| return schid_equal(&schid, &console_subchannel.schid); |
| } |
| |
| struct subchannel * |
| cio_get_console_subchannel(void) |
| { |
| if (!console_subchannel_in_use) |
| return NULL; |
| return &console_subchannel; |
| } |
| |
| #endif |
| static int |
| __disable_subchannel_easy(struct subchannel_id schid, struct schib *schib) |
| { |
| int retry, cc; |
| |
| cc = 0; |
| for (retry=0;retry<3;retry++) { |
| schib->pmcw.ena = 0; |
| cc = msch(schid, schib); |
| if (cc) |
| return (cc==3?-ENODEV:-EBUSY); |
| stsch(schid, schib); |
| if (!schib->pmcw.ena) |
| return 0; |
| } |
| return -EBUSY; /* uhm... */ |
| } |
| |
| /* we can't use the normal udelay here, since it enables external interrupts */ |
| |
| static void udelay_reset(unsigned long usecs) |
| { |
| uint64_t start_cc, end_cc; |
| |
| asm volatile ("STCK %0" : "=m" (start_cc)); |
| do { |
| cpu_relax(); |
| asm volatile ("STCK %0" : "=m" (end_cc)); |
| } while (((end_cc - start_cc)/4096) < usecs); |
| } |
| |
| static int |
| __clear_subchannel_easy(struct subchannel_id schid) |
| { |
| int retry; |
| |
| if (csch(schid)) |
| return -ENODEV; |
| for (retry=0;retry<20;retry++) { |
| struct tpi_info ti; |
| |
| if (tpi(&ti)) { |
| tsch(ti.schid, (struct irb *)__LC_IRB); |
| if (schid_equal(&ti.schid, &schid)) |
| return 0; |
| } |
| udelay_reset(100); |
| } |
| return -EBUSY; |
| } |
| |
| static int pgm_check_occured; |
| |
| static void cio_reset_pgm_check_handler(void) |
| { |
| pgm_check_occured = 1; |
| } |
| |
| static int stsch_reset(struct subchannel_id schid, volatile struct schib *addr) |
| { |
| int rc; |
| |
| pgm_check_occured = 0; |
| s390_base_pgm_handler_fn = cio_reset_pgm_check_handler; |
| rc = stsch(schid, addr); |
| s390_base_pgm_handler_fn = NULL; |
| |
| /* The program check handler could have changed pgm_check_occured. */ |
| barrier(); |
| |
| if (pgm_check_occured) |
| return -EIO; |
| else |
| return rc; |
| } |
| |
| static int __shutdown_subchannel_easy(struct subchannel_id schid, void *data) |
| { |
| struct schib schib; |
| |
| if (stsch_reset(schid, &schib)) |
| return -ENXIO; |
| if (!schib.pmcw.ena) |
| return 0; |
| switch(__disable_subchannel_easy(schid, &schib)) { |
| case 0: |
| case -ENODEV: |
| break; |
| default: /* -EBUSY */ |
| if (__clear_subchannel_easy(schid)) |
| break; /* give up... */ |
| stsch(schid, &schib); |
| __disable_subchannel_easy(schid, &schib); |
| } |
| return 0; |
| } |
| |
| static atomic_t chpid_reset_count; |
| |
| static void s390_reset_chpids_mcck_handler(void) |
| { |
| struct crw crw; |
| struct mci *mci; |
| |
| /* Check for pending channel report word. */ |
| mci = (struct mci *)&S390_lowcore.mcck_interruption_code; |
| if (!mci->cp) |
| return; |
| /* Process channel report words. */ |
| while (stcrw(&crw) == 0) { |
| /* Check for responses to RCHP. */ |
| if (crw.slct && crw.rsc == CRW_RSC_CPATH) |
| atomic_dec(&chpid_reset_count); |
| } |
| } |
| |
| #define RCHP_TIMEOUT (30 * USEC_PER_SEC) |
| static void css_reset(void) |
| { |
| int i, ret; |
| unsigned long long timeout; |
| struct chp_id chpid; |
| |
| /* Reset subchannels. */ |
| for_each_subchannel(__shutdown_subchannel_easy, NULL); |
| /* Reset channel paths. */ |
| s390_base_mcck_handler_fn = s390_reset_chpids_mcck_handler; |
| /* Enable channel report machine checks. */ |
| __ctl_set_bit(14, 28); |
| /* Temporarily reenable machine checks. */ |
| local_mcck_enable(); |
| chp_id_init(&chpid); |
| for (i = 0; i <= __MAX_CHPID; i++) { |
| chpid.id = i; |
| ret = rchp(chpid); |
| if ((ret == 0) || (ret == 2)) |
| /* |
| * rchp either succeeded, or another rchp is already |
| * in progress. In either case, we'll get a crw. |
| */ |
| atomic_inc(&chpid_reset_count); |
| } |
| /* Wait for machine check for all channel paths. */ |
| timeout = get_clock() + (RCHP_TIMEOUT << 12); |
| while (atomic_read(&chpid_reset_count) != 0) { |
| if (get_clock() > timeout) |
| break; |
| cpu_relax(); |
| } |
| /* Disable machine checks again. */ |
| local_mcck_disable(); |
| /* Disable channel report machine checks. */ |
| __ctl_clear_bit(14, 28); |
| s390_base_mcck_handler_fn = NULL; |
| } |
| |
| static struct reset_call css_reset_call = { |
| .fn = css_reset, |
| }; |
| |
| static int __init init_css_reset_call(void) |
| { |
| atomic_set(&chpid_reset_count, 0); |
| register_reset_call(&css_reset_call); |
| return 0; |
| } |
| |
| arch_initcall(init_css_reset_call); |
| |
| struct sch_match_id { |
| struct subchannel_id schid; |
| struct ccw_dev_id devid; |
| int rc; |
| }; |
| |
| static int __reipl_subchannel_match(struct subchannel_id schid, void *data) |
| { |
| struct schib schib; |
| struct sch_match_id *match_id = data; |
| |
| if (stsch_reset(schid, &schib)) |
| return -ENXIO; |
| if ((schib.pmcw.st == SUBCHANNEL_TYPE_IO) && schib.pmcw.dnv && |
| (schib.pmcw.dev == match_id->devid.devno) && |
| (schid.ssid == match_id->devid.ssid)) { |
| match_id->schid = schid; |
| match_id->rc = 0; |
| return 1; |
| } |
| return 0; |
| } |
| |
| static int reipl_find_schid(struct ccw_dev_id *devid, |
| struct subchannel_id *schid) |
| { |
| struct sch_match_id match_id; |
| |
| match_id.devid = *devid; |
| match_id.rc = -ENODEV; |
| for_each_subchannel(__reipl_subchannel_match, &match_id); |
| if (match_id.rc == 0) |
| *schid = match_id.schid; |
| return match_id.rc; |
| } |
| |
| extern void do_reipl_asm(__u32 schid); |
| |
| /* Make sure all subchannels are quiet before we re-ipl an lpar. */ |
| void reipl_ccw_dev(struct ccw_dev_id *devid) |
| { |
| struct subchannel_id schid; |
| |
| s390_reset_system(); |
| if (reipl_find_schid(devid, &schid) != 0) |
| panic("IPL Device not found\n"); |
| do_reipl_asm(*((__u32*)&schid)); |
| } |
| |
| int __init cio_get_iplinfo(struct cio_iplinfo *iplinfo) |
| { |
| struct subchannel_id schid; |
| struct schib schib; |
| |
| schid = *(struct subchannel_id *)__LC_SUBCHANNEL_ID; |
| if (!schid.one) |
| return -ENODEV; |
| if (stsch(schid, &schib)) |
| return -ENODEV; |
| if (schib.pmcw.st != SUBCHANNEL_TYPE_IO) |
| return -ENODEV; |
| if (!schib.pmcw.dnv) |
| return -ENODEV; |
| iplinfo->devno = schib.pmcw.dev; |
| iplinfo->is_qdio = schib.pmcw.qf; |
| return 0; |
| } |