Jeff Garzik | 1fdffbc | 2006-02-09 05:15:27 -0500 | [diff] [blame] | 1 | /* |
| 2 | * libata-bmdma.c - helper library for PCI IDE BMDMA |
| 3 | * |
| 4 | * Maintained by: Jeff Garzik <jgarzik@pobox.com> |
| 5 | * Please ALWAYS copy linux-ide@vger.kernel.org |
| 6 | * on emails. |
| 7 | * |
| 8 | * Copyright 2003-2006 Red Hat, Inc. All rights reserved. |
| 9 | * Copyright 2003-2006 Jeff Garzik |
| 10 | * |
| 11 | * |
| 12 | * This program is free software; you can redistribute it and/or modify |
| 13 | * it under the terms of the GNU General Public License as published by |
| 14 | * the Free Software Foundation; either version 2, or (at your option) |
| 15 | * any later version. |
| 16 | * |
| 17 | * This program is distributed in the hope that it will be useful, |
| 18 | * but WITHOUT ANY WARRANTY; without even the implied warranty of |
| 19 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the |
| 20 | * GNU General Public License for more details. |
| 21 | * |
| 22 | * You should have received a copy of the GNU General Public License |
| 23 | * along with this program; see the file COPYING. If not, write to |
| 24 | * the Free Software Foundation, 675 Mass Ave, Cambridge, MA 02139, USA. |
| 25 | * |
| 26 | * |
| 27 | * libata documentation is available via 'make {ps|pdf}docs', |
| 28 | * as Documentation/DocBook/libata.* |
| 29 | * |
| 30 | * Hardware documentation available from http://www.t13.org/ and |
| 31 | * http://www.sata-io.org/ |
| 32 | * |
| 33 | */ |
| 34 | |
Jeff Garzik | 1fdffbc | 2006-02-09 05:15:27 -0500 | [diff] [blame] | 35 | #include <linux/kernel.h> |
| 36 | #include <linux/pci.h> |
| 37 | #include <linux/libata.h> |
| 38 | |
| 39 | #include "libata.h" |
| 40 | |
| 41 | /** |
Tejun Heo | 90088bb | 2006-10-09 11:10:26 +0900 | [diff] [blame] | 42 | * ata_irq_on - Enable interrupts on a port. |
| 43 | * @ap: Port on which interrupts are enabled. |
| 44 | * |
| 45 | * Enable interrupts on a legacy IDE device using MMIO or PIO, |
| 46 | * wait for idle, clear any pending interrupts. |
| 47 | * |
| 48 | * LOCKING: |
| 49 | * Inherited from caller. |
| 50 | */ |
| 51 | u8 ata_irq_on(struct ata_port *ap) |
| 52 | { |
| 53 | struct ata_ioports *ioaddr = &ap->ioaddr; |
| 54 | u8 tmp; |
| 55 | |
| 56 | ap->ctl &= ~ATA_NIEN; |
| 57 | ap->last_ctl = ap->ctl; |
| 58 | |
Tejun Heo | 0d5ff56 | 2007-02-01 15:06:36 +0900 | [diff] [blame] | 59 | iowrite8(ap->ctl, ioaddr->ctl_addr); |
Tejun Heo | 90088bb | 2006-10-09 11:10:26 +0900 | [diff] [blame] | 60 | tmp = ata_wait_idle(ap); |
| 61 | |
| 62 | ap->ops->irq_clear(ap); |
| 63 | |
| 64 | return tmp; |
| 65 | } |
| 66 | |
Akira Iguchi | 8362500 | 2007-01-26 16:27:32 +0900 | [diff] [blame] | 67 | u8 ata_dummy_irq_on (struct ata_port *ap) { return 0; } |
| 68 | |
| 69 | /** |
| 70 | * ata_irq_ack - Acknowledge a device interrupt. |
| 71 | * @ap: Port on which interrupts are enabled. |
| 72 | * |
| 73 | * Wait up to 10 ms for legacy IDE device to become idle (BUSY |
| 74 | * or BUSY+DRQ clear). Obtain dma status and port status from |
| 75 | * device. Clear the interrupt. Return port status. |
| 76 | * |
| 77 | * LOCKING: |
| 78 | */ |
| 79 | |
| 80 | u8 ata_irq_ack(struct ata_port *ap, unsigned int chk_drq) |
| 81 | { |
| 82 | unsigned int bits = chk_drq ? ATA_BUSY | ATA_DRQ : ATA_BUSY; |
| 83 | u8 host_stat, post_stat, status; |
| 84 | |
| 85 | status = ata_busy_wait(ap, bits, 1000); |
| 86 | if (status & bits) |
| 87 | if (ata_msg_err(ap)) |
| 88 | printk(KERN_ERR "abnormal status 0x%X\n", status); |
| 89 | |
| 90 | /* get controller status; clear intr, err bits */ |
| 91 | host_stat = ioread8(ap->ioaddr.bmdma_addr + ATA_DMA_STATUS); |
| 92 | iowrite8(host_stat | ATA_DMA_INTR | ATA_DMA_ERR, |
| 93 | ap->ioaddr.bmdma_addr + ATA_DMA_STATUS); |
| 94 | |
| 95 | post_stat = ioread8(ap->ioaddr.bmdma_addr + ATA_DMA_STATUS); |
| 96 | |
| 97 | if (ata_msg_intr(ap)) |
| 98 | printk(KERN_INFO "%s: irq ack: host_stat 0x%X, new host_stat 0x%X, drv_stat 0x%X\n", |
| 99 | __FUNCTION__, |
| 100 | host_stat, post_stat, status); |
| 101 | |
| 102 | return status; |
| 103 | } |
| 104 | |
| 105 | u8 ata_dummy_irq_ack(struct ata_port *ap, unsigned int chk_drq) { return 0; } |
| 106 | |
Tejun Heo | 90088bb | 2006-10-09 11:10:26 +0900 | [diff] [blame] | 107 | /** |
Tejun Heo | 0d5ff56 | 2007-02-01 15:06:36 +0900 | [diff] [blame] | 108 | * ata_tf_load - send taskfile registers to host controller |
Jeff Garzik | 1fdffbc | 2006-02-09 05:15:27 -0500 | [diff] [blame] | 109 | * @ap: Port to which output is sent |
| 110 | * @tf: ATA taskfile register set |
| 111 | * |
| 112 | * Outputs ATA taskfile to standard ATA host controller. |
| 113 | * |
| 114 | * LOCKING: |
| 115 | * Inherited from caller. |
| 116 | */ |
| 117 | |
Jeff Garzik | 1fdffbc | 2006-02-09 05:15:27 -0500 | [diff] [blame] | 118 | void ata_tf_load(struct ata_port *ap, const struct ata_taskfile *tf) |
| 119 | { |
Tejun Heo | 0d5ff56 | 2007-02-01 15:06:36 +0900 | [diff] [blame] | 120 | struct ata_ioports *ioaddr = &ap->ioaddr; |
| 121 | unsigned int is_addr = tf->flags & ATA_TFLAG_ISADDR; |
| 122 | |
| 123 | if (tf->ctl != ap->last_ctl) { |
| 124 | iowrite8(tf->ctl, ioaddr->ctl_addr); |
| 125 | ap->last_ctl = tf->ctl; |
| 126 | ata_wait_idle(ap); |
| 127 | } |
| 128 | |
| 129 | if (is_addr && (tf->flags & ATA_TFLAG_LBA48)) { |
| 130 | iowrite8(tf->hob_feature, ioaddr->feature_addr); |
| 131 | iowrite8(tf->hob_nsect, ioaddr->nsect_addr); |
| 132 | iowrite8(tf->hob_lbal, ioaddr->lbal_addr); |
| 133 | iowrite8(tf->hob_lbam, ioaddr->lbam_addr); |
| 134 | iowrite8(tf->hob_lbah, ioaddr->lbah_addr); |
| 135 | VPRINTK("hob: feat 0x%X nsect 0x%X, lba 0x%X 0x%X 0x%X\n", |
| 136 | tf->hob_feature, |
| 137 | tf->hob_nsect, |
| 138 | tf->hob_lbal, |
| 139 | tf->hob_lbam, |
| 140 | tf->hob_lbah); |
| 141 | } |
| 142 | |
| 143 | if (is_addr) { |
| 144 | iowrite8(tf->feature, ioaddr->feature_addr); |
| 145 | iowrite8(tf->nsect, ioaddr->nsect_addr); |
| 146 | iowrite8(tf->lbal, ioaddr->lbal_addr); |
| 147 | iowrite8(tf->lbam, ioaddr->lbam_addr); |
| 148 | iowrite8(tf->lbah, ioaddr->lbah_addr); |
| 149 | VPRINTK("feat 0x%X nsect 0x%X lba 0x%X 0x%X 0x%X\n", |
| 150 | tf->feature, |
| 151 | tf->nsect, |
| 152 | tf->lbal, |
| 153 | tf->lbam, |
| 154 | tf->lbah); |
| 155 | } |
| 156 | |
| 157 | if (tf->flags & ATA_TFLAG_DEVICE) { |
| 158 | iowrite8(tf->device, ioaddr->device_addr); |
| 159 | VPRINTK("device 0x%X\n", tf->device); |
| 160 | } |
| 161 | |
| 162 | ata_wait_idle(ap); |
Jeff Garzik | 1fdffbc | 2006-02-09 05:15:27 -0500 | [diff] [blame] | 163 | } |
| 164 | |
| 165 | /** |
Jeff Garzik | 1fdffbc | 2006-02-09 05:15:27 -0500 | [diff] [blame] | 166 | * ata_exec_command - issue ATA command to host controller |
| 167 | * @ap: port to which command is being issued |
| 168 | * @tf: ATA taskfile register set |
| 169 | * |
Tejun Heo | 0d5ff56 | 2007-02-01 15:06:36 +0900 | [diff] [blame] | 170 | * Issues ATA command, with proper synchronization with interrupt |
| 171 | * handler / other threads. |
Jeff Garzik | 1fdffbc | 2006-02-09 05:15:27 -0500 | [diff] [blame] | 172 | * |
| 173 | * LOCKING: |
Jeff Garzik | cca3974 | 2006-08-24 03:19:22 -0400 | [diff] [blame] | 174 | * spin_lock_irqsave(host lock) |
Jeff Garzik | 1fdffbc | 2006-02-09 05:15:27 -0500 | [diff] [blame] | 175 | */ |
| 176 | void ata_exec_command(struct ata_port *ap, const struct ata_taskfile *tf) |
| 177 | { |
Tejun Heo | 44877b4 | 2007-02-21 01:06:51 +0900 | [diff] [blame] | 178 | DPRINTK("ata%u: cmd 0x%X\n", ap->print_id, tf->command); |
Tejun Heo | 0d5ff56 | 2007-02-01 15:06:36 +0900 | [diff] [blame] | 179 | |
| 180 | iowrite8(tf->command, ap->ioaddr.command_addr); |
| 181 | ata_pause(ap); |
Jeff Garzik | 1fdffbc | 2006-02-09 05:15:27 -0500 | [diff] [blame] | 182 | } |
| 183 | |
| 184 | /** |
Jeff Garzik | 1fdffbc | 2006-02-09 05:15:27 -0500 | [diff] [blame] | 185 | * ata_tf_read - input device's ATA taskfile shadow registers |
| 186 | * @ap: Port from which input is read |
| 187 | * @tf: ATA taskfile register set for storing input |
| 188 | * |
| 189 | * Reads ATA taskfile registers for currently-selected device |
| 190 | * into @tf. |
| 191 | * |
Jeff Garzik | 1fdffbc | 2006-02-09 05:15:27 -0500 | [diff] [blame] | 192 | * LOCKING: |
| 193 | * Inherited from caller. |
| 194 | */ |
| 195 | void ata_tf_read(struct ata_port *ap, struct ata_taskfile *tf) |
| 196 | { |
Tejun Heo | 0d5ff56 | 2007-02-01 15:06:36 +0900 | [diff] [blame] | 197 | struct ata_ioports *ioaddr = &ap->ioaddr; |
Jeff Garzik | 1fdffbc | 2006-02-09 05:15:27 -0500 | [diff] [blame] | 198 | |
Tejun Heo | 0d5ff56 | 2007-02-01 15:06:36 +0900 | [diff] [blame] | 199 | tf->command = ata_check_status(ap); |
| 200 | tf->feature = ioread8(ioaddr->error_addr); |
| 201 | tf->nsect = ioread8(ioaddr->nsect_addr); |
| 202 | tf->lbal = ioread8(ioaddr->lbal_addr); |
| 203 | tf->lbam = ioread8(ioaddr->lbam_addr); |
| 204 | tf->lbah = ioread8(ioaddr->lbah_addr); |
| 205 | tf->device = ioread8(ioaddr->device_addr); |
Jeff Garzik | 1fdffbc | 2006-02-09 05:15:27 -0500 | [diff] [blame] | 206 | |
Tejun Heo | 0d5ff56 | 2007-02-01 15:06:36 +0900 | [diff] [blame] | 207 | if (tf->flags & ATA_TFLAG_LBA48) { |
| 208 | iowrite8(tf->ctl | ATA_HOB, ioaddr->ctl_addr); |
| 209 | tf->hob_feature = ioread8(ioaddr->error_addr); |
| 210 | tf->hob_nsect = ioread8(ioaddr->nsect_addr); |
| 211 | tf->hob_lbal = ioread8(ioaddr->lbal_addr); |
| 212 | tf->hob_lbam = ioread8(ioaddr->lbam_addr); |
| 213 | tf->hob_lbah = ioread8(ioaddr->lbah_addr); |
| 214 | } |
Jeff Garzik | 1fdffbc | 2006-02-09 05:15:27 -0500 | [diff] [blame] | 215 | } |
| 216 | |
Jeff Garzik | 1fdffbc | 2006-02-09 05:15:27 -0500 | [diff] [blame] | 217 | /** |
| 218 | * ata_check_status - Read device status reg & clear interrupt |
| 219 | * @ap: port where the device is |
| 220 | * |
| 221 | * Reads ATA taskfile status register for currently-selected device |
| 222 | * and return its value. This also clears pending interrupts |
| 223 | * from this device |
| 224 | * |
Jeff Garzik | 1fdffbc | 2006-02-09 05:15:27 -0500 | [diff] [blame] | 225 | * LOCKING: |
| 226 | * Inherited from caller. |
| 227 | */ |
| 228 | u8 ata_check_status(struct ata_port *ap) |
| 229 | { |
Tejun Heo | 0d5ff56 | 2007-02-01 15:06:36 +0900 | [diff] [blame] | 230 | return ioread8(ap->ioaddr.status_addr); |
Jeff Garzik | 1fdffbc | 2006-02-09 05:15:27 -0500 | [diff] [blame] | 231 | } |
| 232 | |
Jeff Garzik | 1fdffbc | 2006-02-09 05:15:27 -0500 | [diff] [blame] | 233 | /** |
| 234 | * ata_altstatus - Read device alternate status reg |
| 235 | * @ap: port where the device is |
| 236 | * |
| 237 | * Reads ATA taskfile alternate status register for |
| 238 | * currently-selected device and return its value. |
| 239 | * |
| 240 | * Note: may NOT be used as the check_altstatus() entry in |
| 241 | * ata_port_operations. |
| 242 | * |
| 243 | * LOCKING: |
| 244 | * Inherited from caller. |
| 245 | */ |
| 246 | u8 ata_altstatus(struct ata_port *ap) |
| 247 | { |
| 248 | if (ap->ops->check_altstatus) |
| 249 | return ap->ops->check_altstatus(ap); |
| 250 | |
Tejun Heo | 0d5ff56 | 2007-02-01 15:06:36 +0900 | [diff] [blame] | 251 | return ioread8(ap->ioaddr.altstatus_addr); |
Jeff Garzik | 1fdffbc | 2006-02-09 05:15:27 -0500 | [diff] [blame] | 252 | } |
| 253 | |
Jeff Garzik | 2cc432e | 2006-03-23 00:32:03 -0500 | [diff] [blame] | 254 | /** |
Tejun Heo | 0d5ff56 | 2007-02-01 15:06:36 +0900 | [diff] [blame] | 255 | * ata_bmdma_setup - Set up PCI IDE BMDMA transaction |
Jeff Garzik | 2cc432e | 2006-03-23 00:32:03 -0500 | [diff] [blame] | 256 | * @qc: Info associated with this ATA transaction. |
| 257 | * |
| 258 | * LOCKING: |
Jeff Garzik | cca3974 | 2006-08-24 03:19:22 -0400 | [diff] [blame] | 259 | * spin_lock_irqsave(host lock) |
Jeff Garzik | 2cc432e | 2006-03-23 00:32:03 -0500 | [diff] [blame] | 260 | */ |
Tejun Heo | 0d5ff56 | 2007-02-01 15:06:36 +0900 | [diff] [blame] | 261 | void ata_bmdma_setup(struct ata_queued_cmd *qc) |
Jeff Garzik | 2cc432e | 2006-03-23 00:32:03 -0500 | [diff] [blame] | 262 | { |
| 263 | struct ata_port *ap = qc->ap; |
| 264 | unsigned int rw = (qc->tf.flags & ATA_TFLAG_WRITE); |
| 265 | u8 dmactl; |
Jeff Garzik | 2cc432e | 2006-03-23 00:32:03 -0500 | [diff] [blame] | 266 | |
| 267 | /* load PRD table addr. */ |
| 268 | mb(); /* make sure PRD table writes are visible to controller */ |
Tejun Heo | 0d5ff56 | 2007-02-01 15:06:36 +0900 | [diff] [blame] | 269 | iowrite32(ap->prd_dma, ap->ioaddr.bmdma_addr + ATA_DMA_TABLE_OFS); |
Jeff Garzik | 2cc432e | 2006-03-23 00:32:03 -0500 | [diff] [blame] | 270 | |
| 271 | /* specify data direction, triple-check start bit is clear */ |
Tejun Heo | 0d5ff56 | 2007-02-01 15:06:36 +0900 | [diff] [blame] | 272 | dmactl = ioread8(ap->ioaddr.bmdma_addr + ATA_DMA_CMD); |
Jeff Garzik | 2cc432e | 2006-03-23 00:32:03 -0500 | [diff] [blame] | 273 | dmactl &= ~(ATA_DMA_WR | ATA_DMA_START); |
| 274 | if (!rw) |
| 275 | dmactl |= ATA_DMA_WR; |
Tejun Heo | 0d5ff56 | 2007-02-01 15:06:36 +0900 | [diff] [blame] | 276 | iowrite8(dmactl, ap->ioaddr.bmdma_addr + ATA_DMA_CMD); |
Jeff Garzik | 2cc432e | 2006-03-23 00:32:03 -0500 | [diff] [blame] | 277 | |
| 278 | /* issue r/w command */ |
| 279 | ap->ops->exec_command(ap, &qc->tf); |
| 280 | } |
| 281 | |
| 282 | /** |
Tejun Heo | 0d5ff56 | 2007-02-01 15:06:36 +0900 | [diff] [blame] | 283 | * ata_bmdma_start - Start a PCI IDE BMDMA transaction |
Jeff Garzik | 2cc432e | 2006-03-23 00:32:03 -0500 | [diff] [blame] | 284 | * @qc: Info associated with this ATA transaction. |
| 285 | * |
| 286 | * LOCKING: |
Jeff Garzik | cca3974 | 2006-08-24 03:19:22 -0400 | [diff] [blame] | 287 | * spin_lock_irqsave(host lock) |
Jeff Garzik | 2cc432e | 2006-03-23 00:32:03 -0500 | [diff] [blame] | 288 | */ |
Tejun Heo | 0d5ff56 | 2007-02-01 15:06:36 +0900 | [diff] [blame] | 289 | void ata_bmdma_start (struct ata_queued_cmd *qc) |
Jeff Garzik | 2cc432e | 2006-03-23 00:32:03 -0500 | [diff] [blame] | 290 | { |
| 291 | struct ata_port *ap = qc->ap; |
Jeff Garzik | 2cc432e | 2006-03-23 00:32:03 -0500 | [diff] [blame] | 292 | u8 dmactl; |
| 293 | |
| 294 | /* start host DMA transaction */ |
Tejun Heo | 0d5ff56 | 2007-02-01 15:06:36 +0900 | [diff] [blame] | 295 | dmactl = ioread8(ap->ioaddr.bmdma_addr + ATA_DMA_CMD); |
| 296 | iowrite8(dmactl | ATA_DMA_START, ap->ioaddr.bmdma_addr + ATA_DMA_CMD); |
Jeff Garzik | 2cc432e | 2006-03-23 00:32:03 -0500 | [diff] [blame] | 297 | |
| 298 | /* Strictly, one may wish to issue a readb() here, to |
| 299 | * flush the mmio write. However, control also passes |
| 300 | * to the hardware at this point, and it will interrupt |
| 301 | * us when we are to resume control. So, in effect, |
| 302 | * we don't care when the mmio write flushes. |
| 303 | * Further, a read of the DMA status register _immediately_ |
| 304 | * following the write may not be what certain flaky hardware |
| 305 | * is expected, so I think it is best to not add a readb() |
| 306 | * without first all the MMIO ATA cards/mobos. |
| 307 | * Or maybe I'm just being paranoid. |
| 308 | */ |
| 309 | } |
| 310 | |
| 311 | /** |
Jeff Garzik | 2cc432e | 2006-03-23 00:32:03 -0500 | [diff] [blame] | 312 | * ata_bmdma_irq_clear - Clear PCI IDE BMDMA interrupt. |
| 313 | * @ap: Port associated with this ATA transaction. |
| 314 | * |
| 315 | * Clear interrupt and error flags in DMA status register. |
| 316 | * |
| 317 | * May be used as the irq_clear() entry in ata_port_operations. |
| 318 | * |
| 319 | * LOCKING: |
Jeff Garzik | cca3974 | 2006-08-24 03:19:22 -0400 | [diff] [blame] | 320 | * spin_lock_irqsave(host lock) |
Jeff Garzik | 2cc432e | 2006-03-23 00:32:03 -0500 | [diff] [blame] | 321 | */ |
Jeff Garzik | 2cc432e | 2006-03-23 00:32:03 -0500 | [diff] [blame] | 322 | void ata_bmdma_irq_clear(struct ata_port *ap) |
| 323 | { |
Tejun Heo | 0d5ff56 | 2007-02-01 15:06:36 +0900 | [diff] [blame] | 324 | void __iomem *mmio = ap->ioaddr.bmdma_addr; |
| 325 | |
| 326 | if (!mmio) |
Jeff Garzik | 2cc432e | 2006-03-23 00:32:03 -0500 | [diff] [blame] | 327 | return; |
| 328 | |
Tejun Heo | 0d5ff56 | 2007-02-01 15:06:36 +0900 | [diff] [blame] | 329 | iowrite8(ioread8(mmio + ATA_DMA_STATUS), mmio + ATA_DMA_STATUS); |
Jeff Garzik | 2cc432e | 2006-03-23 00:32:03 -0500 | [diff] [blame] | 330 | } |
| 331 | |
Jeff Garzik | 2cc432e | 2006-03-23 00:32:03 -0500 | [diff] [blame] | 332 | /** |
| 333 | * ata_bmdma_status - Read PCI IDE BMDMA status |
| 334 | * @ap: Port associated with this ATA transaction. |
| 335 | * |
| 336 | * Read and return BMDMA status register. |
| 337 | * |
| 338 | * May be used as the bmdma_status() entry in ata_port_operations. |
| 339 | * |
| 340 | * LOCKING: |
Jeff Garzik | cca3974 | 2006-08-24 03:19:22 -0400 | [diff] [blame] | 341 | * spin_lock_irqsave(host lock) |
Jeff Garzik | 2cc432e | 2006-03-23 00:32:03 -0500 | [diff] [blame] | 342 | */ |
Jeff Garzik | 2cc432e | 2006-03-23 00:32:03 -0500 | [diff] [blame] | 343 | u8 ata_bmdma_status(struct ata_port *ap) |
| 344 | { |
Tejun Heo | 0d5ff56 | 2007-02-01 15:06:36 +0900 | [diff] [blame] | 345 | return ioread8(ap->ioaddr.bmdma_addr + ATA_DMA_STATUS); |
Jeff Garzik | 2cc432e | 2006-03-23 00:32:03 -0500 | [diff] [blame] | 346 | } |
| 347 | |
Jeff Garzik | 2cc432e | 2006-03-23 00:32:03 -0500 | [diff] [blame] | 348 | /** |
| 349 | * ata_bmdma_stop - Stop PCI IDE BMDMA transfer |
| 350 | * @qc: Command we are ending DMA for |
| 351 | * |
| 352 | * Clears the ATA_DMA_START flag in the dma control register |
| 353 | * |
| 354 | * May be used as the bmdma_stop() entry in ata_port_operations. |
| 355 | * |
| 356 | * LOCKING: |
Jeff Garzik | cca3974 | 2006-08-24 03:19:22 -0400 | [diff] [blame] | 357 | * spin_lock_irqsave(host lock) |
Jeff Garzik | 2cc432e | 2006-03-23 00:32:03 -0500 | [diff] [blame] | 358 | */ |
Jeff Garzik | 2cc432e | 2006-03-23 00:32:03 -0500 | [diff] [blame] | 359 | void ata_bmdma_stop(struct ata_queued_cmd *qc) |
| 360 | { |
| 361 | struct ata_port *ap = qc->ap; |
Tejun Heo | 0d5ff56 | 2007-02-01 15:06:36 +0900 | [diff] [blame] | 362 | void __iomem *mmio = ap->ioaddr.bmdma_addr; |
Jeff Garzik | 2cc432e | 2006-03-23 00:32:03 -0500 | [diff] [blame] | 363 | |
Tejun Heo | 0d5ff56 | 2007-02-01 15:06:36 +0900 | [diff] [blame] | 364 | /* clear start/stop bit */ |
| 365 | iowrite8(ioread8(mmio + ATA_DMA_CMD) & ~ATA_DMA_START, |
| 366 | mmio + ATA_DMA_CMD); |
Jeff Garzik | 2cc432e | 2006-03-23 00:32:03 -0500 | [diff] [blame] | 367 | |
| 368 | /* one-PIO-cycle guaranteed wait, per spec, for HDMA1:0 transition */ |
| 369 | ata_altstatus(ap); /* dummy read */ |
| 370 | } |
| 371 | |
Tejun Heo | 6d97dbd | 2006-05-15 20:58:24 +0900 | [diff] [blame] | 372 | /** |
| 373 | * ata_bmdma_freeze - Freeze BMDMA controller port |
| 374 | * @ap: port to freeze |
| 375 | * |
| 376 | * Freeze BMDMA controller port. |
| 377 | * |
| 378 | * LOCKING: |
| 379 | * Inherited from caller. |
| 380 | */ |
| 381 | void ata_bmdma_freeze(struct ata_port *ap) |
| 382 | { |
| 383 | struct ata_ioports *ioaddr = &ap->ioaddr; |
| 384 | |
| 385 | ap->ctl |= ATA_NIEN; |
| 386 | ap->last_ctl = ap->ctl; |
| 387 | |
Tejun Heo | 0d5ff56 | 2007-02-01 15:06:36 +0900 | [diff] [blame] | 388 | iowrite8(ap->ctl, ioaddr->ctl_addr); |
Tejun Heo | 0f0a3ad | 2006-11-17 12:24:22 +0900 | [diff] [blame] | 389 | |
| 390 | /* Under certain circumstances, some controllers raise IRQ on |
| 391 | * ATA_NIEN manipulation. Also, many controllers fail to mask |
| 392 | * previously pending IRQ on ATA_NIEN assertion. Clear it. |
| 393 | */ |
| 394 | ata_chk_status(ap); |
| 395 | |
| 396 | ap->ops->irq_clear(ap); |
Tejun Heo | 6d97dbd | 2006-05-15 20:58:24 +0900 | [diff] [blame] | 397 | } |
| 398 | |
| 399 | /** |
| 400 | * ata_bmdma_thaw - Thaw BMDMA controller port |
| 401 | * @ap: port to thaw |
| 402 | * |
| 403 | * Thaw BMDMA controller port. |
| 404 | * |
| 405 | * LOCKING: |
| 406 | * Inherited from caller. |
| 407 | */ |
| 408 | void ata_bmdma_thaw(struct ata_port *ap) |
| 409 | { |
| 410 | /* clear & re-enable interrupts */ |
| 411 | ata_chk_status(ap); |
| 412 | ap->ops->irq_clear(ap); |
Akira Iguchi | 8362500 | 2007-01-26 16:27:32 +0900 | [diff] [blame] | 413 | ap->ops->irq_on(ap); |
Tejun Heo | 6d97dbd | 2006-05-15 20:58:24 +0900 | [diff] [blame] | 414 | } |
| 415 | |
| 416 | /** |
| 417 | * ata_bmdma_drive_eh - Perform EH with given methods for BMDMA controller |
| 418 | * @ap: port to handle error for |
Tejun Heo | f5914a4 | 2006-05-31 18:27:48 +0900 | [diff] [blame] | 419 | * @prereset: prereset method (can be NULL) |
Tejun Heo | 6d97dbd | 2006-05-15 20:58:24 +0900 | [diff] [blame] | 420 | * @softreset: softreset method (can be NULL) |
| 421 | * @hardreset: hardreset method (can be NULL) |
| 422 | * @postreset: postreset method (can be NULL) |
| 423 | * |
| 424 | * Handle error for ATA BMDMA controller. It can handle both |
| 425 | * PATA and SATA controllers. Many controllers should be able to |
| 426 | * use this EH as-is or with some added handling before and |
| 427 | * after. |
| 428 | * |
| 429 | * This function is intended to be used for constructing |
| 430 | * ->error_handler callback by low level drivers. |
| 431 | * |
| 432 | * LOCKING: |
| 433 | * Kernel thread context (may sleep) |
| 434 | */ |
Tejun Heo | f5914a4 | 2006-05-31 18:27:48 +0900 | [diff] [blame] | 435 | void ata_bmdma_drive_eh(struct ata_port *ap, ata_prereset_fn_t prereset, |
| 436 | ata_reset_fn_t softreset, ata_reset_fn_t hardreset, |
| 437 | ata_postreset_fn_t postreset) |
Tejun Heo | 6d97dbd | 2006-05-15 20:58:24 +0900 | [diff] [blame] | 438 | { |
Tejun Heo | 6d97dbd | 2006-05-15 20:58:24 +0900 | [diff] [blame] | 439 | struct ata_queued_cmd *qc; |
| 440 | unsigned long flags; |
| 441 | int thaw = 0; |
| 442 | |
| 443 | qc = __ata_qc_from_tag(ap, ap->active_tag); |
| 444 | if (qc && !(qc->flags & ATA_QCFLAG_FAILED)) |
| 445 | qc = NULL; |
| 446 | |
| 447 | /* reset PIO HSM and stop DMA engine */ |
Jeff Garzik | ba6a130 | 2006-06-22 23:46:10 -0400 | [diff] [blame] | 448 | spin_lock_irqsave(ap->lock, flags); |
Tejun Heo | 6d97dbd | 2006-05-15 20:58:24 +0900 | [diff] [blame] | 449 | |
Tejun Heo | 6d97dbd | 2006-05-15 20:58:24 +0900 | [diff] [blame] | 450 | ap->hsm_task_state = HSM_ST_IDLE; |
| 451 | |
| 452 | if (qc && (qc->tf.protocol == ATA_PROT_DMA || |
| 453 | qc->tf.protocol == ATA_PROT_ATAPI_DMA)) { |
| 454 | u8 host_stat; |
| 455 | |
Robert Hancock | fbbb262 | 2006-10-27 19:08:41 -0700 | [diff] [blame] | 456 | host_stat = ap->ops->bmdma_status(ap); |
Tejun Heo | 6d97dbd | 2006-05-15 20:58:24 +0900 | [diff] [blame] | 457 | |
Tejun Heo | 6d97dbd | 2006-05-15 20:58:24 +0900 | [diff] [blame] | 458 | /* BMDMA controllers indicate host bus error by |
| 459 | * setting DMA_ERR bit and timing out. As it wasn't |
| 460 | * really a timeout event, adjust error mask and |
| 461 | * cancel frozen state. |
| 462 | */ |
Alan | 18d90de | 2007-01-24 11:42:38 +0000 | [diff] [blame] | 463 | if (qc->err_mask == AC_ERR_TIMEOUT && (host_stat & ATA_DMA_ERR)) { |
Tejun Heo | 6d97dbd | 2006-05-15 20:58:24 +0900 | [diff] [blame] | 464 | qc->err_mask = AC_ERR_HOST_BUS; |
| 465 | thaw = 1; |
| 466 | } |
| 467 | |
| 468 | ap->ops->bmdma_stop(qc); |
| 469 | } |
| 470 | |
| 471 | ata_altstatus(ap); |
| 472 | ata_chk_status(ap); |
| 473 | ap->ops->irq_clear(ap); |
| 474 | |
Jeff Garzik | ba6a130 | 2006-06-22 23:46:10 -0400 | [diff] [blame] | 475 | spin_unlock_irqrestore(ap->lock, flags); |
Tejun Heo | 6d97dbd | 2006-05-15 20:58:24 +0900 | [diff] [blame] | 476 | |
| 477 | if (thaw) |
| 478 | ata_eh_thaw_port(ap); |
| 479 | |
| 480 | /* PIO and DMA engines have been stopped, perform recovery */ |
Tejun Heo | f5914a4 | 2006-05-31 18:27:48 +0900 | [diff] [blame] | 481 | ata_do_eh(ap, prereset, softreset, hardreset, postreset); |
Tejun Heo | 6d97dbd | 2006-05-15 20:58:24 +0900 | [diff] [blame] | 482 | } |
| 483 | |
| 484 | /** |
| 485 | * ata_bmdma_error_handler - Stock error handler for BMDMA controller |
| 486 | * @ap: port to handle error for |
| 487 | * |
| 488 | * Stock error handler for BMDMA controller. |
| 489 | * |
| 490 | * LOCKING: |
| 491 | * Kernel thread context (may sleep) |
| 492 | */ |
| 493 | void ata_bmdma_error_handler(struct ata_port *ap) |
| 494 | { |
| 495 | ata_reset_fn_t hardreset; |
| 496 | |
| 497 | hardreset = NULL; |
| 498 | if (sata_scr_valid(ap)) |
| 499 | hardreset = sata_std_hardreset; |
| 500 | |
Tejun Heo | f5914a4 | 2006-05-31 18:27:48 +0900 | [diff] [blame] | 501 | ata_bmdma_drive_eh(ap, ata_std_prereset, ata_std_softreset, hardreset, |
| 502 | ata_std_postreset); |
Tejun Heo | 6d97dbd | 2006-05-15 20:58:24 +0900 | [diff] [blame] | 503 | } |
| 504 | |
| 505 | /** |
| 506 | * ata_bmdma_post_internal_cmd - Stock post_internal_cmd for |
| 507 | * BMDMA controller |
| 508 | * @qc: internal command to clean up |
| 509 | * |
| 510 | * LOCKING: |
| 511 | * Kernel thread context (may sleep) |
| 512 | */ |
| 513 | void ata_bmdma_post_internal_cmd(struct ata_queued_cmd *qc) |
| 514 | { |
Alan | 61dd08c | 2007-01-25 15:09:05 +0000 | [diff] [blame] | 515 | if (qc->ap->ioaddr.bmdma_addr) |
| 516 | ata_bmdma_stop(qc); |
Tejun Heo | 6d97dbd | 2006-05-15 20:58:24 +0900 | [diff] [blame] | 517 | } |
| 518 | |
Jeff Garzik | 1fdffbc | 2006-02-09 05:15:27 -0500 | [diff] [blame] | 519 | #ifdef CONFIG_PCI |
Alan | 4112e16 | 2007-01-08 12:10:05 +0000 | [diff] [blame] | 520 | |
| 521 | static int ata_resources_present(struct pci_dev *pdev, int port) |
| 522 | { |
| 523 | int i; |
Jeff Garzik | a84471f | 2007-02-26 05:51:33 -0500 | [diff] [blame] | 524 | |
Alan | 4112e16 | 2007-01-08 12:10:05 +0000 | [diff] [blame] | 525 | /* Check the PCI resources for this channel are enabled */ |
| 526 | port = port * 2; |
| 527 | for (i = 0; i < 2; i ++) { |
| 528 | if (pci_resource_start(pdev, port + i) == 0 || |
Tejun Heo | 55a6ade | 2007-03-09 19:43:35 +0900 | [diff] [blame] | 529 | pci_resource_len(pdev, port + i) == 0) |
| 530 | return 0; |
Alan | 4112e16 | 2007-01-08 12:10:05 +0000 | [diff] [blame] | 531 | } |
| 532 | return 1; |
| 533 | } |
Jeff Garzik | a84471f | 2007-02-26 05:51:33 -0500 | [diff] [blame] | 534 | |
Jeff Garzik | 1fdffbc | 2006-02-09 05:15:27 -0500 | [diff] [blame] | 535 | /** |
| 536 | * ata_pci_init_native_mode - Initialize native-mode driver |
| 537 | * @pdev: pci device to be initialized |
| 538 | * @port: array[2] of pointers to port info structures. |
| 539 | * @ports: bitmap of ports present |
| 540 | * |
| 541 | * Utility function which allocates and initializes an |
| 542 | * ata_probe_ent structure for a standard dual-port |
| 543 | * PIO-based IDE controller. The returned ata_probe_ent |
| 544 | * structure can be passed to ata_device_add(). The returned |
| 545 | * ata_probe_ent structure should then be freed with kfree(). |
| 546 | * |
| 547 | * The caller need only pass the address of the primary port, the |
| 548 | * secondary will be deduced automatically. If the device has non |
| 549 | * standard secondary port mappings this function can be called twice, |
| 550 | * once for each interface. |
| 551 | */ |
| 552 | |
| 553 | struct ata_probe_ent * |
| 554 | ata_pci_init_native_mode(struct pci_dev *pdev, struct ata_port_info **port, int ports) |
| 555 | { |
Tejun Heo | 0d5ff56 | 2007-02-01 15:06:36 +0900 | [diff] [blame] | 556 | struct ata_probe_ent *probe_ent; |
Tejun Heo | 55a6ade | 2007-03-09 19:43:35 +0900 | [diff] [blame] | 557 | int i; |
Tejun Heo | 0d5ff56 | 2007-02-01 15:06:36 +0900 | [diff] [blame] | 558 | void __iomem * const *iomap; |
Jeff Garzik | 1fdffbc | 2006-02-09 05:15:27 -0500 | [diff] [blame] | 559 | |
Tejun Heo | 55a6ade | 2007-03-09 19:43:35 +0900 | [diff] [blame] | 560 | /* Discard disabled ports. Some controllers show their unused |
| 561 | * channels this way. Disabled ports will be made dummy. |
| 562 | */ |
| 563 | if (ata_resources_present(pdev, 0) == 0) |
| 564 | ports &= ~ATA_PORT_PRIMARY; |
| 565 | if (ata_resources_present(pdev, 1) == 0) |
| 566 | ports &= ~ATA_PORT_SECONDARY; |
| 567 | |
| 568 | if (!ports) { |
| 569 | dev_printk(KERN_ERR, &pdev->dev, "no available port\n"); |
| 570 | return NULL; |
| 571 | } |
| 572 | |
Tejun Heo | 0d5ff56 | 2007-02-01 15:06:36 +0900 | [diff] [blame] | 573 | /* iomap BARs */ |
| 574 | for (i = 0; i < 4; i++) { |
Tejun Heo | 55a6ade | 2007-03-09 19:43:35 +0900 | [diff] [blame] | 575 | if (!(ports & (1 << (i / 2)))) |
| 576 | continue; |
Tejun Heo | 0d5ff56 | 2007-02-01 15:06:36 +0900 | [diff] [blame] | 577 | if (pcim_iomap(pdev, i, 0) == NULL) { |
| 578 | dev_printk(KERN_ERR, &pdev->dev, |
| 579 | "failed to iomap PCI BAR %d\n", i); |
| 580 | return NULL; |
| 581 | } |
| 582 | } |
| 583 | |
| 584 | pcim_iomap(pdev, 4, 0); /* may fail */ |
| 585 | iomap = pcim_iomap_table(pdev); |
| 586 | |
| 587 | /* alloc and init probe_ent */ |
| 588 | probe_ent = ata_probe_ent_alloc(pci_dev_to_dev(pdev), port[0]); |
Jeff Garzik | 1fdffbc | 2006-02-09 05:15:27 -0500 | [diff] [blame] | 589 | if (!probe_ent) |
| 590 | return NULL; |
| 591 | |
Tejun Heo | 55a6ade | 2007-03-09 19:43:35 +0900 | [diff] [blame] | 592 | probe_ent->n_ports = 2; |
Jeff Garzik | 1fdffbc | 2006-02-09 05:15:27 -0500 | [diff] [blame] | 593 | probe_ent->irq = pdev->irq; |
Thomas Gleixner | 1d6f359 | 2006-07-01 19:29:42 -0700 | [diff] [blame] | 594 | probe_ent->irq_flags = IRQF_SHARED; |
Jeff Garzik | a84471f | 2007-02-26 05:51:33 -0500 | [diff] [blame] | 595 | |
Jeff Garzik | 1fdffbc | 2006-02-09 05:15:27 -0500 | [diff] [blame] | 596 | if (ports & ATA_PORT_PRIMARY) { |
Tejun Heo | 55a6ade | 2007-03-09 19:43:35 +0900 | [diff] [blame] | 597 | probe_ent->port[0].cmd_addr = iomap[0]; |
| 598 | probe_ent->port[0].altstatus_addr = |
| 599 | probe_ent->port[0].ctl_addr = (void __iomem *) |
Tejun Heo | 0d5ff56 | 2007-02-01 15:06:36 +0900 | [diff] [blame] | 600 | ((unsigned long)iomap[1] | ATA_PCI_CTL_OFS); |
| 601 | if (iomap[4]) { |
Tejun Heo | 55a6ade | 2007-03-09 19:43:35 +0900 | [diff] [blame] | 602 | if ((!(port[0]->flags & ATA_FLAG_IGN_SIMPLEX)) && |
Tejun Heo | 0d5ff56 | 2007-02-01 15:06:36 +0900 | [diff] [blame] | 603 | (ioread8(iomap[4] + 2) & 0x80)) |
Jeff Garzik | cca3974 | 2006-08-24 03:19:22 -0400 | [diff] [blame] | 604 | probe_ent->_host_flags |= ATA_HOST_SIMPLEX; |
Tejun Heo | 55a6ade | 2007-03-09 19:43:35 +0900 | [diff] [blame] | 605 | probe_ent->port[0].bmdma_addr = iomap[4]; |
Alan Cox | 4e5ec5d | 2006-03-27 18:42:40 +0100 | [diff] [blame] | 606 | } |
Tejun Heo | 55a6ade | 2007-03-09 19:43:35 +0900 | [diff] [blame] | 607 | ata_std_ports(&probe_ent->port[0]); |
| 608 | } else |
| 609 | probe_ent->dummy_port_mask |= ATA_PORT_PRIMARY; |
Jeff Garzik | 1fdffbc | 2006-02-09 05:15:27 -0500 | [diff] [blame] | 610 | |
| 611 | if (ports & ATA_PORT_SECONDARY) { |
Tejun Heo | 55a6ade | 2007-03-09 19:43:35 +0900 | [diff] [blame] | 612 | probe_ent->port[1].cmd_addr = iomap[2]; |
| 613 | probe_ent->port[1].altstatus_addr = |
| 614 | probe_ent->port[1].ctl_addr = (void __iomem *) |
Tejun Heo | 0d5ff56 | 2007-02-01 15:06:36 +0900 | [diff] [blame] | 615 | ((unsigned long)iomap[3] | ATA_PCI_CTL_OFS); |
| 616 | if (iomap[4]) { |
Tejun Heo | 55a6ade | 2007-03-09 19:43:35 +0900 | [diff] [blame] | 617 | if ((!(port[1]->flags & ATA_FLAG_IGN_SIMPLEX)) && |
Tejun Heo | 0d5ff56 | 2007-02-01 15:06:36 +0900 | [diff] [blame] | 618 | (ioread8(iomap[4] + 10) & 0x80)) |
Jeff Garzik | cca3974 | 2006-08-24 03:19:22 -0400 | [diff] [blame] | 619 | probe_ent->_host_flags |= ATA_HOST_SIMPLEX; |
Tejun Heo | 55a6ade | 2007-03-09 19:43:35 +0900 | [diff] [blame] | 620 | probe_ent->port[1].bmdma_addr = iomap[4] + 8; |
Alan Cox | 4e5ec5d | 2006-03-27 18:42:40 +0100 | [diff] [blame] | 621 | } |
Tejun Heo | 55a6ade | 2007-03-09 19:43:35 +0900 | [diff] [blame] | 622 | ata_std_ports(&probe_ent->port[1]); |
Tejun Heo | fea63e3 | 2006-09-16 03:04:15 +0900 | [diff] [blame] | 623 | probe_ent->pinfo2 = port[1]; |
Tejun Heo | 55a6ade | 2007-03-09 19:43:35 +0900 | [diff] [blame] | 624 | } else |
| 625 | probe_ent->dummy_port_mask |= ATA_PORT_SECONDARY; |
Jeff Garzik | 1fdffbc | 2006-02-09 05:15:27 -0500 | [diff] [blame] | 626 | |
Jeff Garzik | 1fdffbc | 2006-02-09 05:15:27 -0500 | [diff] [blame] | 627 | return probe_ent; |
| 628 | } |
| 629 | |
Tejun Heo | 0f834de | 2007-04-17 23:44:07 +0900 | [diff] [blame^] | 630 | /** |
| 631 | * ata_pci_init_bmdma - acquire PCI BMDMA resources and init ATA host |
| 632 | * @host: target ATA host |
| 633 | * |
| 634 | * Acquire PCI BMDMA resources and initialize @host accordingly. |
| 635 | * |
| 636 | * LOCKING: |
| 637 | * Inherited from calling layer (may sleep). |
| 638 | * |
| 639 | * RETURNS: |
| 640 | * 0 on success, -errno otherwise. |
| 641 | */ |
| 642 | static int ata_pci_init_bmdma(struct ata_host *host) |
Jeff Garzik | 1fdffbc | 2006-02-09 05:15:27 -0500 | [diff] [blame] | 643 | { |
Tejun Heo | 0f834de | 2007-04-17 23:44:07 +0900 | [diff] [blame^] | 644 | struct device *gdev = host->dev; |
| 645 | struct pci_dev *pdev = to_pci_dev(gdev); |
| 646 | int i, rc; |
Jeff Garzik | 1fdffbc | 2006-02-09 05:15:27 -0500 | [diff] [blame] | 647 | |
Tejun Heo | 0f834de | 2007-04-17 23:44:07 +0900 | [diff] [blame^] | 648 | /* TODO: If we get no DMA mask we should fall back to PIO */ |
| 649 | rc = pci_set_dma_mask(pdev, ATA_DMA_MASK); |
| 650 | if (rc) |
| 651 | return rc; |
| 652 | rc = pci_set_consistent_dma_mask(pdev, ATA_DMA_MASK); |
| 653 | if (rc) |
| 654 | return rc; |
| 655 | |
| 656 | /* request and iomap DMA region */ |
| 657 | rc = pcim_iomap_regions(pdev, 1 << 4, DRV_NAME); |
| 658 | if (rc) { |
| 659 | dev_printk(KERN_ERR, gdev, "failed to request/iomap BAR4\n"); |
| 660 | return -ENOMEM; |
| 661 | } |
| 662 | host->iomap = pcim_iomap_table(pdev); |
| 663 | |
| 664 | for (i = 0; i < 2; i++) { |
| 665 | struct ata_port *ap = host->ports[i]; |
| 666 | struct ata_ioports *ioaddr = &ap->ioaddr; |
| 667 | void __iomem *bmdma = host->iomap[4] + 8 * i; |
| 668 | |
| 669 | if (ata_port_is_dummy(ap)) |
| 670 | continue; |
| 671 | |
| 672 | ioaddr->bmdma_addr = bmdma; |
| 673 | if ((!(ap->flags & ATA_FLAG_IGN_SIMPLEX)) && |
| 674 | (ioread8(bmdma + 2) & 0x80)) |
| 675 | host->flags |= ATA_HOST_SIMPLEX; |
Tejun Heo | 0d5ff56 | 2007-02-01 15:06:36 +0900 | [diff] [blame] | 676 | } |
| 677 | |
Tejun Heo | 0f834de | 2007-04-17 23:44:07 +0900 | [diff] [blame^] | 678 | return 0; |
Jeff Garzik | 1fdffbc | 2006-02-09 05:15:27 -0500 | [diff] [blame] | 679 | } |
| 680 | |
Tejun Heo | 0f834de | 2007-04-17 23:44:07 +0900 | [diff] [blame^] | 681 | struct ata_legacy_devres { |
| 682 | unsigned int mask; |
| 683 | unsigned long cmd_port[2]; |
| 684 | void __iomem * cmd_addr[2]; |
| 685 | void __iomem * ctl_addr[2]; |
| 686 | unsigned int irq[2]; |
| 687 | void * irq_dev_id[2]; |
| 688 | }; |
| 689 | |
| 690 | static void ata_legacy_free_irqs(struct ata_legacy_devres *legacy_dr) |
| 691 | { |
| 692 | int i; |
| 693 | |
| 694 | for (i = 0; i < 2; i++) { |
| 695 | if (!legacy_dr->irq[i]) |
| 696 | continue; |
| 697 | |
| 698 | free_irq(legacy_dr->irq[i], legacy_dr->irq_dev_id[i]); |
| 699 | legacy_dr->irq[i] = 0; |
| 700 | legacy_dr->irq_dev_id[i] = NULL; |
| 701 | } |
| 702 | } |
| 703 | |
| 704 | static void ata_legacy_release(struct device *gdev, void *res) |
| 705 | { |
| 706 | struct ata_legacy_devres *this = res; |
| 707 | int i; |
| 708 | |
| 709 | ata_legacy_free_irqs(this); |
| 710 | |
| 711 | for (i = 0; i < 2; i++) { |
| 712 | if (this->cmd_addr[i]) |
| 713 | ioport_unmap(this->cmd_addr[i]); |
| 714 | if (this->ctl_addr[i]) |
| 715 | ioport_unmap(this->ctl_addr[i]); |
| 716 | if (this->cmd_port[i]) |
| 717 | release_region(this->cmd_port[i], 8); |
| 718 | } |
| 719 | } |
| 720 | |
| 721 | static int ata_init_legacy_port(struct ata_port *ap, |
| 722 | struct ata_legacy_devres *legacy_dr) |
| 723 | { |
| 724 | struct ata_host *host = ap->host; |
| 725 | int port_no = ap->port_no; |
| 726 | unsigned long cmd_port, ctl_port; |
| 727 | |
| 728 | if (port_no == 0) { |
| 729 | cmd_port = ATA_PRIMARY_CMD; |
| 730 | ctl_port = ATA_PRIMARY_CTL; |
| 731 | } else { |
| 732 | cmd_port = ATA_SECONDARY_CMD; |
| 733 | ctl_port = ATA_SECONDARY_CTL; |
| 734 | } |
| 735 | |
| 736 | /* request cmd_port */ |
| 737 | if (request_region(cmd_port, 8, "libata")) |
| 738 | legacy_dr->cmd_port[port_no] = cmd_port; |
| 739 | else { |
| 740 | dev_printk(KERN_WARNING, host->dev, |
| 741 | "0x%0lX IDE port busy\n", cmd_port); |
| 742 | return -EBUSY; |
| 743 | } |
| 744 | |
| 745 | /* iomap cmd and ctl ports */ |
| 746 | legacy_dr->cmd_addr[port_no] = ioport_map(cmd_port, 8); |
| 747 | legacy_dr->ctl_addr[port_no] = ioport_map(ctl_port, 1); |
| 748 | if (!legacy_dr->cmd_addr[port_no] || !legacy_dr->ctl_addr[port_no]) |
| 749 | return -ENOMEM; |
| 750 | |
| 751 | /* init IO addresses */ |
| 752 | ap->ioaddr.cmd_addr = legacy_dr->cmd_addr[port_no]; |
| 753 | ap->ioaddr.altstatus_addr = legacy_dr->ctl_addr[port_no]; |
| 754 | ap->ioaddr.ctl_addr = legacy_dr->ctl_addr[port_no]; |
| 755 | ata_std_ports(&ap->ioaddr); |
| 756 | |
| 757 | return 0; |
| 758 | } |
| 759 | |
| 760 | /** |
| 761 | * ata_init_legacy_host - acquire legacy ATA resources and init ATA host |
| 762 | * @host: target ATA host |
| 763 | * @legacy_mask: out parameter, mask indicating ports is in legacy mode |
| 764 | * @was_busy: out parameter, indicates whether any port was busy |
| 765 | * |
| 766 | * Acquire legacy ATA resources for ports. |
| 767 | * |
| 768 | * LOCKING: |
| 769 | * Inherited from calling layer (may sleep). |
| 770 | * |
| 771 | * RETURNS: |
| 772 | * 0 on success, -errno otherwise. |
| 773 | */ |
| 774 | static int ata_init_legacy_host(struct ata_host *host, |
| 775 | unsigned int *legacy_mask, int *was_busy) |
| 776 | { |
| 777 | struct device *gdev = host->dev; |
| 778 | struct ata_legacy_devres *legacy_dr; |
| 779 | int i, rc; |
| 780 | |
| 781 | if (!devres_open_group(gdev, NULL, GFP_KERNEL)) |
| 782 | return -ENOMEM; |
| 783 | |
| 784 | rc = -ENOMEM; |
| 785 | legacy_dr = devres_alloc(ata_legacy_release, sizeof(*legacy_dr), |
| 786 | GFP_KERNEL); |
| 787 | if (!legacy_dr) |
| 788 | goto err_out; |
| 789 | devres_add(gdev, legacy_dr); |
| 790 | |
| 791 | for (i = 0; i < 2; i++) { |
| 792 | *legacy_mask &= ~(1 << i); |
| 793 | rc = ata_init_legacy_port(host->ports[i], legacy_dr); |
| 794 | if (rc == 0) |
| 795 | legacy_dr->mask |= 1 << i; |
| 796 | else if (rc == -EBUSY) |
| 797 | (*was_busy)++; |
| 798 | } |
| 799 | |
| 800 | if (!legacy_dr->mask) |
| 801 | return -EBUSY; |
| 802 | |
| 803 | for (i = 0; i < 2; i++) |
| 804 | if (!(legacy_dr->mask & (1 << i))) |
| 805 | host->ports[i]->ops = &ata_dummy_port_ops; |
| 806 | |
| 807 | *legacy_mask |= legacy_dr->mask; |
| 808 | |
| 809 | devres_remove_group(gdev, NULL); |
| 810 | return 0; |
| 811 | |
| 812 | err_out: |
| 813 | devres_release_group(gdev, NULL); |
| 814 | return rc; |
| 815 | } |
| 816 | |
| 817 | /** |
| 818 | * ata_request_legacy_irqs - request legacy ATA IRQs |
| 819 | * @host: target ATA host |
| 820 | * @handler: array of IRQ handlers |
| 821 | * @irq_flags: array of IRQ flags |
| 822 | * @dev_id: array of IRQ dev_ids |
| 823 | * |
| 824 | * Request legacy IRQs for non-dummy legacy ports in @host. All |
| 825 | * IRQ parameters are passed as array to allow ports to have |
| 826 | * separate IRQ handlers. |
| 827 | * |
| 828 | * LOCKING: |
| 829 | * Inherited from calling layer (may sleep). |
| 830 | * |
| 831 | * RETURNS: |
| 832 | * 0 on success, -errno otherwise. |
| 833 | */ |
| 834 | static int ata_request_legacy_irqs(struct ata_host *host, |
| 835 | irq_handler_t const *handler, |
| 836 | const unsigned int *irq_flags, |
| 837 | void * const *dev_id) |
| 838 | { |
| 839 | struct device *gdev = host->dev; |
| 840 | struct ata_legacy_devres *legacy_dr; |
| 841 | int i, rc; |
| 842 | |
| 843 | legacy_dr = devres_find(host->dev, ata_legacy_release, NULL, NULL); |
| 844 | BUG_ON(!legacy_dr); |
| 845 | |
| 846 | for (i = 0; i < 2; i++) { |
| 847 | unsigned int irq; |
| 848 | |
| 849 | /* FIXME: ATA_*_IRQ() should take generic device not pci_dev */ |
| 850 | if (i == 0) |
| 851 | irq = ATA_PRIMARY_IRQ(to_pci_dev(gdev)); |
| 852 | else |
| 853 | irq = ATA_SECONDARY_IRQ(to_pci_dev(gdev)); |
| 854 | |
| 855 | if (!(legacy_dr->mask & (1 << i))) |
| 856 | continue; |
| 857 | |
| 858 | if (!handler[i]) { |
| 859 | dev_printk(KERN_ERR, gdev, |
| 860 | "NULL handler specified for port %d\n", i); |
| 861 | rc = -EINVAL; |
| 862 | goto err_out; |
| 863 | } |
| 864 | |
| 865 | rc = request_irq(irq, handler[i], irq_flags[i], DRV_NAME, |
| 866 | dev_id[i]); |
| 867 | if (rc) { |
| 868 | dev_printk(KERN_ERR, gdev, |
| 869 | "irq %u request failed (errno=%d)\n", irq, rc); |
| 870 | goto err_out; |
| 871 | } |
| 872 | |
| 873 | /* record irq allocation in legacy_dr */ |
| 874 | legacy_dr->irq[i] = irq; |
| 875 | legacy_dr->irq_dev_id[i] = dev_id[i]; |
| 876 | |
| 877 | /* only used to print info */ |
| 878 | if (i == 0) |
| 879 | host->irq = irq; |
| 880 | else |
| 881 | host->irq2 = irq; |
| 882 | } |
| 883 | |
| 884 | return 0; |
| 885 | |
| 886 | err_out: |
| 887 | ata_legacy_free_irqs(legacy_dr); |
| 888 | return rc; |
| 889 | } |
Jeff Garzik | 1fdffbc | 2006-02-09 05:15:27 -0500 | [diff] [blame] | 890 | |
| 891 | /** |
| 892 | * ata_pci_init_one - Initialize/register PCI IDE host controller |
| 893 | * @pdev: Controller to be initialized |
| 894 | * @port_info: Information from low-level host driver |
| 895 | * @n_ports: Number of ports attached to host controller |
| 896 | * |
| 897 | * This is a helper function which can be called from a driver's |
| 898 | * xxx_init_one() probe function if the hardware uses traditional |
| 899 | * IDE taskfile registers. |
| 900 | * |
| 901 | * This function calls pci_enable_device(), reserves its register |
| 902 | * regions, sets the dma mask, enables bus master mode, and calls |
| 903 | * ata_device_add() |
| 904 | * |
Alan Cox | 2ec7df0 | 2006-08-10 16:59:10 +0900 | [diff] [blame] | 905 | * ASSUMPTION: |
| 906 | * Nobody makes a single channel controller that appears solely as |
| 907 | * the secondary legacy port on PCI. |
| 908 | * |
Jeff Garzik | 1fdffbc | 2006-02-09 05:15:27 -0500 | [diff] [blame] | 909 | * LOCKING: |
| 910 | * Inherited from PCI layer (may sleep). |
| 911 | * |
| 912 | * RETURNS: |
| 913 | * Zero on success, negative on errno-based value on error. |
| 914 | */ |
| 915 | |
| 916 | int ata_pci_init_one (struct pci_dev *pdev, struct ata_port_info **port_info, |
| 917 | unsigned int n_ports) |
| 918 | { |
Tejun Heo | f0d36ef | 2007-01-20 16:00:28 +0900 | [diff] [blame] | 919 | struct device *dev = &pdev->dev; |
Alan Cox | 2ec7df0 | 2006-08-10 16:59:10 +0900 | [diff] [blame] | 920 | struct ata_probe_ent *probe_ent = NULL; |
Tejun Heo | 0f834de | 2007-04-17 23:44:07 +0900 | [diff] [blame^] | 921 | struct ata_host *host = NULL; |
| 922 | const struct ata_port_info *port[2]; |
Jeff Garzik | c791c30 | 2006-09-28 03:40:11 -0400 | [diff] [blame] | 923 | u8 mask; |
Jeff Garzik | 1fdffbc | 2006-02-09 05:15:27 -0500 | [diff] [blame] | 924 | unsigned int legacy_mode = 0; |
Jeff Garzik | 1fdffbc | 2006-02-09 05:15:27 -0500 | [diff] [blame] | 925 | int rc; |
| 926 | |
| 927 | DPRINTK("ENTER\n"); |
| 928 | |
Tejun Heo | f0d36ef | 2007-01-20 16:00:28 +0900 | [diff] [blame] | 929 | if (!devres_open_group(dev, NULL, GFP_KERNEL)) |
| 930 | return -ENOMEM; |
| 931 | |
Jeff Garzik | c791c30 | 2006-09-28 03:40:11 -0400 | [diff] [blame] | 932 | BUG_ON(n_ports < 1 || n_ports > 2); |
| 933 | |
Jeff Garzik | 1fdffbc | 2006-02-09 05:15:27 -0500 | [diff] [blame] | 934 | port[0] = port_info[0]; |
| 935 | if (n_ports > 1) |
| 936 | port[1] = port_info[1]; |
| 937 | else |
| 938 | port[1] = port[0]; |
| 939 | |
Jeff Garzik | 1fdffbc | 2006-02-09 05:15:27 -0500 | [diff] [blame] | 940 | /* FIXME: Really for ATA it isn't safe because the device may be |
| 941 | multi-purpose and we want to leave it alone if it was already |
| 942 | enabled. Secondly for shared use as Arjan says we want refcounting |
| 943 | |
| 944 | Checking dev->is_enabled is insufficient as this is not set at |
| 945 | boot for the primary video which is BIOS enabled |
| 946 | */ |
| 947 | |
Tejun Heo | f0d36ef | 2007-01-20 16:00:28 +0900 | [diff] [blame] | 948 | rc = pcim_enable_device(pdev); |
Jeff Garzik | 1fdffbc | 2006-02-09 05:15:27 -0500 | [diff] [blame] | 949 | if (rc) |
Tejun Heo | f0d36ef | 2007-01-20 16:00:28 +0900 | [diff] [blame] | 950 | goto err_out; |
Jeff Garzik | 1fdffbc | 2006-02-09 05:15:27 -0500 | [diff] [blame] | 951 | |
Jeff Garzik | c791c30 | 2006-09-28 03:40:11 -0400 | [diff] [blame] | 952 | if ((pdev->class >> 8) == PCI_CLASS_STORAGE_IDE) { |
| 953 | u8 tmp8; |
| 954 | |
| 955 | /* TODO: What if one channel is in native mode ... */ |
| 956 | pci_read_config_byte(pdev, PCI_CLASS_PROG, &tmp8); |
| 957 | mask = (1 << 2) | (1 << 0); |
| 958 | if ((tmp8 & mask) != mask) |
| 959 | legacy_mode = (1 << 3); |
Alan Cox | 8eb166b | 2006-10-16 16:24:50 +0100 | [diff] [blame] | 960 | #if defined(CONFIG_NO_ATA_LEGACY) |
| 961 | /* Some platforms with PCI limits cannot address compat |
| 962 | port space. In that case we punt if their firmware has |
| 963 | left a device in compatibility mode */ |
| 964 | if (legacy_mode) { |
| 965 | printk(KERN_ERR "ata: Compatibility mode ATA is not supported on this platform, skipping.\n"); |
Tejun Heo | f0d36ef | 2007-01-20 16:00:28 +0900 | [diff] [blame] | 966 | rc = -EOPNOTSUPP; |
| 967 | goto err_out; |
Alan Cox | 8eb166b | 2006-10-16 16:24:50 +0100 | [diff] [blame] | 968 | } |
| 969 | #endif |
Jeff Garzik | c791c30 | 2006-09-28 03:40:11 -0400 | [diff] [blame] | 970 | } |
| 971 | |
Alan | dc3c337 | 2007-01-02 11:58:34 +0000 | [diff] [blame] | 972 | if (!legacy_mode) { |
| 973 | rc = pci_request_regions(pdev, DRV_NAME); |
| 974 | if (rc) { |
Tejun Heo | f0d36ef | 2007-01-20 16:00:28 +0900 | [diff] [blame] | 975 | pcim_pin_device(pdev); |
Alan | dc3c337 | 2007-01-02 11:58:34 +0000 | [diff] [blame] | 976 | goto err_out; |
| 977 | } |
Tejun Heo | 0f834de | 2007-04-17 23:44:07 +0900 | [diff] [blame^] | 978 | |
| 979 | /* TODO: If we get no DMA mask we should fall back to PIO */ |
| 980 | rc = pci_set_dma_mask(pdev, ATA_DMA_MASK); |
| 981 | if (rc) |
| 982 | goto err_out; |
| 983 | rc = pci_set_consistent_dma_mask(pdev, ATA_DMA_MASK); |
| 984 | if (rc) |
| 985 | goto err_out; |
| 986 | |
| 987 | pci_set_master(pdev); |
Alan | dc3c337 | 2007-01-02 11:58:34 +0000 | [diff] [blame] | 988 | } else { |
Tejun Heo | 0f834de | 2007-04-17 23:44:07 +0900 | [diff] [blame^] | 989 | int was_busy = 0; |
| 990 | |
| 991 | rc = -ENOMEM; |
| 992 | host = ata_host_alloc_pinfo(dev, port, 2); |
| 993 | if (!host) |
| 994 | goto err_out; |
| 995 | |
| 996 | rc = ata_init_legacy_host(host, &legacy_mode, &was_busy); |
| 997 | if (was_busy) |
Jeff Garzik | 8cdfb29 | 2007-03-09 10:54:42 -0500 | [diff] [blame] | 998 | pcim_pin_device(pdev); |
Tejun Heo | 0f834de | 2007-04-17 23:44:07 +0900 | [diff] [blame^] | 999 | if (rc) |
| 1000 | goto err_out; |
Jeff Garzik | 1fdffbc | 2006-02-09 05:15:27 -0500 | [diff] [blame] | 1001 | |
Tejun Heo | 0f834de | 2007-04-17 23:44:07 +0900 | [diff] [blame^] | 1002 | /* request respective PCI regions, may fail */ |
| 1003 | rc = pci_request_region(pdev, 1, DRV_NAME); |
| 1004 | rc = pci_request_region(pdev, 3, DRV_NAME); |
Alan | dc3c337 | 2007-01-02 11:58:34 +0000 | [diff] [blame] | 1005 | |
Tejun Heo | 0f834de | 2007-04-17 23:44:07 +0900 | [diff] [blame^] | 1006 | /* init bmdma */ |
| 1007 | ata_pci_init_bmdma(host); |
| 1008 | pci_set_master(pdev); |
Jeff Garzik | 1fdffbc | 2006-02-09 05:15:27 -0500 | [diff] [blame] | 1009 | } |
| 1010 | |
Jeff Garzik | 1fdffbc | 2006-02-09 05:15:27 -0500 | [diff] [blame] | 1011 | if (legacy_mode) { |
Tejun Heo | 0f834de | 2007-04-17 23:44:07 +0900 | [diff] [blame^] | 1012 | irq_handler_t handler[2] = { host->ops->irq_handler, |
| 1013 | host->ops->irq_handler }; |
| 1014 | unsigned int irq_flags[2] = { IRQF_SHARED, IRQF_SHARED }; |
| 1015 | void *dev_id[2] = { host, host }; |
| 1016 | |
| 1017 | rc = ata_host_start(host); |
| 1018 | if (rc) |
| 1019 | goto err_out; |
| 1020 | |
| 1021 | rc = ata_request_legacy_irqs(host, handler, irq_flags, dev_id); |
| 1022 | if (rc) |
| 1023 | goto err_out; |
| 1024 | |
| 1025 | rc = ata_host_register(host, port_info[0]->sht); |
| 1026 | if (rc) |
| 1027 | goto err_out; |
Jeff Garzik | 1fdffbc | 2006-02-09 05:15:27 -0500 | [diff] [blame] | 1028 | } else { |
| 1029 | if (n_ports == 2) |
Tejun Heo | 0f834de | 2007-04-17 23:44:07 +0900 | [diff] [blame^] | 1030 | probe_ent = ata_pci_init_native_mode(pdev, (struct ata_port_info **)port, ATA_PORT_PRIMARY | ATA_PORT_SECONDARY); |
Jeff Garzik | 1fdffbc | 2006-02-09 05:15:27 -0500 | [diff] [blame] | 1031 | else |
Tejun Heo | 0f834de | 2007-04-17 23:44:07 +0900 | [diff] [blame^] | 1032 | probe_ent = ata_pci_init_native_mode(pdev, (struct ata_port_info **)port, ATA_PORT_PRIMARY); |
Jeff Garzik | 1fdffbc | 2006-02-09 05:15:27 -0500 | [diff] [blame] | 1033 | |
Tejun Heo | 0f834de | 2007-04-17 23:44:07 +0900 | [diff] [blame^] | 1034 | if (!probe_ent) { |
| 1035 | rc = -ENOMEM; |
| 1036 | goto err_out; |
| 1037 | } |
Jeff Garzik | 1fdffbc | 2006-02-09 05:15:27 -0500 | [diff] [blame] | 1038 | |
Tejun Heo | 0f834de | 2007-04-17 23:44:07 +0900 | [diff] [blame^] | 1039 | if (!ata_device_add(probe_ent)) { |
| 1040 | rc = -ENODEV; |
| 1041 | goto err_out; |
| 1042 | } |
| 1043 | |
| 1044 | devm_kfree(dev, probe_ent); |
Jeff Garzik | c791c30 | 2006-09-28 03:40:11 -0400 | [diff] [blame] | 1045 | } |
Tejun Heo | f0d36ef | 2007-01-20 16:00:28 +0900 | [diff] [blame] | 1046 | devres_remove_group(dev, NULL); |
Jeff Garzik | 1fdffbc | 2006-02-09 05:15:27 -0500 | [diff] [blame] | 1047 | return 0; |
| 1048 | |
Jeff Garzik | 1fdffbc | 2006-02-09 05:15:27 -0500 | [diff] [blame] | 1049 | err_out: |
Tejun Heo | f0d36ef | 2007-01-20 16:00:28 +0900 | [diff] [blame] | 1050 | devres_release_group(dev, NULL); |
Jeff Garzik | 1fdffbc | 2006-02-09 05:15:27 -0500 | [diff] [blame] | 1051 | return rc; |
| 1052 | } |
| 1053 | |
Alan Cox | d33d44f | 2006-03-21 15:59:57 +0000 | [diff] [blame] | 1054 | /** |
| 1055 | * ata_pci_clear_simplex - attempt to kick device out of simplex |
| 1056 | * @pdev: PCI device |
| 1057 | * |
| 1058 | * Some PCI ATA devices report simplex mode but in fact can be told to |
Jeff Garzik | 2e9edbf | 2006-03-24 09:56:57 -0500 | [diff] [blame] | 1059 | * enter non simplex mode. This implements the neccessary logic to |
Alan Cox | d33d44f | 2006-03-21 15:59:57 +0000 | [diff] [blame] | 1060 | * perform the task on such devices. Calling it on other devices will |
| 1061 | * have -undefined- behaviour. |
| 1062 | */ |
| 1063 | |
| 1064 | int ata_pci_clear_simplex(struct pci_dev *pdev) |
| 1065 | { |
| 1066 | unsigned long bmdma = pci_resource_start(pdev, 4); |
| 1067 | u8 simplex; |
| 1068 | |
| 1069 | if (bmdma == 0) |
| 1070 | return -ENOENT; |
| 1071 | |
| 1072 | simplex = inb(bmdma + 0x02); |
| 1073 | outb(simplex & 0x60, bmdma + 0x02); |
| 1074 | simplex = inb(bmdma + 0x02); |
| 1075 | if (simplex & 0x80) |
| 1076 | return -EOPNOTSUPP; |
| 1077 | return 0; |
| 1078 | } |
| 1079 | |
Alan Cox | a76b62c | 2007-03-09 09:34:07 -0500 | [diff] [blame] | 1080 | unsigned long ata_pci_default_filter(struct ata_device *adev, unsigned long xfer_mask) |
Alan Cox | d33d44f | 2006-03-21 15:59:57 +0000 | [diff] [blame] | 1081 | { |
| 1082 | /* Filter out DMA modes if the device has been configured by |
| 1083 | the BIOS as PIO only */ |
Jeff Garzik | 2e9edbf | 2006-03-24 09:56:57 -0500 | [diff] [blame] | 1084 | |
Alan Cox | a76b62c | 2007-03-09 09:34:07 -0500 | [diff] [blame] | 1085 | if (adev->ap->ioaddr.bmdma_addr == 0) |
Alan Cox | d33d44f | 2006-03-21 15:59:57 +0000 | [diff] [blame] | 1086 | xfer_mask &= ~(ATA_MASK_MWDMA | ATA_MASK_UDMA); |
| 1087 | return xfer_mask; |
| 1088 | } |
| 1089 | |
Jeff Garzik | 1fdffbc | 2006-02-09 05:15:27 -0500 | [diff] [blame] | 1090 | #endif /* CONFIG_PCI */ |
| 1091 | |