icside.c revision 35c9b4daf4c94b30e5cede597d98016ebf31b5ad
1/* 2 * Copyright (c) 1996-2004 Russell King. 3 * 4 * Please note that this platform does not support 32-bit IDE IO. 5 */ 6 7#include <linux/string.h> 8#include <linux/module.h> 9#include <linux/ioport.h> 10#include <linux/slab.h> 11#include <linux/blkdev.h> 12#include <linux/errno.h> 13#include <linux/ide.h> 14#include <linux/dma-mapping.h> 15#include <linux/device.h> 16#include <linux/init.h> 17#include <linux/scatterlist.h> 18#include <linux/io.h> 19 20#include <asm/dma.h> 21#include <asm/ecard.h> 22 23#define DRV_NAME "icside" 24 25#define ICS_IDENT_OFFSET 0x2280 26 27#define ICS_ARCIN_V5_INTRSTAT 0x0000 28#define ICS_ARCIN_V5_INTROFFSET 0x0004 29#define ICS_ARCIN_V5_IDEOFFSET 0x2800 30#define ICS_ARCIN_V5_IDEALTOFFSET 0x2b80 31#define ICS_ARCIN_V5_IDESTEPPING 6 32 33#define ICS_ARCIN_V6_IDEOFFSET_1 0x2000 34#define ICS_ARCIN_V6_INTROFFSET_1 0x2200 35#define ICS_ARCIN_V6_INTRSTAT_1 0x2290 36#define ICS_ARCIN_V6_IDEALTOFFSET_1 0x2380 37#define ICS_ARCIN_V6_IDEOFFSET_2 0x3000 38#define ICS_ARCIN_V6_INTROFFSET_2 0x3200 39#define ICS_ARCIN_V6_INTRSTAT_2 0x3290 40#define ICS_ARCIN_V6_IDEALTOFFSET_2 0x3380 41#define ICS_ARCIN_V6_IDESTEPPING 6 42 43struct cardinfo { 44 unsigned int dataoffset; 45 unsigned int ctrloffset; 46 unsigned int stepping; 47}; 48 49static struct cardinfo icside_cardinfo_v5 = { 50 .dataoffset = ICS_ARCIN_V5_IDEOFFSET, 51 .ctrloffset = ICS_ARCIN_V5_IDEALTOFFSET, 52 .stepping = ICS_ARCIN_V5_IDESTEPPING, 53}; 54 55static struct cardinfo icside_cardinfo_v6_1 = { 56 .dataoffset = ICS_ARCIN_V6_IDEOFFSET_1, 57 .ctrloffset = ICS_ARCIN_V6_IDEALTOFFSET_1, 58 .stepping = ICS_ARCIN_V6_IDESTEPPING, 59}; 60 61static struct cardinfo icside_cardinfo_v6_2 = { 62 .dataoffset = ICS_ARCIN_V6_IDEOFFSET_2, 63 .ctrloffset = ICS_ARCIN_V6_IDEALTOFFSET_2, 64 .stepping = ICS_ARCIN_V6_IDESTEPPING, 65}; 66 67struct icside_state { 68 unsigned int channel; 69 unsigned int enabled; 70 void __iomem *irq_port; 71 void __iomem *ioc_base; 72 unsigned int sel; 73 unsigned int type; 74 struct ide_host *host; 75}; 76 77#define ICS_TYPE_A3IN 0 78#define ICS_TYPE_A3USER 1 79#define ICS_TYPE_V6 3 80#define ICS_TYPE_V5 15 81#define ICS_TYPE_NOTYPE ((unsigned int)-1) 82 83/* ---------------- Version 5 PCB Support Functions --------------------- */ 84/* Prototype: icside_irqenable_arcin_v5 (struct expansion_card *ec, int irqnr) 85 * Purpose : enable interrupts from card 86 */ 87static void icside_irqenable_arcin_v5 (struct expansion_card *ec, int irqnr) 88{ 89 struct icside_state *state = ec->irq_data; 90 91 writeb(0, state->irq_port + ICS_ARCIN_V5_INTROFFSET); 92} 93 94/* Prototype: icside_irqdisable_arcin_v5 (struct expansion_card *ec, int irqnr) 95 * Purpose : disable interrupts from card 96 */ 97static void icside_irqdisable_arcin_v5 (struct expansion_card *ec, int irqnr) 98{ 99 struct icside_state *state = ec->irq_data; 100 101 readb(state->irq_port + ICS_ARCIN_V5_INTROFFSET); 102} 103 104static const expansioncard_ops_t icside_ops_arcin_v5 = { 105 .irqenable = icside_irqenable_arcin_v5, 106 .irqdisable = icside_irqdisable_arcin_v5, 107}; 108 109 110/* ---------------- Version 6 PCB Support Functions --------------------- */ 111/* Prototype: icside_irqenable_arcin_v6 (struct expansion_card *ec, int irqnr) 112 * Purpose : enable interrupts from card 113 */ 114static void icside_irqenable_arcin_v6 (struct expansion_card *ec, int irqnr) 115{ 116 struct icside_state *state = ec->irq_data; 117 void __iomem *base = state->irq_port; 118 119 state->enabled = 1; 120 121 switch (state->channel) { 122 case 0: 123 writeb(0, base + ICS_ARCIN_V6_INTROFFSET_1); 124 readb(base + ICS_ARCIN_V6_INTROFFSET_2); 125 break; 126 case 1: 127 writeb(0, base + ICS_ARCIN_V6_INTROFFSET_2); 128 readb(base + ICS_ARCIN_V6_INTROFFSET_1); 129 break; 130 } 131} 132 133/* Prototype: icside_irqdisable_arcin_v6 (struct expansion_card *ec, int irqnr) 134 * Purpose : disable interrupts from card 135 */ 136static void icside_irqdisable_arcin_v6 (struct expansion_card *ec, int irqnr) 137{ 138 struct icside_state *state = ec->irq_data; 139 140 state->enabled = 0; 141 142 readb(state->irq_port + ICS_ARCIN_V6_INTROFFSET_1); 143 readb(state->irq_port + ICS_ARCIN_V6_INTROFFSET_2); 144} 145 146/* Prototype: icside_irqprobe(struct expansion_card *ec) 147 * Purpose : detect an active interrupt from card 148 */ 149static int icside_irqpending_arcin_v6(struct expansion_card *ec) 150{ 151 struct icside_state *state = ec->irq_data; 152 153 return readb(state->irq_port + ICS_ARCIN_V6_INTRSTAT_1) & 1 || 154 readb(state->irq_port + ICS_ARCIN_V6_INTRSTAT_2) & 1; 155} 156 157static const expansioncard_ops_t icside_ops_arcin_v6 = { 158 .irqenable = icside_irqenable_arcin_v6, 159 .irqdisable = icside_irqdisable_arcin_v6, 160 .irqpending = icside_irqpending_arcin_v6, 161}; 162 163/* 164 * Handle routing of interrupts. This is called before 165 * we write the command to the drive. 166 */ 167static void icside_maskproc(ide_drive_t *drive, int mask) 168{ 169 ide_hwif_t *hwif = drive->hwif; 170 struct expansion_card *ec = ECARD_DEV(hwif->dev); 171 struct icside_state *state = ecard_get_drvdata(ec); 172 unsigned long flags; 173 174 local_irq_save(flags); 175 176 state->channel = hwif->channel; 177 178 if (state->enabled && !mask) { 179 switch (hwif->channel) { 180 case 0: 181 writeb(0, state->irq_port + ICS_ARCIN_V6_INTROFFSET_1); 182 readb(state->irq_port + ICS_ARCIN_V6_INTROFFSET_2); 183 break; 184 case 1: 185 writeb(0, state->irq_port + ICS_ARCIN_V6_INTROFFSET_2); 186 readb(state->irq_port + ICS_ARCIN_V6_INTROFFSET_1); 187 break; 188 } 189 } else { 190 readb(state->irq_port + ICS_ARCIN_V6_INTROFFSET_2); 191 readb(state->irq_port + ICS_ARCIN_V6_INTROFFSET_1); 192 } 193 194 local_irq_restore(flags); 195} 196 197static const struct ide_port_ops icside_v6_no_dma_port_ops = { 198 .maskproc = icside_maskproc, 199}; 200 201#ifdef CONFIG_BLK_DEV_IDEDMA_ICS 202/* 203 * SG-DMA support. 204 * 205 * Similar to the BM-DMA, but we use the RiscPCs IOMD DMA controllers. 206 * There is only one DMA controller per card, which means that only 207 * one drive can be accessed at one time. NOTE! We do not enforce that 208 * here, but we rely on the main IDE driver spotting that both 209 * interfaces use the same IRQ, which should guarantee this. 210 */ 211 212/* 213 * Configure the IOMD to give the appropriate timings for the transfer 214 * mode being requested. We take the advice of the ATA standards, and 215 * calculate the cycle time based on the transfer mode, and the EIDE 216 * MW DMA specs that the drive provides in the IDENTIFY command. 217 * 218 * We have the following IOMD DMA modes to choose from: 219 * 220 * Type Active Recovery Cycle 221 * A 250 (250) 312 (550) 562 (800) 222 * B 187 250 437 223 * C 125 (125) 125 (375) 250 (500) 224 * D 62 125 187 225 * 226 * (figures in brackets are actual measured timings) 227 * 228 * However, we also need to take care of the read/write active and 229 * recovery timings: 230 * 231 * Read Write 232 * Mode Active -- Recovery -- Cycle IOMD type 233 * MW0 215 50 215 480 A 234 * MW1 80 50 50 150 C 235 * MW2 70 25 25 120 C 236 */ 237static void icside_set_dma_mode(ide_drive_t *drive, const u8 xfer_mode) 238{ 239 int cycle_time, use_dma_info = 0; 240 241 switch (xfer_mode) { 242 case XFER_MW_DMA_2: 243 cycle_time = 250; 244 use_dma_info = 1; 245 break; 246 247 case XFER_MW_DMA_1: 248 cycle_time = 250; 249 use_dma_info = 1; 250 break; 251 252 case XFER_MW_DMA_0: 253 cycle_time = 480; 254 break; 255 256 case XFER_SW_DMA_2: 257 case XFER_SW_DMA_1: 258 case XFER_SW_DMA_0: 259 cycle_time = 480; 260 break; 261 } 262 263 /* 264 * If we're going to be doing MW_DMA_1 or MW_DMA_2, we should 265 * take care to note the values in the ID... 266 */ 267 if (use_dma_info && drive->id[ATA_ID_EIDE_DMA_TIME] > cycle_time) 268 cycle_time = drive->id[ATA_ID_EIDE_DMA_TIME]; 269 270 drive->drive_data = cycle_time; 271 272 printk("%s: %s selected (peak %dMB/s)\n", drive->name, 273 ide_xfer_verbose(xfer_mode), 2000 / drive->drive_data); 274} 275 276static const struct ide_port_ops icside_v6_port_ops = { 277 .set_dma_mode = icside_set_dma_mode, 278 .maskproc = icside_maskproc, 279}; 280 281static void icside_dma_host_set(ide_drive_t *drive, int on) 282{ 283} 284 285static int icside_dma_end(ide_drive_t *drive) 286{ 287 ide_hwif_t *hwif = drive->hwif; 288 struct expansion_card *ec = ECARD_DEV(hwif->dev); 289 290 drive->waiting_for_dma = 0; 291 292 disable_dma(ec->dma); 293 294 /* Teardown mappings after DMA has completed. */ 295 ide_destroy_dmatable(drive); 296 297 return get_dma_residue(ec->dma) != 0; 298} 299 300static void icside_dma_start(ide_drive_t *drive) 301{ 302 ide_hwif_t *hwif = drive->hwif; 303 struct expansion_card *ec = ECARD_DEV(hwif->dev); 304 305 /* We can not enable DMA on both channels simultaneously. */ 306 BUG_ON(dma_channel_active(ec->dma)); 307 enable_dma(ec->dma); 308} 309 310static int icside_dma_setup(ide_drive_t *drive, struct ide_cmd *cmd) 311{ 312 ide_hwif_t *hwif = drive->hwif; 313 struct expansion_card *ec = ECARD_DEV(hwif->dev); 314 struct icside_state *state = ecard_get_drvdata(ec); 315 unsigned int dma_mode; 316 317 if (cmd->tf_flags & IDE_TFLAG_WRITE) 318 dma_mode = DMA_MODE_WRITE; 319 else 320 dma_mode = DMA_MODE_READ; 321 322 /* 323 * We can not enable DMA on both channels. 324 */ 325 BUG_ON(dma_channel_active(ec->dma)); 326 327 /* 328 * Ensure that we have the right interrupt routed. 329 */ 330 icside_maskproc(drive, 0); 331 332 /* 333 * Route the DMA signals to the correct interface. 334 */ 335 writeb(state->sel | hwif->channel, state->ioc_base); 336 337 /* 338 * Select the correct timing for this drive. 339 */ 340 set_dma_speed(ec->dma, drive->drive_data); 341 342 /* 343 * Tell the DMA engine about the SG table and 344 * data direction. 345 */ 346 set_dma_sg(ec->dma, hwif->sg_table, cmd->sg_nents); 347 set_dma_mode(ec->dma, dma_mode); 348 349 drive->waiting_for_dma = 1; 350 351 return 0; 352} 353 354static int icside_dma_test_irq(ide_drive_t *drive) 355{ 356 ide_hwif_t *hwif = drive->hwif; 357 struct expansion_card *ec = ECARD_DEV(hwif->dev); 358 struct icside_state *state = ecard_get_drvdata(ec); 359 360 return readb(state->irq_port + 361 (hwif->channel ? 362 ICS_ARCIN_V6_INTRSTAT_2 : 363 ICS_ARCIN_V6_INTRSTAT_1)) & 1; 364} 365 366static int icside_dma_init(ide_hwif_t *hwif, const struct ide_port_info *d) 367{ 368 hwif->dmatable_cpu = NULL; 369 hwif->dmatable_dma = 0; 370 371 return 0; 372} 373 374static const struct ide_dma_ops icside_v6_dma_ops = { 375 .dma_host_set = icside_dma_host_set, 376 .dma_setup = icside_dma_setup, 377 .dma_start = icside_dma_start, 378 .dma_end = icside_dma_end, 379 .dma_test_irq = icside_dma_test_irq, 380 .dma_lost_irq = ide_dma_lost_irq, 381}; 382#else 383#define icside_v6_dma_ops NULL 384#endif 385 386static int icside_dma_off_init(ide_hwif_t *hwif, const struct ide_port_info *d) 387{ 388 return -EOPNOTSUPP; 389} 390 391static void icside_setup_ports(hw_regs_t *hw, void __iomem *base, 392 struct cardinfo *info, struct expansion_card *ec) 393{ 394 unsigned long port = (unsigned long)base + info->dataoffset; 395 396 hw->io_ports.data_addr = port; 397 hw->io_ports.error_addr = port + (1 << info->stepping); 398 hw->io_ports.nsect_addr = port + (2 << info->stepping); 399 hw->io_ports.lbal_addr = port + (3 << info->stepping); 400 hw->io_ports.lbam_addr = port + (4 << info->stepping); 401 hw->io_ports.lbah_addr = port + (5 << info->stepping); 402 hw->io_ports.device_addr = port + (6 << info->stepping); 403 hw->io_ports.status_addr = port + (7 << info->stepping); 404 hw->io_ports.ctl_addr = (unsigned long)base + info->ctrloffset; 405 406 hw->irq = ec->irq; 407 hw->dev = &ec->dev; 408 hw->chipset = ide_acorn; 409} 410 411static const struct ide_port_info icside_v5_port_info = { 412 .host_flags = IDE_HFLAG_NO_DMA, 413}; 414 415static int __devinit 416icside_register_v5(struct icside_state *state, struct expansion_card *ec) 417{ 418 void __iomem *base; 419 struct ide_host *host; 420 hw_regs_t hw, *hws[] = { &hw, NULL, NULL, NULL }; 421 int ret; 422 423 base = ecardm_iomap(ec, ECARD_RES_MEMC, 0, 0); 424 if (!base) 425 return -ENOMEM; 426 427 state->irq_port = base; 428 429 ec->irqaddr = base + ICS_ARCIN_V5_INTRSTAT; 430 ec->irqmask = 1; 431 432 ecard_setirq(ec, &icside_ops_arcin_v5, state); 433 434 /* 435 * Be on the safe side - disable interrupts 436 */ 437 icside_irqdisable_arcin_v5(ec, 0); 438 439 icside_setup_ports(&hw, base, &icside_cardinfo_v5, ec); 440 441 host = ide_host_alloc(&icside_v5_port_info, hws); 442 if (host == NULL) 443 return -ENODEV; 444 445 state->host = host; 446 447 ecard_set_drvdata(ec, state); 448 449 ret = ide_host_register(host, &icside_v5_port_info, hws); 450 if (ret) 451 goto err_free; 452 453 return 0; 454err_free: 455 ide_host_free(host); 456 ecard_set_drvdata(ec, NULL); 457 return ret; 458} 459 460static const struct ide_port_info icside_v6_port_info __initdata = { 461 .init_dma = icside_dma_off_init, 462 .port_ops = &icside_v6_no_dma_port_ops, 463 .dma_ops = &icside_v6_dma_ops, 464 .host_flags = IDE_HFLAG_SERIALIZE | IDE_HFLAG_MMIO, 465 .mwdma_mask = ATA_MWDMA2, 466 .swdma_mask = ATA_SWDMA2, 467}; 468 469static int __devinit 470icside_register_v6(struct icside_state *state, struct expansion_card *ec) 471{ 472 void __iomem *ioc_base, *easi_base; 473 struct ide_host *host; 474 unsigned int sel = 0; 475 int ret; 476 hw_regs_t hw[2], *hws[] = { &hw[0], NULL, NULL, NULL }; 477 struct ide_port_info d = icside_v6_port_info; 478 479 ioc_base = ecardm_iomap(ec, ECARD_RES_IOCFAST, 0, 0); 480 if (!ioc_base) { 481 ret = -ENOMEM; 482 goto out; 483 } 484 485 easi_base = ioc_base; 486 487 if (ecard_resource_flags(ec, ECARD_RES_EASI)) { 488 easi_base = ecardm_iomap(ec, ECARD_RES_EASI, 0, 0); 489 if (!easi_base) { 490 ret = -ENOMEM; 491 goto out; 492 } 493 494 /* 495 * Enable access to the EASI region. 496 */ 497 sel = 1 << 5; 498 } 499 500 writeb(sel, ioc_base); 501 502 ecard_setirq(ec, &icside_ops_arcin_v6, state); 503 504 state->irq_port = easi_base; 505 state->ioc_base = ioc_base; 506 state->sel = sel; 507 508 /* 509 * Be on the safe side - disable interrupts 510 */ 511 icside_irqdisable_arcin_v6(ec, 0); 512 513 icside_setup_ports(&hw[0], easi_base, &icside_cardinfo_v6_1, ec); 514 icside_setup_ports(&hw[1], easi_base, &icside_cardinfo_v6_2, ec); 515 516 host = ide_host_alloc(&d, hws); 517 if (host == NULL) 518 return -ENODEV; 519 520 state->host = host; 521 522 ecard_set_drvdata(ec, state); 523 524 if (ec->dma != NO_DMA && !request_dma(ec->dma, DRV_NAME)) { 525 d.init_dma = icside_dma_init; 526 d.port_ops = &icside_v6_port_ops; 527 d.dma_ops = NULL; 528 } 529 530 ret = ide_host_register(host, &d, hws); 531 if (ret) 532 goto err_free; 533 534 return 0; 535err_free: 536 ide_host_free(host); 537 if (d.dma_ops) 538 free_dma(ec->dma); 539 ecard_set_drvdata(ec, NULL); 540out: 541 return ret; 542} 543 544static int __devinit 545icside_probe(struct expansion_card *ec, const struct ecard_id *id) 546{ 547 struct icside_state *state; 548 void __iomem *idmem; 549 int ret; 550 551 ret = ecard_request_resources(ec); 552 if (ret) 553 goto out; 554 555 state = kzalloc(sizeof(struct icside_state), GFP_KERNEL); 556 if (!state) { 557 ret = -ENOMEM; 558 goto release; 559 } 560 561 state->type = ICS_TYPE_NOTYPE; 562 563 idmem = ecardm_iomap(ec, ECARD_RES_IOCFAST, 0, 0); 564 if (idmem) { 565 unsigned int type; 566 567 type = readb(idmem + ICS_IDENT_OFFSET) & 1; 568 type |= (readb(idmem + ICS_IDENT_OFFSET + 4) & 1) << 1; 569 type |= (readb(idmem + ICS_IDENT_OFFSET + 8) & 1) << 2; 570 type |= (readb(idmem + ICS_IDENT_OFFSET + 12) & 1) << 3; 571 ecardm_iounmap(ec, idmem); 572 573 state->type = type; 574 } 575 576 switch (state->type) { 577 case ICS_TYPE_A3IN: 578 dev_warn(&ec->dev, "A3IN unsupported\n"); 579 ret = -ENODEV; 580 break; 581 582 case ICS_TYPE_A3USER: 583 dev_warn(&ec->dev, "A3USER unsupported\n"); 584 ret = -ENODEV; 585 break; 586 587 case ICS_TYPE_V5: 588 ret = icside_register_v5(state, ec); 589 break; 590 591 case ICS_TYPE_V6: 592 ret = icside_register_v6(state, ec); 593 break; 594 595 default: 596 dev_warn(&ec->dev, "unknown interface type\n"); 597 ret = -ENODEV; 598 break; 599 } 600 601 if (ret == 0) 602 goto out; 603 604 kfree(state); 605 release: 606 ecard_release_resources(ec); 607 out: 608 return ret; 609} 610 611static void __devexit icside_remove(struct expansion_card *ec) 612{ 613 struct icside_state *state = ecard_get_drvdata(ec); 614 615 switch (state->type) { 616 case ICS_TYPE_V5: 617 /* FIXME: tell IDE to stop using the interface */ 618 619 /* Disable interrupts */ 620 icside_irqdisable_arcin_v5(ec, 0); 621 break; 622 623 case ICS_TYPE_V6: 624 /* FIXME: tell IDE to stop using the interface */ 625 if (ec->dma != NO_DMA) 626 free_dma(ec->dma); 627 628 /* Disable interrupts */ 629 icside_irqdisable_arcin_v6(ec, 0); 630 631 /* Reset the ROM pointer/EASI selection */ 632 writeb(0, state->ioc_base); 633 break; 634 } 635 636 ecard_set_drvdata(ec, NULL); 637 638 kfree(state); 639 ecard_release_resources(ec); 640} 641 642static void icside_shutdown(struct expansion_card *ec) 643{ 644 struct icside_state *state = ecard_get_drvdata(ec); 645 unsigned long flags; 646 647 /* 648 * Disable interrupts from this card. We need to do 649 * this before disabling EASI since we may be accessing 650 * this register via that region. 651 */ 652 local_irq_save(flags); 653 ec->ops->irqdisable(ec, 0); 654 local_irq_restore(flags); 655 656 /* 657 * Reset the ROM pointer so that we can read the ROM 658 * after a soft reboot. This also disables access to 659 * the IDE taskfile via the EASI region. 660 */ 661 if (state->ioc_base) 662 writeb(0, state->ioc_base); 663} 664 665static const struct ecard_id icside_ids[] = { 666 { MANU_ICS, PROD_ICS_IDE }, 667 { MANU_ICS2, PROD_ICS2_IDE }, 668 { 0xffff, 0xffff } 669}; 670 671static struct ecard_driver icside_driver = { 672 .probe = icside_probe, 673 .remove = __devexit_p(icside_remove), 674 .shutdown = icside_shutdown, 675 .id_table = icside_ids, 676 .drv = { 677 .name = "icside", 678 }, 679}; 680 681static int __init icside_init(void) 682{ 683 return ecard_register_driver(&icside_driver); 684} 685 686static void __exit icside_exit(void) 687{ 688 ecard_remove_driver(&icside_driver); 689} 690 691MODULE_AUTHOR("Russell King <rmk@arm.linux.org.uk>"); 692MODULE_LICENSE("GPL"); 693MODULE_DESCRIPTION("ICS IDE driver"); 694 695module_init(icside_init); 696module_exit(icside_exit); 697