icside.c
来自「优龙2410linux2.6.8内核源代码」· C语言 代码 · 共 955 行 · 第 1/2 页
C
955 行
/* * linux/drivers/ide/arm/icside.c * * Copyright (c) 1996-2003 Russell King. */#include <linux/config.h>#include <linux/string.h>#include <linux/module.h>#include <linux/ioport.h>#include <linux/slab.h>#include <linux/blkdev.h>#include <linux/errno.h>#include <linux/hdreg.h>#include <linux/ide.h>#include <linux/dma-mapping.h>#include <linux/device.h>#include <linux/init.h>#include <asm/dma.h>#include <asm/ecard.h>#include <asm/io.h>#define ICS_IDENT_OFFSET 0x2280#define ICS_ARCIN_V5_INTRSTAT 0x000#define ICS_ARCIN_V5_INTROFFSET 0x001#define ICS_ARCIN_V5_IDEOFFSET 0xa00#define ICS_ARCIN_V5_IDEALTOFFSET 0xae0#define ICS_ARCIN_V5_IDESTEPPING 4#define ICS_ARCIN_V6_IDEOFFSET_1 0x800#define ICS_ARCIN_V6_INTROFFSET_1 0x880#define ICS_ARCIN_V6_INTRSTAT_1 0x8a4#define ICS_ARCIN_V6_IDEALTOFFSET_1 0x8e0#define ICS_ARCIN_V6_IDEOFFSET_2 0xc00#define ICS_ARCIN_V6_INTROFFSET_2 0xc80#define ICS_ARCIN_V6_INTRSTAT_2 0xca4#define ICS_ARCIN_V6_IDEALTOFFSET_2 0xce0#define ICS_ARCIN_V6_IDESTEPPING 4struct cardinfo { unsigned int dataoffset; unsigned int ctrloffset; unsigned int stepping;};static struct cardinfo icside_cardinfo_v5 = { ICS_ARCIN_V5_IDEOFFSET, ICS_ARCIN_V5_IDEALTOFFSET, ICS_ARCIN_V5_IDESTEPPING};static struct cardinfo icside_cardinfo_v6_1 = { ICS_ARCIN_V6_IDEOFFSET_1, ICS_ARCIN_V6_IDEALTOFFSET_1, ICS_ARCIN_V6_IDESTEPPING};static struct cardinfo icside_cardinfo_v6_2 = { ICS_ARCIN_V6_IDEOFFSET_2, ICS_ARCIN_V6_IDEALTOFFSET_2, ICS_ARCIN_V6_IDESTEPPING};struct icside_state { unsigned int channel; unsigned int enabled; unsigned long irq_port; unsigned long slot_port; unsigned int type; /* parent device... until the IDE core gets one of its own */ struct device *dev; ide_hwif_t *hwif[2];};#define ICS_TYPE_A3IN 0#define ICS_TYPE_A3USER 1#define ICS_TYPE_V6 3#define ICS_TYPE_V5 15#define ICS_TYPE_NOTYPE ((unsigned int)-1)/* ---------------- Version 5 PCB Support Functions --------------------- *//* Prototype: icside_irqenable_arcin_v5 (struct expansion_card *ec, int irqnr) * Purpose : enable interrupts from card */static void icside_irqenable_arcin_v5 (struct expansion_card *ec, int irqnr){ struct icside_state *state = ec->irq_data; unsigned int base = state->irq_port; outb(0, base + ICS_ARCIN_V5_INTROFFSET);}/* Prototype: icside_irqdisable_arcin_v5 (struct expansion_card *ec, int irqnr) * Purpose : disable interrupts from card */static void icside_irqdisable_arcin_v5 (struct expansion_card *ec, int irqnr){ struct icside_state *state = ec->irq_data; unsigned int base = state->irq_port; inb(base + ICS_ARCIN_V5_INTROFFSET);}static const expansioncard_ops_t icside_ops_arcin_v5 = { .irqenable = icside_irqenable_arcin_v5, .irqdisable = icside_irqdisable_arcin_v5,};/* ---------------- Version 6 PCB Support Functions --------------------- *//* Prototype: icside_irqenable_arcin_v6 (struct expansion_card *ec, int irqnr) * Purpose : enable interrupts from card */static void icside_irqenable_arcin_v6 (struct expansion_card *ec, int irqnr){ struct icside_state *state = ec->irq_data; unsigned int base = state->irq_port; state->enabled = 1; switch (state->channel) { case 0: outb(0, base + ICS_ARCIN_V6_INTROFFSET_1); inb(base + ICS_ARCIN_V6_INTROFFSET_2); break; case 1: outb(0, base + ICS_ARCIN_V6_INTROFFSET_2); inb(base + ICS_ARCIN_V6_INTROFFSET_1); break; }}/* Prototype: icside_irqdisable_arcin_v6 (struct expansion_card *ec, int irqnr) * Purpose : disable interrupts from card */static void icside_irqdisable_arcin_v6 (struct expansion_card *ec, int irqnr){ struct icside_state *state = ec->irq_data; state->enabled = 0; inb (state->irq_port + ICS_ARCIN_V6_INTROFFSET_1); inb (state->irq_port + ICS_ARCIN_V6_INTROFFSET_2);}/* Prototype: icside_irqprobe(struct expansion_card *ec) * Purpose : detect an active interrupt from card */static int icside_irqpending_arcin_v6(struct expansion_card *ec){ struct icside_state *state = ec->irq_data; return inb(state->irq_port + ICS_ARCIN_V6_INTRSTAT_1) & 1 || inb(state->irq_port + ICS_ARCIN_V6_INTRSTAT_2) & 1;}static const expansioncard_ops_t icside_ops_arcin_v6 = { .irqenable = icside_irqenable_arcin_v6, .irqdisable = icside_irqdisable_arcin_v6, .irqpending = icside_irqpending_arcin_v6,};/* * Handle routing of interrupts. This is called before * we write the command to the drive. */static void icside_maskproc(ide_drive_t *drive, int mask){ ide_hwif_t *hwif = HWIF(drive); struct icside_state *state = hwif->hwif_data; unsigned long flags; local_irq_save(flags); state->channel = hwif->channel; if (state->enabled && !mask) { switch (hwif->channel) { case 0: outb(0, state->irq_port + ICS_ARCIN_V6_INTROFFSET_1); inb(state->irq_port + ICS_ARCIN_V6_INTROFFSET_2); break; case 1: outb(0, state->irq_port + ICS_ARCIN_V6_INTROFFSET_2); inb(state->irq_port + ICS_ARCIN_V6_INTROFFSET_1); break; } } else { inb(state->irq_port + ICS_ARCIN_V6_INTROFFSET_2); inb(state->irq_port + ICS_ARCIN_V6_INTROFFSET_1); } local_irq_restore(flags);}#ifdef CONFIG_BLK_DEV_IDEDMA_ICS/* * SG-DMA support. * * Similar to the BM-DMA, but we use the RiscPCs IOMD DMA controllers. * There is only one DMA controller per card, which means that only * one drive can be accessed at one time. NOTE! We do not enforce that * here, but we rely on the main IDE driver spotting that both * interfaces use the same IRQ, which should guarantee this. */#define NR_ENTRIES 256#define TABLE_SIZE (NR_ENTRIES * 8)static void icside_build_sglist(ide_drive_t *drive, struct request *rq){ ide_hwif_t *hwif = drive->hwif; struct icside_state *state = hwif->hwif_data; struct scatterlist *sg = hwif->sg_table; int nents; BUG_ON(hwif->sg_dma_active); if (rq->flags & REQ_DRIVE_TASKFILE) { ide_task_t *args = rq->special; if (args->command_type == IDE_DRIVE_TASK_RAW_WRITE) hwif->sg_dma_direction = DMA_TO_DEVICE; else hwif->sg_dma_direction = DMA_FROM_DEVICE; memset(sg, 0, sizeof(*sg)); sg->page = virt_to_page(rq->buffer); sg->offset = offset_in_page(rq->buffer); sg->length = rq->nr_sectors * SECTOR_SIZE; nents = 1; } else { nents = blk_rq_map_sg(drive->queue, rq, sg); if (rq_data_dir(rq) == READ) hwif->sg_dma_direction = DMA_FROM_DEVICE; else hwif->sg_dma_direction = DMA_TO_DEVICE; } nents = dma_map_sg(state->dev, sg, nents, hwif->sg_dma_direction); hwif->sg_nents = nents;}/* * Configure the IOMD to give the appropriate timings for the transfer * mode being requested. We take the advice of the ATA standards, and * calculate the cycle time based on the transfer mode, and the EIDE * MW DMA specs that the drive provides in the IDENTIFY command. * * We have the following IOMD DMA modes to choose from: * * Type Active Recovery Cycle * A 250 (250) 312 (550) 562 (800) * B 187 250 437 * C 125 (125) 125 (375) 250 (500) * D 62 125 187 * * (figures in brackets are actual measured timings) * * However, we also need to take care of the read/write active and * recovery timings: * * Read Write * Mode Active -- Recovery -- Cycle IOMD type * MW0 215 50 215 480 A * MW1 80 50 50 150 C * MW2 70 25 25 120 C */static int icside_set_speed(ide_drive_t *drive, u8 xfer_mode){ int on = 0, cycle_time = 0, use_dma_info = 0; /* * Limit the transfer speed to MW_DMA_2. */ if (xfer_mode > XFER_MW_DMA_2) xfer_mode = XFER_MW_DMA_2; switch (xfer_mode) { case XFER_MW_DMA_2: cycle_time = 250; use_dma_info = 1; break; case XFER_MW_DMA_1: cycle_time = 250; use_dma_info = 1; break; case XFER_MW_DMA_0: cycle_time = 480; break; case XFER_SW_DMA_2: case XFER_SW_DMA_1: case XFER_SW_DMA_0: cycle_time = 480; break; } /* * If we're going to be doing MW_DMA_1 or MW_DMA_2, we should * take care to note the values in the ID... */ if (use_dma_info && drive->id->eide_dma_time > cycle_time) cycle_time = drive->id->eide_dma_time; drive->drive_data = cycle_time; if (cycle_time && ide_config_drive_speed(drive, xfer_mode) == 0) on = 1; else drive->drive_data = 480; printk("%s: %s selected (peak %dMB/s)\n", drive->name, ide_xfer_verbose(xfer_mode), 2000 / drive->drive_data); drive->current_speed = xfer_mode; return on;}static int icside_dma_host_off(ide_drive_t *drive){ return 0;}static int icside_dma_off_quietly(ide_drive_t *drive){ drive->using_dma = 0; return icside_dma_host_off(drive);}static int icside_dma_host_on(ide_drive_t *drive){ return 0;}static int icside_dma_on(ide_drive_t *drive){ drive->using_dma = 1; return icside_dma_host_on(drive);}static int icside_dma_check(ide_drive_t *drive){ struct hd_driveid *id = drive->id; ide_hwif_t *hwif = HWIF(drive); int xfer_mode = XFER_PIO_2; int on; if (!(id->capability & 1) || !hwif->autodma) goto out; /* * Consult the list of known "bad" drives */ if (__ide_dma_bad_drive(drive)) goto out; /* * Enable DMA on any drive that has multiword DMA */ if (id->field_valid & 2) { xfer_mode = ide_dma_speed(drive, 0); goto out; } /* * Consult the list of known "good" drives */ if (__ide_dma_good_drive(drive)) { if (id->eide_dma_time > 150) goto out; xfer_mode = XFER_MW_DMA_1; }out: on = icside_set_speed(drive, xfer_mode); if (on) return icside_dma_on(drive); else return icside_dma_off_quietly(drive);}static int icside_dma_end(ide_drive_t *drive){ ide_hwif_t *hwif = HWIF(drive); struct icside_state *state = hwif->hwif_data; drive->waiting_for_dma = 0; disable_dma(hwif->hw.dma); /* Teardown mappings after DMA has completed. */ dma_unmap_sg(state->dev, hwif->sg_table, hwif->sg_nents, hwif->sg_dma_direction); hwif->sg_dma_active = 0; return get_dma_residue(hwif->hw.dma) != 0;}static int icside_dma_begin(ide_drive_t *drive){ ide_hwif_t *hwif = HWIF(drive); /* We can not enable DMA on both channels simultaneously. */ BUG_ON(dma_channel_active(hwif->hw.dma)); enable_dma(hwif->hw.dma); return 0;}/* * dma_intr() is the handler for disk read/write DMA interrupts */static ide_startstop_t icside_dmaintr(ide_drive_t *drive){ unsigned int stat; int dma_stat; dma_stat = icside_dma_end(drive); stat = HWIF(drive)->INB(IDE_STATUS_REG); if (OK_STAT(stat, DRIVE_READY, drive->bad_wstat | DRQ_STAT)) { if (!dma_stat) { struct request *rq = HWGROUP(drive)->rq; int i; for (i = rq->nr_sectors; i > 0; ) { i -= rq->current_nr_sectors; DRIVER(drive)->end_request(drive, 1, rq->nr_sectors); } return ide_stopped; } printk(KERN_ERR "%s: bad DMA status (dma_stat=%x)\n", drive->name, dma_stat); } return DRIVER(drive)->error(drive, __FUNCTION__, stat);}static inticside_dma_common(ide_drive_t *drive, struct request *rq, unsigned int dma_mode){ ide_hwif_t *hwif = HWIF(drive); /* * We can not enable DMA on both channels. */ BUG_ON(hwif->sg_dma_active); BUG_ON(dma_channel_active(hwif->hw.dma)); icside_build_sglist(drive, rq); /* * Ensure that we have the right interrupt routed. */ icside_maskproc(drive, 0); /* * Route the DMA signals to the correct interface. */ outb(hwif->select_data, hwif->config_data); /* * Select the correct timing for this drive. */ set_dma_speed(hwif->hw.dma, drive->drive_data); /* * Tell the DMA engine about the SG table and
⌨️ 快捷键说明
复制代码Ctrl + C
搜索代码Ctrl + F
全屏模式F11
增大字号Ctrl + =
减小字号Ctrl + -
显示快捷键?