spd_sdram.c
来自「U-boot源码 ARM7启动代码」· C语言 代码 · 共 913 行 · 第 1/2 页
C
913 行
if (spd.mem_type == SPD_MEMTYPE_DDR) { twr_clk = 3; /* Clocks */ twtr_clk = 1; /* Clocks */ } else { twr_clk = picos_to_clk(spd.twr * 250); twtr_clk = picos_to_clk(spd.twtr * 250); } /* * Calculate Trfc, in picos. * DDR I: Byte 42 straight up in ns. * DDR II: Byte 40 and 42 swizzled some, in ns. */ if (spd.mem_type == SPD_MEMTYPE_DDR) { trfc = spd.trfc * 1000; /* up to ps */ } else { unsigned int byte40_table_ps[8] = { 0, 250, 330, 500, 660, 750, 0, 0 }; trfc = (((spd.trctrfc_ext & 0x1) * 256) + spd.trfc) * 1000 + byte40_table_ps[(spd.trctrfc_ext >> 1) & 0x7]; } trfc_clk = picos_to_clk(trfc); /* * Trcd, Byte 29, from quarter nanos to ps and clocks. */ trcd_clk = picos_to_clk(spd.trcd * 250) & 0x7; /* * Convert trfc_clk to DDR controller fields. DDR I should * fit in the REFREC field (16-19) of TIMING_CFG_1, but the * 83xx controller has an extended REFREC field of three bits. * The controller automatically adds 8 clocks to this value, * so preadjust it down 8 first before splitting it up. */ trfc_low = (trfc_clk - 8) & 0xf; trfc_high = ((trfc_clk - 8) >> 4) & 0x3; ddr->timing_cfg_1 = (((picos_to_clk(spd.trp * 250) & 0x07) << 28 ) | /* PRETOACT */ ((picos_to_clk(spd.tras * 1000) & 0x0f ) << 24 ) | /* ACTTOPRE */ (trcd_clk << 20 ) | /* ACTTORW */ (caslat_ctrl << 16 ) | /* CASLAT */ (trfc_low << 12 ) | /* REFEC */ ((twr_clk & 0x07) << 8) | /* WRRREC */ ((picos_to_clk(spd.trrd * 250) & 0x07) << 4) | /* ACTTOACT */ ((twtr_clk & 0x07) << 0) /* WRTORD */ ); /* * Additive Latency * For DDR I, 0. * For DDR II, with ODT enabled, use "a value" less than ACTTORW, * which comes from Trcd, and also note that: * add_lat + caslat must be >= 4 */ add_lat = 0; if (spd.mem_type == SPD_MEMTYPE_DDR2 && (odt_wr_cfg || odt_rd_cfg) && (caslat < 4)) { add_lat = trcd_clk - 1; if ((add_lat + caslat) < 4) { add_lat = 0; } } /* * Write Data Delay * Historically 0x2 == 4/8 clock delay. * Empirically, 0x3 == 6/8 clock delay is suggested for DDR I 266. */ wr_data_delay = 2; /* * Write Latency * Read to Precharge * Minimum CKE Pulse Width. * Four Activate Window */ if (spd.mem_type == SPD_MEMTYPE_DDR) { /* * This is a lie. It should really be 1, but if it is * set to 1, bits overlap into the old controller's * otherwise unused ACSM field. If we leave it 0, then * the HW will magically treat it as 1 for DDR 1. Oh Yea. */ wr_lat = 0; trtp_clk = 2; /* By the book. */ cke_min_clk = 1; /* By the book. */ four_act = 1; /* By the book. */ } else { wr_lat = caslat - 1; /* Convert SPD value from quarter nanos to picos. */ trtp_clk = picos_to_clk(spd.trtp * 250); cke_min_clk = 3; /* By the book. */ four_act = picos_to_clk(37500); /* By the book. 1k pages? */ } /* * Empirically set ~MCAS-to-preamble override for DDR 2. * Your milage will vary. */ cpo = 0; if (spd.mem_type == SPD_MEMTYPE_DDR2) { if (effective_data_rate == 266 || effective_data_rate == 333) { cpo = 0x7; /* READ_LAT + 5/4 */ } else if (effective_data_rate == 400) { cpo = 0x9; /* READ_LAT + 7/4 */ } else { /* Automatic calibration */ cpo = 0x1f; } } ddr->timing_cfg_2 = (0 | ((add_lat & 0x7) << 28) /* ADD_LAT */ | ((cpo & 0x1f) << 23) /* CPO */ | ((wr_lat & 0x7) << 19) /* WR_LAT */ | ((trtp_clk & 0x7) << 13) /* RD_TO_PRE */ | ((wr_data_delay & 0x7) << 10) /* WR_DATA_DELAY */ | ((cke_min_clk & 0x7) << 6) /* CKE_PLS */ | ((four_act & 0x1f) << 0) /* FOUR_ACT */ ); debug("DDR:timing_cfg_1=0x%08x\n", ddr->timing_cfg_1); debug("DDR:timing_cfg_2=0x%08x\n", ddr->timing_cfg_2); /* Check DIMM data bus width */ if (spd.dataw_lsb == 0x20) { if (spd.mem_type == SPD_MEMTYPE_DDR) burstlen = 0x03; /* 32 bit data bus, burst len is 8 */ else burstlen = 0x02; /* 32 bit data bus, burst len is 4 */ debug("\n DDR DIMM: data bus width is 32 bit"); } else { burstlen = 0x02; /* Others act as 64 bit bus, burst len is 4 */ debug("\n DDR DIMM: data bus width is 64 bit"); } /* Is this an ECC DDR chip? */ if (spd.config == 0x02) debug(" with ECC\n"); else debug(" without ECC\n"); /* Burst length is always 4 for 64 bit data bus, 8 for 32 bit data bus, Burst type is sequential */ if (spd.mem_type == SPD_MEMTYPE_DDR) { switch (caslat) { case 1: ddr->sdram_mode = 0x50 | burstlen; /* CL=1.5 */ break; case 2: ddr->sdram_mode = 0x20 | burstlen; /* CL=2.0 */ break; case 3: ddr->sdram_mode = 0x60 | burstlen; /* CL=2.5 */ break; case 4: ddr->sdram_mode = 0x30 | burstlen; /* CL=3.0 */ break; default: printf("DDR:only CL 1.5, 2.0, 2.5, 3.0 is supported\n"); return 0; } } else { mode_odt_enable = 0x0; /* Default disabled */ if (odt_wr_cfg || odt_rd_cfg) { /* * Bits 6 and 2 in Extended MRS(1) * Bit 2 == 0x04 == 75 Ohm, with 2 DIMM modules. * Bit 6 == 0x40 == 150 Ohm, with 1 DIMM module. */ mode_odt_enable = 0x40; /* 150 Ohm */ } ddr->sdram_mode = (0 | (1 << (16 + 10)) /* DQS Differential disable */ | (add_lat << (16 + 3)) /* Additive Latency in EMRS1 */ | (mode_odt_enable << 16) /* ODT Enable in EMRS1 */ | ((twr_clk - 1) << 9) /* Write Recovery Autopre */ | (caslat << 4) /* caslat */ | (burstlen << 0) /* Burst length */ ); } debug("DDR:sdram_mode=0x%08x\n", ddr->sdram_mode); /* * Clear EMRS2 and EMRS3. */ ddr->sdram_mode2 = 0; debug("DDR: sdram_mode2 = 0x%08x\n", ddr->sdram_mode2); switch (spd.refresh) { case 0x00: case 0x80: refresh_clk = picos_to_clk(15625000); break; case 0x01: case 0x81: refresh_clk = picos_to_clk(3900000); break; case 0x02: case 0x82: refresh_clk = picos_to_clk(7800000); break; case 0x03: case 0x83: refresh_clk = picos_to_clk(31300000); break; case 0x04: case 0x84: refresh_clk = picos_to_clk(62500000); break; case 0x05: case 0x85: refresh_clk = picos_to_clk(125000000); break; default: refresh_clk = 0x512; break; } /* * Set BSTOPRE to 0x100 for page mode * If auto-charge is used, set BSTOPRE = 0 */ ddr->sdram_interval = ((refresh_clk & 0x3fff) << 16) | 0x100; debug("DDR:sdram_interval=0x%08x\n", ddr->sdram_interval); /* * SDRAM Cfg 2 */ odt_cfg = 0; if (odt_rd_cfg | odt_wr_cfg) { odt_cfg = 0x2; /* ODT to IOs during reads */ } if (spd.mem_type == SPD_MEMTYPE_DDR2) { ddr->sdram_cfg2 = (0 | (0 << 26) /* True DQS */ | (odt_cfg << 21) /* ODT only read */ | (1 << 12) /* 1 refresh at a time */ ); debug("DDR: sdram_cfg2 = 0x%08x\n", ddr->sdram_cfg2); }#ifdef CFG_DDR_SDRAM_CLK_CNTL /* Optional platform specific value */ ddr->sdram_clk_cntl = CFG_DDR_SDRAM_CLK_CNTL;#endif debug("DDR:sdram_clk_cntl=0x%08x\n", ddr->sdram_clk_cntl); asm("sync;isync"); udelay(600); /* * Figure out the settings for the sdram_cfg register. Build up * the value in 'sdram_cfg' before writing since the write into * the register will actually enable the memory controller, and all * settings must be done before enabling. * * sdram_cfg[0] = 1 (ddr sdram logic enable) * sdram_cfg[1] = 1 (self-refresh-enable) * sdram_cfg[5:7] = (SDRAM type = DDR SDRAM) * 010 DDR 1 SDRAM * 011 DDR 2 SDRAM * sdram_cfg[12] = 0 (32_BE =0 , 64 bit bus mode) * sdram_cfg[13] = 0 (8_BE =0, 4-beat bursts) */ if (spd.mem_type == SPD_MEMTYPE_DDR) sdram_type = SDRAM_CFG_SDRAM_TYPE_DDR1; else sdram_type = SDRAM_CFG_SDRAM_TYPE_DDR2; sdram_cfg = (0 | SDRAM_CFG_MEM_EN /* DDR enable */ | SDRAM_CFG_SREN /* Self refresh */ | sdram_type /* SDRAM type */ ); /* sdram_cfg[3] = RD_EN - registered DIMM enable */ if (spd.mod_attr & 0x02) sdram_cfg |= SDRAM_CFG_RD_EN; /* The DIMM is 32bit width */ if (spd.dataw_lsb == 0x20) { if (spd.mem_type == SPD_MEMTYPE_DDR) sdram_cfg |= SDRAM_CFG_32_BE | SDRAM_CFG_8_BE; if (spd.mem_type == SPD_MEMTYPE_DDR2) sdram_cfg |= SDRAM_CFG_32_BE; } ddrc_ecc_enable = 0;#if defined(CONFIG_DDR_ECC) /* Enable ECC with sdram_cfg[2] */ if (spd.config == 0x02) { sdram_cfg |= 0x20000000; ddrc_ecc_enable = 1; /* disable error detection */ ddr->err_disable = ~ECC_ERROR_ENABLE; /* set single bit error threshold to maximum value, * reset counter to zero */ ddr->err_sbe = (255 << ECC_ERROR_MAN_SBET_SHIFT) | (0 << ECC_ERROR_MAN_SBEC_SHIFT); } debug("DDR:err_disable=0x%08x\n", ddr->err_disable); debug("DDR:err_sbe=0x%08x\n", ddr->err_sbe);#endif debug(" DDRC ECC mode: %s\n", ddrc_ecc_enable ? "ON":"OFF");#if defined(CONFIG_DDR_2T_TIMING) /* * Enable 2T timing by setting sdram_cfg[16]. */ sdram_cfg |= SDRAM_CFG_2T_EN;#endif /* Enable controller, and GO! */ ddr->sdram_cfg = sdram_cfg; asm("sync;isync"); udelay(500); debug("DDR:sdram_cfg=0x%08x\n", ddr->sdram_cfg); return memsize; /*in MBytes*/}#endif /* CONFIG_SPD_EEPROM */#if defined(CONFIG_DDR_ECC) && !defined(CONFIG_ECC_INIT_VIA_DDRC)/* * Use timebase counter, get_timer() is not availabe * at this point of initialization yet. */static __inline__ unsigned long get_tbms (void){ unsigned long tbl; unsigned long tbu1, tbu2; unsigned long ms; unsigned long long tmp; ulong tbclk = get_tbclk(); /* get the timebase ticks */ do { asm volatile ("mftbu %0":"=r" (tbu1):); asm volatile ("mftb %0":"=r" (tbl):); asm volatile ("mftbu %0":"=r" (tbu2):); } while (tbu1 != tbu2); /* convert ticks to ms */ tmp = (unsigned long long)(tbu1); tmp = (tmp << 32); tmp += (unsigned long long)(tbl); ms = tmp/(tbclk/1000); return ms;}/* * Initialize all of memory for ECC, then enable errors. *//* #define CONFIG_DDR_ECC_INIT_VIA_DMA */void ddr_enable_ecc(unsigned int dram_size){ volatile immap_t *immap = (immap_t *)CFG_IMMR; volatile ddr83xx_t *ddr= &immap->ddr; unsigned long t_start, t_end; register u64 *p; register uint size; unsigned int pattern[2];#if defined(CONFIG_DDR_ECC_INIT_VIA_DMA) uint i;#endif icache_enable(); t_start = get_tbms(); pattern[0] = 0xdeadbeef; pattern[1] = 0xdeadbeef;#if !defined(CONFIG_DDR_ECC_INIT_VIA_DMA) debug("ddr init: CPU FP write method\n"); size = dram_size; for (p = 0; p < (u64*)(size); p++) { ppcDWstore((u32*)p, pattern); } __asm__ __volatile__ ("sync");#else debug("ddr init: DMA method\n"); size = 0x2000; for (p = 0; p < (u64*)(size); p++) { ppcDWstore((u32*)p, pattern); } __asm__ __volatile__ ("sync"); /* Initialise DMA for direct transfer */ dma_init(); /* Start DMA to transfer */ dma_xfer((uint *)0x2000, 0x2000, (uint *)0); /* 8K */ dma_xfer((uint *)0x4000, 0x4000, (uint *)0); /* 16K */ dma_xfer((uint *)0x8000, 0x8000, (uint *)0); /* 32K */ dma_xfer((uint *)0x10000, 0x10000, (uint *)0); /* 64K */ dma_xfer((uint *)0x20000, 0x20000, (uint *)0); /* 128K */ dma_xfer((uint *)0x40000, 0x40000, (uint *)0); /* 256K */ dma_xfer((uint *)0x80000, 0x80000, (uint *)0); /* 512K */ dma_xfer((uint *)0x100000, 0x100000, (uint *)0); /* 1M */ dma_xfer((uint *)0x200000, 0x200000, (uint *)0); /* 2M */ dma_xfer((uint *)0x400000, 0x400000, (uint *)0); /* 4M */ for (i = 1; i < dram_size / 0x800000; i++) { dma_xfer((uint *)(0x800000*i), 0x800000, (uint *)0); }#endif t_end = get_tbms(); icache_disable(); debug("\nREADY!!\n"); debug("ddr init duration: %ld ms\n", t_end - t_start); /* Clear All ECC Errors */ if ((ddr->err_detect & ECC_ERROR_DETECT_MME) == ECC_ERROR_DETECT_MME) ddr->err_detect |= ECC_ERROR_DETECT_MME; if ((ddr->err_detect & ECC_ERROR_DETECT_MBE) == ECC_ERROR_DETECT_MBE) ddr->err_detect |= ECC_ERROR_DETECT_MBE; if ((ddr->err_detect & ECC_ERROR_DETECT_SBE) == ECC_ERROR_DETECT_SBE) ddr->err_detect |= ECC_ERROR_DETECT_SBE; if ((ddr->err_detect & ECC_ERROR_DETECT_MSE) == ECC_ERROR_DETECT_MSE) ddr->err_detect |= ECC_ERROR_DETECT_MSE; /* Disable ECC-Interrupts */ ddr->err_int_en &= ECC_ERR_INT_DISABLE; /* Enable errors for ECC */ ddr->err_disable &= ECC_ERROR_ENABLE; __asm__ __volatile__ ("sync"); __asm__ __volatile__ ("isync");}#endif /* CONFIG_DDR_ECC */
⌨️ 快捷键说明
复制代码Ctrl + C
搜索代码Ctrl + F
全屏模式F11
增大字号Ctrl + =
减小字号Ctrl + -
显示快捷键?