time.c

来自「linux2.6.16版本」· C语言 代码 · 共 963 行 · 第 1/2 页

C
963
字号
/* * Common time routines among all ppc machines. * * Written by Cort Dougan (cort@cs.nmt.edu) to merge * Paul Mackerras' version and mine for PReP and Pmac. * MPC8xx/MBX changes by Dan Malek (dmalek@jlc.net). * Converted for 64-bit by Mike Corrigan (mikejc@us.ibm.com) * * First round of bugfixes by Gabriel Paubert (paubert@iram.es) * to make clock more stable (2.4.0-test5). The only thing * that this code assumes is that the timebases have been synchronized * by firmware on SMP and are never stopped (never do sleep * on SMP then, nap and doze are OK). *  * Speeded up do_gettimeofday by getting rid of references to * xtime (which required locks for consistency). (mikejc@us.ibm.com) * * TODO (not necessarily in this file): * - improve precision and reproducibility of timebase frequency * measurement at boot time. (for iSeries, we calibrate the timebase * against the Titan chip's clock.) * - for astronomical applications: add a new function to get * non ambiguous timestamps even around leap seconds. This needs * a new timestamp format and a good name. * * 1997-09-10  Updated NTP code according to technical memorandum Jan '96 *             "A Kernel Model for Precision Timekeeping" by Dave Mills * *      This program is free software; you can redistribute it and/or *      modify it under the terms of the GNU General Public License *      as published by the Free Software Foundation; either version *      2 of the License, or (at your option) any later version. */#include <linux/config.h>#include <linux/errno.h>#include <linux/module.h>#include <linux/sched.h>#include <linux/kernel.h>#include <linux/param.h>#include <linux/string.h>#include <linux/mm.h>#include <linux/interrupt.h>#include <linux/timex.h>#include <linux/kernel_stat.h>#include <linux/time.h>#include <linux/init.h>#include <linux/profile.h>#include <linux/cpu.h>#include <linux/security.h>#include <linux/percpu.h>#include <linux/rtc.h>#include <linux/jiffies.h>#include <asm/io.h>#include <asm/processor.h>#include <asm/nvram.h>#include <asm/cache.h>#include <asm/machdep.h>#include <asm/uaccess.h>#include <asm/time.h>#include <asm/prom.h>#include <asm/irq.h>#include <asm/div64.h>#include <asm/smp.h>#include <asm/vdso_datapage.h>#ifdef CONFIG_PPC64#include <asm/firmware.h>#endif#ifdef CONFIG_PPC_ISERIES#include <asm/iseries/it_lp_queue.h>#include <asm/iseries/hv_call_xm.h>#endif#include <asm/smp.h>/* keep track of when we need to update the rtc */time_t last_rtc_update;extern int piranha_simulator;#ifdef CONFIG_PPC_ISERIESunsigned long iSeries_recal_titan = 0;unsigned long iSeries_recal_tb = 0; static unsigned long first_settimeofday = 1;#endif/* The decrementer counts down by 128 every 128ns on a 601. */#define DECREMENTER_COUNT_601	(1000000000 / HZ)#define XSEC_PER_SEC (1024*1024)#ifdef CONFIG_PPC64#define SCALE_XSEC(xsec, max)	(((xsec) * max) / XSEC_PER_SEC)#else/* compute ((xsec << 12) * max) >> 32 */#define SCALE_XSEC(xsec, max)	mulhwu((xsec) << 12, max)#endifunsigned long tb_ticks_per_jiffy;unsigned long tb_ticks_per_usec = 100; /* sane default */EXPORT_SYMBOL(tb_ticks_per_usec);unsigned long tb_ticks_per_sec;u64 tb_to_xs;unsigned tb_to_us;#define TICKLEN_SCALE	(SHIFT_SCALE - 10)u64 last_tick_len;	/* units are ns / 2^TICKLEN_SCALE */u64 ticklen_to_xs;	/* 0.64 fraction *//* If last_tick_len corresponds to about 1/HZ seconds, then   last_tick_len << TICKLEN_SHIFT will be about 2^63. */#define TICKLEN_SHIFT	(63 - 30 - TICKLEN_SCALE + SHIFT_HZ)DEFINE_SPINLOCK(rtc_lock);EXPORT_SYMBOL_GPL(rtc_lock);u64 tb_to_ns_scale;unsigned tb_to_ns_shift;struct gettimeofday_struct do_gtod;extern unsigned long wall_jiffies;extern struct timezone sys_tz;static long timezone_offset;unsigned long ppc_proc_freq;unsigned long ppc_tb_freq;u64 tb_last_jiffy __cacheline_aligned_in_smp;unsigned long tb_last_stamp;/* * Note that on ppc32 this only stores the bottom 32 bits of * the timebase value, but that's enough to tell when a jiffy * has passed. */DEFINE_PER_CPU(unsigned long, last_jiffy);void __delay(unsigned long loops){	unsigned long start;	int diff;	if (__USE_RTC()) {		start = get_rtcl();		do {			/* the RTCL register wraps at 1000000000 */			diff = get_rtcl() - start;			if (diff < 0)				diff += 1000000000;		} while (diff < loops);	} else {		start = get_tbl();		while (get_tbl() - start < loops)			HMT_low();		HMT_medium();	}}EXPORT_SYMBOL(__delay);void udelay(unsigned long usecs){	__delay(tb_ticks_per_usec * usecs);}EXPORT_SYMBOL(udelay);static __inline__ void timer_check_rtc(void){        /*         * update the rtc when needed, this should be performed on the         * right fraction of a second. Half or full second ?         * Full second works on mk48t59 clocks, others need testing.         * Note that this update is basically only used through          * the adjtimex system calls. Setting the HW clock in         * any other way is a /dev/rtc and userland business.         * This is still wrong by -0.5/+1.5 jiffies because of the         * timer interrupt resolution and possible delay, but here we          * hit a quantization limit which can only be solved by higher         * resolution timers and decoupling time management from timer         * interrupts. This is also wrong on the clocks         * which require being written at the half second boundary.         * We should have an rtc call that only sets the minutes and         * seconds like on Intel to avoid problems with non UTC clocks.         */        if (ppc_md.set_rtc_time && ntp_synced() &&	    xtime.tv_sec - last_rtc_update >= 659 &&	    abs((xtime.tv_nsec/1000) - (1000000-1000000/HZ)) < 500000/HZ) {		struct rtc_time tm;		to_tm(xtime.tv_sec + 1 + timezone_offset, &tm);		tm.tm_year -= 1900;		tm.tm_mon -= 1;		if (ppc_md.set_rtc_time(&tm) == 0)			last_rtc_update = xtime.tv_sec + 1;		else			/* Try again one minute later */			last_rtc_update += 60;        }}/* * This version of gettimeofday has microsecond resolution. */static inline void __do_gettimeofday(struct timeval *tv, u64 tb_val){	unsigned long sec, usec;	u64 tb_ticks, xsec;	struct gettimeofday_vars *temp_varp;	u64 temp_tb_to_xs, temp_stamp_xsec;	/*	 * These calculations are faster (gets rid of divides)	 * if done in units of 1/2^20 rather than microseconds.	 * The conversion to microseconds at the end is done	 * without a divide (and in fact, without a multiply)	 */	temp_varp = do_gtod.varp;	tb_ticks = tb_val - temp_varp->tb_orig_stamp;	temp_tb_to_xs = temp_varp->tb_to_xs;	temp_stamp_xsec = temp_varp->stamp_xsec;	xsec = temp_stamp_xsec + mulhdu(tb_ticks, temp_tb_to_xs);	sec = xsec / XSEC_PER_SEC;	usec = (unsigned long)xsec & (XSEC_PER_SEC - 1);	usec = SCALE_XSEC(usec, 1000000);	tv->tv_sec = sec;	tv->tv_usec = usec;}void do_gettimeofday(struct timeval *tv){	if (__USE_RTC()) {		/* do this the old way */		unsigned long flags, seq;		unsigned int sec, nsec, usec;		do {			seq = read_seqbegin_irqsave(&xtime_lock, flags);			sec = xtime.tv_sec;			nsec = xtime.tv_nsec + tb_ticks_since(tb_last_stamp);		} while (read_seqretry_irqrestore(&xtime_lock, seq, flags));		usec = nsec / 1000;		while (usec >= 1000000) {			usec -= 1000000;			++sec;		}		tv->tv_sec = sec;		tv->tv_usec = usec;		return;	}	__do_gettimeofday(tv, get_tb());}EXPORT_SYMBOL(do_gettimeofday);/* * There are two copies of tb_to_xs and stamp_xsec so that no * lock is needed to access and use these values in * do_gettimeofday.  We alternate the copies and as long as a * reasonable time elapses between changes, there will never * be inconsistent values.  ntpd has a minimum of one minute * between updates. */static inline void update_gtod(u64 new_tb_stamp, u64 new_stamp_xsec,			       u64 new_tb_to_xs){	unsigned temp_idx;	struct gettimeofday_vars *temp_varp;	temp_idx = (do_gtod.var_idx == 0);	temp_varp = &do_gtod.vars[temp_idx];	temp_varp->tb_to_xs = new_tb_to_xs;	temp_varp->tb_orig_stamp = new_tb_stamp;	temp_varp->stamp_xsec = new_stamp_xsec;	smp_mb();	do_gtod.varp = temp_varp;	do_gtod.var_idx = temp_idx;	/*	 * tb_update_count is used to allow the userspace gettimeofday code	 * to assure itself that it sees a consistent view of the tb_to_xs and	 * stamp_xsec variables.  It reads the tb_update_count, then reads	 * tb_to_xs and stamp_xsec and then reads tb_update_count again.  If	 * the two values of tb_update_count match and are even then the	 * tb_to_xs and stamp_xsec values are consistent.  If not, then it	 * loops back and reads them again until this criteria is met.	 * We expect the caller to have done the first increment of	 * vdso_data->tb_update_count already.	 */	vdso_data->tb_orig_stamp = new_tb_stamp;	vdso_data->stamp_xsec = new_stamp_xsec;	vdso_data->tb_to_xs = new_tb_to_xs;	vdso_data->wtom_clock_sec = wall_to_monotonic.tv_sec;	vdso_data->wtom_clock_nsec = wall_to_monotonic.tv_nsec;	smp_wmb();	++(vdso_data->tb_update_count);}/* * When the timebase - tb_orig_stamp gets too big, we do a manipulation * between tb_orig_stamp and stamp_xsec. The goal here is to keep the * difference tb - tb_orig_stamp small enough to always fit inside a * 32 bits number. This is a requirement of our fast 32 bits userland * implementation in the vdso. If we "miss" a call to this function * (interrupt latency, CPU locked in a spinlock, ...) and we end up * with a too big difference, then the vdso will fallback to calling * the syscall */static __inline__ void timer_recalc_offset(u64 cur_tb){	unsigned long offset;	u64 new_stamp_xsec;	u64 tlen, t2x;	u64 tb, xsec_old, xsec_new;	struct gettimeofday_vars *varp;	if (__USE_RTC())		return;	tlen = current_tick_length();	offset = cur_tb - do_gtod.varp->tb_orig_stamp;	if (tlen == last_tick_len && offset < 0x80000000u)		return;	if (tlen != last_tick_len) {		t2x = mulhdu(tlen << TICKLEN_SHIFT, ticklen_to_xs);		last_tick_len = tlen;	} else		t2x = do_gtod.varp->tb_to_xs;	new_stamp_xsec = (u64) xtime.tv_nsec * XSEC_PER_SEC;	do_div(new_stamp_xsec, 1000000000);	new_stamp_xsec += (u64) xtime.tv_sec * XSEC_PER_SEC;	++vdso_data->tb_update_count;	smp_mb();	/*	 * Make sure time doesn't go backwards for userspace gettimeofday.	 */	tb = get_tb();	varp = do_gtod.varp;	xsec_old = mulhdu(tb - varp->tb_orig_stamp, varp->tb_to_xs)		+ varp->stamp_xsec;	xsec_new = mulhdu(tb - cur_tb, t2x) + new_stamp_xsec;	if (xsec_new < xsec_old)		new_stamp_xsec += xsec_old - xsec_new;	update_gtod(cur_tb, new_stamp_xsec, t2x);}#ifdef CONFIG_SMPunsigned long profile_pc(struct pt_regs *regs){	unsigned long pc = instruction_pointer(regs);	if (in_lock_functions(pc))		return regs->link;	return pc;}EXPORT_SYMBOL(profile_pc);#endif#ifdef CONFIG_PPC_ISERIES/*  * This function recalibrates the timebase based on the 49-bit time-of-day * value in the Titan chip.  The Titan is much more accurate than the value * returned by the service processor for the timebase frequency.   */static void iSeries_tb_recal(void){	struct div_result divres;	unsigned long titan, tb;	tb = get_tb();	titan = HvCallXm_loadTod();	if ( iSeries_recal_titan ) {		unsigned long tb_ticks = tb - iSeries_recal_tb;		unsigned long titan_usec = (titan - iSeries_recal_titan) >> 12;		unsigned long new_tb_ticks_per_sec   = (tb_ticks * USEC_PER_SEC)/titan_usec;		unsigned long new_tb_ticks_per_jiffy = (new_tb_ticks_per_sec+(HZ/2))/HZ;		long tick_diff = new_tb_ticks_per_jiffy - tb_ticks_per_jiffy;		char sign = '+';				/* make sure tb_ticks_per_sec and tb_ticks_per_jiffy are consistent */		new_tb_ticks_per_sec = new_tb_ticks_per_jiffy * HZ;		if ( tick_diff < 0 ) {			tick_diff = -tick_diff;			sign = '-';		}		if ( tick_diff ) {			if ( tick_diff < tb_ticks_per_jiffy/25 ) {				printk( "Titan recalibrate: new tb_ticks_per_jiffy = %lu (%c%ld)\n",						new_tb_ticks_per_jiffy, sign, tick_diff );				tb_ticks_per_jiffy = new_tb_ticks_per_jiffy;				tb_ticks_per_sec   = new_tb_ticks_per_sec;				div128_by_32( XSEC_PER_SEC, 0, tb_ticks_per_sec, &divres );				do_gtod.tb_ticks_per_sec = tb_ticks_per_sec;				tb_to_xs = divres.result_low;				do_gtod.varp->tb_to_xs = tb_to_xs;				vdso_data->tb_ticks_per_sec = tb_ticks_per_sec;				vdso_data->tb_to_xs = tb_to_xs;			}			else {				printk( "Titan recalibrate: FAILED (difference > 4 percent)\n"					"                   new tb_ticks_per_jiffy = %lu\n"					"                   old tb_ticks_per_jiffy = %lu\n",					new_tb_ticks_per_jiffy, tb_ticks_per_jiffy );			}		}	}	iSeries_recal_titan = titan;	iSeries_recal_tb = tb;}#endif/* * For iSeries shared processors, we have to let the hypervisor * set the hardware decrementer.  We set a virtual decrementer * in the lppaca and call the hypervisor if the virtual * decrementer is less than the current value in the hardware * decrementer. (almost always the new decrementer value will * be greater than the current hardware decementer so the hypervisor * call will not be needed) *//* * timer_interrupt - gets called when the decrementer overflows, * with interrupts disabled. */void timer_interrupt(struct pt_regs * regs){	int next_dec;	int cpu = smp_processor_id();	unsigned long ticks;#ifdef CONFIG_PPC32	if (atomic_read(&ppc_n_lost_interrupts) != 0)		do_IRQ(regs);#endif	irq_enter();	profile_tick(CPU_PROFILING, regs);#ifdef CONFIG_PPC_ISERIES	get_lppaca()->int_dword.fields.decr_int = 0;#endif	while ((ticks = tb_ticks_since(per_cpu(last_jiffy, cpu)))	       >= tb_ticks_per_jiffy) {		/* Update last_jiffy */		per_cpu(last_jiffy, cpu) += tb_ticks_per_jiffy;		/* Handle RTCL overflow on 601 */		if (__USE_RTC() && per_cpu(last_jiffy, cpu) >= 1000000000)			per_cpu(last_jiffy, cpu) -= 1000000000;		/*		 * We cannot disable the decrementer, so in the period		 * between this cpu's being marked offline in cpu_online_map		 * and calling stop-self, it is taking timer interrupts.		 * Avoid calling into the scheduler rebalancing code if this		 * is the case.		 */		if (!cpu_is_offline(cpu))			update_process_times(user_mode(regs));		/*		 * No need to check whether cpu is offline here; boot_cpuid		 * should have been fixed up by now.		 */		if (cpu != boot_cpuid)			continue;		write_seqlock(&xtime_lock);		tb_last_jiffy += tb_ticks_per_jiffy;		tb_last_stamp = per_cpu(last_jiffy, cpu);		do_timer(regs);		timer_recalc_offset(tb_last_jiffy);		timer_check_rtc();		write_sequnlock(&xtime_lock);	}		next_dec = tb_ticks_per_jiffy - ticks;

⌨️ 快捷键说明

复制代码Ctrl + C
搜索代码Ctrl + F
全屏模式F11
增大字号Ctrl + =
减小字号Ctrl + -
显示快捷键?