⭐ 欢迎来到虫虫下载站! | 📦 资源下载 📁 资源专辑 ℹ️ 关于我们
⭐ 虫虫下载站

📄 irq.c

📁 是关于linux2.5.1的完全源码
💻 C
📖 第 1 页 / 共 2 页
字号:
	   Since we set PENDING, if another processor is handling	   a different instance of this same irq, the other processor	   will take care of it.	 */	if (!action)		goto out;	/*	 * Edge triggered interrupts need to remember	 * pending events.	 * This applies to any hw interrupts that allow a second	 * instance of the same irq to arrive while we are in do_IRQ	 * or in the handler. But the code here only handles the _second_	 * instance of the irq, not the third or fourth. So it is mostly	 * useful for irq hardware that does not mask cleanly in an	 * SMP environment.	 */	for (;;) {		spin_unlock(&desc->lock);		handle_irq_event(irq, regs, action);		spin_lock(&desc->lock);				if (!(desc->status & IRQ_PENDING))			break;		desc->status &= ~IRQ_PENDING;	}	desc->status &= ~IRQ_INPROGRESS;out:	/*	 * The ->end() handler has to deal with interrupts which got	 * disabled while the handler was running.	 */	if (irq_desc[irq].handler) {		if (irq_desc[irq].handler->end)			irq_desc[irq].handler->end(irq);		else if (irq_desc[irq].handler->enable)			irq_desc[irq].handler->enable(irq);	}	spin_unlock(&desc->lock);}int do_IRQ(struct pt_regs *regs, int isfake){	int cpu = smp_processor_id();	int irq;	struct Paca * paca;	struct ItLpQueue * lpq;	/* if(cpu) udbg_printf("Entering do_IRQ\n");  */        irq_enter(cpu);	if ( _machine != _MACH_iSeries ) {			/* every arch is required to have a get_irq -- Cort */		irq = ppc_md.get_irq( regs );		if ( irq >= 0 ) {			ppc_irq_dispatch_handler( regs, irq );			if (ppc_md.post_irq)				ppc_md.post_irq( regs, irq );		} else {			/* -2 means ignore, already handled */			if (irq != -2) {				printk(KERN_DEBUG "Bogus interrupt %d from PC = %lx\n",					irq, regs->nip);				ppc_spurious_interrupts++;			}		}	}	/* if on iSeries partition */	else {		paca = (struct Paca *)mfspr(SPRG3);#ifdef CONFIG_SMP		if ( paca->xLpPaca.xIntDword.xFields.xIpiCnt ) {			paca->xLpPaca.xIntDword.xFields.xIpiCnt = 0;			iSeries_smp_message_recv( regs );		}#endif /* CONFIG_SMP */		lpq = paca->lpQueuePtr;		if ( lpq && ItLpQueue_isLpIntPending( lpq ) )			lpEvent_count += ItLpQueue_process( lpq, regs );	}		        irq_exit(cpu);	if ( _machine == _MACH_iSeries ) {		if ( paca->xLpPaca.xIntDword.xFields.xDecrInt ) {			paca->xLpPaca.xIntDword.xFields.xDecrInt = 0;			/* Signal a fake decrementer interrupt */			timer_interrupt( regs );		}	}	if (softirq_pending(cpu))		do_softirq();	return 1; /* lets ret_from_int know we can do checks */}unsigned long probe_irq_on (void){	return 0;}int probe_irq_off (unsigned long irqs){	return 0;}unsigned int probe_irq_mask(unsigned long irqs){	return 0;}void __init init_IRQ(void){	static int once = 0;	if ( once )		return;	else		once++;		ppc_md.init_IRQ();	if(ppc_md.init_ras_IRQ) ppc_md.init_ras_IRQ(); }#ifdef CONFIG_SMPunsigned char global_irq_holder = NO_PROC_ID;static void show(char * str){	int cpu = smp_processor_id();	int i;	printk("\n%s, CPU %d:\n", str, cpu);	printk("irq:  %d [ ", irqs_running());	for (i = 0; i < smp_num_cpus; i++)		printk("%u ", __brlock_array[i][BR_GLOBALIRQ_LOCK]);	printk("]\nbh:   %d [ ",		(spin_is_locked(&global_bh_lock) ? 1 : 0));	for (i = 0; i < smp_num_cpus; i++)		printk("%u ", local_bh_count(i));	printk("]\n");}#define MAXCOUNT 10000000void synchronize_irq(void){	if (irqs_running()) {		cli();		sti();	}}static inline void get_irqlock(int cpu){        int count;        if ((unsigned char)cpu == global_irq_holder)                return;        count = MAXCOUNT;again:        br_write_lock(BR_GLOBALIRQ_LOCK);        for (;;) {                spinlock_t *lock;                if (!irqs_running() &&                    (local_bh_count(smp_processor_id()) || !spin_is_locked(&global_bh_lock)))                        break;                br_write_unlock(BR_GLOBALIRQ_LOCK);                lock = &__br_write_locks[BR_GLOBALIRQ_LOCK].lock;                while (irqs_running() ||                       spin_is_locked(lock) ||                       (!local_bh_count(smp_processor_id()) && spin_is_locked(&global_bh_lock))) {                        if (!--count) {                                show("get_irqlock");                                count = (~0 >> 1);                        }                        __sti();                        barrier();                        __cli();                }                goto again;        }        global_irq_holder = cpu;}/* * A global "cli()" while in an interrupt context * turns into just a local cli(). Interrupts * should use spinlocks for the (very unlikely) * case that they ever want to protect against * each other. * * If we already have local interrupts disabled, * this will not turn a local disable into a * global one (problems with spinlocks: this makes * save_flags+cli+sti usable inside a spinlock). */void __global_cli(void){	unsigned long flags;		__save_flags(flags);	if (flags & (1UL << 15)) {		int cpu = smp_processor_id();		__cli();		if (!local_irq_count(cpu))			get_irqlock(cpu);	}}void __global_sti(void){	int cpu = smp_processor_id();	if (!local_irq_count(cpu))		release_irqlock(cpu);	__sti();}/* * SMP flags value to restore to: * 0 - global cli * 1 - global sti * 2 - local cli * 3 - local sti */unsigned long __global_save_flags(void){	int retval;	int local_enabled;	unsigned long flags;	__save_flags(flags);	local_enabled = (flags >> 15) & 1;	/* default to local */	retval = 2 + local_enabled;	/* check for global flags if we're not in an interrupt */	if (!local_irq_count(smp_processor_id())) {		if (local_enabled)			retval = 1;		if (global_irq_holder == (unsigned char) smp_processor_id())			retval = 0;	}	return retval;}void __global_restore_flags(unsigned long flags){	switch (flags) {	case 0:		__global_cli();		break;	case 1:		__global_sti();		break;	case 2:		__cli();		break;	case 3:		__sti();		break;	default:		printk("global_restore_flags: %016lx caller %p\n",			flags, __builtin_return_address(0));	}}#endif /* CONFIG_SMP */static struct proc_dir_entry * root_irq_dir;static struct proc_dir_entry * irq_dir [NR_IRQS];static struct proc_dir_entry * smp_affinity_entry [NR_IRQS];#ifdef CONFIG_IRQ_ALL_CPUSunsigned int irq_affinity [NR_IRQS] = { [0 ... NR_IRQS-1] = 0xffffffff};#else  /* CONFIG_IRQ_ALL_CPUS */unsigned int irq_affinity [NR_IRQS] = { [0 ... NR_IRQS-1] = 0x00000000};#endif /* CONFIG_IRQ_ALL_CPUS */#define HEX_DIGITS 8static int irq_affinity_read_proc (char *page, char **start, off_t off,			int count, int *eof, void *data){	if (count < HEX_DIGITS+1)		return -EINVAL;	return sprintf (page, "%08x\n", irq_affinity[(int)(long)data]);}static unsigned int parse_hex_value (const char *buffer,		unsigned long count, unsigned long *ret){	unsigned char hexnum [HEX_DIGITS];	unsigned long value;	int i;	if (!count)		return -EINVAL;	if (count > HEX_DIGITS)		count = HEX_DIGITS;	if (copy_from_user(hexnum, buffer, count))		return -EFAULT;	/*	 * Parse the first 8 characters as a hex string, any non-hex char	 * is end-of-string. '00e1', 'e1', '00E1', 'E1' are all the same.	 */	value = 0;	for (i = 0; i < count; i++) {		unsigned int c = hexnum[i];		switch (c) {			case '0' ... '9': c -= '0'; break;			case 'a' ... 'f': c -= 'a'-10; break;			case 'A' ... 'F': c -= 'A'-10; break;		default:			goto out;		}		value = (value << 4) | c;	}out:	*ret = value;	return 0;}static int irq_affinity_write_proc (struct file *file, const char *buffer,					unsigned long count, void *data){	int irq = (int)(long) data, full_count = count, err;	unsigned long new_value;	if (!irq_desc[irq].handler->set_affinity)		return -EIO;	err = parse_hex_value(buffer, count, &new_value);/* Why is this disabled ? --BenH */#if 0/*CONFIG_SMP*/	/*	 * Do not allow disabling IRQs completely - it's a too easy	 * way to make the system unusable accidentally :-) At least	 * one online CPU still has to be targeted.	 */	if (!(new_value & cpu_online_map))		return -EINVAL;#endif	irq_affinity[irq] = new_value;	irq_desc[irq].handler->set_affinity(irq, new_value);	return full_count;}static int prof_cpu_mask_read_proc (char *page, char **start, off_t off,			int count, int *eof, void *data){	unsigned long *mask = (unsigned long *) data;	if (count < HEX_DIGITS+1)		return -EINVAL;	return sprintf (page, "%08lx\n", *mask);}static int prof_cpu_mask_write_proc (struct file *file, const char *buffer,					unsigned long count, void *data){	unsigned long *mask = (unsigned long *) data, full_count = count, err;	unsigned long new_value;	err = parse_hex_value(buffer, count, &new_value);	if (err)		return err;	*mask = new_value;#ifdef CONFIG_PPC_ISERIES	{		unsigned i;		for (i=0; i<maxPacas; ++i) {			if ( xPaca[i].prof_buffer && (new_value & 1) )				xPaca[i].prof_enabled = 1;			else				xPaca[i].prof_enabled = 0;			new_value >>= 1;		}	}#endif	return full_count;}#define MAX_NAMELEN 10static void register_irq_proc (unsigned int irq){	struct proc_dir_entry *entry;	char name [MAX_NAMELEN];	if (!root_irq_dir || (irq_desc[irq].handler == NULL))		return;	memset(name, 0, MAX_NAMELEN);	sprintf(name, "%d", irq);	/* create /proc/irq/1234 */	irq_dir[irq] = proc_mkdir(name, root_irq_dir);	/* create /proc/irq/1234/smp_affinity */	entry = create_proc_entry("smp_affinity", 0600, irq_dir[irq]);	entry->nlink = 1;	entry->data = (void *)(long)irq;	entry->read_proc = irq_affinity_read_proc;	entry->write_proc = irq_affinity_write_proc;	smp_affinity_entry[irq] = entry;}unsigned long prof_cpu_mask = -1;void init_irq_proc (void){	struct proc_dir_entry *entry;	int i;	/* create /proc/irq */	root_irq_dir = proc_mkdir("irq", 0);	/* create /proc/irq/prof_cpu_mask */	entry = create_proc_entry("prof_cpu_mask", 0600, root_irq_dir);	entry->nlink = 1;	entry->data = (void *)&prof_cpu_mask;	entry->read_proc = prof_cpu_mask_read_proc;	entry->write_proc = prof_cpu_mask_write_proc;	/*	 * Create entries for all existing IRQs.	 */	for (i = 0; i < NR_IRQS; i++) {		if (irq_desc[i].handler == NULL)			continue;		register_irq_proc(i);	}}void no_action(int irq, void *dev, struct pt_regs *regs){}

⌨️ 快捷键说明

复制代码 Ctrl + C
搜索代码 Ctrl + F
全屏模式 F11
切换主题 Ctrl + Shift + D
显示快捷键 ?
增大字号 Ctrl + =
减小字号 Ctrl + -