⭐ 欢迎来到虫虫下载站! | 📦 资源下载 📁 资源专辑 ℹ️ 关于我们
⭐ 虫虫下载站

📄 smp.c

📁 这个linux源代码是很全面的~基本完整了~使用c编译的~由于时间问题我没有亲自测试~但就算用来做参考资料也是非常好的
💻 C
📖 第 1 页 / 共 2 页
字号:
/* *  arch/s390/kernel/smp.c * *  S390 version *    Copyright (C) 1999,2000 IBM Deutschland Entwicklung GmbH, IBM Corporation *    Author(s): Denis Joseph Barrow (djbarrow@de.ibm.com,barrow_dj@yahoo.com), *               Martin Schwidefsky (schwidefsky@de.ibm.com) * *  based on other smp stuff by  *    (c) 1995 Alan Cox, CymruNET Ltd  <alan@cymru.net> *    (c) 1998 Ingo Molnar * * We work with logical cpu numbering everywhere we can. The only * functions using the real cpu address (got from STAP) are the sigp * functions. For all other functions we use the identity mapping. * That means that cpu_number_map[i] == i for every cpu. cpu_number_map is * used e.g. to find the idle task belonging to a logical cpu. Every array * in the kernel is sorted by the logical cpu number and not by the physical * one which is causing all the confusion with __cpu_logical_map and * cpu_number_map in other architectures. */#include <linux/module.h>#include <linux/init.h>#include <linux/mm.h>#include <linux/spinlock.h>#include <linux/kernel_stat.h>#include <linux/smp_lock.h>#include <linux/delay.h>#include <linux/cache.h>#include <asm/sigp.h>#include <asm/pgalloc.h>#include <asm/irq.h>#include <asm/s390_ext.h>#include <asm/cpcmd.h>/* prototypes */extern int cpu_idle(void * unused);extern __u16 boot_cpu_addr;extern volatile int __cpu_logical_map[];/* * An array with a pointer the lowcore of every CPU. */static int       max_cpus = NR_CPUS;	  /* Setup configured maximum number of CPUs to activate	*/int              smp_num_cpus;struct _lowcore *lowcore_ptr[NR_CPUS];cycles_t         cacheflush_time=0;int              smp_threads_ready=0;      /* Set when the idlers are all forked. */static atomic_t  smp_commenced = ATOMIC_INIT(0);spinlock_t       kernel_flag __cacheline_aligned_in_smp = SPIN_LOCK_UNLOCKED;unsigned long	 cpu_online_map;/* *      Setup routine for controlling SMP activation * *      Command-line option of "nosmp" or "maxcpus=0" will disable SMP *      activation entirely (the MPS table probe still happens, though). * *      Command-line option of "maxcpus=<NUM>", where <NUM> is an integer *      greater than 0, limits the maximum number of CPUs activated in *      SMP mode to <NUM>. */static int __init nosmp(char *str){	max_cpus = 0;	return 1;}__setup("nosmp", nosmp);static int __init maxcpus(char *str){	get_option(&str, &max_cpus);	return 1;}__setup("maxcpus=", maxcpus);/* * Reboot, halt and power_off routines for SMP. */extern char vmhalt_cmd[];extern char vmpoff_cmd[];extern void reipl(unsigned long devno);static sigp_ccode smp_ext_bitcall(int, ec_bit_sig);static void smp_ext_bitcall_others(ec_bit_sig);/* * Structure and data for smp_call_function(). This is designed to minimise * static memory requirements. It also looks cleaner. */static spinlock_t call_lock = SPIN_LOCK_UNLOCKED;struct call_data_struct {	void (*func) (void *info);	void *info;	atomic_t started;	atomic_t finished;	int wait;};static struct call_data_struct * call_data;/* * 'Call function' interrupt callback */static void do_call_function(void){	void (*func) (void *info) = call_data->func;	void *info = call_data->info;	int wait = call_data->wait;	atomic_inc(&call_data->started);	(*func)(info);	if (wait)		atomic_inc(&call_data->finished);}/* * this function sends a 'generic call function' IPI to all other CPUs * in the system. */int smp_call_function (void (*func) (void *info), void *info, int nonatomic,			int wait)/* * [SUMMARY] Run a function on all other CPUs. * <func> The function to run. This must be fast and non-blocking. * <info> An arbitrary pointer to pass to the function. * <nonatomic> currently unused. * <wait> If true, wait (atomically) until function has completed on other CPUs. * [RETURNS] 0 on success, else a negative status code. Does not return until * remote CPUs are nearly ready to execute <<func>> or are or have executed. * * You must not call this function with disabled interrupts or from a * hardware interrupt handler, you may call it from a bottom half handler. */{	struct call_data_struct data;	int cpus = smp_num_cpus-1;	if (!cpus || !atomic_read(&smp_commenced))		return 0;	data.func = func;	data.info = info;	atomic_set(&data.started, 0);	data.wait = wait;	if (wait)		atomic_set(&data.finished, 0);	spin_lock_bh(&call_lock);	call_data = &data;	/* Send a message to all other CPUs and wait for them to respond */        smp_ext_bitcall_others(ec_call_function);	/* Wait for response */	while (atomic_read(&data.started) != cpus)		barrier();	if (wait)		while (atomic_read(&data.finished) != cpus)			barrier();	spin_unlock_bh(&call_lock);	return 0;}static inline void do_send_stop(void){        u32 dummy;        int i;        /* stop all processors */        for (i =  0; i < smp_num_cpus; i++) {                if (smp_processor_id() != i) {                        int ccode;                        do {                                ccode = signal_processor_ps(                                   &dummy,                                   0,                                   i,                                   sigp_stop);                        } while(ccode == sigp_busy);                }        }}static inline void do_store_status(void){        unsigned long low_core_addr;        u32 dummy;        int i;        /* store status of all processors in their lowcores (real 0) */        for (i =  0; i < smp_num_cpus; i++) {                if (smp_processor_id() != i) {                        int ccode;                        low_core_addr = (unsigned long)get_cpu_lowcore(i);                        do {                                ccode = signal_processor_ps(                                   &dummy,                                   low_core_addr,                                   i,                                   sigp_store_status_at_address);                        } while(ccode == sigp_busy);                }        }}/* * this function sends a 'stop' sigp to all other CPUs in the system. * it goes straight through. */void smp_send_stop(void){        /* write magic number to zero page (absolute 0) */        get_cpu_lowcore(smp_processor_id())->panic_magic = __PANIC_MAGIC;	/* stop other processors. */	do_send_stop();	/* store status of other processors. */	do_store_status();}/* * Reboot, halt and power_off routines for SMP. */static volatile unsigned long cpu_restart_map;static void do_machine_restart(void * __unused){	clear_bit(smp_processor_id(), &cpu_restart_map);	if (smp_processor_id() == 0) {		/* Wait for all other cpus to enter do_machine_restart. */		while (cpu_restart_map != 0);		/* Store status of other cpus. */		do_store_status();		/*		 * Finally call reipl. Because we waited for all other		 * cpus to enter this function we know that they do		 * not hold any s390irq-locks (the cpus have been		 * interrupted by an external interrupt and s390irq		 * locks are always held disabled).		 */		reipl(S390_lowcore.ipl_device);	}	signal_processor(smp_processor_id(), sigp_stop);}void machine_restart_smp(char * __unused) {	cpu_restart_map = cpu_online_map;        smp_call_function(do_machine_restart, NULL, 0, 0);	do_machine_restart(NULL);}static void do_machine_halt(void * __unused){	if (smp_processor_id() == 0) {		smp_send_stop();		if (MACHINE_IS_VM && strlen(vmhalt_cmd) > 0)			cpcmd(vmhalt_cmd, NULL, 0);		signal_processor(smp_processor_id(),				 sigp_stop_and_store_status);	}	for (;;)		enabled_wait();}void machine_halt_smp(void){        smp_call_function(do_machine_halt, NULL, 0, 0);	do_machine_halt(NULL);}static void do_machine_power_off(void * __unused){	if (smp_processor_id() == 0) {		smp_send_stop();		if (MACHINE_IS_VM && strlen(vmpoff_cmd) > 0)			cpcmd(vmpoff_cmd, NULL, 0);		signal_processor(smp_processor_id(),				 sigp_stop_and_store_status);	}	for (;;)		enabled_wait();}void machine_power_off_smp(void){        smp_call_function(do_machine_power_off, NULL, 0, 0);	do_machine_power_off(NULL);}/* * This is the main routine where commands issued by other * cpus are handled. */void do_ext_call_interrupt(struct pt_regs *regs, __u16 code){        int bits;        /*         * handle bit signal external calls         *         * For the ec_schedule signal we have to do nothing. All the work         * is done automatically when we return from the interrupt.         */        do {                bits = atomic_read(&S390_lowcore.ext_call_fast);        } while (atomic_compare_and_swap(bits,0,&S390_lowcore.ext_call_fast));	if (test_bit(ec_call_function, &bits)) 		do_call_function();}/* * Send an external call sigp to another cpu and return without waiting * for its completion. */

⌨️ 快捷键说明

复制代码 Ctrl + C
搜索代码 Ctrl + F
全屏模式 F11
切换主题 Ctrl + Shift + D
显示快捷键 ?
增大字号 Ctrl + =
减小字号 Ctrl + -