⭐ 欢迎来到虫虫下载站! | 📦 资源下载 📁 资源专辑 ℹ️ 关于我们
⭐ 虫虫下载站

📄 archive.txt

📁 mpi并行计算的c++代码 可用vc或gcc编译通过 可以用来搭建并行计算试验环境
💻 TXT
字号:
int MPIDI_CH3I_Progress(int blocking, MPID_Progress_state *state);#define MPIDI_CH3_Progress_test() MPIDI_CH3I_Progress(FALSE, NULL)#define MPIDI_CH3_Progress_wait(state) MPIDI_CH3I_Progress(TRUE, state)/*****************************//*                           *//*   naive progress engine   *//*                           *//*****************************/#undef FUNCNAME#define FUNCNAME MPIDI_CH3_Progress#undef FCNAME#define FCNAME MPIDI_QUOTE(FUNCNAME)int MPIDI_CH3I_Progress(int is_blocking, MPID_Progress_state *state){    int mpi_errno = MPI_SUCCESS;    int rc;    int register count, bShmProgressMade;    MPIDU_Sock_event_t event;    unsigned completions = MPIDI_CH3I_progress_completion_count;    MPIDI_CH3I_Shmem_queue_info info;    int num_bytes;    MPIDI_VC_t *vc_ptr;    static int spin_count = 1;    static int msg_queue_count = 0;    MPIDI_STATE_DECL(MPID_STATE_MPIDI_CH3_PROGRESS);    MPIDI_STATE_DECL(MPID_STATE_MPIDU_YIELD);    MPIDI_FUNC_ENTER(MPID_STATE_MPIDI_CH3_PROGRESS);#ifdef MPICH_DBG_OUTPUT    if (is_blocking)    {	MPIDI_DBG_PRINTF((50, FCNAME, "entering, blocking=%s", is_blocking ? "true" : "false"));    }#endif    do    {	/* make progress on the shared memory queues */	bShmProgressMade = FALSE;	if (MPIDI_CH3I_Process.shm_reading_list)	{	    rc = MPIDI_CH3I_SHM_read_progress(MPIDI_CH3I_Process.shm_reading_list, 0, &vc_ptr, &num_bytes);	    if (rc == MPI_SUCCESS)	    {		MPIDI_DBG_PRINTF((50, FCNAME, "MPIDI_CH3I_SHM_read_progress reported %d bytes read", num_bytes));		mpi_errno = MPIDI_CH3I_Handle_shm_read(vc_ptr, num_bytes);		if (mpi_errno != MPI_SUCCESS)		{		    mpi_errno = MPIR_Err_create_code(mpi_errno, MPIR_ERR_FATAL, FCNAME, __LINE__, MPI_ERR_OTHER, "**progress", 0);		    goto fn_exit;		}		bShmProgressMade = TRUE;	    }	    else	    {		if (rc != SHM_WAIT_TIMEOUT)		{		    /*MPIDI_err_printf("MPIDI_CH3_Progress", "MPIDI_CH3I_SHM_read_progress returned error %d\n", rc);*/		    mpi_errno = MPIR_Err_create_code(rc, MPIR_ERR_FATAL, FCNAME, __LINE__, MPI_ERR_OTHER, "**shm_read_progress", 0);		    goto fn_exit;		}	    }	}	if (MPIDI_CH3I_Process.shm_writing_list)	{	    vc_ptr = MPIDI_CH3I_Process.shm_writing_list;	    while (vc_ptr)	    {		if (vc_ptr->ch.send_active != NULL)		{		    rc = MPIDI_CH3I_SHM_write_progress(vc_ptr);		    if (rc == MPI_SUCCESS)		    {			bShmProgressMade = TRUE;		    }		    else if (rc != SHM_WAIT_TIMEOUT)		    {			mpi_errno = MPIR_Err_create_code(rc, MPIR_ERR_FATAL, FCNAME, __LINE__, MPI_ERR_OTHER, "**progress", 0);			goto fn_exit;		    }		}		vc_ptr = vc_ptr->ch.shm_next_writer;	    }	}	MPIDU_Yield();	/* make progress on the sockets */	mpi_errno = MPIDU_Sock_wait(MPIDI_CH3I_sock_set, 0, &event);	if (mpi_errno == MPI_SUCCESS)	{	    mpi_errno = MPIDI_CH3I_Progress_handle_sock_event(&event);	    if (mpi_errno != MPI_SUCCESS)	    {		mpi_errno = MPIR_Err_create_code(mpi_errno, MPIR_ERR_FATAL, FCNAME, __LINE__, MPI_ERR_OTHER, "**progress_handle_sock_op", 0);		goto fn_exit;	    }	}	else	{	    if (MPIR_ERR_GET_CLASS(mpi_errno) != MPIDU_SOCK_ERR_TIMEOUT)	    {		mpi_errno = MPIR_Err_create_code(mpi_errno, MPIR_ERR_FATAL, FCNAME, __LINE__, MPI_ERR_OTHER, "**progress_sock_wait", 0);		goto fn_exit;	    }	    mpi_errno = MPI_SUCCESS;	    MPIDU_Yield();	}	/* The following check is commented out because it leads to long delays in establishing connections */	/*if ((msg_queue_count++ % MPIDI_CH3I_MSGQ_ITERATIONS) == 0)*/	{	    /* check for new shmem queue connection requests */	    rc = MPIDI_CH3I_BootstrapQ_recv_msg(MPIDI_Process.my_pg->ch.bootstrapQ, &info, sizeof(info), &num_bytes, FALSE);	    if (rc != MPI_SUCCESS)	    {		mpi_errno = MPIR_Err_create_code(rc, MPIR_ERR_FATAL, FCNAME, __LINE__, MPI_ERR_OTHER, "**boot_recv", 0);		goto fn_exit;	    }#ifdef MPICH_DBG_OUTPUT	    /*MPIU_Assert(num_bytes == 0 || num_bytes == sizeof(info));*/	    if (num_bytes != 0 && num_bytes != sizeof(info))	    {		mpi_errno = MPIR_Err_create_code(MPI_SUCCESS, MPIR_ERR_FATAL, FCNAME, __LINE__, MPI_ERR_OTHER, "**bootqmsg", "**bootqmsg %d", num_bytes);		goto fn_exit;	    }#endif	    if (num_bytes)	    {		MPIDI_PG_t *pg;		MPIDI_PG_Find(info.pg_id, &pg);		MPIDI_PG_Get_vc(pg, info.pg_rank, &vc_ptr);		/*vc_ptr = &MPIDI_Process.my_pg->ch.vc_table[info.pg_rank];*/		rc = MPIDI_CH3I_SHM_Attach_to_mem(&info.info, &vc_ptr->ch.shm_read_queue_info);		if (rc != MPI_SUCCESS)		{		    mpi_errno = MPIR_Err_create_code(rc, MPIR_ERR_FATAL, FCNAME, __LINE__, MPI_ERR_OTHER, "**attach_to_mem", "**attach_to_mem %d", vc_ptr->ch.shm_read_queue_info.error);		    goto fn_exit;		}		MPIU_DBG_PRINTF(("attached to queue from process %d\n", info.pg_rank));		/*vc_ptr->ch.state = MPIDI_CH3I_VC_STATE_CONNECTED;*/ /* we are read connected but not write connected */		vc_ptr->ch.shm_read_connected = 1;		vc_ptr->ch.bShm = TRUE;		vc_ptr->ch.read_shmq = vc_ptr->ch.shm_read_queue_info.addr;/*info.info.addr;*/		MPIU_DBG_PRINTF(("read_shmq = %p\n", vc_ptr->ch.read_shmq));		vc_ptr->ch.shm_reading_pkt = TRUE;		/* add this VC to the global list to be shm_waited on */		MPIDI_CH3I_SHM_Add_to_reader_list(vc_ptr);	    }	}    }    while (completions == MPIDI_CH3I_progress_completion_count && is_blocking);fn_exit:#ifdef MPICH_DBG_OUTPUT    count = MPIDI_CH3I_progress_completion_count - completions;    if (is_blocking)    {	MPIDI_DBG_PRINTF((50, FCNAME, "exiting, count=%d", count));    }    else    {	if (count > 0)	{	    MPIDI_DBG_PRINTF((50, FCNAME, "exiting (non-blocking), count=%d", count));	}    }#endif    MPIDI_FUNC_EXIT(MPID_STATE_MPIDI_CH3_PROGRESS);    return mpi_errno;}----------#if 0/* Here's a clip of the SHARED_PROCESS_READ code *//* allocation */#ifdef HAVE_WINDOWS_H            pSharedProcessHandles[i] =                OpenProcess(STANDARD_RIGHTS_REQUIRED | PROCESS_VM_READ | PROCESS_VM_WRITE | PROCESS_VM_OPERATION,                             FALSE, pSharedProcess[i].nPid);            if (pSharedProcessHandles[i] == NULL)            {                int err = GetLastError();                mpi_errno = MPIR_Err_create_code(MPI_SUCCESS, MPIR_ERR_FATAL, FCNAME, __LINE__, MPI_ERR_OTHER, "**OpenProcess", "**OpenProcess %d %d", i, err); /*"unable to open process %d, error %d", i, err);*/            }#else            MPIU_Snprintf(filename, 256, "/proc/%d/mem", pSharedProcess[i].nPid);            pSharedProcessIDs[i] = pSharedProcess[i].nPid;            pSharedProcessFileDescriptors[i] = open(filename, O_RDONLY);            if (pSharedProcessFileDescriptors[i] == -1)	    {                mpi_errno = MPIR_Err_create_code(MPI_SUCCESS, MPIR_ERR_FATAL, FCNAME, __LINE__, MPI_ERR_OTHER, "**open", "**open %s %d %d", filename, pSharedProcess[i].nPid, errno); /*"failed to open mem file, '%s', for process %d", filename, pSharedProcess[i].nPid);*/	    }#endif/* deallocation */#ifdef HAVE_SHARED_PROCESS_READ#ifdef HAVE_WINDOWS_H	for (i=0; i<pg->size; i++)	    CloseHandle(pg->pSharedProcessHandles[i]);	MPIU_Free(pg->pSharedProcessHandles);	pg->pSharedProcessHandles = NULL;#else	for (i=0; i<pg->size; i++)	    close(pg->pSharedProcessFileDescriptors[i]);	MPIU_Free(pg->pSharedProcessFileDescriptors);	MPIU_Free(pg->pSharedProcessIDs);	pg->pSharedProcessFileDescriptors = NULL;	pg->pSharedProcessIDs = NULL;#endif#endif#endif /* 0 */

⌨️ 快捷键说明

复制代码 Ctrl + C
搜索代码 Ctrl + F
全屏模式 F11
切换主题 Ctrl + Shift + D
显示快捷键 ?
增大字号 Ctrl + =
减小字号 Ctrl + -