uvm_pdaemon.c

来自「基于组件方式开发操作系统的OSKIT源代码」· C语言 代码 · 共 1,116 行 · 第 1/2 页

C
1,116
字号
/*	$NetBSD: uvm_pdaemon.c,v 1.25 2000/11/30 11:04:44 simonb Exp $	*//*  * Copyright (c) 1997 Charles D. Cranor and Washington University. * Copyright (c) 1991, 1993, The Regents of the University of California.   * * All rights reserved. * * This code is derived from software contributed to Berkeley by * The Mach Operating System project at Carnegie-Mellon University. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions * are met: * 1. Redistributions of source code must retain the above copyright *    notice, this list of conditions and the following disclaimer. * 2. Redistributions in binary form must reproduce the above copyright *    notice, this list of conditions and the following disclaimer in the *    documentation and/or other materials provided with the distribution. * 3. All advertising materials mentioning features or use of this software *    must display the following acknowledgement: *	This product includes software developed by Charles D. Cranor, *      Washington University, the University of California, Berkeley and  *      its contributors. * 4. Neither the name of the University nor the names of its contributors *    may be used to endorse or promote products derived from this software *    without specific prior written permission. * * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE * ARE DISCLAIMED.  IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF * SUCH DAMAGE. * *	@(#)vm_pageout.c        8.5 (Berkeley) 2/14/94 * from: Id: uvm_pdaemon.c,v 1.1.2.32 1998/02/06 05:26:30 chs Exp * * * Copyright (c) 1987, 1990 Carnegie-Mellon University. * All rights reserved. *  * Permission to use, copy, modify and distribute this software and * its documentation is hereby granted, provided that both the copyright * notice and this permission notice appear in all copies of the * software, derivative works or modified versions, and any portions * thereof, and that both notices appear in supporting documentation. *  * CARNEGIE MELLON ALLOWS FREE USE OF THIS SOFTWARE IN ITS "AS IS"  * CONDITION.  CARNEGIE MELLON DISCLAIMS ANY LIABILITY OF ANY KIND  * FOR ANY DAMAGES WHATSOEVER RESULTING FROM THE USE OF THIS SOFTWARE. *  * Carnegie Mellon requests users of this software to return to * *  Software Distribution Coordinator  or  Software.Distribution@CS.CMU.EDU *  School of Computer Science *  Carnegie Mellon University *  Pittsburgh PA 15213-3890 * * any improvements or extensions that they make and grant Carnegie the * rights to redistribute these changes. */#include "opt_uvmhist.h"/* * uvm_pdaemon.c: the page daemon */#include <sys/param.h>#include <sys/proc.h>#include <sys/systm.h>#include <sys/kernel.h>#include <sys/pool.h>#include <sys/buf.h>#include <uvm/uvm.h>#ifdef OSKIT#include "oskit_uvm_internal.h"#endifextern struct uvm_pagerops uvm_vnodeops;/* * UVMPD_NUMDIRTYREACTS is how many dirty pages the pagedeamon will reactivate * in a pass thru the inactive list when swap is full.  the value should be * "small"... if it's too large we'll cycle the active pages thru the inactive * queue too quickly to for them to be referenced and avoid being freed. */#define UVMPD_NUMDIRTYREACTS 16/* * local prototypes */static void		uvmpd_scan __P((void));static boolean_t	uvmpd_scan_inactive __P((struct pglist *));static void		uvmpd_tune __P((void));/* * uvm_wait: wait (sleep) for the page daemon to free some pages * * => should be called with all locks released * => should _not_ be called by the page daemon (to avoid deadlock) */voiduvm_wait(wmsg)	const char *wmsg;{	int timo = 0;	int s = splbio();	XPRINTF(OSKIT_DEBUG_SYNC, __FUNCTION__": wmsg = %s\n", wmsg);	/*	 * check for page daemon going to sleep (waiting for itself)	 */	if (curproc == uvm.pagedaemon_proc) {		/*		 * now we have a problem: the pagedaemon wants to go to		 * sleep until it frees more memory.   but how can it		 * free more memory if it is asleep?  that is a deadlock.		 * we have two options:		 *  [1] panic now		 *  [2] put a timeout on the sleep, thus causing the		 *      pagedaemon to only pause (rather than sleep forever)		 *		 * note that option [2] will only help us if we get lucky		 * and some other process on the system breaks the deadlock		 * by exiting or freeing memory (thus allowing the pagedaemon		 * to continue).  for now we panic if DEBUG is defined,		 * otherwise we hope for the best with option [2] (better		 * yet, this should never happen in the first place!).		 */		printf("pagedaemon: deadlock detected!\n");		timo = hz >> 3;		/* set timeout */#if defined(DEBUG)		/* DEBUG: panic so we can debug it */		panic("pagedaemon deadlock");#endif	}	simple_lock(&uvm.pagedaemon_lock);	wakeup(&uvm.pagedaemon);		/* wake the daemon! */	UVM_UNLOCK_AND_WAIT(&uvmexp.free, &uvm.pagedaemon_lock, FALSE, wmsg,	    timo);	splx(s);}/* * uvmpd_tune: tune paging parameters * * => called when ever memory is added (or removed?) to the system * => caller must call with page queues locked */static voiduvmpd_tune(){	UVMHIST_FUNC("uvmpd_tune"); UVMHIST_CALLED(pdhist);	uvmexp.freemin = uvmexp.npages / 20;	/* between 16k and 256k */	/* XXX:  what are these values good for? */	uvmexp.freemin = max(uvmexp.freemin, (16*1024) >> PAGE_SHIFT);	uvmexp.freemin = min(uvmexp.freemin, (256*1024) >> PAGE_SHIFT);	/* Make sure there's always a user page free. */	if (uvmexp.freemin < uvmexp.reserve_kernel + 1)		uvmexp.freemin = uvmexp.reserve_kernel + 1;	uvmexp.freetarg = (uvmexp.freemin * 4) / 3;	if (uvmexp.freetarg <= uvmexp.freemin)		uvmexp.freetarg = uvmexp.freemin + 1;	/* uvmexp.inactarg: computed in main daemon loop */	uvmexp.wiredmax = uvmexp.npages / 3;	UVMHIST_LOG(pdhist, "<- done, freemin=%d, freetarg=%d, wiredmax=%d",	      uvmexp.freemin, uvmexp.freetarg, uvmexp.wiredmax, 0);}/* * uvm_pageout: the main loop for the pagedaemon */voiduvm_pageout(void *arg){	int npages = 0;	UVMHIST_FUNC("uvm_pageout"); UVMHIST_CALLED(pdhist);	UVMHIST_LOG(pdhist,"<starting uvm pagedaemon>", 0, 0, 0, 0);	/*	 * ensure correct priority and set paging parameters...	 */	uvm.pagedaemon_proc = curproc;	(void) spl0();	uvm_lock_pageq();	npages = uvmexp.npages;	uvmpd_tune();	uvm_unlock_pageq();	/*	 * main loop	 */	for (;;) {		simple_lock(&uvm.pagedaemon_lock);		UVMHIST_LOG(pdhist,"  <<SLEEPING>>",0,0,0,0);		UVM_UNLOCK_AND_WAIT(&uvm.pagedaemon,		    &uvm.pagedaemon_lock, FALSE, "pgdaemon", 0);		uvmexp.pdwoke++;		UVMHIST_LOG(pdhist,"  <<WOKE UP>>",0,0,0,0);		/* drain pool resources */		pool_drain(0);		/*		 * now lock page queues and recompute inactive count		 */		uvm_lock_pageq();		if (npages != uvmexp.npages) {	/* check for new pages? */			npages = uvmexp.npages;			uvmpd_tune();		}		uvmexp.inactarg = (uvmexp.active + uvmexp.inactive) / 3;		if (uvmexp.inactarg <= uvmexp.freetarg) {			uvmexp.inactarg = uvmexp.freetarg + 1;		}		UVMHIST_LOG(pdhist,"  free/ftarg=%d/%d, inact/itarg=%d/%d",		    uvmexp.free, uvmexp.freetarg, uvmexp.inactive,		    uvmexp.inactarg);		/*		 * scan if needed		 */		if (uvmexp.free + uvmexp.paging < uvmexp.freetarg ||		    uvmexp.inactive < uvmexp.inactarg ||		    uvmexp.vnodepages >		    (uvmexp.active + uvmexp.inactive + uvmexp.wired +		     uvmexp.free) * 13 / 16) {			uvmpd_scan();		}		/*		 * if there's any free memory to be had,		 * wake up any waiters.		 */		if (uvmexp.free > uvmexp.reserve_kernel ||		    uvmexp.paging == 0) {			wakeup(&uvmexp.free);		}		/*		 * scan done.  unlock page queues (the only lock we are holding)		 */		uvm_unlock_pageq();	}	/*NOTREACHED*/#ifdef OSKIT	osenv_process_unlock();	/* never executed */#endif}/* * uvm_aiodone_daemon:  main loop for the aiodone daemon. */voiduvm_aiodone_daemon(void *arg){	int s, free;	struct buf *bp, *nbp;	UVMHIST_FUNC("uvm_aiodoned"); UVMHIST_CALLED(pdhist);	for (;;) {		/*		 * carefully attempt to go to sleep (without losing "wakeups"!).		 * we need splbio because we want to make sure the aio_done list		 * is totally empty before we go to sleep.		 */		s = splbio();		simple_lock(&uvm.aiodoned_lock);		if (TAILQ_FIRST(&uvm.aio_done) == NULL) {			UVMHIST_LOG(pdhist,"  <<SLEEPING>>",0,0,0,0);			UVM_UNLOCK_AND_WAIT(&uvm.aiodoned,			    &uvm.aiodoned_lock, FALSE, "aiodoned", 0);			UVMHIST_LOG(pdhist,"  <<WOKE UP>>",0,0,0,0);			/* relock aiodoned_lock, still at splbio */			simple_lock(&uvm.aiodoned_lock);		}		/*		 * check for done aio structures		 */		bp = TAILQ_FIRST(&uvm.aio_done);		if (bp) {			TAILQ_INIT(&uvm.aio_done);		}		simple_unlock(&uvm.aiodoned_lock);		splx(s);		/*		 * process each i/o that's done.		 */		free = uvmexp.free;		while (bp != NULL) {			if (bp->b_flags & B_PDAEMON) {				uvmexp.paging -= bp->b_bufsize >> PAGE_SHIFT;			}			nbp = TAILQ_NEXT(bp, b_freelist);			(*bp->b_iodone)(bp);			bp = nbp;		}		if (free <= uvmexp.reserve_kernel) {			s = uvm_lock_fpageq();			wakeup(&uvm.pagedaemon);			uvm_unlock_fpageq(s);		} else {			simple_lock(&uvm.pagedaemon_lock);			wakeup(&uvmexp.free);			simple_unlock(&uvm.pagedaemon_lock);		}	}}/* * uvmpd_scan_inactive: scan an inactive list for pages to clean or free. * * => called with page queues locked * => we work on meeting our free target by converting inactive pages *    into free pages. * => we handle the building of swap-backed clusters * => we return TRUE if we are exiting because we met our target */static boolean_tuvmpd_scan_inactive(pglst)	struct pglist *pglst;{	boolean_t retval = FALSE;	/* assume we haven't hit target */	int s, free, result;	struct vm_page *p, *nextpg;	struct uvm_object *uobj;	struct vm_page *pps[MAXBSIZE >> PAGE_SHIFT], **ppsp;	int npages;	struct vm_page *swpps[MAXBSIZE >> PAGE_SHIFT]; 	/* XXX: see below */	int swnpages, swcpages;				/* XXX: see below */	int swslot;	struct vm_anon *anon;	boolean_t swap_backed, vnode_only;	vaddr_t start;	int dirtyreacts, vpgs;	UVMHIST_FUNC("uvmpd_scan_inactive"); UVMHIST_CALLED(pdhist);	/*	 * note: we currently keep swap-backed pages on a seperate inactive	 * list from object-backed pages.   however, merging the two lists	 * back together again hasn't been ruled out.   thus, we keep our	 * swap cluster in "swpps" rather than in pps (allows us to mix	 * clustering types in the event of a mixed inactive queue).	 */	/*	 * swslot is non-zero if we are building a swap cluster.  we want	 * to stay in the loop while we have a page to scan or we have	 * a swap-cluster to build.	 */	swslot = 0;	swnpages = swcpages = 0;	free = 0;	dirtyreacts = 0;	vnode_only = FALSE;	for (p = TAILQ_FIRST(pglst); p != NULL || swslot != 0; p = nextpg) {		/*		 * note that p can be NULL iff we have traversed the whole		 * list and need to do one final swap-backed clustered pageout.		 */		uobj = NULL;		anon = NULL;		if (p) {			/*			 * update our copy of "free" and see if we've met			 * our target			 */			s = uvm_lock_fpageq();			free = uvmexp.free;			uvm_unlock_fpageq(s);			/* XXXUBC */			vpgs = uvmexp.vnodepages -				(uvmexp.active + uvmexp.inactive +				 uvmexp.wired + uvmexp.free) * 13 / 16;			if (free + uvmexp.paging >= uvmexp.freetarg << 2 ||			    dirtyreacts == UVMPD_NUMDIRTYREACTS) {				if (vpgs <= 0) {					UVMHIST_LOG(pdhist,"  met free target: "						    "exit loop", 0, 0, 0, 0);					retval = TRUE;					if (swslot == 0)						/* exit now if no                                                   swap-i/o pending */						break;					/* set p to null to signal final                                           swap i/o */					p = NULL;				} else {					vnode_only = TRUE;				}			}		}		if (p) {	/* if (we have a new page to consider) */			/*			 * we are below target and have a new page to consider.			 */			uvmexp.pdscans++;			nextpg = TAILQ_NEXT(p, pageq);			/*			 * first we attempt to lock the object that this page			 * belongs to.  if our attempt fails we skip on to			 * the next page (no harm done).  it is important to			 * "try" locking the object as we are locking in the			 * wrong order (pageq -> object) and we don't want to			 * deadlock.			 *			 * the only time we expect to see an ownerless page			 * (i.e. a page with no uobject and !PQ_ANON) is if an			 * anon has loaned a page from a uvm_object and the			 * uvm_object has dropped the ownership.  in that			 * case, the anon can "take over" the loaned page			 * and make it its own.			 */					/* is page part of an anon or ownerless ? */			if ((p->pqflags & PQ_ANON) || p->uobject == NULL) {				if (vnode_only) {					uvm_pageactivate(p);					continue;				}				anon = p->uanon;				KASSERT(anon != NULL);				if (!simple_lock_try(&anon->an_lock))					/* lock failed, skip this page */					continue;				/*				 * if the page is ownerless, claim it in the				 * name of "anon"!				 */				if ((p->pqflags & PQ_ANON) == 0) {					KASSERT(p->loan_count > 0);					p->loan_count--;					p->pqflags |= PQ_ANON;					/* anon now owns it */				}				if (p->flags & PG_BUSY) {					simple_unlock(&anon->an_lock);					uvmexp.pdbusy++;					/* someone else owns page, skip it */					continue;				}				uvmexp.pdanscan++;			} else {				uobj = p->uobject;				KASSERT(uobj != NULL);				if (vnode_only &&				    uobj->pgops != &uvm_vnodeops) {					uvm_pageactivate(p);					continue;				}				if (!simple_lock_try(&uobj->vmobjlock))					/* lock failed, skip this page */					continue;				if (p->flags & PG_BUSY) {					simple_unlock(&uobj->vmobjlock);					uvmexp.pdbusy++;					/* someone else owns page, skip it */					continue;				}				uvmexp.pdobscan++;			}			/*			 * we now have the object and the page queues locked.			 * the page is not busy.   if the page is clean we			 * can free it now and continue.			 */			if (p->flags & PG_CLEAN) {				if (p->pqflags & PQ_SWAPBACKED) {					/* this page now lives only in swap */					simple_lock(&uvm.swap_data_lock);					uvmexp.swpgonly++;					simple_unlock(&uvm.swap_data_lock);				}				uvm_pagefree(p);				uvmexp.pdfreed++;				if (anon) {					/*					 * an anonymous page can only be clean					 * if it has backing store assigned.					 */					KASSERT(anon->an_swslot != 0);

⌨️ 快捷键说明

复制代码Ctrl + C
搜索代码Ctrl + F
全屏模式F11
增大字号Ctrl + =
减小字号Ctrl + -
显示快捷键?