⭐ 欢迎来到虫虫下载站! | 📦 资源下载 📁 资源专辑 ℹ️ 关于我们
⭐ 虫虫下载站

📄 buffer.c,v

📁 linux-0.10 对于想了解linux内核,而又不想花太多精力的人,最好就是这种低版本,而具有核心价值的程序
💻 C,V
字号:
head     1.2;branch   ;access   ;symbols  ;locks    ; strict;comment  @ * @;1.2date     91.11.20.00.10.40;  author tytso;  state Exp;branches ;next     1.1;1.1date     91.11.12.15.49.42;  author tytso;  state Exp;branches ;next     ;desc@@1.2log@Used new version of buffer.c supplied from Linus.@text@/* *  linux/fs/buffer.c * *  (C) 1991  Linus Torvalds *//* *  'buffer.c' implements the buffer-cache functions. Race-conditions have * been avoided by NEVER letting a interrupt change a buffer (except for the * data, of course), but instead letting the caller do it. NOTE! As interrupts * can wake up a caller, some cli-sti sequences are needed to check for * sleep-on-calls. These should be extremely quick, though (I hope). *//* * NOTE! There is one discordant note here: checking floppies for * disk change. This is where it fits best, I think, as it should * invalidate changed floppy-disk-caches. */#include <stdarg.h> #include <linux/config.h>#include <linux/sched.h>#include <linux/kernel.h>#include <asm/system.h>#include <asm/io.h>extern int end;struct buffer_head * start_buffer = (struct buffer_head *) &end;struct buffer_head * hash_table[NR_HASH];static struct buffer_head * free_list;static struct task_struct * buffer_wait = NULL;int NR_BUFFERS = 0;static inline void wait_on_buffer(struct buffer_head * bh){	cli();	while (bh->b_lock)		sleep_on(&bh->b_wait);	sti();}int sys_sync(void){	int i;	struct buffer_head * bh;	sync_inodes();		/* write out inodes into buffers */	bh = start_buffer;	for (i=0 ; i<NR_BUFFERS ; i++,bh++) {		wait_on_buffer(bh);		if (bh->b_dirt)			ll_rw_block(WRITE,bh);	}	return 0;}int sync_dev(int dev){	int i;	struct buffer_head * bh;	bh = start_buffer;	for (i=0 ; i<NR_BUFFERS ; i++,bh++) {		if (bh->b_dev != dev)			continue;		wait_on_buffer(bh);		if (bh->b_dev == dev && bh->b_dirt)			ll_rw_block(WRITE,bh);	}	sync_inodes();	bh = start_buffer;	for (i=0 ; i<NR_BUFFERS ; i++,bh++) {		if (bh->b_dev != dev)			continue;		wait_on_buffer(bh);		if (bh->b_dev == dev && bh->b_dirt)			ll_rw_block(WRITE,bh);	}	return 0;}/* * This routine checks whether a floppy has been changed, and * invalidates all buffer-cache-entries in that case. This * is a relatively slow routine, so we have to try to minimize using * it. Thus it is called only upon a 'mount' or 'open'. This * is the best way of combining speed and utility, I think. * People changing diskettes in the middle of an operation deserve * to loose :-) * * NOTE! Although currently this is only for floppies, the idea is * that any additional removable block-device will use this routine, * and that mount/open needn't know that floppies/whatever are * special. */void check_disk_change(int dev){	int i;	struct buffer_head * bh;	if (MAJOR(dev) != 2)		return;	dev=MINOR(dev) & 0x03;	/* which floppy is it? */	if (!floppy_change(dev))		return;	dev |= 0x200;	for (i=0 ; i<NR_SUPER ; i++)		if ((super_block[i].s_dev & 0xff03)==dev)			put_super(super_block[i].s_dev);	bh = start_buffer;	for (i=0 ; i<NR_BUFFERS ; i++,bh++) {		if ((bh->b_dev & 0xff03) != dev)			continue;		wait_on_buffer(bh);		if ((bh->b_dev & 0xff03) == dev)			bh->b_uptodate = bh->b_dirt = 0;	}}#define _hashfn(dev,block) (((unsigned)(dev^block))%NR_HASH)#define hash(dev,block) hash_table[_hashfn(dev,block)]static inline void remove_from_queues(struct buffer_head * bh){/* remove from hash-queue */	if (bh->b_next)		bh->b_next->b_prev = bh->b_prev;	if (bh->b_prev)		bh->b_prev->b_next = bh->b_next;	if (hash(bh->b_dev,bh->b_blocknr) == bh)		hash(bh->b_dev,bh->b_blocknr) = bh->b_next;/* remove from free list */	if (!(bh->b_prev_free) || !(bh->b_next_free))		panic("Free block list corrupted");	bh->b_prev_free->b_next_free = bh->b_next_free;	bh->b_next_free->b_prev_free = bh->b_prev_free;	if (free_list == bh)		free_list = bh->b_next_free;}static inline void insert_into_queues(struct buffer_head * bh){/* put at end of free list */	bh->b_next_free = free_list;	bh->b_prev_free = free_list->b_prev_free;	free_list->b_prev_free->b_next_free = bh;	free_list->b_prev_free = bh;/* put the buffer in new hash-queue if it has a device */	bh->b_prev = NULL;	bh->b_next = NULL;	if (!bh->b_dev)		return;	bh->b_next = hash(bh->b_dev,bh->b_blocknr);	hash(bh->b_dev,bh->b_blocknr) = bh;	bh->b_next->b_prev = bh;}static struct buffer_head * find_buffer(int dev, int block){			struct buffer_head * tmp;	for (tmp = hash(dev,block) ; tmp != NULL ; tmp = tmp->b_next)		if (tmp->b_dev==dev && tmp->b_blocknr==block)			return tmp;	return NULL;}/* * Why like this, I hear you say... The reason is race-conditions. * As we don't lock buffers (unless we are readint them, that is), * something might happen to it while we sleep (ie a read-error * will force it bad). This shouldn't really happen currently, but * the code is ready. */struct buffer_head * get_hash_table(int dev, int block){	struct buffer_head * bh;	for (;;) {		if (!(bh=find_buffer(dev,block)))			return NULL;		bh->b_count++;		wait_on_buffer(bh);		if (bh->b_dev == dev && bh->b_blocknr == block)			return bh;		bh->b_count--;	}}/* * Ok, this is getblk, and it isn't very clear, again to hinder * race-conditions. Most of the code is seldom used, (ie repeating), * so it should be much more efficient than it looks. * * The algoritm is changed: better, and an elusive bug removed. *		LBT 11.11.91 */#define BADNESS(bh) (((bh)->b_dirt<<1)+(bh)->b_lock)struct buffer_head * getblk(int dev,int block){	struct buffer_head * tmp, * bh;repeat:	if (bh = get_hash_table(dev,block))		return bh;	tmp = free_list;	do {		if (tmp->b_count)			continue;		if (!bh || BADNESS(tmp)<BADNESS(bh)) {			bh = tmp;			if (!BADNESS(tmp))				break;		}	} while ((tmp = tmp->b_next_free) != free_list);	if (!bh) {		sleep_on(&buffer_wait);		goto repeat;	}	wait_on_buffer(bh);	if (bh->b_count)		goto repeat;	while (bh->b_dirt) {		sync_dev(bh->b_dev);		wait_on_buffer(bh);		if (bh->b_count)			goto repeat;	}/* NOTE!! While we slept waiting for this block, somebody else might *//* already have added "this" block to the cache. check it */	if (find_buffer(dev,block))		goto repeat;/* OK, FINALLY we know that this buffer is the only one of it's kind, *//* and that it's unused (b_count=0), unlocked (b_lock=0), and clean */	bh->b_count=1;	bh->b_dirt=0;	bh->b_uptodate=0;	remove_from_queues(bh);	bh->b_dev=dev;	bh->b_blocknr=block;	insert_into_queues(bh);	return bh;}void brelse(struct buffer_head * buf){	if (!buf)		return;	wait_on_buffer(buf);	if (!(buf->b_count--))		panic("Trying to free free buffer");	wake_up(&buffer_wait);}/* * bread() reads a specified block and returns the buffer that contains * it. It returns NULL if the block was unreadable. */struct buffer_head * bread(int dev,int block){	struct buffer_head * bh;	if (!(bh=getblk(dev,block)))		panic("bread: getblk returned NULL\n");	if (bh->b_uptodate)		return bh;	ll_rw_block(READ,bh);	wait_on_buffer(bh);	if (bh->b_uptodate)		return bh;	brelse(bh);	return NULL;}/* * Ok, breada can be used as bread, but additionally to mark other * blocks for reading as well. End the argument list with a negative * number. */struct buffer_head * breada(int dev,int first, ...){	va_list args;	struct buffer_head * bh, *tmp;	va_start(args,first);	if (!(bh=getblk(dev,first)))		panic("bread: getblk returned NULL\n");	if (!bh->b_uptodate)		ll_rw_block(READ,bh);	while ((first=va_arg(args,int))>=0) {		tmp=getblk(dev,first);		if (tmp) {			if (!tmp->b_uptodate)				ll_rw_block(READA,bh);			tmp->b_count--;		}	}	va_end(args);	wait_on_buffer(bh);	if (bh->b_uptodate)		return bh;	brelse(bh);	return (NULL);}void buffer_init(long buffer_end){	struct buffer_head * h = start_buffer;	void * b;	int i;	if (buffer_end == 1<<20)		b = (void *) (640*1024);	else		b = (void *) buffer_end;	while ( (b -= BLOCK_SIZE) >= ((void *) (h+1)) ) {		h->b_dev = 0;		h->b_dirt = 0;		h->b_count = 0;		h->b_lock = 0;		h->b_uptodate = 0;		h->b_wait = NULL;		h->b_next = NULL;		h->b_prev = NULL;		h->b_data = (char *) b;		h->b_prev_free = h-1;		h->b_next_free = h+1;		h++;		NR_BUFFERS++;		if (b == (void *) 0x100000)			b = (void *) 0xA0000;	}	h--;	free_list = start_buffer;	free_list->b_prev_free = h;	h->b_next_free = free_list;	for (i=0;i<NR_HASH;i++)		hash_table[i]=NULL;}	@1.1log@Initial revision@text@d72 9d196 3d200 1d203 1a203 1	struct buffer_head * tmp;d206 2a207 2	if (tmp=get_hash_table(dev,block))		return tmp;d209 6a214 4	for (;;) {		if (!tmp->b_count) {			wait_on_buffer(tmp);	/* we still have to wait */			if (!tmp->b_count)	/* on it, it might be dirty */d217 12a228 3		tmp = tmp->b_next_free;		if (tmp == free_list) {			sleep_on(&buffer_wait);a229 1		}d231 3a233 23	tmp->b_count++;	remove_from_queues(tmp);/* * Now, when we know nobody can get to this node (as it's removed from the * free list), we write it out. We can sleep here without fear of race- * conditions. */ 	while (tmp->b_dirt) {		sync_dev(tmp->b_dev);		wait_on_buffer(tmp);	}/* update buffer contents */	tmp->b_dev=dev;	tmp->b_blocknr=block;	tmp->b_dirt=0;	tmp->b_uptodate=0;/* NOTE!! While we possibly slept in sync_dev(), somebody else might have *//* added "this" block already, so check for that. Thank God for goto's. */	if (find_buffer(dev,block)) {		tmp->b_dev=0;		/* ok, someone else has beaten us */		tmp->b_blocknr=0;	/* to it - free this block and */		tmp->b_count=0;		/* try again */		insert_into_queues(tmp);d235 10a244 4	}/* and then insert into correct position */	insert_into_queues(tmp);	return tmp;@

⌨️ 快捷键说明

复制代码 Ctrl + C
搜索代码 Ctrl + F
全屏模式 F11
切换主题 Ctrl + Shift + D
显示快捷键 ?
增大字号 Ctrl + =
减小字号 Ctrl + -