⭐ 欢迎来到虫虫下载站! | 📦 资源下载 📁 资源专辑 ℹ️ 关于我们
⭐ 虫虫下载站

📄 mmap.c

📁 linux1.1源代码
💻 C
字号:
/* *	linux/mm/mmap.c * * Written by obz. */#include <linux/stat.h>#include <linux/sched.h>#include <linux/kernel.h>#include <linux/mm.h>#include <linux/shm.h>#include <linux/errno.h>#include <linux/mman.h>#include <linux/string.h>#include <linux/malloc.h>#include <asm/segment.h>#include <asm/system.h>static int anon_map(struct inode *, struct file *,		    unsigned long, size_t, int,		    unsigned long);/* * description of effects of mapping type and prot in current implementation. * this is due to the current handling of page faults in memory.c. the expected * behavior is in parens: * * map_type	prot *		PROT_NONE	PROT_READ	PROT_WRITE	PROT_EXEC * MAP_SHARED	r: (no) yes	r: (yes) yes	r: (no) yes	r: (no) no *		w: (no) yes	w: (no) copy	w: (yes) yes	w: (no) no *		x: (no) no	x: (no) no	x: (no) no	x: (yes) no *		 * MAP_PRIVATE	r: (no) yes	r: (yes) yes	r: (no) yes	r: (no) no *		w: (no) copy	w: (no) copy	w: (copy) copy	w: (no) no *		x: (no) no	x: (no) no	x: (no) no	x: (yes) no * */#define CODE_SPACE(addr)	\ (PAGE_ALIGN(addr) < current->start_code + current->end_code)int do_mmap(struct file * file, unsigned long addr, unsigned long len,	unsigned long prot, unsigned long flags, unsigned long off){	int mask, error;	if ((len = PAGE_ALIGN(len)) == 0)		return addr;	if (addr > TASK_SIZE || len > TASK_SIZE || addr > TASK_SIZE-len)		return -EINVAL;	/*	 * do simple checking here so the lower-level routines won't have	 * to. we assume access permissions have been handled by the open	 * of the memory object, so we don't do any here.	 */	if (file != NULL)		switch (flags & MAP_TYPE) {		case MAP_SHARED:			if ((prot & PROT_WRITE) && !(file->f_mode & 2))				return -EACCES;			/* fall through */		case MAP_PRIVATE:			if (!(file->f_mode & 1))				return -EACCES;			break;		default:			return -EINVAL;		}	/*	 * obtain the address to map to. we verify (or select) it and ensure	 * that it represents a valid section of the address space.	 */	if (flags & MAP_FIXED) {		if (addr & ~PAGE_MASK)			return -EINVAL;		if (len > TASK_SIZE || addr > TASK_SIZE - len)			return -EINVAL;	} else {		struct vm_area_struct * vmm;		/* Maybe this works.. Ugly it is. */		addr = SHM_RANGE_START;		while (addr+len < SHM_RANGE_END) {			for (vmm = current->mmap ; vmm ; vmm = vmm->vm_next) {				if (addr >= vmm->vm_end)					continue;				if (addr + len <= vmm->vm_start)					continue;				addr = PAGE_ALIGN(vmm->vm_end);				break;			}			if (!vmm)				break;		}		if (addr+len >= SHM_RANGE_END)			return -ENOMEM;	}	/*	 * determine the object being mapped and call the appropriate	 * specific mapper. the address has already been validated, but	 * not unmapped, but the maps are removed from the list.	 */	if (file && (!file->f_op || !file->f_op->mmap))		return -ENODEV;	mask = 0;	if (prot & (PROT_READ | PROT_EXEC))		mask |= PAGE_READONLY;	if (prot & PROT_WRITE)		if ((flags & MAP_TYPE) == MAP_PRIVATE)			mask |= PAGE_COPY;		else			mask |= PAGE_SHARED;	if (!mask)		return -EINVAL;	do_munmap(addr, len);	/* Clear old maps */	if (file)		error = file->f_op->mmap(file->f_inode, file, addr, len, mask, off);	else		error = anon_map(NULL, NULL, addr, len, mask, off);		if (!error)		return addr;	if (!current->errno)		current->errno = -error;	return -1;}asmlinkage int sys_mmap(unsigned long *buffer){	int error;	unsigned long flags;	struct file * file = NULL;	error = verify_area(VERIFY_READ, buffer, 6*4);	if (error)		return error;	flags = get_fs_long(buffer+3);	if (!(flags & MAP_ANONYMOUS)) {		unsigned long fd = get_fs_long(buffer+4);		if (fd >= NR_OPEN || !(file = current->filp[fd]))			return -EBADF;	}	return do_mmap(file, get_fs_long(buffer), get_fs_long(buffer+1),		get_fs_long(buffer+2), flags, get_fs_long(buffer+5));}/* * Normal function to fix up a mapping * This function is the default for when an area has no specific * function.  This may be used as part of a more specific routine. * This function works out what part of an area is affected and * adjusts the mapping information.  Since the actual page * manipulation is done in do_mmap(), none need be done here, * though it would probably be more appropriate. * * By the time this function is called, the area struct has been * removed from the process mapping list, so it needs to be * reinserted if necessary. * * The 4 main cases are: *    Unmapping the whole area *    Unmapping from the start of the segment to a point in it *    Unmapping from an intermediate point to the end *    Unmapping between to intermediate points, making a hole. * * Case 4 involves the creation of 2 new areas, for each side of * the hole. */void unmap_fixup(struct vm_area_struct *area,		 unsigned long addr, size_t len){	struct vm_area_struct *mpnt;	unsigned long end = addr + len;	if (addr < area->vm_start || addr >= area->vm_end ||	    end <= area->vm_start || end > area->vm_end ||	    end < addr)	{		printk("unmap_fixup: area=%lx-%lx, unmap %lx-%lx!!\n",		       area->vm_start, area->vm_end, addr, end);		return;	}	/* Unmapping the whole area */	if (addr == area->vm_start && end == area->vm_end) {		if (area->vm_ops && area->vm_ops->close)			area->vm_ops->close(area);		return;	}	/* Work out to one of the ends */	if (addr >= area->vm_start && end == area->vm_end)		area->vm_end = addr;	if (addr == area->vm_start && end <= area->vm_end) {		area->vm_offset += (end - area->vm_start);		area->vm_start = end;	}	/* Unmapping a hole */	if (addr > area->vm_start && end < area->vm_end)	{		/* Add end mapping -- leave beginning for below */		mpnt = (struct vm_area_struct *)kmalloc(sizeof(*mpnt), GFP_KERNEL);		*mpnt = *area;		mpnt->vm_offset += (end - area->vm_start);		mpnt->vm_start = end;		if (mpnt->vm_inode)			mpnt->vm_inode->i_count++;		insert_vm_struct(current, mpnt);		area->vm_end = addr;	/* Truncate area */	}	/* construct whatever mapping is needed */	mpnt = (struct vm_area_struct *)kmalloc(sizeof(*mpnt), GFP_KERNEL);	*mpnt = *area;	insert_vm_struct(current, mpnt);}asmlinkage int sys_mprotect(unsigned long addr, size_t len, unsigned long prot){	return -EINVAL; /* Not implemented yet */}asmlinkage int sys_munmap(unsigned long addr, size_t len){	return do_munmap(addr, len);}/* * Munmap is split into 2 main parts -- this part which finds * what needs doing, and the areas themselves, which do the * work.  This now handles partial unmappings. * Jeremy Fitzhardine <jeremy@sw.oz.au> */int do_munmap(unsigned long addr, size_t len){	struct vm_area_struct *mpnt, **npp, *free;	if ((addr & ~PAGE_MASK) || addr > TASK_SIZE || len > TASK_SIZE-addr)		return -EINVAL;	if ((len = PAGE_ALIGN(len)) == 0)		return 0;	/*	 * Check if this memory area is ok - put it on the temporary	 * list if so..  The checks here are pretty simple --	 * every area affected in some way (by any overlap) is put	 * on the list.  If nothing is put on, nothing is affected.	 */	npp = &current->mmap;	free = NULL;	for (mpnt = *npp; mpnt != NULL; mpnt = *npp) {		unsigned long end = addr+len;		if ((addr < mpnt->vm_start && end <= mpnt->vm_start) ||		    (addr >= mpnt->vm_end && end > mpnt->vm_end))		{			npp = &mpnt->vm_next;			continue;		}		*npp = mpnt->vm_next;		mpnt->vm_next = free;		free = mpnt;	}	if (free == NULL)		return 0;	/*	 * Ok - we have the memory areas we should free on the 'free' list,	 * so release them, and unmap the page range..	 * If the one of the segments is only being partially unmapped,	 * it will put new vm_area_struct(s) into the address space.	 */	while (free) {		unsigned long st, end;		mpnt = free;		free = free->vm_next;		st = addr < mpnt->vm_start ? mpnt->vm_start : addr;		end = addr+len;		end = end > mpnt->vm_end ? mpnt->vm_end : end;		if (mpnt->vm_ops && mpnt->vm_ops->unmap)			mpnt->vm_ops->unmap(mpnt, st, end-st);		else			unmap_fixup(mpnt, st, end-st);		kfree(mpnt);	}	unmap_page_range(addr, len);	return 0;}/* This is used for a general mmap of a disk file */int generic_mmap(struct inode * inode, struct file * file,	unsigned long addr, size_t len, int prot, unsigned long off){  	struct vm_area_struct * mpnt;	extern struct vm_operations_struct file_mmap;	struct buffer_head * bh;	if (prot & PAGE_RW)	/* only PAGE_COW or read-only supported right now */		return -EINVAL;	if (off & (inode->i_sb->s_blocksize - 1))		return -EINVAL;	if (!inode->i_sb || !S_ISREG(inode->i_mode))		return -EACCES;	if (!inode->i_op || !inode->i_op->bmap)		return -ENOEXEC;	if (!(bh = bread(inode->i_dev,bmap(inode,0),inode->i_sb->s_blocksize)))		return -EACCES;	if (!IS_RDONLY(inode)) {		inode->i_atime = CURRENT_TIME;		inode->i_dirt = 1;	}	brelse(bh);	mpnt = (struct vm_area_struct * ) kmalloc(sizeof(struct vm_area_struct), GFP_KERNEL);	if (!mpnt)		return -ENOMEM;	unmap_page_range(addr, len);		mpnt->vm_task = current;	mpnt->vm_start = addr;	mpnt->vm_end = addr + len;	mpnt->vm_page_prot = prot;	mpnt->vm_share = NULL;	mpnt->vm_inode = inode;	inode->i_count++;	mpnt->vm_offset = off;	mpnt->vm_ops = &file_mmap;	insert_vm_struct(current, mpnt);	merge_segments(current->mmap, NULL, NULL);		return 0;}/* * Insert vm structure into process list * This makes sure the list is sorted by start address, and * some some simple overlap checking. * JSGF */void insert_vm_struct(struct task_struct *t, struct vm_area_struct *vmp){	struct vm_area_struct **nxtpp, *mpnt;	nxtpp = &t->mmap;		for(mpnt = t->mmap; mpnt != NULL; mpnt = mpnt->vm_next)	{		if (mpnt->vm_start > vmp->vm_start)			break;		nxtpp = &mpnt->vm_next;		if ((vmp->vm_start >= mpnt->vm_start &&		     vmp->vm_start < mpnt->vm_end) ||		    (vmp->vm_end >= mpnt->vm_start &&		     vmp->vm_end < mpnt->vm_end))			printk("insert_vm_struct: ins area %lx-%lx in area %lx-%lx\n",			       vmp->vm_start, vmp->vm_end,			       mpnt->vm_start, vmp->vm_end);	}		vmp->vm_next = mpnt;	*nxtpp = vmp;}/* * Merge a list of memory segments if possible. * Redundant vm_area_structs are freed. * This assumes that the list is ordered by address. */void merge_segments(struct vm_area_struct *mpnt,		    map_mergep_fnp mergep, void *mpd){	struct vm_area_struct *prev, *next;	if (mpnt == NULL)		return;		for(prev = mpnt, mpnt = mpnt->vm_next;	    mpnt != NULL;	    prev = mpnt, mpnt = next)	{		int mp;		next = mpnt->vm_next;				if (mergep == NULL)		{			unsigned long psz = prev->vm_end - prev->vm_start;			mp = prev->vm_offset + psz == mpnt->vm_offset;		}		else			mp = (*mergep)(prev, mpnt, mpd);		/*		 * Check they are compatible.		 * and the like...		 * What does the share pointer mean?		 */		if (prev->vm_ops != mpnt->vm_ops ||		    prev->vm_page_prot != mpnt->vm_page_prot ||		    prev->vm_inode != mpnt->vm_inode ||		    prev->vm_end != mpnt->vm_start ||		    !mp ||		    prev->vm_share != mpnt->vm_share ||		/* ?? */		    prev->vm_next != mpnt)			/* !!! */			continue;		/*		 * merge prev with mpnt and set up pointers so the new		 * big segment can possibly merge with the next one.		 * The old unused mpnt is freed.		 */		prev->vm_end = mpnt->vm_end;		prev->vm_next = mpnt->vm_next;		kfree_s(mpnt, sizeof(*mpnt));		mpnt = prev;	}}/* * Map memory not associated with any file into a process * address space.  Adjecent memory is merged. */static int anon_map(struct inode *ino, struct file * file,		    unsigned long addr, size_t len, int mask,		    unsigned long off){  	struct vm_area_struct * mpnt;	if (zeromap_page_range(addr, len, mask))		return -ENOMEM;	mpnt = (struct vm_area_struct * ) kmalloc(sizeof(struct vm_area_struct), GFP_KERNEL);	if (!mpnt)		return -ENOMEM;	mpnt->vm_task = current;	mpnt->vm_start = addr;	mpnt->vm_end = addr + len;	mpnt->vm_page_prot = mask;	mpnt->vm_share = NULL;	mpnt->vm_inode = NULL;	mpnt->vm_offset = 0;	mpnt->vm_ops = NULL;	insert_vm_struct(current, mpnt);	merge_segments(current->mmap, ignoff_mergep, NULL);	return 0;}/* Merge, ignoring offsets */int ignoff_mergep(const struct vm_area_struct *m1,		  const struct vm_area_struct *m2,		  void *data){	if (m1->vm_inode != m2->vm_inode)	/* Just to be sure */		return 0;	return (struct inode *)data == m1->vm_inode;}

⌨️ 快捷键说明

复制代码 Ctrl + C
搜索代码 Ctrl + F
全屏模式 F11
切换主题 Ctrl + Shift + D
显示快捷键 ?
增大字号 Ctrl + =
减小字号 Ctrl + -