📄 direct.c
字号:
/* * linux/fs/nfs/direct.c * * Copyright (C) 2003 by Chuck Lever <cel@netapp.com> * * High-performance uncached I/O for the Linux NFS client * * There are important applications whose performance or correctness * depends on uncached access to file data. Database clusters * (multiple copies of the same instance running on separate hosts) * implement their own cache coherency protocol that subsumes file * system cache protocols. Applications that process datasets * considerably larger than the client's memory do not always benefit * from a local cache. A streaming video server, for instance, has no * need to cache the contents of a file. * * When an application requests uncached I/O, all read and write requests * are made directly to the server; data stored or fetched via these * requests is not cached in the Linux page cache. The client does not * correct unaligned requests from applications. All requested bytes are * held on permanent storage before a direct write system call returns to * an application. * * Solaris implements an uncached I/O facility called directio() that * is used for backups and sequential I/O to very large files. Solaris * also supports uncaching whole NFS partitions with "-o forcedirectio," * an undocumented mount option. * * Designed by Jeff Kimmel, Chuck Lever, and Trond Myklebust, with * help from Andrew Morton. * * 18 Dec 2001 Initial implementation for 2.4 --cel * 08 Jul 2002 Version for 2.4.19, with bug fixes --trondmy * 08 Jun 2003 Port to 2.5 APIs --cel * 31 Mar 2004 Handle direct I/O without VFS support --cel * */#include <linux/config.h>#include <linux/errno.h>#include <linux/sched.h>#include <linux/kernel.h>#include <linux/smp_lock.h>#include <linux/file.h>#include <linux/pagemap.h>#include <linux/nfs_fs.h>#include <linux/nfs_page.h>#include <linux/sunrpc/clnt.h>#include <asm/system.h>#include <asm/uaccess.h>#define NFSDBG_FACILITY NFSDBG_VFS#define VERF_SIZE (2 * sizeof(__u32))#define MAX_DIRECTIO_SIZE (4096UL << PAGE_SHIFT)/** * nfs_get_user_pages - find and set up pages underlying user's buffer * rw: direction (read or write) * user_addr: starting address of this segment of user's buffer * count: size of this segment * @pages: returned array of page struct pointers underlying user's buffer */static inline intnfs_get_user_pages(int rw, unsigned long user_addr, size_t size, struct page ***pages){ int result = -ENOMEM; unsigned long page_count; size_t array_size; /* set an arbitrary limit to prevent arithmetic overflow */ if (size > MAX_DIRECTIO_SIZE) return -EFBIG; page_count = (user_addr + size + PAGE_SIZE - 1) >> PAGE_SHIFT; page_count -= user_addr >> PAGE_SHIFT; array_size = (page_count * sizeof(struct page *)); *pages = kmalloc(array_size, GFP_KERNEL); if (*pages) { down_read(¤t->mm->mmap_sem); result = get_user_pages(current, current->mm, user_addr, page_count, (rw == READ), 0, *pages, NULL); up_read(¤t->mm->mmap_sem); } return result;}/** * nfs_free_user_pages - tear down page struct array * @pages: array of page struct pointers underlying target buffer */static voidnfs_free_user_pages(struct page **pages, int npages, int do_dirty){ int i; for (i = 0; i < npages; i++) { if (do_dirty) set_page_dirty_lock(pages[i]); page_cache_release(pages[i]); } kfree(pages);}/** * nfs_direct_read_seg - Read in one iov segment. Generate separate * read RPCs for each "rsize" bytes. * @inode: target inode * @ctx: target file open context * user_addr: starting address of this segment of user's buffer * count: size of this segment * file_offset: offset in file to begin the operation * @pages: array of addresses of page structs defining user's buffer * nr_pages: size of pages array */static intnfs_direct_read_seg(struct inode *inode, struct nfs_open_context *ctx, unsigned long user_addr, size_t count, loff_t file_offset, struct page **pages, int nr_pages){ const unsigned int rsize = NFS_SERVER(inode)->rsize; int tot_bytes = 0; int curpage = 0; struct nfs_read_data rdata = { .inode = inode, .cred = ctx->cred, .args = { .fh = NFS_FH(inode), .context = ctx, }, .res = { .fattr = &rdata.fattr, }, }; rdata.args.pgbase = user_addr & ~PAGE_MASK; rdata.args.offset = file_offset; do { int result; rdata.args.count = count; if (rdata.args.count > rsize) rdata.args.count = rsize; rdata.args.pages = &pages[curpage]; dprintk("NFS: direct read: c=%u o=%Ld ua=%lu, pb=%u, cp=%u\n", rdata.args.count, (long long) rdata.args.offset, user_addr + tot_bytes, rdata.args.pgbase, curpage); lock_kernel(); result = NFS_PROTO(inode)->read(&rdata); unlock_kernel(); if (result <= 0) { if (tot_bytes > 0) break; if (result == -EISDIR) result = -EINVAL; return result; } tot_bytes += result; if (rdata.res.eof) break; rdata.args.offset += result; rdata.args.pgbase += result; curpage += rdata.args.pgbase >> PAGE_SHIFT; rdata.args.pgbase &= ~PAGE_MASK; count -= result; } while (count != 0); /* XXX: should we zero the rest of the user's buffer if we * hit eof? */ return tot_bytes;}/** * nfs_direct_read - For each iov segment, map the user's buffer * then generate read RPCs. * @inode: target inode * @ctx: target file open context * @iov: array of vectors that define I/O buffer * file_offset: offset in file to begin the operation * nr_segs: size of iovec array * * generic_file_direct_IO has already pushed out any non-direct * writes so that this read will see them when we read from the * server. */static ssize_tnfs_direct_read(struct inode *inode, struct nfs_open_context *ctx, const struct iovec *iov, loff_t file_offset, unsigned long nr_segs){ ssize_t tot_bytes = 0; unsigned long seg = 0; while ((seg < nr_segs) && (tot_bytes >= 0)) { ssize_t result; int page_count; struct page **pages; const struct iovec *vec = &iov[seg++]; unsigned long user_addr = (unsigned long) vec->iov_base; size_t size = vec->iov_len; page_count = nfs_get_user_pages(READ, user_addr, size, &pages); if (page_count < 0) { nfs_free_user_pages(pages, 0, 0); if (tot_bytes > 0) break; return page_count; } result = nfs_direct_read_seg(inode, ctx, user_addr, size, file_offset, pages, page_count); nfs_free_user_pages(pages, page_count, 1); if (result <= 0) { if (tot_bytes > 0) break; return result; } tot_bytes += result; file_offset += result; if (result < size) break; } return tot_bytes;}/** * nfs_direct_write_seg - Write out one iov segment. Generate separate * write RPCs for each "wsize" bytes, then commit. * @inode: target inode * @ctx: target file open context * user_addr: starting address of this segment of user's buffer * count: size of this segment * file_offset: offset in file to begin the operation * @pages: array of addresses of page structs defining user's buffer * nr_pages: size of pages array */static intnfs_direct_write_seg(struct inode *inode, struct nfs_open_context *ctx, unsigned long user_addr, size_t count, loff_t file_offset, struct page **pages, int nr_pages){ const unsigned int wsize = NFS_SERVER(inode)->wsize; size_t request; int curpage, need_commit, result, tot_bytes; struct nfs_writeverf first_verf; struct nfs_write_data wdata = { .inode = inode, .cred = ctx->cred, .args = { .fh = NFS_FH(inode), .context = ctx, }, .res = { .fattr = &wdata.fattr, .verf = &wdata.verf, }, }; wdata.args.stable = NFS_UNSTABLE; if (IS_SYNC(inode) || NFS_PROTO(inode)->version == 2 || count <= wsize) wdata.args.stable = NFS_FILE_SYNC; nfs_begin_data_update(inode);retry: need_commit = 0; tot_bytes = 0; curpage = 0; request = count; wdata.args.pgbase = user_addr & ~PAGE_MASK; wdata.args.offset = file_offset; do { wdata.args.count = request; if (wdata.args.count > wsize) wdata.args.count = wsize; wdata.args.pages = &pages[curpage]; dprintk("NFS: direct write: c=%u o=%Ld ua=%lu, pb=%u, cp=%u\n", wdata.args.count, (long long) wdata.args.offset, user_addr + tot_bytes, wdata.args.pgbase, curpage); lock_kernel(); result = NFS_PROTO(inode)->write(&wdata); unlock_kernel(); if (result <= 0) { if (tot_bytes > 0) break; goto out; } if (tot_bytes == 0)
⌨️ 快捷键说明
复制代码
Ctrl + C
搜索代码
Ctrl + F
全屏模式
F11
切换主题
Ctrl + Shift + D
显示快捷键
?
增大字号
Ctrl + =
减小字号
Ctrl + -