📄 http:^^www.cs.wisc.edu^~david^david.html
字号:
<p><!WA22><!WA22><!WA22><!WA22><!WA22><!WA22><!WA22><!WA22><!WA22><!WA22><!WA22><!WA22><!WA22><!WA22><!WA22><!WA22><!WA22><!WA22><!WA22><!WA22><!WA22><!WA22><!WA22><!WA22><!WA22><!WA22><!WA22><!WA22><!WA22><!WA22><!WA22><A HREF="ftp://ftp.cs.wisc.edu/wwt/ipps95_netsim.ps"><cite>Accuracy vs. Performance in Parallel Simulation of Interconnection Networks</cite></A>,<br>Douglas C. Burger and David A. Wood.<br>In the proceedings of the 9th International Parallel Processing Symposium, April, 1995.<br><P><!WA23><!WA23><!WA23><!WA23><!WA23><!WA23><!WA23><!WA23><!WA23><!WA23><!WA23><!WA23><!WA23><!WA23><!WA23><!WA23><!WA23><!WA23><!WA23><!WA23><!WA23><!WA23><!WA23><!WA23><!WA23><!WA23><!WA23><!WA23><!WA23><!WA23><!WA23><A HREF="ftp://ftp.cs.wisc.edu/wwt/sc94_protocols.ps"><CITE>Application-Specific Protocols for User-Level Shared Memory,</CITE></A>Babak Falsafi, Alvin Lebeck, Steven Reinhardt, Ioannis Schoinas,Mark Hill, James Larus, Anne Rogers, and David Wood,In Proceedings of Supercomputing '94.<P><!WA24><!WA24><!WA24><!WA24><!WA24><!WA24><!WA24><!WA24><!WA24><!WA24><!WA24><!WA24><!WA24><!WA24><!WA24><!WA24><!WA24><!WA24><!WA24><!WA24><!WA24><!WA24><!WA24><!WA24><!WA24><!WA24><!WA24><!WA24><!WA24><!WA24><!WA24><A HREF="ftp://ftp.cs.wisc.edu/wwt/asplos6_fine_grain.ps"><CITE>Fine-grain Access Control for Distributed Shared Memory,</CITE></A>Ioannis Schoinas, Babak Falsafi, Alvin Lebeck, Steven Reinhardt,James Larus, and David Wood,Proceedings of ASPLOS VI.<P><!WA25><!WA25><!WA25><!WA25><!WA25><!WA25><!WA25><!WA25><!WA25><!WA25><!WA25><!WA25><!WA25><!WA25><!WA25><!WA25><!WA25><!WA25><!WA25><!WA25><!WA25><!WA25><!WA25><!WA25><!WA25><!WA25><!WA25><!WA25><!WA25><!WA25><!WA25><A HREF="ftp://ftp.cs.wisc.edu/wwt/isca94_typhoon.ps"><CITE>Tempest and Typhoon: User-Level Shared Memory,</CITE></A>Steven Reinhardt, James Larus, and David Wood,Proceedings of Int'l Symposium on Computer Architecture, 1994.<P><!WA26><!WA26><!WA26><!WA26><!WA26><!WA26><!WA26><!WA26><!WA26><!WA26><!WA26><!WA26><!WA26><!WA26><!WA26><!WA26><!WA26><!WA26><!WA26><!WA26><!WA26><!WA26><!WA26><!WA26><!WA26><!WA26><!WA26><!WA26><!WA26><!WA26><!WA26><A HREF="http://www.cs.wisc.edu/~alvy/papers/cprof.ps"><cite>Cache Profiling and the SPEC Benchmarks: A Case Study,</cite></A>Alvin R. Lebeck andDavid A. Wood,pages 15-26,IEEE COMPUTER,October 1994<P><!WA27><!WA27><!WA27><!WA27><!WA27><!WA27><!WA27><!WA27><!WA27><!WA27><!WA27><!WA27><!WA27><!WA27><!WA27><!WA27><!WA27><!WA27><!WA27><!WA27><!WA27><!WA27><!WA27><!WA27><!WA27><!WA27><!WA27><!WA27><!WA27><!WA27><!WA27><A HREF="ftp://ftp.cs.wisc.edu/wwt/tocs93_csm.ps"><CITE>Cooperative Shared Memory: Software and Hardware for Scalable Multiprocessors,</CITE></A>Mark D. Hill, James R. Larus, Steven K. Reinhardt, David A. Wood,ACM Transactions on Computer Systems (TOCS), November 1993.<P><!WA28><!WA28><!WA28><!WA28><!WA28><!WA28><!WA28><!WA28><!WA28><!WA28><!WA28><!WA28><!WA28><!WA28><!WA28><!WA28><!WA28><!WA28><!WA28><!WA28><!WA28><!WA28><!WA28><!WA28><!WA28><!WA28><!WA28><!WA28><!WA28><!WA28><!WA28><A HREF="ftp://ftp.cs.wisc.edu/wwt/annobib.ps"><CITE>The Wisconsin Wind Tunnel Project: An Annotated Bibliography,</CITE></A>Mark D. Hill, James R. Larus, David A. Wood,Computer Architecture News, v. 22, n. 5, December 1994.On-line version revised frequently.<P><!WA29><!WA29><!WA29><!WA29><!WA29><!WA29><!WA29><!WA29><!WA29><!WA29><!WA29><!WA29><!WA29><!WA29><!WA29><!WA29><!WA29><!WA29><!WA29><!WA29><!WA29><!WA29><!WA29><!WA29><!WA29><!WA29><!WA29><!WA29><!WA29><!WA29><!WA29><A HREF="ftp://ftp.cs.wisc.edu/markhill/Papers/can93_warts.ps"><CITE>Wisconsin Architectural Research Tool Set (WARTS),</CITE></A>Mark D. Hill, James R. Larus, Alvin R. Lebeck, Madhusudhan Talluri,David A. Wood,Computer Architecture News (CAN), August 1993.</UL><HR><hr><h2> Research Summary </h2> My main research goals lie in developing cost-effective computerarchitectures that take advantage of rapidly changing technologies. Myresearch program has two major thrusts: <ul><li> evaluating the performance,feasibility, and correctness of new architectures, and<li> developing new tools and techniques to facilitate this evaluation.</ul>Currently, this research focusses on the following three areas:<ul><li> multi-paradigm multiprocessors,which efficiently integrate shared-memory, message-passing, and hybridprogramming paradigms,<li> a virtual prototyping system, which exploits the similaritesof an existing parallel machine to simulate a hypothetical parallel machine,<li> and, techniques for understanding and tuning program performance.</ul>Recent results include developing a new interface---calledTempest---between user-level protocol handlers and system-suppliedmechanisms. Tempest provides the mechanisms that allow programmers,compilers, and program libraries to implement and use message passing,transparent shared memory, and hybrid combinations of the two. Tempestmechanisms are low-overhead messages, bulk data transfer, virtualmemory management, and fine-grain access control. The most novelmechanism---fine-grain access control---allows user software to tagblocks (e.g., 32 bytes) as read-write, read-only, or invalid, so thelocal memory can be used to transparently cache remote data.<br><br>We are exploring alternative ways to support this interface.The first---called Typhoon---isa proposed hardwareplatform that implements the Tempest mechanisms with a fully-programmable,user-level processor in the network interface. A reverse-translationtable (RTLB) invokes the network processor when it detects a fine-grainaccess fault.We have simulated Typhoon on the Wisconsin Wind Tunnel and found thata transparent shared-memory protocol running on Typhoon performscomparably +/- 30% to anall-hardware Dir{N}NB cache-coherence protocol for five shared-memoryprograms.<br><br>We have also developed a new memory system simulation method thatoptimizes the common case---cache hits---significantly reducingsimulation time.Fast-Cache tightly integrates reference generation and simulation byproviding the abstraction of tagged memory blocks: each referenceinvokes a user-specified function depending upon the reference type andmemory block state. The simulator controls how references are processedby manipulating memory block states, specifying a special NULL functionfor no action cases. Fast-Cache implements this abstraction by usingbinary-rewriting to perform a table lookup before each memoryreference. On a SPARCStation 10, Fast-Cache simulation times are two tothree times faster than a conventional trace-driven simulator thatcalls a procedure on each memory reference; simulation times are onlythree to six times slower than the original, un-instrumented program.We are also investigating using Fast-Cache's binary rewriting techniquesto support the Tempest interface on existing hardware platforms.<hr><address> Last Updated: July 11, 1996 </address><hr></body></html>
⌨️ 快捷键说明
复制代码
Ctrl + C
搜索代码
Ctrl + F
全屏模式
F11
切换主题
Ctrl + Shift + D
显示快捷键
?
增大字号
Ctrl + =
减小字号
Ctrl + -