⭐ 欢迎来到虫虫下载站! | 📦 资源下载 📁 资源专辑 ℹ️ 关于我们
⭐ 虫虫下载站

📄 rfc1273.txt

📁 中、英文RFC文档大全打包下载完全版 .
💻 TXT
📖 第 1 页 / 共 2 页
字号:
Network Working Group                                        M. SchwartzRequest for Comments: 1273                        University of Colorado                                                           November 1991                   A Measurement Study of Changes in                Service-Level Reachability in the Global              TCP/IP Internet: Goals, Experimental Design,               Implementation, and Policy ConsiderationsStatus of this Memo   This memo provides information for the Internet community.  It does   not specify an Internet standard.  Distribution of this memo is   unlimited.Abstract   In this report we discuss plans to carry out a longitudinal   measurement study of changes in service-level reachability in the   global TCP/IP Internet.  We overview our experimental design,   considerations of network and remote site load, mechanisms used to   control the measurement collection process, and network appropriate   use and privacy issues, including our efforts to inform sites   measured by this study.  A list of references and information on how   to contact the Principal Investigator are included.Introduction   The global TCP/IP Internet interconnects millions of individuals at   thousands of institutions worldwide, offering the potential for   significant collaboration through network services and electronic   information exchange.  At the same time, such powerful connectivity   offers many avenues for security violations, as evidenced by a number   of well publicized events over the past few years.  In response, many   sites have imposed mechanisms to limit their exposure to security   intrusions, ranging from disabling certain inter-site services, to   using external gateways that only allow electronic mail delivery, to   gateways that limit remote interactions via access control lists, to   disconnection from the Internet.  While these measures are preferable   to the damage that could occur from security violations, taken to an   extreme they could eventually reduce the Internet to little more than   a means of supporting certain pre-approved point-to-point data   transfers.  Such diminished functionality could hinder or prevent the   deployment of important new types of network services, impeding both   research and commercial advancement.   To understand the evolution of this situation, we have designed aSchwartz                                                        [Page 1]RFC 1273                  A Measurement Study              November 1991   study to measure changes in Internet service-level reachability over   a period of one year.  The study considers upper layer service   reachability instead of basic IP connectivity because the former   indicates the willingness of organizations to participate in inter-   organizational computing, which will be an important component of   future wide area distributed applications.   The data we gather will contribute to Internet research and   engineering planning activities in a number of ways.  The data will   indicate the mechanisms sites use to distance themselves from   Internet connectivity, the types of services that sites are willing   to run (and hence the type of distributed collaboration they are   willing to support), and variations in these characteristics as a   function of geographic location and type of institution (commercial,   educational, etc.).  Understanding these trends will allow   application designers and network builders to more realistically plan   for how to support future wide area distributed applications such as   digital library systems, information services, wide area distributed   file systems, and conferencing and other collaboration-support   systems.  The measurements will also be of general interest, as they   represent direct measurements of the evolution of a global electronic   society.   Clearly, a study of this nature and magnitude raises a number of   potential concerns.  In this note we overview our experimental   design, considerations of network and remote site load, mechanisms   used to control the measurement collection process, and our efforts   to inform sites measured by this study, along with concomitant   network appropriate use and privacy issues.   A point we wish to stress from the outset is that this is not a study   of network security.  The experiments do not attempt to probe the   security mechanisms of any machine on the network.  The study is   concerned solely with the evolution of network connectivity and   service reachability.Experimental Design   The study consists of a set of runs of a program over the span of one   to two days each month, repeated bimonthly for a period of one year   (in January 1992, March 1992, May 1992, July 1992, September 1992,   and November 1992).  Each program run attempts to connect to 13   different TCP services at each of approximately 12,700 Internet   domains worldwide, recording the failure/success status of each   attempt.  The program will attempt no data transfers in either   direction.  If a connection is successful, it is simply closed and   counted.  (Note in particular that this means that the security   mechanism behind individual network services will not be tested.)Schwartz                                                        [Page 2]RFC 1273                  A Measurement Study              November 1991   The machines on which connections are attempted will be selected at   random from a large list of machines in the Internet, constrained   such that at most 1 to 3 machines is contacted in any particular   domain.   The services to which connections will be attempted are:    __________________________________________________________________      Port Number   Service                Port Number   Service    ------------------------------------------------------------------          13        daytime                    111       Sun portmap          15        netstat                    513       rlogin          21        FTP                        514       rsh          23        telnet                     540       UUCP          25        SMTP                       543       klogin          53        Domain Naming System       544       krcmd, kshell          79        finger     _________________________________________________________________   This list was chosen to span a representative range of  service   types, each of which can be expected to be found on any machine in a   site (so that probing random machines is meaningful).  The one   exception  is  the  Domain  Naming  System,  for which the machines   to probe are selected from information  obtained  from the  Domain   system itself.  Only TCP services are tested, since the TCP   connection mechanism  allows  one  to  determine  if  a server is   running in an application-independent fashion.   As an aside, it would be possible  to  retrieve  "Well  Known   Service"  records  from the Domain Naming System, as a somewhat less   "invasive" measurement approach.  However,  these  records are  not   required  for proper network operation, and hence are far from   complete or consistent in the  Domain  Naming  System.  The  only way   to collect the data we want is to measure them in the fashion   described above.Network and Remote Site Load   The measurement software is quite careful to avoid generating   unnecessary internet packets, and to avoid congesting the internet   with too much concurrent activity.  Once it has successfully   connected to a particular service in a domain, the software never   attempts to connect to that service on any machine in that domain   again, for the duration of the current measurement run (i.e., the   current 60 days).  Once it has recorded 3 connection refusals at any   machines in that domain for a service, it does not try that service   at that domain again during the current measurement run.  If it   experiences 3 timeouts on any machine in a domain, it gives up on theSchwartz                                                        [Page 3]RFC 1273                  A Measurement Study              November 1991   domain, possibly to be retried again a day later (to overcome   transient network problems).  In the worst case there will be 3   connection failures for each service at 3 different machines, which   amounts to 37 connection requests per domain (3 for each of the 12   services other than the Domain Naming System, and one for the Domain   Naming System).  However, the average will be much less than this.   To quantify the actual Internet load, we now present some   measurements from test runs of the measurement software that were   performed in August 1991.  In total, 50,549 Domain Naming System   lookups were performed, and 73,760 connections were attempted.  This   measurement run completed in approximately 10 hours, never initiating   more than 20 network operations (name lookups or connection attempts)   concurrently.  The total NSFNET backbone load from all traffic   sources that month was approximately 5 billion packets.  Therefore,   the traffic from our measurement study amounted to less than .5% of   this volume on the day that the measurements were collected.  Since   the Internet contains several other backbones besides NSFNET, the   proportionate increase in total Internet traffic was significantly   less than .5%.   The cost to a remote site being measured is effectively zero.  From   the above measurements, on average we attempted 5.7 connections per   remote domain.  The cost of a connection open/close sequence is quite   small, particularly when compared to the cost of the many electronic   mail and news transmissions that most sites experience on a given   day.Control Over Measurement Collection Process   The measurement software evolved from an earlier set of experiments   used to measure the reach of an experimental Internet white pages   tool called netfind [Schwartz & Tsirigotis 1991b], and has been   evolved and tested extensively over a period of two years.  During   this time it has been used in a number of experiments of increasing   scale.  The software uses several redundant checks and other   mechanisms to ensure that careful control is maintained over the   network operations that are performed [Schwartz & Tsirigotis 1991a].   In addition, we monitor the progress and network loading of the   measurements during the measurement runs, observing the log of   connection requests in progress as well as physical and transport   level network status (which indicate the amount of concurrent network   activity in progress).  Finally, because the measurements are   controlled from a single centralized location, it is quite easy to   stop the measurements at any time.Schwartz                                                        [Page 4]

⌨️ 快捷键说明

复制代码 Ctrl + C
搜索代码 Ctrl + F
全屏模式 F11
切换主题 Ctrl + Shift + D
显示快捷键 ?
增大字号 Ctrl + =
减小字号 Ctrl + -