⭐ 欢迎来到虫虫下载站! | 📦 资源下载 📁 资源专辑 ℹ️ 关于我们
⭐ 虫虫下载站

📄 rfc2775.txt

📁 中、英文RFC文档大全打包下载完全版 .
💻 TXT
📖 第 1 页 / 共 3 页
字号:
RFC 2775                 Internet Transparency             February 2000   selected traffic, usually belonging to a very limited set of   applications). Firewalls, by their nature, fundamentally limit   transparency.3.3.2 SOCKS   A footnote to the effect of firewalls is the SOCKS mechanism [RFC   1928] by which untrusted applications such as telnet and ftp can   punch through a firewall.  SOCKS requires a shim library in the   Intranet client, and a server in the firewall which is essentially an   application level relay. As a result, the remote server does not see   the real client; it believes that the firewall is the client.3.4 Private addresses   When the threat of IPv4 address exhaustion first arose, and in some   cases user sites were known to be "pirating" addresses for private   use, a set of official private addresses were hurriedly allocated   [RFC 1597] and later more carefully defined [BCP 5].  The legitimate   existence of such an address allocation proved to very appealing, so   Intranets with large numbers of non-global addresses came into   existence. Unfortunately, such addresses by their nature cannot be   used for communication across the public Internet; without special   measures, hosts using private addresses are cut off from the world.   Note that private address space is sometimes asserted to be a   security feature, based on the notion that outside knowledge of   internal addresses might help intruders. This is a false argument,   since it is trivial to hide addresses by suitable access control   lists, even if they are globally unique - indeed that is a basic   feature of a filtering router, the simplest form of firewall. A   system with a hidden address is just as private as a system with a   private address.  There is of course no possible point in hiding the   addresses of servers to which outside access is required.   It is also worth noting that the IPv6 equivalent of private   addresses, i.e. site-local addresses, have similar characteristics to   BCP 5 addresses, but their use will not be forced by a lack of   globally unique IPv6 addresses.3.5 Network address translators   Network address translators (NATs) are an almost inevitable   consequence of the existence of Intranets using private addresses yet   needing to communicate with the Internet at large. Their   architectural implications are discussed at length in [NAT-ARCH], the   fundamental point being that address translation on the fly destroys   end-to-end address transparency and breaks any middleware orCarpenter                    Informational                      [Page 7]RFC 2775                 Internet Transparency             February 2000   applications that depend on it. Numerous protocols, for example   H.323, carry IP addresses at application level and fail to traverse a   simple NAT box correctly. If the full range of Internet applications   is to be used, NATs have to be coupled with application level   gateways (ALGs) or proxies. Furthermore, the ALG or proxy must be   updated whenever a new address-dependent application comes along.  In   practice, NAT functionality is built into many firewall products, and   all useful NATs have associated ALGs, so it is difficult to   disentangle their various impacts.3.6 Application level gateways, relays, proxies, and caches   It is reasonable to position application level gateways, relays,   proxies, and caches at certain critical topological points,   especially the Intranet/Internet boundary.  For example, if an   Intranet does not use SMTP as its mail protocol, an SMTP gateway is   needed. Even in the normal case, an SMTP relay is common, and can   perform useful mail routing functions, spam filtering, etc. (It may   be observed that spam filtering is in some ways a firewall function,   but the store-and-forward nature of electronic mail and the   availability of MX records allow it to be a distinct and separate   function.)   Similarly, for a protocol such as HTTP with a well-defined voluntary   proxy mechanism, application proxies also serving as caches are very   useful. Although these devices interfere with transparency, they do   so in a precise way, correctly terminating network, transport and   application protocols on both sides. They can however exhibit some   shortfalls in ease of configuration and failover.   However, there appear to be cases of "involuntary" applications level   devices such as proxies that grab and modify HTTP traffic without   using the appropriate mechanisms, sometimes known as "transparent   caches", or mail relays that purport to remove undesirable words.   These devices are by definition not transparent, and may have totally   unforeseeable side effects.  (A possible conclusion is that even for   non-store-and-forward protocols, a generic diversion mechanism   analogous to the MX record would be of benefit. The SRV record [RFC   2052] is a step in this direction.)3.7 Voluntary isolation and peer networks   There are communities that think of themselves as being so different   that they require isolation via an explicit proxy, and even by using   proprietary protocols and addressing schemes within their network. An   example is the WAP Forum which targets very small phone-like devices   with some capabilities for Internet connectivity. However, it's notCarpenter                    Informational                      [Page 8]RFC 2775                 Internet Transparency             February 2000   the Internet they're connecting directly to. They have to go through   a proxy. This could potentially mean that millions of devices will   never know the benefits of end-to-end connectivity to the Internet.   A similar effect arises when applications such as telephony span both   an IP network and a peer network layer using a different technology.   Although the application may work end-to-end, there is no possibility   of end-to-end packet transmission.3.8 Split DNS   Another consequence of the Intranet/Internet split is "split DNS" or   "two faced DNS", where a corporate network serves up partly or   completely different DNS inside and outside its firewall. There are   many possible variants on this; the basic point is that the   correspondence between a given FQDN (fully qualified domain name) and   a given IPv4 address is no longer universal and stable over long   periods.3.9 Various load-sharing tricks   IPv4 was not designed to support anycast [RFC 1546], so there is no   natural approach to load-sharing when one server cannot do the job.   Various tricks have been used to resolve this (multicast to find a   free server, the DNS returns different addresses for the same FQDN in   a round-robin, a router actually routes packets sent to the same   address automatically to different servers, etc.). While these tricks   are not particularly harmful in the overall picture, they can be   implemented in such a way as to interfere with name or address   transparency.4. Summary of current status and impact   It is impossible to estimate with any numerical reliability how   widely the above inventions have been deployed. Since many of them   preserve the illusion of transparency while actually interfering with   it, they are extremely difficult to measure.   However it is certain that all the mechanisms just described are in   very widespread use; they are not a marginal phenomenon. In corporate   networks, many of them are the norm. Some of them (firewalls, relays,   proxies and caches) clearly have intrinsic value given the Intranet   concept. The others are largely artefacts and pragmatic responses to   various pressures in the operational Internet, and they must be   costing the industry very dearly in constant administration and   complex fault diagnosis.Carpenter                    Informational                      [Page 9]RFC 2775                 Internet Transparency             February 2000   In particular, the decline of transparency is having a severe effect   on deployment of end-to-end IP security. The Internet security model   [SECMECH] calls for security at several levels (roughly, network,   applications, and object levels).  The current network level security   model [RFC 2401] was constructed prior to the recognition that end-   to-end address transparency was under severe threat.  Although   alternative proposals have begun to emerge [HIP] the current reality   is that IPSEC cannot be deployed end-to-end in the general case.   Tunnel-mode IPSEC can be deployed between corporate gateways or   firewalls.  Transport-mode IPSEC can be deployed within a corporate   network in some cases, but it cannot span from Intranet to Internet   and back to another Intranet if there is any chance of a NAT along   the way.   Indeed, NAT breaks other security mechanisms as well, such as DNSSEC   and Kerberos, since they rely on address values.   The loss of transparency brought about by private addresses and NATs   affects many applications protocols to a greater or lesser extent.   This is explored in detail in [NAT-PROT]. A more subtle effect is   that the prevalence of dynamic addresses (from DHCP, SLIP and PPP)   has fed upon the trend towards client/server computing.  Today it is   largely true that servers have fixed addresses, clients have dynamic   addresses, and servers can in no way assume that a client's IP   address identifies the client. On the other hand, clients rely on   servers having stable addresses since a DNS lookup is the only   generally deployed mechanism by which a client can find a server and   solicit service.  In this environment, there is little scope for true   peer-to-peer applications protocols, and no easy solution for mobile   servers. Indeed, the very limited demand for Mobile IP might be   partly attributed to the market dominance of client/server   applications in which the client's address is of transient   significance. We also see a trend towards single points of failure   such as media gateways, again resulting from the difficulty of   implementing peer-to-peer solutions directly between clients with no   fixed address.   The notion that servers can use precious globally unique addresses   from a small pool, because there will always be fewer servers than   clients, may become anachronistic when most electrical devices become   network-manageable and thus become servers for a management protocol.   Similarly, if every PC becomes a telephone (or the converse), capable   of receiving unsolicited incoming calls, the lack of stable IP   addresses for PCs will be an issue. Another impending paradigm shift   is when domestic and small-office subscribers move from dial-up to   always-on Internet connectivity, at which point transient address   assignment from a pool becomes much less appropriate.Carpenter                    Informational                     [Page 10]RFC 2775                 Internet Transparency             February 2000   Many of the inventions described in the previous section lead to the   datagram traffic between two hosts being directly or indirectly   mediated by at least one other host. For example a client may depend   on a DHCP server, a server may depend on a NAT, and any host may   depend on a firewall. This violates the fate-sharing principle of   [Saltzer] and introduces single points of failure. Worse, most of   these points of failure require configuration data, yet another   source of operational risk. The original notion that datagrams would   find their way around failures, especially around failed routers, has   been lost; indeed the overloading of border routers with additional   functions has turned them into critical rather than redundant   components, even for multihomed sites.   The loss of address transparency has other negative effects.  For   example, large scale servers may use heuristics or even formal   policies that assign different priorities to service for different   clients, based on their addresses. As addresses lose their global   meaning, this mechanism will fail. Similarly, any anti-spam or anti-   spoofing techniques that rely on reverse DNS lookup of address values   can be confused by translated addresses. (Uncoordinated renumbering   can have similar disadvantages.)   The above issues are not academic. They add up to complexity in   applications design, complexity in network configuration, complexity   in security mechanisms, and complexity in network management.   Specifically, they make fault diagnosis much harder, and by   introducing more single points of failure, they make faults more   likely to occur.5. Possible future directions5.1 Successful migration to IPv6   In this scenario, IPv6 becomes fully implemented on all hosts and   routers, including the adaptation of middleware, applications, and   management systems. Since the address space then becomes big enough   for all conceivable needs, address transparency can be restored.   Transport-mode IPSEC can in principle deploy, given adequate security   policy tools and a key infrastructure.  However, it is widely   believed that the Intranet/firewall model will certainly persist.   Note that it is a basic assumption of IPv6 that no artificial   constraints will be placed on the supply of addresses, given that   there are so many of them. Current practices by which some ISPs   strongly limit the number of IPv4 addresses per client will have no   reason to exist for IPv6. (However, addresses will still be assigned   prudently, according to guidelines designed to favour hierarchical   routing.)Carpenter                    Informational                     [Page 11]RFC 2775                 Internet Transparency             February 2000   Clearly this is in any case a very long term scenario, since it   assumes that IPv4 has declined to the point where IPv6 is required   for universal connectivity. Thus, a viable version of Scenario 5.3 is   a prerequisite for Scenario 5.1.5.2 Complete failure of IPv6   In this scenario, IPv6 fails to reach any significant level of   operational deployment, IPv4 addressing is the only available   mechanism, and address transparency cannot be restored. IPSEC cannot   be deployed globally in its current form. In the very long term, the   pool of globally unique IPv4 addresses will be nearly totally   allocated, and new addresses will generally not be available for any   purpose.   It is unclear exactly what is likely to happen if the Internet   continues to rely exclusively on IPv4, because in that eventuality a   variety of schemes are likely to promulgated, with a view toward   providing an acceptable evolutionary path for the network. However,   we can examine two of the more simplistic sub-scenarios which are   possible, while realising that the future would be unlikely to match   either one exactly:5.2.1 Re-allocating the IPv4 address space   Suppose that a mechanism is created to continuously recover and re-   allocate IPv4 addresses. A single global address space is maintained,   with all sites progressively moving to an Intranet private address   model, with global addresses being assigned temporarily from a pool   of several billion.   5.2.1.1 A sub-sub-scenario of this is generalised use of NAT and NAPT           (NAT with port number translation). This has the disadvantage           that the host is unaware of the unique address being used for           its traffic, being aware only of its ambiguous private           address, with all the problems that this generates. See           [NAT-ARCH].   5.2.1.2 Another sub-sub-scenario is the use of realm-specific IP           addressing implemented at the host rather than by a NAT box.           See [RSIP]. In this case the host is aware of its unique           address, allowing for substantial restoration of the end-to-           end usefulness of addresses, e.g. for TCP or cryptographic           checksums.Carpenter                    Informational                     [Page 12]RFC 2775                 Internet Transparency             February 2000

⌨️ 快捷键说明

复制代码 Ctrl + C
搜索代码 Ctrl + F
全屏模式 F11
切换主题 Ctrl + Shift + D
显示快捷键 ?
增大字号 Ctrl + =
减小字号 Ctrl + -