diff options
Diffstat (limited to 'doc/rfc/rfc4948.txt')
-rw-r--r-- | doc/rfc/rfc4948.txt | 2411 |
1 files changed, 2411 insertions, 0 deletions
diff --git a/doc/rfc/rfc4948.txt b/doc/rfc/rfc4948.txt new file mode 100644 index 0000000..1fb0a3e --- /dev/null +++ b/doc/rfc/rfc4948.txt @@ -0,0 +1,2411 @@ + + + + + + +Network Working Group L. Andersson +Request for Comments: 4948 Acreo AB +Category: Informational E. Davies + Folly Consulting + L. Zhang + UCLA + August 2007 + + + Report from the IAB workshop on Unwanted Traffic March 9-10, 2006 + +Status of This Memo + + This memo provides information for the Internet community. It does + not specify an Internet standard of any kind. Distribution of this + memo is unlimited. + +Copyright Notice + + Copyright (C) The IETF Trust (2007). + +Abstract + + This document reports the outcome of a workshop held by the Internet + Architecture Board (IAB) on Unwanted Internet Traffic. The workshop + was held on March 9-10, 2006 at USC/ISI in Marina del Rey, CA, USA. + The primary goal of the workshop was to foster interchange between + the operator, standards, and research communities on the topic of + unwanted traffic, as manifested in, for example, Distributed Denial + of Service (DDoS) attacks, spam, and phishing, to gain understandings + on the ultimate sources of these unwanted traffic, and to assess + their impact and the effectiveness of existing solutions. It was + also a goal of the workshop to identify engineering and research + topics that could be undertaken by the IAB, the IETF, the IRTF, and + the network research and development community at large to develop + effective countermeasures against the unwanted traffic. + + + + + + + + + + + + + + + +Andersson, et al. Informational [Page 1] + +RFC 4948 Unwanted Traffic August 2007 + + +Table of Contents + + 1. Introduction . . . . . . . . . . . . . . . . . . . . . . . . . 3 + 2. The Root of All Evils: An Underground Economy . . . . . . . . 4 + 2.1. The Underground Economy . . . . . . . . . . . . . . . . . 5 + 2.2. Our Enemy Using Our Networks, Our Tools . . . . . . . . . 6 + 2.3. Compromised Systems Being a Major Source of Problems . . . 7 + 2.4. Lack of Meaningful Deterrence . . . . . . . . . . . . . . 8 + 2.5. Consequences . . . . . . . . . . . . . . . . . . . . . . . 10 + 3. How Bad Is The Problem? . . . . . . . . . . . . . . . . . . . 10 + 3.1. Backbone Providers . . . . . . . . . . . . . . . . . . . . 10 + 3.1.1. DDoS Traffic . . . . . . . . . . . . . . . . . . . . . 10 + 3.1.2. Problem Mitigation . . . . . . . . . . . . . . . . . . 11 + 3.2. Access Providers . . . . . . . . . . . . . . . . . . . . . 12 + 3.3. Enterprise Networks: Perspective from a Large + Enterprise . . . . . . . . . . . . . . . . . . . . . . . . 13 + 3.4. Domain Name Services . . . . . . . . . . . . . . . . . . . 14 + 4. Current Vulnerabilities and Existing Solutions . . . . . . . . 15 + 4.1. Internet Vulnerabilities . . . . . . . . . . . . . . . . . 15 + 4.2. Existing Solutions . . . . . . . . . . . . . . . . . . . . 16 + 4.2.1. Existing Solutions for Backbone Providers . . . . . . 16 + 4.2.2. Existing Solutions for Enterprise Networks . . . . . . 17 + 4.3. Shortfalls in the Existing Network Protection . . . . . . 18 + 4.3.1. Inadequate Tools . . . . . . . . . . . . . . . . . . . 18 + 4.3.2. Inadequate Deployments . . . . . . . . . . . . . . . . 18 + 4.3.3. Inadequate Education . . . . . . . . . . . . . . . . . 19 + 4.3.4. Is Closing Down Open Internet Access Necessary? . . . 19 + 5. Active and Potential Solutions in the Pipeline . . . . . . . . 20 + 5.1. Central Policy Repository . . . . . . . . . . . . . . . . 20 + 5.2. Flow Based Tools . . . . . . . . . . . . . . . . . . . . . 21 + 5.3. Internet Motion Sensor (IMS) . . . . . . . . . . . . . . . 21 + 5.4. BCP 38 . . . . . . . . . . . . . . . . . . . . . . . . . . 22 + 5.5. Layer 5 to 7 Awareness . . . . . . . . . . . . . . . . . . 22 + 5.6. How To's . . . . . . . . . . . . . . . . . . . . . . . . . 22 + 5.7. SHRED . . . . . . . . . . . . . . . . . . . . . . . . . . 23 + 6. Research in Progress . . . . . . . . . . . . . . . . . . . . . 23 + 6.1. Ongoing Research . . . . . . . . . . . . . . . . . . . . . 23 + 6.1.1. Exploited Hosts . . . . . . . . . . . . . . . . . . . 23 + 6.1.2. Distributed Denial of Service (DDoS) Attacks . . . . . 25 + 6.1.3. Spyware . . . . . . . . . . . . . . . . . . . . . . . 26 + 6.1.4. Forensic Aids . . . . . . . . . . . . . . . . . . . . 26 + 6.1.5. Measurements . . . . . . . . . . . . . . . . . . . . . 27 + 6.1.6. Traffic Analysis . . . . . . . . . . . . . . . . . . . 27 + 6.1.7. Protocol and Software Security . . . . . . . . . . . . 27 + 6.2. Research on the Internet . . . . . . . . . . . . . . . . . 27 + 6.2.1. Research and Standards . . . . . . . . . . . . . . . . 28 + 6.2.2. Research and the Bad Guys . . . . . . . . . . . . . . 29 + + + + +Andersson, et al. Informational [Page 2] + +RFC 4948 Unwanted Traffic August 2007 + + + 7. Aladdin's Lamp . . . . . . . . . . . . . . . . . . . . . . . . 30 + 7.1. Security Improvements . . . . . . . . . . . . . . . . . . 30 + 7.2. Unwanted Traffic . . . . . . . . . . . . . . . . . . . . . 31 + 8. Workshop Summary . . . . . . . . . . . . . . . . . . . . . . . 31 + 8.1. Hard Questions . . . . . . . . . . . . . . . . . . . . . . 31 + 8.2. Medium or Long Term Steps . . . . . . . . . . . . . . . . 32 + 8.3. Immediately Actionable Steps . . . . . . . . . . . . . . . 33 + 9. Terminology . . . . . . . . . . . . . . . . . . . . . . . . . 33 + 10. Security Considerations . . . . . . . . . . . . . . . . . . . 38 + 11. Acknowledgements . . . . . . . . . . . . . . . . . . . . . . . 38 + 12. Informative References . . . . . . . . . . . . . . . . . . . . 39 + Appendix A. Participants in the Workshop . . . . . . . . . . . . 40 + Appendix B. Workshop Agenda . . . . . . . . . . . . . . . . . . . 41 + Appendix C. Slides . . . . . . . . . . . . . . . . . . . . . . . 41 + +1. Introduction + + The Internet carries a lot of unwanted traffic today. To gain a + better understanding of the driving forces behind such unwanted + traffic and to assess existing countermeasures, the IAB organized an + "Unwanted Internet Traffic" workshop and invited experts on different + aspects of unwanted traffic from operator, vendor, and research + communities to the workshop. The intention was to share information + among people from different fields and organizations, fostering an + interchange of experiences, views, and ideas between the various + communities on this important topic. The major goal of this workshop + was to stimulate discussion at a deep technical level to assess + today's situation in regards to: + + o the kinds of unwanted traffic that are seen on the Internet, + + o how bad the picture looks, + + o who and where are the major sources of the problem, + + o which solutions work and which do not, and + + o what needs to be done. + + The workshop was very successful. Over one and half days of + intensive discussions, the major sources of the unwanted traffic were + identified, and a critical assessment of the existing mitigation + tools was conducted. However, due to the limitation of available + time, it was impossible to cover the topic of unwanted traffic in its + entirety. Thus, for some of the important issues, only the surface + was touched. Furthermore, because the primary focus of the workshop + was to collect and share information on the current state of affairs, + it is left as the next step to attempt to derive solutions to the + + + +Andersson, et al. Informational [Page 3] + +RFC 4948 Unwanted Traffic August 2007 + + + issues identified. This will be done in part as activities within + the IAB, the IETF, and the IRTF. + + During the workshop, a number of product and company names were + cited, which are reflected in the report to a certain extent. + However, a mention of any product in this report should not be taken + as an endorsement of that product; there may well be alternative, + equally relevant or efficacious products in the market place. + + This report is a summary of the contributions by the workshop + participants, and thus it is not an IAB document. The views and + positions documented in the report do not necessarily reflect IAB + views and positions. + + The workshop participant list is attached in Appendix A. The agenda + of the workshop can be found in Appendix B. Links to a subset of the + presentations are provided in Appendix C; the rest of the + presentations are of a sensitive nature, and it has been agreed that + they will not be made public. Definitions of the jargon used in + describing unwanted traffic can be found in Section 9. + +2. The Root of All Evils: An Underground Economy + + The first important message this workshop would like to bring to the + Internet community's attention is the existence of an underground + economy. This underground economy provides an enormous amount of + monetary fuel that drives the generation of unwanted traffic. This + economic incentive feeds on an Internet that is to a large extent + wide open. The open nature of the Internet fosters innovations but + offers virtually no defense against abuses. It connects to millions + of mostly unprotected hosts owned by millions of mostly naive users. + These users explore and benefit from the vast opportunities offered + by the new cyberspace, with little understanding of its vulnerability + to abuse and the potential danger of their computers being + compromised. Moreover, the Internet was designed without built-in + auditing trails. This was an appropriate choice at the time, but now + the lack of traceability makes it difficult to track down malicious + activities. Combined with a legal system that is yet to adapt to the + new challenge of regulating the cyberspace, this means the Internet, + as of today, has no effective deterrent to miscreants. The + unfettered design and freedom from regulation have contributed to the + extraordinary success of the Internet. At the same time, the + combination of these factors has also led to an increasing volume of + unwanted traffic. The rest of this section provides a more detailed + account of each of the above factors. + + + + + + +Andersson, et al. Informational [Page 4] + +RFC 4948 Unwanted Traffic August 2007 + + +2.1. The Underground Economy + + As in any economic system, the underground economy is regulated by a + demand and supply chain. The underground economy, which began as a + barter system, has evolved into a giant shopping mall, commonly + running on IRC (Internet Relay Chat) servers. The IRC servers + provide various online stores selling information about stolen credit + cards and bank accounts, malware, bot code, botnets, root accesses to + compromised hosts and web servers, and much more. There are DDoS + attack stores, credit card stores, PayPal and bank account stores, as + well as Cisco and Juniper router stores that sell access to + compromised routers. Although not everything can be found on every + server, most common tools used to operate in the underground economy + can be found on almost all the servers. + + How do miscreants turn attack tools and compromised machines into + real assets? In the simplest case, miscreants electronically + transfer money from stolen bank accounts directly to an account that + they control, often in another country. In a more sophisticated + example, miscreants use stolen credit cards or PayPal accounts for + online purchases. To hide their trails, they often find remailers + who receive the purchased goods and then repackage them to send to + the miscreants for a fee. The miscreants may also sell the goods + through online merchandising sites such as eBay. They request the + payments be made in cashier checks or money orders to be sent to the + people who provide money laundering services for the miscreants by + receiving the payments and sending them to banks in a different + country, again in exchange for a fee. In either case, the + destination bank accounts are used only for a short period and are + closed as soon as the money is withdrawn by the miscreants. + + The miscreants obtain private and financial information from + compromised hosts and install bots (a.k.a. zombies) on them. They + can also obtain such information from phishing attacks. Spam + messages mislead naive users into accessing spoofed web sites run by + the miscreants where their financial information is extracted and + collected. + + The miscreants in general are not skilled programmers. With money, + however, they can hire professional writers to produce well phrased + spam messages, and hire coders to develop new viruses, worms, + spyware, and botnet control packages, thereby resupplying the + underground market with new tools that produce more unwanted traffic + on the Internet: spam messages that spread phishing attacks, botnets + that are used to launch DDoS attacks, click fraud that "earns" income + by deceiving online commercial advertisers, and new viruses and worms + that compromise more hosts and steal additional financial information + as well as system passwords and personal identity information. + + + +Andersson, et al. Informational [Page 5] + +RFC 4948 Unwanted Traffic August 2007 + + + The income gained from the above illegal activities allows miscreants + to hire spammers, coders, and IRC server providers. Spammers use + botnets. Direct marketing companies set up dirty affiliate programs. + Some less than scrupulous banks are also involved to earn transaction + fees from moving the dirty money around. In the underground market, + everything can be traded, and everything has a value. Thus is + spawned unwanted traffic of all kinds. + + The underground economy has evolved very rapidly over the past few + years. In the early days of bots and botnets, their activities were + largely devoted to DDoS attacks and were relatively easy to detect. + As the underground economy has evolved, so have the botnets. They + have moved from easily detectable behavior to masquerading as normal + user network activity to achieve their goals, making detection very + difficult even by vigilant system administrators. + + The drive for this rapid evolution comes to a large extent from the + change in the intention of miscreant activity. Early virus attacks + and botnets were largely anarchic activities. Many were done by + "script kiddies" to disrupt systems without a real purpose or to + demonstrate the prowess of the attacker, for example in compromising + systems that were touted as "secure". Mirroring the + commercialization of the Internet and its increasing use for + e-business, miscreant activity is now mostly focused on conventional + criminal lines. Systems are quietly subverted with the goal of + obtaining illicit financial gain in the future, rather than causing + visible disruptions as was often the aim of the early hackers. + +2.2. Our Enemy Using Our Networks, Our Tools + + Internet Relay Chat (IRC) servers are commonly used as the command + and control channel for the underground market. These servers are + paid for by miscreants and are professionally supported. They are + advertised widely to attract potential consumers, and thus are easy + to find. The miscreants first talk to each other on the servers to + find out who is offering what on the market, then exchange encrypted + private messages to settle the deals. + + The miscreants are not afraid of network operators seeing their + actions. If their activities are interrupted, they simply move to + another venue. When ISPs take actions to protect their customers, + revenge attacks are uncommon as long as the miscreants' cash flow is + not disturbed. When a botnet is taken out, they move on to the next + one, as there is a plentiful supply. However, if an IRC server is + taken out that disturbs their cash flow, miscreants can be ruthless + and severe attacks may follow. They currently have no fear, as they + know the chances of their being caught are minimal. + + + + +Andersson, et al. Informational [Page 6] + +RFC 4948 Unwanted Traffic August 2007 + + + Our enemies make good use of the Internet's global connectivity as + well as all the tools the Internet has developed. IRC servers + provide a job market for the miscreants and shopping malls of attack + tools. Networking research has produced abundant results making it + easier to build large scale distributed systems, and these have been + adopted by miscreants to build large size, well-controlled botnets. + Powerful search engines also enable one to quickly find readily + available tools and resources. The sophistication of attacks has + increased with time, while the skills required to launch effective + attacks have become minimal. Attackers can be hiding anywhere in the + Internet while attacks get launched on a global scale. + +2.3. Compromised Systems Being a Major Source of Problems + + The current Internet provides a field ripe for exploitation. The + majority of end hosts run vulnerable platforms. People from all + walks of life eagerly jump into the newly discovered online world, + yet without the proper training needed to understand the full + implications. This is at least partially due to most users failing + to anticipate how such a great invention can be readily abused. As a + result, the Internet has ended up with a huge number of compromised + hosts, without their owners being aware that it has happened. + + Unprotected hosts can be compromised in multiple ways. Viruses and + worms can get into the system through exploiting bugs in the existing + operating systems or applications, sometimes even in anti-virus + programs. A phishing site may also take the opportunity to install + malware on a victim's computer when a user is lured to the site. + More recently, viruses have also started being propagated through + popular peer-to-peer file sharing applications. With multiple + channels of propagation, malware has become wide-spread, and infected + machines include not only home PCs (although they do represent a + large percentage), but also corporate servers, and even government + firewalls. + + News of new exploits of vulnerabilities of Microsoft Windows + platforms is all too frequent, which leads to a common perception + that the Microsoft Windows platform is a major source of + vulnerability. One of the reasons for the frequent vulnerability + exploits of the Windows system is its popularity in the market place; + thus, a miscreant's investment in each exploit can gain big returns + from infecting millions of machines. As a result, each incident is + also likely to make headlines in the news. In reality, all other + platforms such as Linux, Solaris, and MAC OS for example, are also + vulnerable to compromises. Routers are not exempt from security + break-ins either, and using a high-end router as a DoS launchpad can + be a lot more effective than using a bundle of home PCs. + + + + +Andersson, et al. Informational [Page 7] + +RFC 4948 Unwanted Traffic August 2007 + + + Quietly subverting large numbers of hosts and making them part of a + botnet, while leaving their normal functionality and connectivity + essentially unimpaired, is now a major aim of miscreants and it + appears that they are being all too successful. Bots and the + functions they perform are often hard to detect and most of the time + their existence are not known to system operators or owners (hence, + the alternative name for hosts infected with bots controlled by + miscreants - zombies); by the time they are detected, it might very + well be too late as they have carried out the intended + (mal-)function. + + The existence of a large number of compromised hosts is a + particularly challenging problem to the Internet's security. Not + only does the stolen financial information lead to enormous economic + losses, but also there has been no quick fix to the problem. As + noted above, in many cases the owners of the compromised computers + are unaware of the problem. Even after being notified, some owners + still do not care about fixing the problem as long as their own + interest, such as playing online games, is not affected, even though + the public interest is endangered --- large botnets can use multi- + millions of such compromised hosts to launch DDoS attacks, with each + host sending an insignificant amount of traffic but the aggregate + exceeding the capacity of the best engineered systems. + +2.4. Lack of Meaningful Deterrence + + One of the Internet's big strengths is its ability to provide + seamless interconnection among an effectively unlimited number of + parties. However, the other side of the same coin is that there may + not be clear ways to assign responsibilities when something goes + wrong. Taking DDoS attacks as an example, an attack is normally + launched from a large number of compromised hosts, the attack traffic + travels across the Internet backbone to the access network(s) linking + to the victims. As one can see, there are a number of independent + stake-holders involved, and it is not immediately clear which party + should take responsibility for resolving the problem. + + Furthermore, tracking down an attack is an extremely difficult task. + The Internet architecture enables any IP host to communicate with any + other hosts, and it provides no audit trails. As a result, not only + is there no limit to what a host may do, but also there is no trace + after the event of what a host may have done. At this time, there is + virtually no effective tool available for problem diagnosis or packet + trace back. Thus, tracking down an attack is labor intensive and + requires sophisticated skills. As will be mentioned in the next + section, there is also a lack of incentive to report security + attacks. Compounded with the high cost, these factors make forensic + tracing of an attack to its root a rare event. + + + +Andersson, et al. Informational [Page 8] + +RFC 4948 Unwanted Traffic August 2007 + + + In human society, the legal systems provide protection against + criminals. However, in the cyberspace, the legal systems are lagging + behind in establishing regulations. The laws and regulations aim at + penalizing the conduct after the fact. If the likelihood of + detection is low, the deterrence would be minimal. Many national + jurisdictions have regulations about acts of computer fraud and + abuse, and they often carry significant criminal penalties. In the + US (and many other places), it is illegal to access government + computers without authorization, illegal to damage protected + government computers, and illegal to access confidential information + on protected computers. However, the definition of "access" can be + difficult to ascertain. For example, is sending an ICMP (Internet + Control Messaging Protocol) packet to a protected computer considered + illegal access? There is a lack of technical understanding among + lawmakers that would be needed to specify the laws precisely and + provide effective targeting limited to undesirable acts. Computer + fraud and liabilities laws provide a forum to address illegal access + activities and enable prosecution of cybercriminals. However, one + difficulty in prosecuting affiliate programs using bot infrastructure + is that they are either borderline legal, or there is little + evidence. There is also the mentality of taking legal action only + when the measurable monetary damage exceeds a high threshold, while + it is often difficult to quantify the monetary damage in individual + cases of cyberspace crimes. + + There is a coalition between countries on collecting cybercriminal + evidence across the world, but there is no rigorous way to trace + across borders. Laws and rules are mostly local to a country, + policies (when they exist) are mostly enacted and enforced locally, + while the Internet itself, that carries the unwanted traffic, + respects no borders. One estimate suggests that most players in the + underground economy are outside the US, yet most IRC servers + supporting the underground market may be running in US network + providers, enjoying the reliable service and wide connectivity to the + rest of the world provided by the networks. + + In addition, the definition of "unwanted" traffic also varies between + different countries. For example, China bans certain types of + network traffic that are considered legitimate elsewhere. Yet + another major difficulty is the trade-off and blurred line between + having audit trails to facilitate forensic analysis and to enforce + censorship. The greater ability we build into the network to control + traffic, the stronger would be the monitoring requirements coming + from the legislators. + + It should be emphasized that, while a legal system is necessary to + create effective deterrence and sanctions against miscreants, it is + by no means sufficient on its own. Rather, it must be accompanied by + + + +Andersson, et al. Informational [Page 9] + +RFC 4948 Unwanted Traffic August 2007 + + + technical solutions to unwanted traffic detection and damage + recovery. It is also by no means a substitute for user education. + Only a well informed user community can collectively establish an + effective defense in the cyberspace. + +2.5. Consequences + + What we have today is not a rosy picture: there are + + o big economic incentives and a rich environment to exploit, + + o no specific party to carry responsibility, + + o no auditing system to trace back to the sources of attacks, and + + o no well established legal regulations to punish offenders. + + The combination of these factors inevitably leads to ever increasing + types and volume of unwanted traffic. However, our real threats are + not the bots or DDoS attacks, but the criminals behind them. + Unwanted traffic is no longer only aiming for maximal disruption; in + many cases, it is now a means to illicit ends with the specific + purpose of generating financial gains for the miscreants. Their + crimes cause huge economic losses, counted in multiple billions of + dollars and continuing. + +3. How Bad Is The Problem? + + There are quite a number of different kinds of unwanted traffic on + the Internet today; the discussions at this workshop were mainly + around DDoS traffic and spam. The impact of DDoS and spam on + different parts of the network differs. Below, we summarize the + impact on backbone providers, access providers, and enterprise + customers, respectively. + +3.1. Backbone Providers + + Since backbone providers' main line of business is packet forwarding, + the impact of unwanted traffic is mainly measured by whether DDoS + traffic affects network availability. Spam or malware is not a major + concern because backbone networks do not directly support end users. + Router compromises may exist, but they are rare events at this time. + +3.1.1. DDoS Traffic + + Observation shows that, in the majority of DDoS attacks, attack + traffic can originate from almost anywhere in the Internet. In + particular, those regions with high speed user connectivity but + + + +Andersson, et al. Informational [Page 10] + +RFC 4948 Unwanted Traffic August 2007 + + + poorly managed end hosts are often the originating sites of DDoS + attacks. The miscreants tend to find targets that offer maximal + returns with minimal efforts. + + Backbone networks in general are well-provisioned in regard to + traffic capacities. Therefore, core routers and backbone link + capacity do not get affected much by most DDoS attacks; a 5Gbps + attack could be easily absorbed without causing noticeable impact on + the performance of backbone networks. However, DDoS attacks often + saturate access networks and make a significant impact on customers. + In particular, multihomed customers who have multiple well- + provisioned connections for high throughput and performance may + suffer from aggregated DDoS traffic coming in from all directions. + +3.1.2. Problem Mitigation + + Currently, backbone networks do not have effective diagnosis or + mitigation tools against DDoS attacks. The foremost problem is a + lack of incentives to deploy security solutions. Because IP transit + services are a commodity, controlling cost is essential to surviving + the competition. Thus, any expenditure tends to require a clearly + identified return-on-investment (ROI). Even when new security + solutions become available, providers do not necessarily upgrade + their infrastructure to deploy the solutions, as security solutions + are often prevention mechanisms that may not have an easily + quantifiable ROI. To survive in the competitive environment in which + they find themselves, backbone providers also try to recruit more + customers. Thus, a provider's reputation is important. Due to the + large number of attacks and inadequate security solution deployment, + effective attacks and security glitches can be expected. However, it + is not in a provider's best interest to report all the observed + attacks. Instead, the provider's first concern is to minimize the + number of publicized security incidents. For example, a "trouble + ticket" acknowledging the problem is issued only after a customer + complains. An informal estimate suggested that only about 10% of + DDoS attacks are actually reported (some other estimates have put the + figure as low as 2%). In short, there is a lack of incentives to + either report problems or deploy solutions. + + Partly as a consequence of the lack of incentive and lack of funding, + there exist few DDoS mitigation tools for backbone providers. + Network operators often work on their own time to fight the battle + against malicious attacks. Their primary mitigation tools today are + Access Control Lists (ACL) and BGP (Border Gateway Protocol) null + routes to black-hole unwanted traffic. These tools can be turned on + locally and do not require coordination across administrative + domains. When done at, or near, DDoS victims, these simple tools can + have an immediate effect in reducing the DDoS traffic volume. + + + +Andersson, et al. Informational [Page 11] + +RFC 4948 Unwanted Traffic August 2007 + + + However, these tools are rather rudimentary and inadequate, as we + will elaborate in Section 4.2.1. + +3.2. Access Providers + + A common issue that access providers share with backbone providers is + the lack of incentive and the shortage of funding needed to deploy + security solutions. As with the situation with security incidents on + the backbone, the number of security incidents reported by access + providers is estimated to be significantly lower than the number of + the actual incidents that occurred. + + Because access providers are directly connected to end customers, + they also face unique problems of their own. From the access + providers' viewpoint, the most severe impact of unwanted traffic is + not the bandwidth exhaustion, but the customer support load it + engenders. The primary impact of unwanted traffic is on end users, + and access providers must respond to incident reports from their + customers. Today, access providers are playing the role of IT help + desk for many of their customers, especially residential users. + According to some access providers, during the Microsoft Blaster worm + attack, the average time taken to handle a customer call was over an + hour. Due to the high cost of staffing the help desks, it is + believed that, if a customer calls the help desk just once, the + provider would lose the profit they would otherwise have otherwise + made over the lifetime of that customer account. + + To reduce the high customer service cost caused by security breaches, + most access providers offer free security software to their + customers. It is much cheaper to give the customer "free" security + software in the hope of preventing system compromises than handling + the system break-ins after the event. However, perhaps due to their + lack of understanding of the possible security problems they may + face, many customers fail to install security software despite the + free offer from their access providers, or even when they do, they + may lack the skill needed to configure a complex system correctly. + + What factors may influence how quickly customers get the security + breaches fixed? Past experience suggests the following observations: + + o Notification has little impact on end user repair behavior. + + o There is no significant difference in terms of repair behavior + between different industries or between business and home users. + + o Users' patching behavior follows an exponential decay pattern with + a time constant of approximately 40% per month. Thus, about 40% + of computers tend to be patched very quickly when a patch is + + + +Andersson, et al. Informational [Page 12] + +RFC 4948 Unwanted Traffic August 2007 + + + released, and approximately 40% of the remaining vulnerable + computers in each following month will show signs of being + patched. This leaves a few percent still unpatched after 6 + months. In the very large population of Internet hosts, this + results in a significant number of hosts that will be vulnerable + for the rest of their life. + + o There is a general lack of user understanding: after being + compromised, unmanaged computers may get replaced rather than + repaired, and this often results in infections occurring during + the installation process on the replacement. + +3.3. Enterprise Networks: Perspective from a Large Enterprise + + The operators of one big enterprise network reported their experience + regarding unwanted traffic to the workshop. Enterprises perceive + many forms of bad traffic including worms, malware, spam, spyware, + Instant Messaging (IM), peer-to-peer (P2P) traffic, and DoS. + Compared to backbone and access providers, enterprise network + operators are more willing to investigate security breaches, although + they may hesitate to pay a high price for security solutions. False + positives are very costly. Most operators prefer false negatives to + false positives. In general, enterprises prefer prevention solutions + to detection solutions. + + Deliberately created unwanted traffic (as opposed to unwanted traffic + that might arise from misconfiguration) in enterprise networks can be + sorted into three categories. The first is "Nuisance", which + includes unwanted traffic such as spam and peer-to-peer file sharing. + Although there were different opinions among the workshop + participants as to whether P2P traffic should, or should not, be + considered as unwanted traffic, enterprise network operators are + concerned not only that P2P traffic represents a significant share of + the total network load, but they are also sensitive to potential + copyright infringement issues that might lead to significant + financial and legal impacts on the company as a whole. In addition, + P2P file sharing applications have also became a popular channel for + malware propagation. + + The second category of unwanted traffic is labeled "Malicious", which + includes the traffic that spreads malware. This class of traffic can + be small in volume but the cost from the resulting damage can be + high. The clean up after an incident also requires highly skilled + operators. + + The third category of unwanted traffic is "Unknown": it is known that + there exists a class of traffic in the network that can be best + described in this way, as no one knows its purpose or the locations + + + +Andersson, et al. Informational [Page 13] + +RFC 4948 Unwanted Traffic August 2007 + + + of the sources. Malicious traffic can be obscured by encryption, + encapsulation, or covered up as legitimate traffic. The existing + detection tools are ineffective for this type of traffic. Noisy + worms are easy to identify, but stealth worms can open a backdoor on + hosts and stay dormant for a long time without causing any noticeable + detrimental effect. This type of bad traffic has the potential to + make the greatest impact on an enterprise from a threat perspective. + + There are more mitigation tools available for enterprise networks + than for backbone and access network providers; one explanation might + be the greater affordability of solutions for enterprise networks. + The costs of damage from a security breach can also have a very + significant impact on the profits of an enterprise. At the same + time, however, the workshop participants also expressed concerns + regarding the ongoing arms race between security exploits and + patching solutions. Up to now, security efforts have, by and large, + been reactive, creating a chain of security exploits and a consequent + stream of "fixes". Such a reactive mode has not only created a big + security market, but also does not enable us to get ahead of + attackers. + +3.4. Domain Name Services + + Different from backbone and access providers, there also exists a + class of Internet service infrastructure providers. Provision of + Domain Name System (DNS) services offers an example here. As + reported by operators from a major DNS hosting company, over time + there have been increasingly significant DDoS attacks on .com, .net + and root servers. + + DNS service operators have witnessed large scale DDoS attacks. The + most recent ones include reflection attacks resulting from queries + using spoofed source addresses. The major damage caused by these + attacks are bandwidth and resource exhaustion, which led to + disruption of critical services. The peak rate of daily DNS + transactions has been growing at a much faster rate than the number + of Internet users, and this trend is expected to continue. The heavy + load on the DNS servers has led to increasing complexity in providing + the services. + + In addition to intentional DDoS Attacks, some other causes of the + heavy DNS load included (1) well known bugs in a small number of DNS + servers that still run an old version of the BIND software, causing + significant load increase at top level servers; and (2) + inappropriately configured firewalls that allow DNS queries to come + out but block returning DNS replies, resulting in big adverse impacts + on the overall system. Most of such issues have been addressed in + the DNS operational guidelines drafted by the IETF DNS Operations + + + +Andersson, et al. Informational [Page 14] + +RFC 4948 Unwanted Traffic August 2007 + + + Working Group; however, many DNS operators have not taken appropriate + actions. + + At this time, the only effective and viable mitigation approach is + over-engineering the DNS service infrastructure by increasing link + bandwidth, the number of servers, and the server processing power, as + well as deploying network anycast. There is a concern about whether + the safety margin gained from over-engineering is, or is not, + adequate in sustaining DNS services over future attacks. Looking + forward, there are also a few new issues looming. Two imminent ones + are the expected widespread deployment of IPv6 whose new DNS software + would inevitably contain new bugs, and the DNS Security Extensions + (DNSSEC), which could potentially be abused to generate DDoS attacks. + +4. Current Vulnerabilities and Existing Solutions + + This section summarizes three aspects of the workshop discussions. + We first collected the major vulnerabilities mentioned in the + workshop, then made a summary of the existing solutions, and followed + up with an examination of the effectiveness, or lack of it, of the + existing solutions. + +4.1. Internet Vulnerabilities + + Below is a list of known Internet vulnerabilities and issues around + unwanted traffic. + + o Packet source address spoofing: there has been speculation that + attacks using spoofed source addresses are decreasing, due to the + proliferation of botnets, which can be used to launch various + attacks without using spoofed source addresses. It is certainly + true that not all the attacks use spoofed addresses; however, many + attacks, especially reflection attacks, do use spoofed source + addresses. + + o BGP route hijacking: in a survey conducted by Arbor Networks, + route hijacking together with source address spoofing are listed + as the two most critical vulnerabilities on the Internet. It has + been observed that miscreants hijack bogon prefixes for spam + message injections. Such hijacks do not affect normal packet + delivery and thus have a low chance of being noticed. + + o Everything over HTTP: port scan attacks occur frequently in + today's Internet, looking for open TCP or UDP ports through which + to gain access to computers. The reaction from computer system + management has been to close down all the unused ports, especially + in firewalls. One result of this reaction is that application + designers have moved to transporting all data communications over + + + +Andersson, et al. Informational [Page 15] + +RFC 4948 Unwanted Traffic August 2007 + + + HTTP to avoid firewall traversal issues. Transporting "everything + over HTTP" does not block attacks but has simply moved the + vulnerability from one place to another. + + o Everyone comes from Everywhere: in the earlier life of the + Internet it had been possible to get some indication of the + authenticity of traffic from a specific sender based for example + on the Time To Live (TTL). The TTL would stay almost constant + when traffic from a certain sender to a specific host entered an + operators network, since the sender will "always" set the TTL to + the same value. If a change in the TTL value occurred without an + accompanying change in the routing, one could draw the conclusion + that this was potential unwanted traffic. However, since hosts + have become mobile, they may be roaming within an operator's + network and the resulting path changes may put more (or less) hops + between the source and the destination. Thus, it is no longer + possible to interpret a change in the TTL value, even if it occurs + without any corresponding change in routing, as an indication that + the traffic has been subverted. + + o Complex Network Authentication: Network authentication as it is + used today is far too complex to be feasible for users to use + effectively. It will also be difficult to make it work with new + wireless access technologies. + + A possible scenario envisages a customers handset that is + initially on a corporate wireless network. If that customer + steps out of the corporate building, the handset may get + connected to the corporate network through a GPRS network. The + handset may then roam to a wireless LAN network when the user + enters a public area with a hotspot. Consequently, we need + authentication tools for cases when the underlying data link + layer technology changes quickly, possibly during a single + application session. + + o Unused Security Tools: Vendors and standards have produced quite a + number of useful security tools; however, not all, or even most, + of them get used extensively. + +4.2. Existing Solutions + +4.2.1. Existing Solutions for Backbone Providers + + Several engineering solutions exist that operators can deploy to + defend the network against unwanted traffic. Adequate provisioning + is one commonly used approach that can diminish the impact of DDoS on + the Internet backbone. The solution that received most mentions at + the workshop was BCP 38 on ingress filtering: universal deployment of + + + +Andersson, et al. Informational [Page 16] + +RFC 4948 Unwanted Traffic August 2007 + + + BCP 38 can effectively block DDoS attacks using spoofed source IP + addresses. At present, Access Control List (ACL) and BGP null + routing are the two tools most commonly used by network operators to + mitigate DDoS attacks. They are effective in blocking DDoS attacks, + especially when being applied at or near a victim's site. + + Unfortunately, BCP 38 is not widely deployed today. BCP 38 may + require device upgrades, and is considered tedious to configure and + maintain. Although widespread deployment of BCP 38 could benefit the + Internet as a whole, deployment by individual sites imposes a certain + amount of cost to the site, and does not provide a direct and + tangible benefit in return. In other words, BCP 38 suffers from a + lack of deployment incentives. + + Both BGP null routing and ACL have the drawback of relying on manual + configuration and thus are labor intensive. In addition, they also + suffer from blocking both attack and legitimate packets. There is + also a potential that some tools could back-fire, e.g., an overly + long ACL list might significantly slow down packet forwarding in a + router. + + Unicast Reverse Path Filtering (uRPF), which is available on some + routers, provides a means of implementing a restricted form of BCP 38 + ingress filtering without the effort of maintaining ACLs. uRPF uses + the routing table to check that a valid path back to the source + exists. However, its effectiveness depends on the specificity of the + routes against which source addresses are compared. The prevalence + of asymmetric routing means that the strict uRPF test (where the + route to the source must leave from the same interface on which the + packet being tested arrived) may have to be replaced by the loose + uRPF test (where the route may leave from any interface). The loose + uRPF test is not a guarantee against all cases of address spoofing, + and it may still be necessary to maintain an ACL to deal with + exceptions. + +4.2.2. Existing Solutions for Enterprise Networks + + A wide variety of commercial products is available for enterprise + network protection. Three popular types of protection mechanisms are + + o Firewalls: firewalls are perhaps the most widely deployed + protection products. However, the effectiveness of firewalls in + protecting enterprise confidential information can be weakened by + spyware installed internally, and they are ineffective against + attacks carried out from inside the perimeter established by the + firewalls. Too often, spyware installation is a byproduct of + installing other applications permitted by end users. + + + + +Andersson, et al. Informational [Page 17] + +RFC 4948 Unwanted Traffic August 2007 + + + o Application level gateways: these are becoming more widely used. + However, because they require application-specific support, and in + many cases they cache all the in-flight documents, configuration + can be difficult and the costs high. Thus, enterprise network + operators prefer network level protections over layer-7 solutions. + + o Anti-spam software: Anti-spam measures consume significant human + resources. Current spam mitigation tools include blacklists and + content filters. The more recent "learning" filters may help + significantly reduce the human effort needed and decrease the + number of both false positives and negatives. + + A more recent development is computer admission control, where a + computer is granted network access if and only if it belongs to a + valid user and appears to have the most recent set of security + patches installed. It is however a more expensive solution. A major + remaining issue facing enterprise network operators is how to solve + the user vulnerability problem and reduce reliance on user's + understanding of the need for security maintenance. + +4.3. Shortfalls in the Existing Network Protection + +4.3.1. Inadequate Tools + + Generally speaking, network and service operators do not have + adequate tools for network problem diagnosis. The current approaches + largely rely on the experience and skills of the operators, and on + time-consuming manual operations. The same is true for mitigation + tools against attacks. + +4.3.2. Inadequate Deployments + + The limited number of existing Internet protection measures have not + been widely deployed. Deployment of security solutions requires + resources which may not be available. It also requires education + among the operational community to recognize the critical importance + of patch installation and software upgrades; for example, a bug in + the BIND packet was discovered and fixed in 2003, yet a number of DNS + servers still run the old software today. Perhaps most importantly, + a security solution must be designed with the right incentives to + promote their deployment. Effective protection also requires + coordination between competing network providers. For the time + being, it is often difficult to even find the contact information for + operators of other networks. + + A number of workshop participants shared the view that, if all the + known engineering approaches and bug fixes were universally deployed, + the Internet could have been enjoying a substantially reduced number + + + +Andersson, et al. Informational [Page 18] + +RFC 4948 Unwanted Traffic August 2007 + + + of security problems today. In particular, the need for, and lack + of, BCP 38 deployment was mentioned numerous times during the + workshop. There is also a lack of enthusiasm about the routing + security requirements document being developed by the IETF RPSEC + (Routing Protocol Security) Working Group, which focuses heavily on + cryptographically-based protection requirements. Not only would + cryptographically-based solutions face the obstacle of funding for + deployment, but also they are likely to bring with them their own set + of problems. + +4.3.3. Inadequate Education + + There exists an educational challenge to disseminate the knowledge + needed for secure Internet usage and operations. Easily guessed + passwords and plaintext password transmission are still common in + many parts of the Internet. One common rumor claims that Cisco + routers were shipped with a default password "cisco" and this was + used by attackers to break into routers. In reality, operators often + configure Cisco routers with that password, perhaps because of the + difficulty of disseminating passwords to multiple maintainers. A + similar problem exists for Juniper routers and other vendors' + products. + + How to provide effective education to the Internet user community at + large remains a great challenge. As mentioned earlier in this + report, the existence of a large number of compromised hosts is one + major source of the unwanted traffic problem, and the ultimate + solution to this problem is a well-informed, vigilant user community. + +4.3.4. Is Closing Down Open Internet Access Necessary? + + One position made at the workshop is that, facing the problems of + millions of vulnerable computers and lack of effective deterrence, + protecting the Internet might require a fundamental change to the + current Internet architecture, by replacing unconstrained open access + to the Internet with strictly controlled access. Although the + participants held different positions on this issue, a rough + consensus was reached that, considering the overall picture, + enforcing controlled access does not seem the best solution to + Internet protection. Instead, the workshop identified a number of + needs that should be satisfied to move towards a well protected + Internet: + + o the need for risk assessment for service providers; at this time, + we lack a commonly agreed bar for security assurance; + + o the need to add traceability to allow tracking of abnormal + behavior in the network, and + + + +Andersson, et al. Informational [Page 19] + +RFC 4948 Unwanted Traffic August 2007 + + + o the need for liability if someone fails to follow recommended + practices. + + Adding traceability has been difficult due to the distributed nature + of the Internet. Collaboration among operators is a necessity in + fighting cybercrimes. We must also pay attention to preparation for + the next cycle of miscreant activity, and not devote all our efforts + to fixing the existing problems. As discussed above, the current + reactive approach to security problems is not a winning strategy. + +5. Active and Potential Solutions in the Pipeline + + This section addresses the issues that vendors recognized as + important and for which there will be solutions available in the near + future. + + There are a number of potential solutions that vendors are working + on, but are not yet offering as part of their product portfolio, that + will allegedly remedy or diagnose the problems described in + Section 4.1. + + Inevitably, when vendors have or are about to make a decision on + implementing new features in their products but have not made any + announcement, the vendors are not willing to talk about the new + features openly, which limits what can be said in this section. + +5.1. Central Policy Repository + + One idea is to build a Central Policy Repository that holds policies + that are known to work properly, e.g., policies controlling from whom + one would accept traffic when under attack. This repository could, + for example, keep information on which neighbor router or AS is doing + proper ingress address filtering. The repository could also hold the + configurations that operators use to upgrade configurations on their + routers. + + If such a repository is to be a shared resource used by multiple + operators, it will necessarily require validation and authentication + of the stored policies to ensure that the repository does not become + the cause of vulnerabilities. Inevitably, this would mean that the + information comes with a cost and it will only be viable if the sum + of the reductions in individual operators' costs is greater than the + costs of maintaining the repository. + + + + + + + + +Andersson, et al. Informational [Page 20] + +RFC 4948 Unwanted Traffic August 2007 + + +5.2. Flow Based Tools + + A set of tools based on flow data is widely used to extract + information from both network and data link layers. Tools have been + built that can be used to find out the sources of almost any type of + traffic, including certain unwanted traffic. These flow-based tools + make it possible to do things like DDoS traceback, traffic/peering + analyses, and detection of botnets, worms, and spyware. + + These tools monitor flows on the network and build baselines for what + is the "normal" behavior. Once the baseline is available, it is + possible to detect anomalous activity. It is easy to detect + variations over time, and decide if the variation is legitimate or + not. It is possible to take this approach further, typically + involving the identification of signatures of particular types of + traffic. + + These flow-based tools are analogous to the "sonar" that is used by + navies to listen for submarines. Once a particular submarine is + identified, it is possible to record its sonar signature to be used + to provide rapid identification in the future when the same submarine + is encountered again. + + Examples of existing tools include + Cisco IOS NetFlow <http://www.cisco.com/en/US/products/ps6601/ + products_ios_protocol_group_home.html>, + sFlow <http://www.sflow.org/>, and + NeTraMet <http://www.caida.org/tools/measurement/netramet/> based on + the IETF RTFM and IPFIX standards. + + There are also tools for working with the output of NetFlow such as + jFlow <http://www.net-track.ch/opensource/jflow/> and + Arbor Networks' Peakflow + <http://www.arbor.net/products_platform.php>. + + The Cooperative Association for Internet Data Analysis (CAIDA) + maintains a taxonomy of available tools on its web site at + <http://www.caida.org/tools/taxonomy/index.xml>. + +5.3. Internet Motion Sensor (IMS) + + The Internet Motion Sensor (IMS) [IMS] may be used to watch traffic + to or from "Darknets" (routable prefixes that don't have end hosts + attached), unassigned address spaces, and unannounced address spaces. + By watching activities in these types of address spaces, one can + understand and detect, e.g., scanning activities, DDoS worms, worm + infected hosts, and misconfigured hosts. + + + + +Andersson, et al. Informational [Page 21] + +RFC 4948 Unwanted Traffic August 2007 + + + Currently, the IMS is used to monitor approximately 17 million + prefixes, about 1.2% of the IPv4 address space. The use of IMS has + highlighted two major characteristics of attacks; malicious attacks + are more targeted than one might have assumed, and a vulnerability in + a system does not necessarily lead to a threat to that system (e.g., + the vulnerability may not be exploited to launch attacks if the + perceived "benefit" to the attacker appears small). Data from IMS + and other sources indicates that attackers are making increased use + of information from social networking sites to target their attacks + and select perceived easy targets, such as computers running very old + versions of systems or new, unpatched vulnerabilities. + + This form of passive data collection is also known as a "Network + Telescope". Links to similar tools can be found on the CAIDA web + site at <http://www.caida.org/data/passive/network_telescope.xml>. + +5.4. BCP 38 + + In the year 2000, the IETF developed a set of recommendations to + limit DOS attacks and Address Spoofing published as BCP 38 [RFC2827], + "Network Ingress Filtering: Defeating Denial of Service Attacks which + employ IP Source Address Spoofing". However, up to now BCP 38 + capabilities still have not been widely deployed, perhaps due to the + incentive issue discussed earlier. + + The IETF has also developed an additional set of recommendations + extending BCP 38 to multihomed networks. These recommendations are + published as BCP 84 [RFC3704]. + +5.5. Layer 5 to 7 Awareness + + Tools are being developed that will make it possible to perform deep + packet inspection at high speed. Some companies are working on + hardware implementation to inspect all layers from 2 to 7 (e.g., + EZchip <http://www.ezchip.com/t_npu_whpaper.htm>). A number of other + companies, including Cisco and Juniper, offer tools capable of + analyzing packets at the transport layer and above. + +5.6. How To's + + One idea that was discussed at the workshop envisaged operators and + standards bodies cooperating to produce a set of "How To" documents + as guidelines on how to configure networks. Dissemination and use of + these "How To's" should be encouraged by vendors, operators, and + standards bodies. + + + + + + +Andersson, et al. Informational [Page 22] + +RFC 4948 Unwanted Traffic August 2007 + + + This type of initiative needs a "sponsor" or "champion" that takes + the lead and starts collecting a set of "How To's" that could be + freely distributed. The workshop did not discuss this further. + +5.7. SHRED + + Methods to discourage the dissemination of spam by punishing the + spammers, such as Spam Harassment Reduction via Economic Disincentive + (SHRED) [SHRED], were discussed. The idea is to make it increasingly + expensive for spammers to use the email system, while normal users + retain what they have come to expect as normal service. There was no + agreement on the effectiveness of this type of system. + +6. Research in Progress + + In preparation for this session, several researchers active in + Internet Research were asked two rather open ended questions: "Where + is the focus on Internet research today?" and "Where should it be?" + + A summary of the answers to these questions is given below. + Section 6.2.2 covers part of the relationship between research and + miscreants. For example, research activities in each area (please + refer to the slide set for Workshop Session 8 which can be found at + the link referred to in Appendix C). + +6.1. Ongoing Research + + Section 6.1 discusses briefly areas where we see active research on + unwanted traffic today. + +6.1.1. Exploited Hosts + + One area where researchers are very active is analyzing situations + where hosts are exploited. This has been a major focus for a long + time, and an abundance of reports have been published. Current + research may be divided into three different categories: prevention, + detection, and defense. + +6.1.1.1. Prevention + + Code quality is crucial when it comes to preventing exploitation of + Internet hosts. Quite a bit of research effort has therefore gone + into improvement of code quality. Researchers are looking into + automated methods for finding bugs and maybe in the end fixes for any + bugs detected. + + A second approach designed to stop hosts from becoming compromised is + to reduce the "attack surface". Researchers are thinking about + + + +Andersson, et al. Informational [Page 23] + +RFC 4948 Unwanted Traffic August 2007 + + + changes or extensions to the Internet architecture. The idea is to + create a strict client server architecture, where the clients only + are allowed to initiate connections, and while servers may only + accept connections. + + Researchers have put a lot of effort into better scaling of honey + pots and honey farms to better understand and neutralize the methods + miscreants are using to exploit hosts. Research also goes into + developing honey monkeys in order to understand how hosts are + vulnerable. Both honey pots/farms and honey monkeys are aimed at + taking measures that prevent further (mis-)use of possible exploits. + +6.1.1.2. Detection + + When an attack is launched against a computer system, the attack + typically leaves evidence of the intrusion in the system logs. Each + type of intrusion leaves a specific kind of footprint or signature. + The signature can be evidence that certain software has been + executed, that logins have failed, that administrative privileges + have been misused, or that particular files and directories have been + accessed. Administrators can document these attack signatures and + use them to detect the same type of attack in the future. This + process can be automated. + + Because each signature is different, it is possible for system + administrators to determine by looking at the intrusion signature + what the intrusion was, how and when it was perpetrated, and even how + skilled the intruder is. + + Once an attack signature is available, it can be used to create a + vulnerability filter, i.e., the stored attack signature is compared + to actual events in real time and an alarm is given when this pattern + is repeated. + + A further step may be taken with automated vulnerability signatures, + i.e., when a new type of attack is found, a vulnerability filter is + automatically created. This vulnerability filter can be made + available for nodes to defend themselves against this new type of + attack. The automated vulnerability signatures may be part of an + Intrusion Detection System (IDS). + +6.1.1.3. Defense + + An IDS can be a part of the defense against actual attacks, e.g., by + using vulnerability filters. An Intrusion Detection System (IDS) + inspects inbound and outbound network activities and detects + signatures that indicate that a system is under attack from someone + attempting to break into or compromise the system. + + + +Andersson, et al. Informational [Page 24] + +RFC 4948 Unwanted Traffic August 2007 + + +6.1.2. Distributed Denial of Service (DDoS) Attacks + + Research on DDoS attacks follows two separate approaches, the first + has the application as its focus, while the second focuses on the + network. + +6.1.2.1. Application Oriented DDoS Research + + The key issue with application oriented research is to distinguish + between legitimate activities and attacks. Today, several tools + exist that can do this and research has moved on to more advanced + things. + + Research today looks into tools that can detect and filter activities + that have been generated by bots and botnets. + + One approach is to set up a tool that sends challenges to senders + that want to send traffic to a certain node. The potential sender + then has to respond correctly to that challenge; otherwise, the + traffic will be filtered out. + + The alternative is to get more capacity between sender and receiver. + This is done primarily by some form of use of peer-to-peer + technology. + + Today, there is "peer-to-peer hype" in the research community; a sure + way of making yourself known as a researcher is to publish something + that solves old problems by means of some peer-to-peer technology. + Proposals now exist for peer-to-peer DNS, peer-to-peer backup + solutions, peer-to-peer web-cast, etc. Whether these proposals can + live up to the hype remains to be seen. + +6.1.2.2. Network Oriented DDoS Research + + Research on DDoS attacks that takes a network oriented focus may be + described by the following oversimplified three steps. + + 1. Find the bad stuff + + 2. Set the "evil bit" on those packets + + 3. Filter out the packets with the "evil bit" set + + This rather uncomplicated scheme has to be carried out on high-speed + links and interfaces. Automation is the only way of achieving this. + + One way of indirectly setting the "evil bit" is to use a normalized + TTL. The logic goes: the TTL for traffic from this sender has always + + + +Andersson, et al. Informational [Page 25] + +RFC 4948 Unwanted Traffic August 2007 + + + been "x", but has now suddenly become "y", without any corresponding + change in routing. The conclusion is that someone is masquerading as + the legitimate sender. Traffic with the "y" TTL is filtered out. + + Another idea is to give traffic received from ISPs that are known to + do source address validation the "red carpet treatment", i.e., to set + the "good bit". When an attack is detected, traffic from everyone + that doesn't have the "good bit" is filtered out. Apart from + reacting to the attack, this also give ISPs an incentive to do source + address validation. If they don't do it, their peers won't set the + "good bit" and the ISP's customers will suffer, dragging down their + reputation. + + Overlay networks can also be used to stop a DDoS attack. The idea + here is that traffic is not routed directly to the destination. + Instead, it is hidden behind some entry points in the overlay. The + entry points make sure the sender is the host he claims he is, and in + that case, marks the packet with a "magic bit". Packets lacking the + "magic bit" are not forwarded on the overlay. This has good scaling + properties; you only need to have enough capacity to tag the amount + of traffic you want to receive, not the amount you actually receive. + +6.1.3. Spyware + + Current research on spyware and measurements of spyware are aiming to + find methods to understand when certain activities associated with + spyware happen and to understand the impact of this activity. + + There are a number of research activities around spyware, e.g., + looking into threats caused by spyware; however, these were only + briefly touched upon at the workshop. + +6.1.4. Forensic Aids + + Lately, research has started to look into tools and support to answer + the "What happened here?" question. These tools are called "forensic + aids", and can be used to "recreate" an illegal activity just as the + police do when working on a crime scene. + + The techniques that these forensic aids take as their starting point + involve the identification of a process or program that should not be + present on a computer. The effort goes into building tools and + methods that can trace the intruder back to its origin. Methods to + understand how a specific output depends on a particular input also + exist. + + + + + + +Andersson, et al. Informational [Page 26] + +RFC 4948 Unwanted Traffic August 2007 + + +6.1.5. Measurements + + Measurements are always interesting for the research community, + because they generate new data. Consequently, lots of effort goes + into specifying how measurements should be performed and into + development of measurement tools. Measurements have been useful in + creating effective counter-measures against worms. Before + measurements gave actual data of how worms behave, actions taken + against worms were generally ineffective. + +6.1.6. Traffic Analysis + + One aspect of research that closely relates to measurements is + analysis. Earlier, it was common to look for the amount of traffic + traversing certain transport ports. Lately, it has become common to + tunnel "everything" over something else, and a shift has occurred + towards looking for behavior and/or content. When you see a certain + behavior or content over a protocol that is not supposed to behave in + this way, it is likely that something bad is going on. + + Since this is an arms race, the miscreants that use tunneling + protocols have started to mimic the pattern of something that is + acceptable. + +6.1.7. Protocol and Software Security + + The general IETF design guidelines for robust Internet protocols + says: "Be liberal in what you receive and conservative in what you + send". The downside is that most protocols believe what they get and + as a consequence also get what they deserve. The IAB is intending to + work on new design guidelines, e.g., rules of thumb and things you do + and things you don't. This is not ready yet, but will be offered as + input to a BCP in due course. + + An area where there is a potential overlap between standards people + and researchers is protocol analysis languages. The protocol + analysis languages could be used, for example, look for + vulnerabilities. + +6.2. Research on the Internet + + The workshop discussed the interface between people working in + standardization organizations in general and IETF in particular on + the one hand and people working with research on the other. The + topic of discussion was broader than just "Unwanted traffic". Three + topics were touched on: what motivates researchers, how to attract + researchers to problems that are hindering or have been discovered in + + + + +Andersson, et al. Informational [Page 27] + +RFC 4948 Unwanted Traffic August 2007 + + + the context of standardization, and the sometimes rocky relations + between the research community and the "bad boys". + +6.2.1. Research and Standards + + The workshop discussed how research and standardization could + mutually support each other. Quite often there is a commonality of + interest between the two groups. The IAB supports the Internet + Research Task Force (IRTF) as a venue for Internet research. The + delta between what is done and what could be is still substantial. + The discussion focused on how standardization in general and the IETF + in particular can get help from researchers. + + Since standardization organizations don't have the economic strength + to simply finance the research they need or want, other means have to + be used. One is to correctly and clearly communicate problems, + another is to supply adequate and relevant information. + + To attract the research community to work with standardization + organizations, it is necessary to identify the real problems and + state them in such a way that they are amenable to solution. General + unspecified problems are of no use, e.g., "This is an impossible + problem!" or "All the problems are because my users behave badly!" + + Instead, saying "This is an absolutely critical problem, and we have + no idea how to solve it!" is much more attractive. + + The potential research problem should also be communicated in a way + that is public. A researcher that wants to take on a problem is + helped if she/he can point at a slide from NANOG or RIPE that + identifies this problem. + + The way researchers go about solving problems is basically to + identify all the existing constraints, and then relax one of the + constraints and see what happens. Therefore, rock solid constraints + are a show stopper, e.g., "We can't do that, because it has to go + into an ASIC!". Real constraints have to be clearly communicated to + and understood by the researcher. + + One reasonable way of fostering cooperation is to entice two or three + people and have them write a paper on the problem. What will happen + then is that this paper will be incrementally improved by other + researchers. The vast majority of all research goes into improving + on someone else's paper. + + A second important factor is to supply sufficient relevant + information. New information that suggests possible ways to address + new problems or improve on old or partial solutions to previously + + + +Andersson, et al. Informational [Page 28] + +RFC 4948 Unwanted Traffic August 2007 + + + investigated problems are attractive. Often, understanding of + important problems comes from the operator community; when trying to + initiate research from a standards perspective, keeping operators in + the loop may be beneficial. + + Today, the research community is largely left on its own, and + consequently tends to generate essentially random, untargeted + results. If the right people in the standards community say the + right things to the right people in the research community, it can + literally focus hundreds of graduate students on a single problem. + Problem statements and data are needed. + +6.2.2. Research and the Bad Guys + + A general problem with all research and development is that what can + be used may also be misused. In some cases, miscreants have received + help from research that was never intended. + + There are several examples of Free Nets, i.e., networks designed to + allow end-users to participate without revealing their identity or + how and where they are connected to the network. The Free Nets are + designed based on technologies such as onion routing or mix networks. + Free Nets create anonymity that allows people to express opinions + without having to reveal their true identity and thus can be used to + promote free speech. However, these are tools that can also work + just as well to hide illegal activities in democracies. + + Mix networks create hard-to-trace communications by using a chain of + proxy servers. A message from a sender to a receiver passes by the + chain of proxies. A message is encrypted with a layered encryption + where each layer is understood by only one of the proxies in the + chain; the actual message is the innermost layer. A mix network will + achieve untraceable communication, even if all but one of the proxies + are compromised by a potential tracer. + + Onion routing is a technique for anonymous communication over a + computer network; it is a technique that encodes routing information + in a set of encrypted layers. Onion routing is a further development + of mix networks. + + Research projects have resulted in methods for distributed command + and control, e.g., in the form of Distributed Hash Tables (DHT) and + gossip protocols. This of course has legitimate uses, e.g., for + security and reliability applications, but it also is extremely + useful for DDoS attacks and unwanted traffic in general. + + A lot of effort has gone into research around worms, the result is + that we have a very good understanding of the characteristics of the + + + +Andersson, et al. Informational [Page 29] + +RFC 4948 Unwanted Traffic August 2007 + + + technology associated with worms and how they behave. This is a very + good basis when we want to protect against worms. The downside is + that researchers also understand how to implement future worms, + including knowledge on how to design faster worms that won't leave a + footprint. + +7. Aladdin's Lamp + + If we had an Aladdin's Lamp and could be granted anything we wanted + in the context of remedying unwanted traffic or effects of such + traffic - what would we wish for? The topic of this session was + wishes, i.e., loosening the constraints that depend on what we have + and focus on what we really want. + + There certainly are lots of "wishes" around, not least of which is + making things simpler and safer. On the other hand, very few of + these wishes are clearly stated. One comment on this lack of clarity + was that we are too busy putting out the fires of today and don't + have the time to be thinking ahead. + +7.1. Security Improvements + + Operators at the workshop expressed a number of wishes that, if + fulfilled, would help to improve and simplify security. The list + below contains a number of examples of actions that ought to improve + security. The content is still at the "wish-level", i.e., no effort + has gone in to trying to understand the feasibility of realizing + these wishes. + + Wish: Reliable point of contact in each administrative domain for + security coordination. + First and foremost, operators would like to see correct and complete + contact information to coordinate security problems across operators. + + The "whois" database of registration details for IP addresses and + Autonomous System numbers held by Regional Internet Registries (e.g., + ARIN, RIPE, APNIC) was intended to be a directory for this type of + information, and RFC 2142 [RFC2142] established common mailbox names + for certain roles and services. There are several reasons why these + tools are largely unused, including unwanted traffic. + + Wish: Organized testing for security. + Today, new hardware and software are extensively tested for + performance. There is almost no testing of this hardware and + software for security. + + + + + + +Andersson, et al. Informational [Page 30] + +RFC 4948 Unwanted Traffic August 2007 + + + Wish: Infrastructure or test bed for security. + It would be good to have an organized infrastructure or test bed for + testing of security for new products. + + Wish: Defaults for security. + Equipment and software should come with a simple and effective + default setting for security. + + Wish: Shared information regarding attacks. + It would be useful to have an automated sharing mechanism for + attacks, vulnerabilities, and sources of threats between network + users and providers in order to meet attacks in a more timely and + efficient manner. + +7.2. Unwanted Traffic + + Wish: Automatic filtering of unwanted traffic. + It would be useful, not least for enterprises, to have mechanisms + that would automatically filter out the unwanted traffic. + + Some filtering of spam, viruses, and malware that is sent by email is + already practicable but inevitably is imperfect because it mainly + relies on "heuristics" to identify the unwanted traffic. This is + another example of the "arms race" between filtering and the + ingenuity of spammers trying to evade the filters. This "wish" needs + to be further discussed and developed to make it something that could + be turned into practical ideas. + + Wish: Fix Spam. + A large fraction of the email traffic coming into enterprises today + is spam, and consequently any fixes to the spam problem are very high + on their priority list. + +8. Workshop Summary + + The workshop spent its last two hours discussing the following + question: What are the engineering (immediate and longer term) and + research issues that might be pursued within the IETF and the IRTF, + and what actions could the IAB take? The suggested actions can be + summarized into three classes. + +8.1. Hard Questions + + The discussions during this concluding section raised a number of + questions that touched upon the overall network architecture designs. + + o What should be the roles of cryptographic mechanisms in the + overall Internet architecture? For example, do we need to apply + + + +Andersson, et al. Informational [Page 31] + +RFC 4948 Unwanted Traffic August 2007 + + + cryptographic mechanisms to harden the shell, or rely on deep + packet inspection to filter out bad traffic? + + o To add effective protection to the Internet, how far are we + willing to go in + + * curtailing its openness, and + + * increasing the system complexity? + + And what architectural principles do we need to preserve as we go + along these paths? + + o A simple risk analysis would suggest that an ideal attack target + of minimal cost but maximal disruption is the core routing + infrastructure. However, do we really need an unlinked and + separately managed control plane to secure it? This requires a + deep understanding of the architectural design trade-offs. + + o Can we, and how do we, change the economic substructure? A + special workshop was suggested as a next step to gain a better + understanding of the question. + +8.2. Medium or Long Term Steps + + While answering the above hard questions may take some time and + effort, several specific steps were suggested as medium or long term + efforts to add protection to the Internet: + + o Tightening the security of the core routing infrastructure. + + o Cleaning up the Internet Routing Registry repository [IRR], and + securing both the database and the access, so that it can be used + for routing verifications. + + o Take down botnets. + + o Although we do not have a magic wand to wave all the unwanted + traffic off the Internet, we should be able to develop effective + measures to reduce the unwanted traffic to a tiny fraction of its + current volume and keep it under control. + + o Community education, to try to ensure people *use* updated host, + router, and ingress filtering BCPs. + + + + + + + +Andersson, et al. Informational [Page 32] + +RFC 4948 Unwanted Traffic August 2007 + + +8.3. Immediately Actionable Steps + + The IETF is recommended to take steps to carry out the following + actions towards enhancing the network protection. + + o Update the host requirements RFC. The Internet host requirements + ([RFC1122], [RFC1123]) were developed in 1989. The Internet has + gone through fundamental changes since then, including the + pervasive security threats. Thus, a new set of requirements is + overdue. + + o Update the router requirements. The original router requirements + [RFC1812] were developed in 1995. As with the host requirements, + it is also overdue for an update. + + o Update ingress filtering (BCP 38 [RFC2827] and BCP 84 [RFC3704]). + + One immediate action that the IAB should carry out is to inform the + community about the existence of the underground economy. + + The IRTF is recommended to take further steps toward understanding + the Underground Economy and to initiate research on developing + effective countermeasures. + + Overall, the workshop attendees wish to raise the community's + awareness of the underground economy. The community as a whole + should undertake a systematic examination of the current situation + and develop both near- and long-term plans. + +9. Terminology + + This section gives an overview of some of the key concepts and + terminology used in this document. It is not intended to be + complete, but is offered as a quick reference for the reader of the + report. + + ACL + Access Control List in the context of Internet networking refers to a + set of IP addresses or routing prefixes (layer 3 or Internet layer + information), possibly combined with transport protocol port numbers + (layer 4 or transport layer information). The layer 3 and/or layer 4 + information in the packets making up a flow entering or leaving a + device in the Internet is matched against the entries in an ACL to + determine whether the packets should, for example, be allowed or + denied access to some resources. The ACL effectively specifies a + filter to be used on a flow of packets. + + + + + +Andersson, et al. Informational [Page 33] + +RFC 4948 Unwanted Traffic August 2007 + + + BGP route hijacking + Attack in which an inappropriate route is injected into the global + routing system with the intent of diverting traffic from its intended + recipient either as a DoS attack (q.v.) where the traffic is just + dropped or as part of some wider attack on the recipient. Injecting + spurious routes specifying addresses used for bogons can, for + example, provide bogus assurance to email systems that spam is coming + from legitimate addresses. + + Bogon + A bogon is an IP packet that has a source address taken for a range + of addresses that has not yet been allocated to legitimate users, or + is a private [RFC1918] or reserved address [RFC3330]. + + Bogon prefix + A bogon prefix is a route that should never appear in the Internet + routing table, e.g., from the private or unallocated address blocks. + + Bot + A bot is common parlance on the Internet for a software program that + is a software agent. A Bot interacts with other network services + intended for people as if it were a real person. One typical use of + bots is to gather information. The term is derived from the word + "robot," reflecting the autonomous character in the "virtual robot"- + ness of the concept. + The most common bots are those that covertly install themselves on + people's computers for malicious purposes, and that have been + described as remote attack tools. Bots are sometimes called + "zombies". + + Botnet + Botnet is a jargon term for a collection of software robots, or bots, + which run autonomously. This can also refer to the network of + computers using distributed computing software. While the term + "botnet" can be used to refer to any group of bots, such as IRC bots, + the word is generally used to refer to a collection of compromised + machines running programs, usually referred to as worms, Trojan + horses, or backdoors, under a common command and control + infrastructure. + + Click fraud + Click fraud occurs in pay per click (PPC) advertising when a person, + automated script, or computer program imitates a legitimate user of a + web browser clicking on an ad for the purpose of generating an + improper charge per click. Pay per click advertising is when + operators of web sites act as publishers and offer clickable links + from advertisers in exchange for a charge per click. + + + + +Andersson, et al. Informational [Page 34] + +RFC 4948 Unwanted Traffic August 2007 + + + Darknet + A Darknet (also known as a Network Telescope, a Blackhole, or an + Internet Sink) is a globally routed network that has no "real" + machines attached and carries only a very small amount of specially + crafted legitimate traffic. It is therefore easily possible to + separate out and analyze unwanted traffic that can arise from a wide + variety of events including misconfiguration (e.g., a human being + mis-typing an IP address), malicious scanning of address space by + hackers looking for vulnerable targets, backscatter from random + source denial-of-service attacks, and the automated spread of + malicious software called Internet worms. + + Dirty affiliate program + Affiliate programs are distributed marketing programs that recruit + agents to promote a product or service. Affiliates get financially + compensated for each sale associated with their unique 'affiliate + ID.' Affiliates are normally instructed by the operator of the + affiliate program to not break any laws while promoting the product + or service. Sanctions (typically loss of unpaid commissions or + removal from the affiliate program) are normally applied if the + affiliate spams or otherwise violates the affiliate program's + policies. + + Dirty affiliate programs allow spamming, or if they do nominally + prohibit spamming, they don't actually sanction violators. Dirty + affiliate programs often promote illegal or deceptive products + (prescription drugs distributed without regard to normal dispensing + requirements, body part enlargement products, etc.), employ anonymous + or untraceable affiliates, offer payment via anonymous online + financial channels, and may fail to follow normal tax withholding and + reporting practices. + + DoS attack + Denial-Of-Service attack, a type of attack on a network that is + designed to bring the network to its knees by flooding it with + useless traffic or otherwise blocking resources necessary to allow + normal traffic flow. + + DDoS attack + Distributed Denial of Service, an attack where multiple compromised + systems are used to target a single system causing a Denial of + Service (DoS) attack. + + Honey farm + A honey farm is a set of honey pots working together. + + + + + + +Andersson, et al. Informational [Page 35] + +RFC 4948 Unwanted Traffic August 2007 + + + Honey monkey + A honey monkey is a honey pot in reverse; instead of sitting and + waiting for miscreants, a honey monkey actively mimics the actions of + a user surfing the Web. The honey monkey runs on virtual machines in + order to detect exploit sites. + + Honey pot + A honey pot is a server attached to the Internet that acts as a + decoy, attracting potential miscreants in order to study their + activities and monitor how they are able to break into a system. + Honeypots are designed to mimic systems that an intruder would like + to break into but limit the intruder from having access to an entire + network. + + IRC + Internet Relay Chat is a form of instant communication over the + Internet. It is mainly designed for group (many-to-many) + communication in discussion forums called channels, but also allows + one-to-one communication, originally standardized by RFC 1459 + [RFC1459] but much improved and extended since its original + invention. IRC clients rendezvous and exchange messages through IRC + servers. IRC servers are run by many organizations for both benign + and nefarious purposes. + + Malware + Malware is software designed to infiltrate or damage a computer + system, without the owner's informed consent. There are + disagreements about the etymology of the term itself, the primary + uncertainty being whether it is a portmanteau word (of "malicious" + and "software") or simply composed of the prefix "mal-" and the + morpheme "ware". Malware references the intent of the creator, + rather than any particular features. It includes computer viruses, + worms, Trojan horses, spyware, adware, and other malicious and + unwanted software. In law, malware is sometimes known as a computer + contaminant. + + Mix networks + Mix networks create hard-to-trace communications by using a chain of + proxy servers [MIX]. Each message is encrypted to each proxy; the + resulting encryption is layered like a Russian doll with the message + as the innermost layer. Even if all but one of the proxies are + compromised by a tracer, untraceability is still achieved. More + information can be found at + <http://www.adastral.ucl.ac.uk/~helger/crypto/link/protocols/ + mix.php>. + + + + + + +Andersson, et al. Informational [Page 36] + +RFC 4948 Unwanted Traffic August 2007 + + + Onion routing + Onion routing is a technique for anonymous communication over a + computer network, it is a technique that encodes routing information + in a set of encrypted layers. Onion routing is based on mix cascades + (see mix networks (q.v.)). More information can be found at + <http://www.onion-router.net/>. + + Phishing + Phishing is a form of criminal activity using social engineering + techniques. It is characterized by attempts to fraudulently acquire + sensitive information, such as passwords and credit card details, by + masquerading as a trustworthy person or business in an apparently + official electronic communication. Phishing is typically carried out + using spoofed websites, email, or an instant message. The term + phishing derives from password harvesting and the use of increasingly + sophisticated lures to "fish" for users' financial information and + passwords. + + Root access + Access to a system with full administrative privileges bypassing any + security restrictions placed on normal users. Derived from the name + traditionally used for the 'superuser' on Unix systems. + + Script kiddy + Derogatory term for an inexperienced hacker who mindlessly uses + scripts and other programs developed by others with the intent of + compromising computers or generating DoS attacks. + + Spam + Spamming is the abuse of electronic messaging systems to send + unsolicited, undesired bulk messages. The individual messages are + refereed to as spam. The term is frequently used to refer + specifically to the electronic mail form of spam. + + Spoofing + (IP) spoofing is a technique where the illegitimate source of IP + packets is obfuscated by contriving to use IP address(es) that the + receiver recognizes as a legitimate source. Spoofing is often used + to gain unauthorized access to computers or mislead filtering + mechanisms, whereby the intruder sends packets into the network with + an IP source address indicating that the message is coming from a + legitimate host. To engage in IP spoofing, a hacker must first use a + variety of techniques to find an IP address of a valid host and then + modify the packet headers so that it appears that the packets are + coming from that host. + + + + + + +Andersson, et al. Informational [Page 37] + +RFC 4948 Unwanted Traffic August 2007 + + + Spyware + Any software that covertly gathers user information through the + user's Internet connection without his or her knowledge, e.g., for + spam purposes. + + UBE + Unsolicited Bulk Email: an official term for spam. + + UCE + Unsolicited Commercial Email: an official term for spam. + + Virus + A program or piece of code that is loaded onto a computer without the + owner's knowledge and runs without their consent. A virus is self- + replicating code that spreads by inserting copies of itself into + other executable code or documents, which are then transferred to + other machines. Typically, the virus has a payload that causes some + harm to the infected machine when the virus code is executed. + + Worm + A computer worm is a self-replicating computer program. It uses a + network to send copies of itself to other systems and it may do so + without any user intervention. Unlike a virus, it does not need to + attach itself to an existing program. Worms always harm the network + (if only by consuming bandwidth), whereas viruses always infect or + corrupt files on a targeted computer. + + Zombie + This is another name for a bot. + +10. Security Considerations + + This document does not specify any protocol or "bits on the wire". + +11. Acknowledgements + + The IAB would like to thank the University of Southern California + Information Sciences Institute (ISI) who hosted the workshop and all + those people at ISI and elsewhere who assisted with the organization + and logistics of the workshop at ISI. + + The IAB would also like to thank the scribes listed in Appendix A who + diligently recorded the proceedings during the workshop. + + A special thanks to all the participants in the workshop, who took + the time, came to the workshop to participate in the discussions, and + who put in the effort to make this workshop a success. The IAB + + + + +Andersson, et al. Informational [Page 38] + +RFC 4948 Unwanted Traffic August 2007 + + + especially appreciates the effort of those that prepared and made + presentations at the workshop. + +12. Informative References + + [IMS] University of Michigan, "Internet Motion Sensor", 2006, + <http://ims.eecs.umich.edu/>. + + [IRR] Merit Network Inc, "Internet Routing Registry Routing + Assets Database", 2006, <http://www.irr.net/>. + + [MIX] Hill, R., Hwang, A., and D. Molnar, "Approaches to Mix + Nets", MIT 6.857 Final Project, December 1999, <http:// + www.mit.edu/afs/athena/course/6/6.857/OldStuff/Fall99/ + papers/mixnet.ps.gz>. + + [RFC1122] Braden, R., "Requirements for Internet Hosts - + Communication Layers", STD 3, RFC 1122, October 1989. + + [RFC1123] Braden, R., "Requirements for Internet Hosts - Application + and Support", STD 3, RFC 1123, October 1989. + + [RFC1459] Oikarinen, J. and D. Reed, "Internet Relay Chat Protocol", + RFC 1459, May 1993. + + [RFC1812] Baker, F., "Requirements for IP Version 4 Routers", + RFC 1812, June 1995. + + [RFC1918] Rekhter, Y., Moskowitz, R., Karrenberg, D., Groot, G., and + E. Lear, "Address Allocation for Private Internets", + BCP 5, RFC 1918, February 1996. + + [RFC2142] Crocker, D., "MAILBOX NAMES FOR COMMON SERVICES, ROLES AND + FUNCTIONS", RFC 2142, May 1997. + + [RFC2827] Ferguson, P. and D. Senie, "Network Ingress Filtering: + Defeating Denial of Service Attacks which employ IP Source + Address Spoofing", BCP 38, RFC 2827, May 2000. + + [RFC3330] IANA, "Special-Use IPv4 Addresses", RFC 3330, + September 2002. + + [RFC3704] Baker, F. and P. Savola, "Ingress Filtering for Multihomed + Networks", BCP 84, RFC 3704, March 2004. + + [SHRED] Krishnamurthy, B. and E. Blackmond, "SHRED: Spam + Harassment Reduction via Economic Disincentives", 2003, + <http://www.research.att.com/~bala/papers/shred-ext.ps>. + + + +Andersson, et al. Informational [Page 39] + +RFC 4948 Unwanted Traffic August 2007 + + +Appendix A. Participants in the Workshop + + Bernard Aboba (IAB) + Loa Andersson (IAB) + Ganesha Bhaskara (scribe) + Bryan Burns + Leslie Daigle (IAB chair) + Sean Donelan + Rich Draves (IAB Executive Director) + Aaron Falk (IAB, IRTF chair) + Robert Geigle + Minas Gjoka (scribe) + Barry Greene + Sam Hartman (IESG, Security Area Director) + Bob Hinden (IAB) + Russ Housely (IESG, Security Area Director) + Craig Huegen + Cullen Jennings + Rodney Joffe + Mark Kosters + Bala Krishnamurthy + Gregory Lebovitz + Ryan McDowell + Danny McPherson + Dave Merrill + David Meyer (IAB) + Alan Mitchell + John Morris + Eric Osterweil (scribe) + Eric Rescorla (IAB) + Pete Resnick (IAB) + Stefan Savage + Joe St Sauver + Michael Sirivianos (scribe) + Rob Thomas + Helen Wang + Lixia Zhang (IAB) + + + + + + + + + + + + + + +Andersson, et al. Informational [Page 40] + +RFC 4948 Unwanted Traffic August 2007 + + +Appendix B. Workshop Agenda + + Session 1: + How bad is the problem? What are the most important symptoms? + + Session 2: + What are the sources of the problem? + + Lunch session (session 3): + Solutions in regulatory and societal space + + Session 4: + The underground economy + + Session 5: + Current countermeasures, what works, what doesn't + + Session 6: + If all our wishes could be granted, what would they be? + + Session 7: + What's in the pipeline, or should be in the pipeline + + Session 8: + What is being actively researched on? + + Session 9: + What are the engineering (immediate and longer term) and + research issues that might be pursued within the IETF/IAB/IRTF? + +Appendix C. Slides + + Links to a subset of the presentations given by the participants at + the workshop can be found via the IAB Workshops page on the IAB web + site at <http://utgard.ietf.org/iab/about/workshops/unwantedtraffic/ + index.html>. As mentioned in Section 1, this is not a complete set + of the presentations because certain of the presentations were of a + sensitive nature which it would be inappropriate to make public at + this time. + + + + + + + + + + + + +Andersson, et al. Informational [Page 41] + +RFC 4948 Unwanted Traffic August 2007 + + +Authors' Addresses + + Loa Andersson + Acreo AB + + EMail: loa@pi.se + + + Elwyn Davies + Folly Consulting + + EMail: elwynd@dial.pipex.com + + + Lixia Zhang + UCLA + + EMail: lixia@cs.ucla.edu + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +Andersson, et al. Informational [Page 42] + +RFC 4948 Unwanted Traffic August 2007 + + +Full Copyright Statement + + Copyright (C) The IETF Trust (2007). + + This document is subject to the rights, licenses and restrictions + contained in BCP 78, and except as set forth therein, the authors + retain all their rights. + + This document and the information contained herein are provided on an + "AS IS" basis and THE CONTRIBUTOR, THE ORGANIZATION HE/SHE REPRESENTS + OR IS SPONSORED BY (IF ANY), THE INTERNET SOCIETY, THE IETF TRUST AND + THE INTERNET ENGINEERING TASK FORCE DISCLAIM ALL WARRANTIES, EXPRESS + OR IMPLIED, INCLUDING BUT NOT LIMITED TO ANY WARRANTY THAT THE USE OF + THE INFORMATION HEREIN WILL NOT INFRINGE ANY RIGHTS OR ANY IMPLIED + WARRANTIES OF MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. + +Intellectual Property + + The IETF takes no position regarding the validity or scope of any + Intellectual Property Rights or other rights that might be claimed to + pertain to the implementation or use of the technology described in + this document or the extent to which any license under such rights + might or might not be available; nor does it represent that it has + made any independent effort to identify any such rights. Information + on the procedures with respect to rights in RFC documents can be + found in BCP 78 and BCP 79. + + Copies of IPR disclosures made to the IETF Secretariat and any + assurances of licenses to be made available, or the result of an + attempt made to obtain a general license or permission for the use of + such proprietary rights by implementers or users of this + specification can be obtained from the IETF on-line IPR repository at + http://www.ietf.org/ipr. + + The IETF invites any interested party to bring to its attention any + copyrights, patents or patent applications, or other proprietary + rights that may cover technology that may be required to implement + this standard. Please address the information to the IETF at + ietf-ipr@ietf.org. + +Acknowledgement + + Funding for the RFC Editor function is currently provided by the + Internet Society. + + + + + + + +Andersson, et al. Informational [Page 43] + |