Univ. of Karlsruhe
K. Wehrle
Univ. of Tuebingen
April 2004
IP Multicast in Differentiated Services (DS) Networks
Status of this Memo
-
This memo provides information for the Internet community. It does not specify an Internet standard of any kind. Distribution of this memo is unlimited.
Copyright Notice
-
Copyright © The Internet Society (2004). All Rights Reserved.
Abstract
-
This document discusses the problems of IP Multicast use in Differentiated Services (DS) networks, expanding on the discussion in RFC 2475 ("An Architecture of Differentiated Services"). It also suggests possible solutions to these problems, describes a potential implementation model, and presents simulation results.
Table of Contents
-
1. Introduction . . . . . . . . . . . . . . . . . . . . . . . . . 2 1.1. Management of Differentiated Services. . . . . . . . . . 2 2. Problems of IP Multicast in DS Domains . . . . . . . . . . . . 3 2.1. Neglected Reservation Subtree Problem (NRS Problem). . . 4 2.2. Heterogeneous Multicast Groups . . . . . . . . . . . . . 12 2.3. Dynamics of Any-Source Multicast . . . . . . . . . . . . 13 3. Solutions for Enabling IP-Multicast in Differentiated Services Networks. . . . . . . . . . . . . . . . . . . . . . . 13 3.1. Solution for the NRS Problem . . . . . . . . . . . . . . 13 3.2. Solution for Supporting Heterogeneous Multicast Groups . 15 3.3. Solution for Any-Source Multicast. . . . . . . . . . . . 16 4. Scalability Considerations . . . . . . . . . . . . . . . . . . 16 5. Deployment Considerations. . . . . . . . . . . . . . . . . . . 17 6. Security Considerations. . . . . . . . . . . . . . . . . . . . 18 7. Implementation Model Example . . . . . . . . . . . . . . . . . 18 8. Proof of the Neglected Reservation Subtree Problem . . . . . . 19 8.1. Implementation of the Proposed Solution. . . . . . . . . 20 8.2. Test Environment and Execution . . . . . . . . . . . . . 21 9. Simulative Study of the NRS Problem and Limited Effort PHB . . 23 9.1. Simulation Scenario. . . . . . . . . . . . . . . . . . . 24 9.2. Simulation Results for Different Router Types. . . . . . 26 10. Acknowledgements . . . . . . . . . . . . . . . . . . . . . . . 31 11. References . . . . . . . . . . . . . . . . . . . . . . . . . . 31 11.1. Normative References . . . . . . . . . . . . . . . . . . 31 11.2. Informative References . . . . . . . . . . . . . . . . . 31 12. Authors' Addresses . . . . . . . . . . . . . . . . . . . . . . 33 13. Full Copyright Statement . . . . . . . . . . . . . . . . . . . 34
1. Introduction
-
This document discusses the problems of IP Multicast use in Differentiated Services (DS) networks, expanding on the discussion in RFC 2475 ("An Architecture of Differentiated Services"). It also suggests possible solutions to these problems, describes a potential implementation model, and presents simulation results.
The "Differentiated Services" (DiffServ or DS) approach [1, 2, 3] defines certain building blocks and mechanisms to offer qualitatively better services than the traditional best-effort delivery service in an IP network. In the DiffServ Architecture [2], scalability is achieved by avoiding complexity and maintenance of per-flow state information in core nodes, and by pushing unavoidable complexity to the network edges. Therefore, individual flows belonging to the same service are aggregated, thereby eliminating the need for complex classification or managing state information per flow in interior nodes.
On the other hand, the reduced complexity in DS nodes makes it more complex to use those "better" services together with IP Multicast (i.e., point-to-multipoint or multipoint-to-multipoint communication). Problems emerging from this fact are described in section 2. Although the basic DS forwarding mechanisms also work with IP Multicast, some facts have to be considered which are related to the provisioning of multicast resources. It is important to integrate IP Multicast functionality into the architecture from the beginning, and to provide simple solutions for those problems that will not defeat the already gained advantages.
1.1. Management of Differentiated Services
-
At least for Per-Domain Behaviors and services based on the EF PHB, admission control and resource reservation are required [14, 15]. Installation and updating of traffic profiles in boundary nodes is necessary. Most network administrators cannot accomplish this task manually, even for long term service level agreements (SLAs). Furthermore, offering services on demand requires some kind of signaling and automatic admission control procedures.
However, no standardized resource management architecture for DiffServ domains exists. The remainder of this document assumes that at least some logical resource management entity is available to perform resource-based admission control and allotment functions. This entity may also be realized in a distributed fashion, e.g., within the routers themselves. Detailed aspects of the resource management realization within a DiffServ domain, as well as the interactions between resource management and routers or end-systems (e.g., signaling for resources), are out of scope of this document.
Protocols for signaling a reservation request to a Differentiated Services Domain are required. For accomplishing end-system signaling to DS domains, RSVP [4] may be used with new DS specific reservation objects [5]. RSVP provides support for multicast scenarios and is already supported by many systems. However, application of RSVP in a DiffServ multicast context may lead to problems that are also described in the next section. The NSIS Working Group is currently defining new signaling protocols that may show a different behavior, but the WG has its current focus more on unicast flows than on multicast flows.
2. Problems of IP Multicast in DS Domains
-
Although potential problems and the complexity of providing multicast with Differentiated Services are considered in a separate section of [2], both aspects have to be discussed in greater detail. The simplicity of the DiffServ Architecture and its DS node types is necessary to reach high scalability, but it also causes fundamental problems in conjunction with the use of IP Multicast in DS domains. The following subsections describe these problems for which a generic solution is proposed in section 3. This solution is as scalable as IP Multicast and needs no resource separation by using different codepoint values for unicast and multicast traffic.
Because Differentiated Services are unidirectional by definition, the point-to-multipoint communication is also considered as unidirectional. In traditional IP Multicast, any node can send packets spontaneously and asynchronously to a multicast group specified by their multicast group address, i.e., traditional IP Multicast offers a multipoint-to-multipoint service, also referred to as Any-Source Multicast. Implications of this feature are discussed in section 2.3.
For subsequent considerations we assume, unless stated otherwise, at least a unidirectional point-to-multipoint communication scenario in which the sender generates packets which experience a "better" Per- Hop-Behavior than the traditional default PHB, resulting in a service of better quality than the default best-effort service. In order to accomplish this, a traffic profile corresponding to the traffic conditioning specification has to be installed in the sender's first DS-capable boundary node. Furthermore, it must be assured that the corresponding resources are available on the path from the sender to all the receivers, possibly requiring adaptation of traffic profiles at involved domain boundaries. Moreover, on demand resource reservations may be receiver-initiated.
2.1. Neglected Reservation Subtree Problem (NRS Problem)
-
Typically, resources for Differentiated Services must be reserved before they are used. But in a multicast scenario, group membership is often highly dynamic, thereby limiting the use of a sender- initiated resource reservation in advance. Unfortunately, dynamic addition of new members of the multicast group using Differentiated Services can adversely affect other existing traffic if resources were not explicitly reserved before use. A practical proof of this problem is given in section 8.
IP Multicast packet replication usually takes place when the packet is handled by the forwarding core (cf. Fig. 1), i.e., when it is forwarded and replicated according to the multicast forwarding table. Thus, a DiffServ capable node would also copy the content of the DS field [1] into the IP packet header of every replicate. Consequently, replicated packets get exactly the same DS codepoint (DSCP) as the original packet, and therefore experience the same forwarding treatment as the incoming packets of this multicast group. This is also illustrated in Fig. 1, where each egress interface comprises functions for (BA-) classification, traffic conditioning (TC), and queueing.
Interface A IP Forwarding Interface B +-----------+ +--------------+ +-----------+ MC-flow | | | replication | | egress | ---->| ingress |---->|------+-------|----->|(class.,TC,|----> | | | | | | queueing) | +-----------+ | | | +-----------+ | | | | | | Interface C | | | +-----------+ | | | | egress | | +-------|----->|(class.,TC,|----> | | | queueing) | +--------------+ +-----------+
Figure 1: Multicast packet replication in a DS node
-
Normally, the replicating node cannot test whether a corresponding resource reservation exists for a particular flow of replicated packets on an output link (i.e., its corresponding interface). This is because flow-specific information (e.g., traffic profiles) is usually not available in every boundary and interior node.
When a new receiver joins an IP Multicast group, a multicast routing protocol (e.g., DVMRP [6], PIM-DM [7] or PIM-SM [8]) grafts a new branch to an existing multicast tree in order to connect the new receiver to the tree. As a result of tree expansion, missing per- flow classification, and policing mechanisms, the new receiver will implicitly use the service of better quality, because of the "better" copied DSCP.
If the additional amount of resources which are consumed by the new part of the multicast tree are not taken into account by the domain resource management (cf. section 1.1), the currently provided quality of service of other receivers (with correct reservations) will be affected adversely or even violated. This negative effect on existing traffic contracts by a neglected resource reservation -- in the following designated as the Neglected Reservation Subtree Problem (NRS Problem) -- must be avoided under all circumstances. Strong admission control policies at the domain boundary will not help to prevent this problem either, because the new flow that inadmissibly consumes resources has its origin inside the domain.
One can distinguish two major cases of the NRS Problem. They show a different behavior depending on the location of the branching point. In order to compare their different effects, a simplistic example of a share of bandwidth is illustrated in Fig. 2 and is used in the following explanations. Neither the specific PHB types nor their assigned bandwidth share are important; however, their relative priority with respect to each other is of importance.
40% 40% 20% +--------------------+---------------------+------------+ |Expedited Forwarding| Assured Forwarding | Best-Effort| +--------------------+---------------------+------------+ ----------------------------------------------------------> output link bandwidth
-
Figure 2:
-
-
-
An example bandwidth share of different behavior aggregates
-
-
The bandwidth of the considered output link is shared by three types of services (i.e., by three behavior aggregates): Expedited Forwarding, Assured Forwarding, and the traditional Best-Effort service. In this example, we assume that routers perform strict priority queueing, where EF has the highest, AF the middle, and Best-Effort the lowest assigned scheduling priority. Though not mandatory for an EF implementation, a strict non-preemptive priority scheduler is one implementation option as described in section 5.1.1 of RFC 3247 [15]. Were Weighted Fair Queueing (WFQ) to be used, the described effects would essentially also occur, but with minor differences. In the following scenarios, it is illustrated that PHBs of equal or lower priority (in comparison to the multicast flow's PHB) are affected by the NRS problem.
The Neglected Reservation Subtree problem appears in two different cases:
- Case 1: If the branching point of the new subtree (at first only a branch) and the previous multicast tree is a (egress) boundary node, as shown in Fig. 3, the additional multicast flow now increases the total amount of used resources for the corresponding behavior aggregate on the affected output link. The total amount will be greater than the originally reserved amount. Consequently, the policing component in the egress boundary node drops packets until the traffic aggregate is in accordance with the traffic contract. But while dropping packets, the router can not identify the responsible flow (because of missing flow classification functionality), and thus randomly discards packets, whether they belong to a correctly behaving flow or not. As a result, there will no longer be any service guarantee for the flows with properly reserved resources.
Sender +---+ | S | DS domains +---+ / \ .||............... / \ ................ . || .<- ->. . . || . . . . +---+ +--+ +--+ *) +--+ +--+ +--+ +------+ . |FHN|===|IN|=====|BN|###########|BN|####|IN|######|BN|####|Recv.B| . +---+ +--+ +--+\\ +--+ +--+ +--+ +------+ . \\ \ . \\ . \ . . +--+ +--+ . \\ . \ . . |IN|-----|IN| . \\ . +--+ . . +--+ +--+ . \\ ..........|BN|.. . || \ . +------+ +--+ . || \ . |Recv.A| .+--+ +--+. +------+ |BN|........|BN| +--+ +--+ ||
S: Sender
Recv.x: Receiver x
FHN: First-Hop Node
BN: Boundary Node
IN: Interior Node
===: Multicast branch with reserved bandwidth
###: Multicast branch without reservation
*) Bandwidth of EF aggregated on the output link is higher than-
actual reservation, EF aggregate will be limited in bandwidth without considering the responsible flow.
-
Figure 3:
-
-
The NRS Problem (case 1) occurs when Receiver B joins
-
In figure 3, it is assumed that receiver A is already attached to the egress boundary node (BN) of the first domain. Furthermore, resources are properly reserved along the path to receiver A and used by correspondingly marked packets. When receiver B joins the same group as receiver A, packets are replicated and forwarded along the new branch towards the second domain with the same PHB as for receiver A. If this PHB is EF, the new branch possibly exhausts allotted resources for the EF PHB, adversely affecting other EF users that receive their packets over the link that is marked with the *). The BN usually ensures that outgoing traffic aggregates to the next domain are conforming to the agreed traffic conditioning specification. The egress BN will, therefore, drop packets of the PHB type that are used for the multicast flow.
Other PHBs of lower or higher priority are not affected adversely in this case. The following example in Fig. 4. illustrates this for two PHBs.
-
+------------------+---------------------+--------------+------+ | Expedited Forw. | Expedited Forw. | Assured Forw.| BE | | | | | | | with reservation | excess flow | with reserv. | | | | without reservation | | | +------------------+---------------------+--------------+------+ | EF with and without reservation share | 40 % | 20% | | 40% of reserved EF aggregate. | | | | -> EF packets with reservation and | | | | without reservation will be | | | | discarded! | | | +------------------+---------------------+--------------+------+
-
-
-
(a) Excess flow has EF codepoint
-
-
+------------------+---------------------+--------------+------+ | Expedited Forw. | Assured Forwarding | Assured Forw.| BE | | | | | | | with reservation | excess flow | with reserv. | | | | without reservation | | | +------------------+---------------------+--------------+------+ | | AF with & without reservation share| 20 % | | | 40% of reserved EF aggregate. | | | 40% | -> EF packets with reservation and | | | | without reservation will be | | | | discarded! | | +------------------+---------------------+--------------+------+
-
-
-
(b) Excess flow has AF codepoint
-
Figure 4:
-
-
-
Resulting share of bandwidth in a egress boundary node with a neglected reservation of (a) an Expedited Forwarding flow or (b) an Assured Forwarding flow.
-
-
Fig. 4 shows the resulting share of bandwidth in cases when (a) Expedited Forwarding and (b) Assured Forwarding is used by the additional multicast branch causing the NRS Problem. Assuming that the additional traffic would use another 30% of the link bandwidth, Fig. 4 (a) illustrates that the resulting aggregate of Expedited Forwarding (70% of the outgoing link bandwidth) is throttled down to its originally reserved 40%. In this case, the amount of dropped EF bandwidth is equal to the amount of excess bandwidth. Consequently, the original Expedited Forwarding aggregate (which had 40% of the link bandwidth reserved) is also affected by packet losses. The other services, e.g., Assured Forwarding or Best-Effort, are not disadvantaged.
Fig. 4 (b) shows the same situation for Assured Forwarding. The only difference is that now Assured Forwarding is solely affected by discards, as the other services will still get their guarantees. In either case, packet losses are restricted to the misbehaving service class by the traffic meter and policing mechanisms in boundary nodes. Moreover, the latter problem (case 1) occurs only in egress boundary nodes because they are responsible for ensuring that the traffic leaving the Differentiated Services domain is not more than the following ingress boundary node will accept. Therefore, those violations of SLAs will already be detected and processed in egress boundary nodes.
-
- Case 2: The Neglected Reservation Subtree problem can also occur if the branching point between the previous multicast tree and the new subtree is located in an interior node (as shown in Fig. 5). In Fig. 5, it is assumed that receivers A and B have already joined the multicast group and have reserved resources accordingly. The interior node in the second domain starts replication of multicast packets as soon as receiver C joins. Because the router is not equipped with metering or policing functions, it will not recognize any amount of excess traffic and will forward the new multicast flow. If the latter belongs to a higher priority service, such as Expedited Forwarding, bandwidth of the aggregate is higher than the aggregate's reservation at the new branch and will use bandwidth from lower priority services.
Sender +---+ | S | DS domains +---+ / \ .||............... / \ ................ . || .<- ->. . . || . . . . +---+ +--+ +--+ +--+ +--+ +--+ +------+ . |FHN|===|IN|=====|BN|===========|BN|====|IN|======|BN|===|Recv.B| . +---+ +--+ +--+\\ +--+ +--+ +--+ +------+ . \\ \ . \\ . # . . +--+ +--+ . \\ . # *) . . |IN|-----|IN| . \\ . +--+ . . +--+ +--+ . \\ ..........|BN|.. . || \ . +------+ +--+ . || \ . |Recv.A| # .+--+ +--+. +------+ # |BN|........|BN| +------+ +--+ +--+ |Recv.C| || +------+
FHN: First-Hop Node, BN: Boundary Node, Recv.x: Receiver x S: Sender, IN: Interior Node
===: Multicast branch with reserved bandwidth
###: Multicast branch without reservation
*) Bandwidth of EF aggregated on the output link is higher than-
actual reservation, EF aggregate will be limited in bandwidth without considering the responsible flow
-
Figure 5:
-
-
Neglected Reservation Subtree problem case 2 after join of receiver C
-
The additional amount of EF without a corresponding reservation is forwarded together with the aggregate which has a reservation. This results in no packet losses for Expedited Forwarding as long as the resulting aggregate is not higher than the output link bandwidth. Because of its higher priority, Expedited Forwarding gets as much bandwidth as needed and as is available. The effects on other PHBs are illustrated by the following example in Fig. 6.
-
+------------------+---------------------+--------------+------+ | Expedited Forw. | Expedited Forw. | Assured Forw.| BE | | | | | | | with reservation | excess flow | with reserv. | | | | without reservation | | | +------------------+---------------------+--------------+------+ | 40% | 30% | 30% | 0% | +------------------+---------------------+--------------+------+ EF with reservation and the excess flow use together 70% of the link bandwidth because EF, with or without reservation, has the highest priority.
-
-
-
(a) Excess flow has EF codepoint
-
-
+------------------+---------------------+--------------+------+ | Expedited Forw. | Assured Forw. | Assured Forw.| BE | | | | | | | with reservation | excess flow | with reserv. | | | | without reservation | | | +------------------+---------------------+--------------+------+ | 40% | 60% | 0% | | | (10% loss) | | +------------------+---------------------+--------------+------+ AF with reservation and the excess flow use together 60% of the link bandwidth because EF has the highest priority (-> 40%). 10% of AF packets will be lost.
-
-
-
(b) Excess flow has AF codepoint
-
Figure 6:
-
-
-
Resulting share of bandwidth in an interior node with a neglected reservation of (a) an Expedited Forwarding flow or (b) an Assured Forwarding flow
-
-
The result of case 2 is that there is no restriction for Expedited Forwarding, but as Fig. 6 (a) shows, other services will be extremely disadvantaged by this use of non-reserved resources. Their bandwidth is used by the new additional flow. In this case, the additional 30% Expedited Forwarding traffic preempts resources from the Assured Forwarding traffic, which in turn preempts resources from the best-effort traffic, resulting in 10% packet losses for the Assured Forwarding aggregate, and a complete loss of best-effort traffic. The example in Fig. 6 (b) shows that this can also happen with lower priority services like Assured Forwarding. When a reservation for a service flow with lower priority is neglected, other services (with even lower priority) can be reduced in their quality (in this case the best-effort service). As shown in the example, the service's aggregate causing the NRS problem can itself be affected by packet losses (10% of the Assured Forwarding aggregate is discarded). Besides the described problems of case 2, case 1 will occur in the DS boundary node of the next DS domain that performs traffic metering and policing for the service aggregate.
-
Directly applying RSVP to Differentiated Services would also result in temporary occurrence of the NRS Problem. A receiver has to join the IP multicast group to receive the sender's PATH messages, before being able to send a resource reservation request (RESV message). Thus, the join message on the link for receiving PATH messages can cause the NRS Problem, if this situation is not handled in a special way (e.g., by marking all PATH messages with codepoint 0 and dropping or re-marking all other data packets of the multicast flow).
-
2.2. Heterogeneous Multicast Groups
-
Heterogeneous multicast groups contain one or more receivers, which would like to get another service or quality of service as the sender provides or other receiver subsets currently use. A very important characteristic which should be supported by Differentiated Services is that participants requesting a best-effort quality only should also be able to participate in a group communication which otherwise utilizes a better service class. The next better support for heterogeneity provides concurrent use of more than two different service classes within a group. Things tend to get even more complex when not only different service classes are required, but also different values for quality parameters within a certain service class.
A further problem is to support heterogeneous groups with different service classes in a consistent way. It is possible that some services will not be comparable to each other so that one service cannot be replaced by the other, and both services have to be provided over the same link within this group.
Because an arbitrary new receiver that wants to get the different service can be grafted to any point of the current multicast delivery tree, even interior nodes may have to replicate packets using the different service. At a first glance, this seems to be a contradiction with respect to simplicity of the interior nodes, because they do not even have a profile available and should now convert the service of quality of individual receivers. Consequently, in order to accomplish this, interior nodes have to change the codepoint value during packet replication.
2.3. Dynamics of Any-Source Multicast
-
Basically, within an IP multicast group, any participant (actually, it can be any host not even receiving packets of this multicast group) can act as a sender. This is an important feature which should also be available in case a specific service other than best- effort is used within the group. Differentiated Services possess, conceptually, a unidirectional character. Therefore, for every multicast tree implied by a sender, resources must be reserved separately if simultaneous sending should be possible with a better service. This is even true if shared multicast delivery trees are used (e.g., with PIM-SM or Core Based Trees). If not enough resources are reserved for a service within a multicast tree allowing simultaneous sending of more than one participant, the NRS problem will occur again. The same argument applies to half-duplex reservations which would share the reserved resources by several senders, because it cannot be ensured by the network that exactly one sender sends packets to the group. Accordingly, the corresponding RSVP reservation styles "Wildcard Filter" and "Shared-Explicit Filter" [4] cannot be supported within Differentiated Services. The Integrated Services approach is able to ensure the half-duplex nature of the traffic, because every router can check each packet for its conformance with the installed reservation state.
3. Solutions for Enabling IP-Multicast in Differentiated Services
Networks
-
The problems described in the previous section are mainly caused by the simplicity of the Differentiated Services architecture. Solutions that do not introduce additional complexity need to be introduced so as to not diminish the scalability of the DiffServ approach. This document suggests a straightforward solution for most of the problems.
3.1. Solution for the NRS Problem
-
The proposed solution consists conceptually of the following three steps that are described in more detail later.
-
- A new receiver joins a multicast group that is using a DiffServ service. Multicast routing protocols accomplish the connection of the new branch to the (possibly already existing) multicast delivery tree as usual.
- The unauthorized use of resources is avoided by re-marking at branching nodes all additional packets departing down the new branch. At first, the new receiver will get all packets of the multicast group without quality of service. The management entity of the correspondent DiffServ domain may get informed about the extension of the multicast tree.
- If a pre-issued reservation is available for the new branch or somebody (receiver, sender or a third party) issues one, the management entity instructs the branching router to set the corresponding codepoint for the demanded service.
Usage of resources which were not previously reserved must be prevented. In the following example, we consider a case where the join of a new receiver to a DS multicast group requires grafting of a new branch to an already existing multicast delivering tree. The connecting node that joins both trees converts the codepoint (and therefore the Per-Hop Behavior) to a codepoint of a PHB which is similar to the default PHB in order to provide a best-effort-like service for the new branch. More specifically, this particular PHB can provide a service that is even worse than the best-effort service of the default PHB. See RFC 3662 [16] for a corresponding Lower Effort Per-Domain Behavior.
The conversion to this specific PHB could be necessary in order to avoid unfairness being introduced within the best-effort service aggregate, and, which results from the higher amount of resource usage of the incoming traffic belonging to the multicast group. If the rate at which re-marked packets are injected into the outgoing aggregate is not reduced, those re-marked packets will probably cause discarding of other flow's packets in the outgoing aggregate if resources are scarce.
Therefore, the re-marked packets from this multicast group should be discarded more aggressively than other packets in this outgoing aggregate. This could be accomplished by using an appropriately configured PHB (and a related DSCP) for those packets. In order to distinguish this kind of PHB from the default PHB, it is referred to as the Limited Effort (LE) PHB (which can be realized by an appropriately configured AF PHB [9] or Class Selector Compliant PHB [1]) throughout this document. Merely dropping packets more aggressively at the re-marking node is not sufficient, because there may be enough resources in the outgoing behavior aggregate (BA) to transmit every re-marked packet without having to discard any other packets within the same BA. However, resources in the next node may be short for this particular BA. Those "excess" packets, therefore, must be identifiable at this node.
Re-marking packets is only required at branching nodes, whereas all other nodes of the multicast tree (such with outdegree 1) replicate packets as usual. Because a branching node may also be an interior node of a domain, re-marking of packets requires conceptually per- flow classification. Though this seems to be in contradiction to the DiffServ philosophy of a core that avoids per-flow states, IP multicast flows are different from unicast flows: traditional IP multicast forwarding and multicast routing are required to install states per multicast group for every outgoing link anyway. Therefore, re-marking in interior nodes is scalable to the same extent as IP multicast (cf. section 4).
Re-marking with standard DiffServ mechanisms [10] for every new branch requires activation of a default traffic profile. The latter accomplishes re-marking by using a combination of an MF-classifier and a marker at an outgoing link that constitutes a new branch. The classifier will direct all replicated packets to a marker that sets the new codepoint. An alternative implementation is described in section 7.
The better service will only be provided if a reservation request was processed and approved by the resource management function. That means an admission control test must be performed before resources are actually used by the new branch. In case the admission test is successful, the re-marking node will be instructed by the resource management to stop re-marking and to use the original codepoint again (conceptually by removing the profile).
In summary, only those receivers will obtain a better service within a DiffServ multicast group, which previously reserved the corresponding resources in the new branch with assistance of the resource management. Otherwise, they get a quality which might be even lower than best-effort.
-
3.2. Solution for Supporting Heterogeneous Multicast Groups
-
In this document, considerations are limited to provisioning different service classes, but not different quality parameters within a certain service class.
The proposed concept from section 3.1 provides a limited solution of the heterogeneity problem. Receivers are allowed to obtain a Limited Effort service without a reservation, so that at least two different service classes within a multicast group are possible. Therefore, it is possible for any receiver to participate in the multicast session without getting any quality of service. This is useful if a receiver just wants to see whether the content of the multicast group is interesting enough, before requesting a better service which must be paid for (like snooping into a group without prior reservation).
Alternatively, a receiver might not be able to receive this better quality of service (e.g., because it is mobile and uses a wireless link of limited capacity), but it may be satisfied with the reduced quality, instead of getting no content at all.
Additionally, applying the RSVP concept of listening for PATH messages before sending any RESV message is feasible again. Without using the proposed solution, this would have caused the NRS Problem.
Theoretically, the proposed approach in section 7 also supports more than two different services within one multicast group, because the additional field in the multicast routing table can store any DSCP value. However, this would work only if PHBs can be ordered, so that the "best" PHB among different required PHBs downstream is chosen to be forwarded on a specific link. This is mainly a management issue and is out of the scope of this document.
More advanced concepts may also support conditional re-marking in dependence on the group address and DSCP value. This is useful if the group uses different PHBs (e.g., for flows to different transport protocol ports) and the re-marking should thus additionally depend on the DSCP value of an incoming packet.
3.3. Solution for Any-Source Multicast
-
Every participant would have to initiate an explicit reservation to ensure the possibility of sending to the group with a better service quality, regardless of whether other senders within the group already use the same service class simultaneously. This would require a separate reservation for each sender-rooted multicast tree.
However, in the specific case of best-effort service (the default PHB), it is nevertheless possible for participants to send packets to the group anytime without requiring any additional mechanisms. The reason for this is that the first DS-capable boundary node will mark those packets with the DSCP of the default PHB because of a missing traffic profile for this particular sender. The first DS capable boundary nodes should therefore always classify multicast packets based on both the sender's address and the multicast group address.
4. Scalability Considerations
-
The proposed solution does not add complexity to the DS architecture or to a DS node, nor does it change the scalability properties of DiffServ. With current IP multicast routing protocols, a multicast router has to manage and hold state information per traversing multicast flow. The suggested solution scales to the same extent as IP multicast itself, because the proposed re-marking may occur per branch of a multicast flow. This re-marking is logically associated with an addition to the multicast routing state that is required anyway. In this respect, re-marking of packets for multicast flows in interior nodes is not considered as a scalability problem or to be in contradiction to the DiffServ approach itself. It is important to distinguish the multicast case from existing justifiable scalability concerns relating to re-marking packets of unicast flows within interior routers. Moreover, the decision of when to change a re- marking policy is not performed by the router, but by some management entity at a time scale which is different from the time scale at the packet forwarding level.
5. Deployment Considerations
-
The solution proposed in section 3.1 can be deployed on most router platforms available today. Architectures that perform routing and forwarding functions in software could be updated by a new software release.
However, there may be some specialized hardware platforms that are currently not able to deploy the proposed solution from section 7. This may be the case when a multicast packet is directly duplicated on the backplane of the router, so that all outgoing interfaces read the packet in parallel. Consequently, the codepoint cannot be changed for a subset of these outgoing interfaces and the NRS problem can not be solved directly in the branching point.
In this case, there exist several alternative solutions:
-
- As mentioned in section 3.1, if traffic conditioning mechanisms can be applied on the outgoing packets at the individual output interfaces, a combination of classifier and marker may be used for each branch.
- The change of the codepoint for subtrees without properly allocated resources could take place in the following downstream router. There, for every incoming packet of the considered multicast group, the codepoint would be changed to the value that the previous router should have set. If a LAN (e.g., a high-speed switching LAN) is attached to the considered outgoing interface, then on every router connected to the LAN, packets of the considered group should be changed
-
on the incoming interface by standard DiffServ mechanisms.
-
Future releases of router architectures may support the change of the codepoint directly in the replication process as proposed in section 7.
6. Security Considerations
-
Basically, the security considerations in [1] apply. The proposed solution does not imply new security aspects. If a join of arbitrary end-systems to a multicast group is not desired (thereby receiving a lower than best-effort quality) the application usually has to exclude these participants. This can be accomplished by using authentication, authorization, or ciphering techniques at the application level -- like in traditional IP multicast scenarios.
Moreover, it is important to consider the security of corresponding management mechanisms, because they are used to activate re-marking of multicast flows. On the one hand, functions for instructing the router to mark or re-mark packets of multicast flows are attractive targets to perform theft of service attacks. On the other hand, their security depends on the router management mechanisms which are used to realize this functionality. Router management should generally be protected against unauthorized use, therefore preventing those attacks as well.
7. Implementation Model Example
-
One possibility of implementing the proposed solution from section 3.1 is described in the following. It has to be emphasized that other realizations are also possible, and this description should not be understood as a restriction on potential implementations. The benefit of the following described implementation is that it does not require any additional classification of multicast groups within an aggregate. It serves as a proof of concept that no additional complexity is necessary to implement the proposed general solution described in section 3.
Because every multicast flow has to be considered by the multicast routing process (in this context, this notion signifies the multicast forwarding part and not the multicast route calculation and maintenance part, cf. Fig. 1), the addition of an extra byte in each multicast routing table entry containing the DS field, and thus its DS codepoint value per output link (resp. virtual interface, see Fig. 8), results in nearly no additional cost. Packets will be replicated by the multicast forwarding process, so this is also the right place for setting the correct DSCP values of the replicated packets. Their DSCP values are not copied from the incoming original packet, but from the additional DS field in the multicasting routing table entry for the corresponding output link (only the DSCP value must be copied, while the two remaining bits are ignored and are present for simplification reasons only). This field initially contains the codepoint of the LE PHB if incoming packets for this specific group do not carry the codepoint of the default PHB.
When a packet arrives with the default PHB, the outgoing replicates should also get the same codepoint in order to retain the behavior of current common multicast groups using the default PHB. A router configuration message changes the DSCP values in the multicast routing table and may also carry the new DSCP value which should be set in the replicated packets. It should be noted that although re- marking may also be performed by interior nodes, the forwarding performance will not be decreased, because the decision of when and what to re-mark is made by the management (control plane).
Multicast Other List Destination Fields of Address virtual Inter- DS interfaces face ID Field +--------------------------------+ +-------------------+ | X | .... | *-------------------->| C | (DSCP,CU) | |--------------------------------| +-------------------+ | Y | .... | *-----------+ | D | (DSCP,CU) | |--------------------------------| | +-------------------+ | ... | .... | ... | | . . . . | +-------------------+ . ... . .... . ... . +-------->| B | (DSCP,CU) | +--------------------------------+ +-------------------+ | ... | .... | ... | | D | (DSCP,CU) | +--------------------------------+ +-------------------+ | ... | ... | . . . . . .
-
-
Figure 8:
-
-
Multicast routing table with additional fields for DSCP values
-
-
-
8. Proof of the Neglected Reservation Subtree Problem
-
In the following, it is shown that the NRS problem actually exists and occurs in reality. Hence, the problem and its solution was investigated using a standard Linux Kernel (v2.4.18) and the Linux- based implementation KIDS [11].
Furthermore, the proposed solution for the NRS problem has been implemented by enhancing the multicast routing table, as well as the multicast routing behavior in the Linux kernel. In the following section, the modifications are briefly described.
Additional measurements with the simulation model simulatedKIDS [12] will be presented in section 9. They show the effects of the NRS problem in more detail and also the behavior of the BAs using or not using the Limited Effort PHB for re-marking.
8.1. Implementation of the Proposed Solution
-
As described in section 3.1, the proposed solution for avoiding the NRS Problem is an extension of each routing table entry in every Multicast router by one byte. In the Linux OS, the multicast routing table is implemented by the "Multicast Forwarding Cache (MFC)". The MFC is a hash table consisting of an "mfc-cache"-entry for each combination of the following three parameters: sender's IP address, multicast group address, and incoming interface.
The routing information in a "mfc-cache"-entry is kept in an array of TTLs for each virtual interface. When the TTL is zero, a packet matching to this "mfc-cache"-entry will not be forwarded on this virtual interface. Otherwise, if the TTL is less than the packet's TTL, the latter will be forwarded on the interface with a decreased TTL.
In order to set an appropriate codepoint if bandwidth is allocated on an outgoing link, we added a second array of bytes -- similar to the TTL array -- for specifying the codepoint that should be used on a particular virtual interface. The first six bits of the byte contain the DSCP that should be used, and the seventh bit indicates whether the original codepoint in the packet has to be changed to the specified one (=0) or has to be left unchanged (=1). The default entry of the codepoint byte is zero; so initially, all packets will be re-marked to the default DSCP.
Furthermore, we modified the multicast forwarding code for considering this information while replicating multicast packets. To change an "mfc-cache"-entry we implemented a daemon for exchanging the control information with a management entity (e.g., a bandwidth broker). Currently, the daemon uses a proprietary protocol, but migration to the COPS protocol (RFC 2748) is planned.
8.2. Test Environment and Execution
-
Sender
+---+ FHN: First Hop Node | S | BN: Boundary Node +---+ +# +# +# +---+ +--+ +------+ |FHN|++++++++++++|BN|+++++++++++| host | | |############| |***********| B | +---+ +--+## +------+ +# # +# # +# # +------+ +------+ |host A| |host C| +------+ +------+ +++ EF flow (group1) with reservation ### EF flow (group2) with reservation *** EF flow (group2) without reservation
-
-
Figure 8.1:
-
-
-
Evaluation of NRS-Problem described in Figure 3
-
-
-
In order to prove case 1 of the NRS problem, as described in section 2.1, the testbed shown in Figure 8.1 was built. It is a reduced version of the network shown in Figure 5 and consists of two DS- capable nodes, an ingress boundary node and an egress boundary node. The absence of interior nodes does not have any effect on to the proof of the described problem.
The testbed is comprised of two Personal Computers (Pentium III at 450 Mhz, 128 MB Ram, 3 network cards Intel eepro100) used as DiffServ nodes, as well as one sender and three receiver systems (also PCs). On the routers, KIDS has been installed and an mrouted (Multicast Routing Daemon) was used to perform multicast routing. The network was completely built of separate 10BaseT Ethernet segments in full- duplex mode. In [11], we evaluated the performance of the software routers and found out that even a PC at 200Mhz had no problem handling up to 10Mbps DS traffic on each link. Therefore, the presented measurements are not a result of performance bottlenecks caused by these software routers.
The sender generated two shaped UDP traffic flows of 500kbps (packets of 1.000 byte constant size) each and sent them to multicast group 1 (233.1.1.1) and 2 (233.2.2.2). In both measurements, receiver A had a reservation along the path to the sender for each flow, receiver B had reserved for flow 1, and C for flow 2. Therefore, two static profiles are installed in the ingress boundary node with 500kbps EF bandwidth and a token bucket size of 10.000 bytes for each flow.
In the egress boundary node, one profile has been installed for the output link to host B and one related for the output link to host C. Each of them permits up to 500kbps Expedited Forwarding, but only the aggregate of Expedited Forwarding traffic carried on the outgoing link is considered.
In measurement 1, the hosts A and B joined to group 1, while A, B, and C joined to group 2. Those joins are using a reservation for the group towards the sender. Only the join of host B to group 2 has no admitted reservation. As described in section 2.1, this will cause the NRS problem (case 1). Metering and policing mechanisms in the egress boundary node throttle down the EF aggregate to the reserved 500kbps, and do not depend on whether or not individual flows have been reserved.
+--------+--------+--------+ | Host A | Host B | Host C | +---------+--------+--------+--------+ | Group 1 | 500kbps| 250kbps| 500kbps| +---------+--------+--------+--------+ | Group 2 | 500kbps| 250kbps| | +---------+--------+--------+--------+
-
-
Figure 8.2:
-
-
-
Results of measurement 1 (without the proposed solution): Average bandwidth of each flow.
--> Flows of group 1 and 2 on the link to host B share the reserved aggregate of group 1.
-
-
-
Figure 8.2 shows the obtained results. Host A and C received their flows without any interference. But host B received data from group 1 with only half of the reserved bandwidth, so one half of the packets have been discarded. Figure 8.2 also shows that receiver B got the total amount of bandwidth for group 1 and 2, that is exactly the reserved 500kbps. Flow 2 got Expedited Forwarding without actually having reserved any bandwidth and additionally violated the guarantee of group 1 on that link.
For measurement 2, the previously presented solution (cf. section 3.1) has been installed in the boundary node. Now, while duplicating the packets, whether the codepoint has to be changed to Best-Effort (or Limited Effort) or whether it can be just duplicated is checked. In this measurement, it changed the codepoint for group 2 on the link to Host B to Best-Effort.
+--------+--------+--------+ | Host A | Host B | Host C | +---------+--------+--------+--------+ | Group 1 | 500kbps| 500kbps| 500kbps| +---------+--------+--------+--------+ | Group 2 | 500kbps| 500kbps| | +---------+--------+--------+--------+
-
-
Figure 8.3:
-
-
-
Results of measurement 1 (with the proposed solution): Average bandwidth of each flow.
--> Flow of group 1 on the link to host B gets the reserved bandwidth of group 1. The flow of group 2 has been re-marked to Best-Effort.
-
-
-
Results of this measurement are presented in Figure 8.3. Each host received its flows with the reserved bandwidth and without any packet loss. Packets from group 2 are re-marked in the boundary node so that they have been treated as best-effort traffic. In this case, they got the same bandwidth as the Expedited Forwarding flow, and because there was not enough other traffic on the link present, there was no need to discard packets.
The above measurements confirm that the Neglected Reservation Subtree problem is to be taken seriously and that the presented solution will solve it.
9. Simulative Study of the NRS Problem and Limited Effort PHB
-
This section shows some results from a simulative study which shows the correctness of the proposed solution and the effect of re-marking the responsible flow to Limited Effort. A proof of the NRS problem has also been given in section 8 and in [13]. This section shows the benefit for the default Best Effort traffic when Limited Effort is used for re-marking instead of Best Effort. The results strongly motivate the use of Limited Effort.
9.1. Simulation Scenario
-
In the following scenario, the boundary nodes had a link speed of 10 Mpbs and Interior Routers had a link speed of 12 Mbps. In boundary nodes, a 5 Mbps aggregate for EF has been reserved.
When Limited Effort was used, LE got 10% capacity (0.5Mpbs) from the original BE aggregate and BE 90% (4.5Mbps) of the original BE aggregate capacity. The bandwidth between LE and BE is shared by using WFQ scheduling.
The following topology was used, where Sx is a sender, BRx a boundary node, IRx an interior node, and Dx a destination/receiver.
+--+ +--+ +---+ +--+ |S1| |S0| /=|BR5|=====|D0| +--+ +--+ // +---+ +--+ \\ || // \\ || // +--+ \+---+ +---+ +---+ +---+ +--+ |S2|===|BR1|=====|IR1|=====|IR2|======|BR3|=====|D1| +--+ +---+ /+---+ +---+ +---+ +--+ // \\ +--+ // \\ /=|D2| +--+ +---+ // \\ // +--+ |S3|===|BR2|=/ +---+/ +--+ +---+ /=|BR4|=\ || +--+ // +---+ \\ +--+ +--+ |D4|=/ \=|D3| |S4| +--+ +--+ +--+
-
-
-
Figure 9.1: Simulation Topology
-
-
The following table shows the flows in the simulation runs, e.g., EF0 is sent from Sender S0 to Destination D0 with a rate of 4 Mbps using an EF reservation.
In the presented cases (I to IV), different amounts of BE traffic were used to show the effects of Limited Effort in different cases. The intention of these four cases is described after the table.
In all simulation models, EF sources generated constant rate traffic with constant packet sizes using UDP.
The BE sources also generated constant rate traffic, where BE0 used UDP, and BE1 used TCP as a transport protocol.
+----+--------+-------+----------+-----------+-----------+---------+ |Flow| Source | Dest. | Case I | Case II | Case III | Case IV | +----+--------+-------+----------+-----------+-----------+---------+ | EF0| S0 | D0 | 4 Mbps | 4 Mbps | 4 Mbps | 4 Mbps | +----+--------+-------+----------+-----------+-----------+---------+ | EF1| S1 | D1 | 2 Mbps | 2 Mbps | 2 Mbps | 2 Mbps | +----+--------+-------+----------+-----------+-----------+---------+ | EF2| S2 | D2 | 5 Mbps | 5 Mbps | 5 Mbps | 5 Mbps | +----+--------+-------+----------+-----------+-----------+---------+ | BE0| S3 | D3 | 1 Mbps | 2.25 Mbps | 0.75 Mbps |3.75 Mbps| +----+--------+-------+----------+-----------+-----------+---------+ | BE1| S4 | D4 | 4 Mbps | 2.25 Mbps | 0.75 Mbps |3.75 Mbps| +----+--------+-------+----------+-----------+-----------+---------+
Table 9.1:
-
-
-
Direction, amount and Codepoint of flows in the four simulation cases (case I to IV)
-
-
The four cases (I to IV) used in the simulation runs had the following characteristics:
Case I: In this scenario, the BE sources sent together exactly 5 Mbps, so there is no congestion in the BE queue. Case II: BE is sending less than 5 Mbps, so there is space available in the BE queue for re-marked traffic. BE0 and BE1 are sending together 4.5 Mbps, which is exactly the share of BE when LE is used. So, when multicast packets are re-marked to LE because of the NRS problem, the LE should get 0.5 Mbps and BE 4.5 Mbps, which is still enough for BE0 and BE1. LE should not show a greedy behavior and should not use resources from BE.
Case III:
-
-
-
In this case, BE is very low. BE0 and BE1 use together only 1.5 Mbps. So when LE is used, it should be able to use the unused bandwidth resources from BE.
-
-
Case IV: BE0 and BE1 send together 7.5 Mbps so there is congestion in the BE queue. In this case, LE should get 0.5 Mbps (not more and not less).
In each scenario, loss rate and throughput of the considered flows and aggregates have been metered.
-
9.2. Simulation Results for Different Router Types
9.2.1. Interior Node
-
When the branching point of a newly added multicast subtree is located in an interior node, the NRS problem can occur as described in section 2.1 (Case 2).
In the simulation runs presented in the following four subsections, D3 joins to the multicast group of sender S0 without making any reservation or resource allocation. Consequently, a new branch is added to the existing multicast tree. The branching point issued by the join of D3 is located in IR2. On the link to BR3, no bandwidth was allocated for the new flow (EF0).
The metered throughput of flows on the link between IR2 and BR3 in the four different cases is shown in the following four subsections. The situation before the new receiver joins is shown in the second column. The situation after the join without the proposed solution is shown in column three. The fourth column presents the results when the proposed solution of section 3.1 is used and the responsible flow is re-marked to LE.
9.2.1.1. Case I:
-
+--------+-----------------+-----------------+------------------+ | | before join | after join |after join, | | | | (no re-marking) |(re-marking to LE)| +--------+-----------------+-----------------+------------------+ | | EF0: --- | EF0: 4.007 Mbps | LE0: 0.504 Mbps | |achieved| EF1: 2.001 Mbps | EF1: 2.003 Mbps | EF1: 2.000 Mbps | |through-| EF2: 5.002 Mbps | EF2: 5.009 Mbps | EF2: 5.000 Mbps | |put | BE0: 1.000 Mbps | BE0: 0.601 Mbps | BE0: 1.000 Mbps | | | BE1: 4.000 Mbps | BE1: 0.399 Mbps | BE1: 3.499 Mbps | +--------+-----------------+-----------------+------------------+ |BA | EF: 7.003 Mbps | EF: 11.019 Mbps | EF: 7.000 Mbps | |through-| BE: 5.000 Mbps | BE: 1.000 Mbps | BE: 4.499 Mbps | |put | LE: --- | LE: --- | LE: 0.504 Mbps | +--------+-----------------+-----------------+------------------+ | | EF0: --- | EF0: 0 % | LE0: 87.4 % | |packet | EF1: 0 % | EF1: 0 % | EF1: 0 % | |loss | EF2: 0 % | EF2: 0 % | EF2: 0 % | |rate | BE0: 0 % | BE0: 34.8 % | BE0: 0 % | | | BE1: 0 % | BE1: 59.1 % | BE1: 0 % | +--------+-----------------+-----------------+------------------+ (*) EF0 is re-marked to LE and signed as LE0
9.2.1.2. Case II:
-
+--------+-----------------+-----------------+------------------+ | | before join | after join |after join, | | | | (no re-marking) |(re-marking to LE)| +--------+-----------------+-----------------+------------------+ | | EF0: --- | EF0: 4.003 Mbps | LE0: 0.500 Mbps | |achieved| EF1: 2.000 Mbps | EF1: 2.001 Mbps | EF1: 2.001 Mbps | |through-| EF2: 5.002 Mbps | EF2: 5.005 Mbps | EF2: 5.002 Mbps | |put | BE0: 2.248 Mbps | BE0: 0.941 Mbps | BE0: 2.253 Mbps | | | BE1: 2.252 Mbps | BE1: 0.069 Mbps | BE1: 2.247 Mbps | +--------+-----------------+-----------------+------------------+ |BA | EF: 7.002 Mbps | EF: 11.009 Mbps | EF: 7.003 Mbps. | |through-| BE: 4.500 Mbps | BE: 1.010 Mbps | BE: 4.500 Mbps | |put | LE: --- | LE: --- | LE: 0.500 Mbps | +--------+-----------------+-----------------+------------------+ | | EF0: --- | EF0: 0 % | LE0: 87.4 % | |packet | EF1: 0 % | EF1: 0 % | EF1: 0 % | |loss | EF2: 0 % | EF2: 0 % | EF2: 0 % | |rate | BE0: 0 % | BE0: 58.0 % | BE0: 0 % | | | BE1: 0 % | BE1: 57.1 % | BE1: 0 % | +--------+-----------------+-----------------+------------------+ (*) EF0 is re-marked to LE and signed as LE0
9.2.1.3. Case III:
-
+--------+-----------------+-----------------+------------------+ | | before join | after join |after join, | | | | (no re-marking) |(re-marking to LE)| +--------+-----------------+-----------------+------------------+ | | EF0: --- | EF0: 3.998 Mbps | LE0: 3.502 Mbps | |achieved| EF1: 2.000 Mbps | EF1: 2.001 Mbps | EF1: 2.001 Mbps | |through-| EF2: 5.000 Mbps | EF2: 5.002 Mbps | EF2: 5.003 Mbps | |put | BE0: 0.749 Mbps | BE0: 0.572 Mbps | BE0: 0.748 Mbps | | | BE1: 0.749 Mbps | BE1: 0.429 Mbps | BE1: 0.748 Mbps | +--------+-----------------+-----------------+------------------+ |BA | EF: 7.000 Mbps | EF: 11.001 Mbps | EF: 7.004 Mbps | |through-| BE: 1.498 Mbps | BE: 1.001 Mbps | BE: 1.496 Mbps | |put | LE: --- | LE: --- | LE: 3.502 Mbps | +--------+-----------------+-----------------+------------------+ | | EF0: --- | EF0: 0 % | LE0: 12.5 % | |packet | EF1: 0 % | EF1: 0 % | EF1: 0 % | |loss | EF2: 0 % | EF2: 0 % | EF2: 0 % | |rate | BE0: 0 % | BE0: 19.7 % | BE0: 0 % | | | BE1: 0 % | BE1: 32.6 % | BE1: 0 % | +--------+-----------------+-----------------+------------------+ (*) EF0 is re-marked to LE and signed as LE0
9.2.1.4. Case IV:
-
+--------+-----------------+-----------------+------------------+ | | before join | after join |after join, | | | | (no re-marking) |(re-marking to LE)| +--------+-----------------+-----------------+------------------+ | | EF0: --- | EF0: 4.001 Mbps | LE0: 0.500 Mbps | |achieved| EF1: 2.018 Mbps | EF1: 2.000 Mbps | EF1: 2.003 Mbps | |through-| EF2: 5.005 Mbps | EF2: 5.001 Mbps | EF2: 5.007 Mbps | |put | BE0: 2.825 Mbps | BE0: 1.000 Mbps | BE0: 3.425 Mbps | | | BE1: 2.232 Mbps | BE1: --- | BE1: 1.074 Mbps | +--------+-----------------+-----------------+------------------+ |BA | EF: 7.023 Mbps | EF: 11.002 Mbps | EF: 7.010 Mbps | |through-| BE: 5.057 Mbps | BE: 1.000 Mbps | BE: 4.499 Mbps | |put | LE: --- | LE: --- | LE: 0.500 Mbps | +--------+-----------------+-----------------+------------------+ | | EF0: --- | EF0: 0 % | LE0: 75.0 % | |packet | EF1: 0 % | EF1: 0 % | EF1: 0 % | |loss | EF2: 0 % | EF2: 0 % | EF2: 0 % | |rate | BE0: 23.9 % | BE0: 73.3 % | BE0: 0 % | | | BE1: 41.5 % | BE1: --- | BE1: 0 % | +--------+-----------------+-----------------+------------------+ (*) EF0 is re-marked to LE and signed as LE0
NOTE: BE1 has undefined throughput and loss in situation "after join (no re-marking)", because TCP is going into retransmission back-off timer phase and closes the connection after 512 seconds.
9.2.2. Boundary Node
-
When the branching point of a newly added multicast subtree is located in a boundary node, the NRS problem can occur as described in section 2.1 (Case 1).
In the simulation runs presented in the following four subsections, D3 joins to the multicast group of sender S1 without making any reservation or resource allocation. Consequently, a new branch is added to the existing multicast tree. The branching point issued by the join of D3 is located in BR3. On the link to BR4, no bandwidth was allocated for the new flow (EF1).
The metered throughput of the flows on the link between BR3 and BR4 in the four different cases is shown in the following four subsections. The situation before the new receiver joins is shown in the second column. The situation after the join but without the proposed solution is shown in column three. The fourth column presents results when the proposed solution of section 3.1 is used and the responsible flow is re-marked to LE.
9.2.2.1. Case I:
-
+--------+-----------------+-----------------+------------------+ | | before join | after join |after join, | | | | (no re-marking) |(re-marking to LE)| +--------+-----------------+-----------------+------------------+ | | EF0: --- | EF0: --- | EF0: --- | |achieved| EF1: --- | EF1: 1.489 Mbps | LE1: 0.504 Mbps | |through-| EF2: 5.002 Mbps | EF2: 3.512 Mbps | EF2: 5.002 Mbps | |put | BE0: 1.000 Mbps | BE0: 1.000 Mbps | BE0: 1.004 Mbps | | | BE1: 4.000 Mbps | BE1: 4.002 Mbps | BE1: 3.493 Mbps | +--------+-----------------+-----------------+------------------+ |BA | EF: 5.002 Mbps | EF: 5.001 Mbps | EF: 5.002 Mbps | |through-| BE: 5.000 Mbps | BE: 5.002 Mbps | BE: 4.497 Mbps | |put | LE: --- | LE: --- | LE: 0.504 Mbps | +--------+-----------------+-----------------+------------------+ | | EF0: --- | EF0: --- | EF0: --- | |packet | EF1: --- | EF1: 25.6 % | LE1: 73.4 % | |loss | EF2: 0 % | EF2: 29.7 % | EF2: 0 % | |rate | BE0: 0 % | BE0: 0 % | BE0: 0 % | | | BE1: 0 % | BE1: 0 % | BE1: 0 % | +--------+-----------------+-----------------+------------------+ (*) EF1 is re-marked to LE and signed as LE1
9.2.2.2. Case II:
-
+--------+-----------------+-----------------+------------------+ | | before join | after join |after join, | | | | (no re-marking) |(re-marking to LE)| +--------+-----------------+-----------------+------------------+ | | EF0: --- | EF0: --- | EF0: --- | |achieved| EF1: --- | EF1: 1.520 Mbps | LE1: 0.504 Mbps | |through-| EF2: 5.003 Mbps | EF2: 3.482 Mbps | EF2: 5.002 Mbps | |put | BE0: 2.249 Mbps | BE0: 2.249 Mbps | BE0: 2.245 Mbps | | | BE1: 2.252 Mbps | BE1: 2.252 Mbps | BE1: 2.252 Mbps | +--------+-----------------+-----------------+------------------+ |BA | EF: 5.003 Mbps | EF: 5.002 Mbps | EF: 5.002 Mbps | |through-| BE: 4.501 Mbps | BE: 4.501 Mbps | BE: 4.497 Mbps | |put | LE: --- | LE: --- | LE: 0.504 Mbps | +--------+-----------------+-----------------+------------------+ | | EF0: --- | EF0: --- | EF0: --- | |packet | EF1: --- | EF1: 24.0 % | LE1: 74.8 % | |loss | EF2: 0 % | EF2: 30.4 % | EF2: 0 % | |rate | BE0: 0 % | BE0: 0 % | BE0: 0 % | | | BE1: 0 % | BE1: 0 % | BE1: 0 % | +--------+-----------------+-----------------+------------------+ (*) EF1 is re-marked to LE and signed as LE1
9.2.2.3. Case III:
-
+--------+-----------------+-----------------+------------------+ | | before join | after join |after join, | | | | (no re-marking) |(re-marking to LE)| +--------+-----------------+-----------------+------------------+ | | EF0: --- | EF0: --- | EF0: --- | |achieved| EF1: --- | EF1: 1.084 Mbps | LE1: 2.000 Mbps | |through-| EF2: 5.001 Mbps | EF2: 3.919 Mbps | EF2: 5.000 Mbps | |put | BE0: 0.749 Mbps | BE0: 0.752 Mbps | BE0: 0.750 Mbps | | | BE1: 0.749 Mbps | BE1: 0.748 Mbps | BE1: 0.750 Mbps | +--------+-----------------+-----------------+------------------+ |BA | EF: 5.001 Mbps | EF: 5.003 Mbps | EF: 5.000 Mbps | |through-| BE: 1.498 Mbps | BE: 1.500 Mbps | BE: 1.500 Mbps | |put | LE: --- | LE: --- | LE: 2.000 Mbps | +--------+-----------------+-----------------+------------------+ | | EF0: --- | EF0: --- | EF0: --- | |packet | EF1: --- | EF1: 45.7 % | LE1: 0 % | |loss | EF2: 0 % | EF2: 21.7 % | EF2: 0 % | |rate | BE0: 0 % | BE0: 0 % | BE0: 0 % | | | BE1: 0 % | BE1: 0 % | BE1: 0 % | +--------+-----------------+-----------------+------------------+ (*) EF1 is re-marked to LE and signed as LE1
9.2.2.4. Case IV:
-
+--------+-----------------+-----------------+------------------+ | | before join | after join |after join, | | | | (no re-marking) |(re-marking to LE)| +--------+-----------------+-----------------+------------------+ | | EF0: --- | EF0: --- | EF0: --- | |achieved| EF1: --- | EF1: 1.201 Mbps | LE1: 0.500 Mbps | |through-| EF2: 5.048 Mbps | EF2: 3.803 Mbps | EF2: 5.004 Mbps | |put | BE0: 2.638 Mbps | BE0: 2.535 Mbps | BE0: 3.473 Mbps | | | BE1: 2.379 Mbps | BE1: 2.536 Mbps | BE1: 1.031 Mbps | +--------+-----------------+-----------------+------------------+ |BA | EF: 5.048 Mbps | EF: 5.004 Mbps | EF: 5.004 Mbps | |through-| BE: 5.017 Mbps | BE: 5.071 Mbps | BE: 4.504 Mbps | |put | LE: --- | LE: --- | LE: 0.500 Mbps | +--------+-----------------+-----------------+------------------+ | | EF0: --- | EF0: --- | EF0: --- | |packet | EF1: --- | EF1: 40.0 % | LE1: 68.6 % | |loss | EF2: 0 % | EF2: 23.0 % | EF2: 0 % | |rate | BE0: 30.3 % | BE0: 32.1 % | BE0: 0 % | | | BE1: 33.3 % | BE1: 32.7 % | BE1: 0 % | +--------+-----------------+-----------------+------------------+ (*) EF1 is re-marked to LE and signed as LE1
10. Acknowledgements
-
The authors wish to thank Mark Handley and Bill Fenner for their valuable comments on this document. Special thanks go to Milena Neumann for her extensive efforts in performing the simulations. We would also like to thank the KIDS simulation team [12].
11. References
11.1. Normative References
-
[1] Nichols, K., Blake, S., Baker, F. and D. Black, "Definition of the Differentiated Services Field (DS Field) in the IPv4 and IPv6 Headers", RFC 2474, December 1998. [2] Blake, S., Black, D., Carlson, M., Davies, E., Wang, Z. and W. Weiss, "An Architecture for Differentiated Services", RFC 2475, December 1998.
11.2. Informative References
-
[3] Nichols, K. and B. Carpenter, "Definition of Differentiated Services Per Domain Behaviors and Rules for their Specification", RFC 3086, April 2001. [4] Braden, R., Ed., Zhang, L., Berson, S., Herzog, S. and S. Jamin, "Resource ReSerVation Protocol (RSVP) -- Version 1", RFC 2205, September 1997. [5] Bernet, Y., "Format of the RSVP DCLASS Object", RFC 2996, November 2000. [6] Waitzman, D., Partridge, C. and S. Deering, "Distance Vector Multicast Routing Protocol", RFC 1075, November 1988. [7] Estrin, D., Farinacci, D., Helmy, A., Thaler, D., Deering, S., Handley, M., Jacobson, V., Liu, L., Sharma, P. and L. Wei, "Protocol Independent Multicast-Sparse Mode (PIM-SM): Protocol Specification", RFC 2362, June 1998. [8] Adams, A., Nicholas, J. and W. Siadak, "Protocol Independent Multicast - Dense Mode (PIM-DM) Protocol Specification (Revised)", Work in Progress. [9] Heinanen, J., Baker, F., Weiss, W. and J. Wroclawski, "Assured Forwarding PHB Group" RFC 2597, June 1999. [10] Bernet, Y., Blake, S., Grossman, D. and A. Smith, "An Informal Management Model for DiffServ Routers", RFC 3290, May 2002. [11] R. Bless, K. Wehrle. Evaluation of Differentiated Services using an Implementation under Linux, Proceedings of the Intern. Workshop on Quality of Service (IWQOS'99), London, 1999. [12] K. Wehrle, J. Reber, V. Kahmann. A simulation suite for Internet nodes with the ability to integrate arbitrary Quality of Service behavior, Proceedings of Communication Networks And Distributed Systems Modeling And Simulation Conference (CNDS 2001), Phoenix (AZ), January 2001. [13] R. Bless, K. Wehrle. Group Communication in Differentiated Services Networks, Internet QoS for the Global Computing 2001 (IQ 2001), IEEE International Symposium on Cluster Computing and the Grid, May 2001, Brisbane, Australia, IEEE Press. [14] Davie, B., Charny, A., Bennett, J.C.R., Benson, K., Le Boudec, J.Y., Courtney, W., Davari, S., Firoiu, V. and D. Stiliadis, "An Expedited Forwarding PHB (Per-Hop Behavior)", RFC 3246, March 2002. [15] Charny, A., Bennett, J.C.R., Benson, K., Le Boudec, J.Y., Chiu, A., Courtney, W., Davari, S., Firoiu, V., Kalmanek, C. and K.K. Ramakrishnan, "Supplemental Information for the New Definition of the EF PHB (Expedited Forwarding Per-Hop Behavior)", RFC 3247, March 2002. [16] Bless, R., Nichols, K. and K. Wehrle, "A Lower Effort Per-Domain Behavior (PDB) for Differentiated Services", RFC 3662, December 2003.
12. Authors' Addresses
-
Comments and questions related to this document can be addressed to one of the authors listed below.
Roland Bless
Institute of Telematics
Universitaet Karlsruhe (TH)
Zirkel 2
76128 Karlsruhe, GermanyPhone: +49 721 608 6413 EMail: bless@tm.uka.de URI: http://www.tm.uka.de/~bless
Klaus Wehrle
University of Tuebingen
WSI - Computer Networks and Internet /
Protocol Engineering & Distributed Systems
Morgenstelle 10c
72076 Tuebingen, GermanyEMail: Klaus.Wehrle@uni-tuebingen.de URI: http://net.informatik.uni-tuebingen.de/~wehrle/
13. Full Copyright Statement
-
Copyright © The Internet Society (2004). This document is subject to the rights, licenses and restrictions contained in BCP 78 and except as set forth therein, the authors retain all their rights.
This document and the information contained herein are provided on an "AS IS" basis and THE CONTRIBUTOR, THE ORGANIZATION HE/SHE REPRESENTS OR IS SPONSORED BY (IF ANY), THE INTERNET SOCIETY AND THE INTERNET ENGINEERING TASK FORCE DISCLAIM ALL WARRANTIES, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO ANY WARRANTY THAT THE USE OF THE INFORMATION HEREIN WILL NOT INFRINGE ANY RIGHTS OR ANY IMPLIED WARRANTIES OF MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE.
Intellectual Property
-
The IETF takes no position regarding the validity or scope of any Intellectual Property Rights or other rights that might be claimed to pertain to the implementation or use of the technology described in this document or the extent to which any license under such rights might or might not be available; nor does it represent that it has made any independent effort to identify any such rights. Information on the procedures with respect to rights in RFC documents can be found in BCP 78 and BCP 79.
Copies of IPR disclosures made to the IETF Secretariat and any assurances of licenses to be made available, or the result of an attempt made to obtain a general license or permission for the use of such proprietary rights by implementers or users of this specification can be obtained from the IETF on-line IPR repository at http://www.ietf.org/ipr.
The IETF invites any interested party to bring to its attention any copyrights, patents or patent applications, or other proprietary rights that may cover technology that may be required to implement this standard. Please address the information to the IETF at ietf- ipr@ietf.org.
Acknowledgement
-
Funding for the RFC Editor function is currently provided by the Internet Society.