From: Michael Zuo (mzuo@ixiacom.com)
Date: Mon Oct 30 2006 - 22:18:17 ART
Hi Ivan,
For third one (multicast rate-limiting), MQC with bandwidth would work
just as well?
thanks
-----Original Message-----
From: nobody@groupstudy.com [mailto:nobody@groupstudy.com] On Behalf Of
Ivan
Sent: Monday, October 30, 2006 10:58 AM
To: ccielab@groupstudy.com; Joe Carr
Subject: Re: QoS Questions
comment inline
On Monday 30 October 2006 20:32, Joe Carr wrote:
> Generator Microsoft Word 11 (filtered medium) I have been studying for
my
> CCIE lab for some time now and I came across a couple QoS scenario
> questions that I had some trouble with. Let me know what you guys
think:
>
>
> Enable fragmentation on a point to point frame link so that the
maximum
> fragmented size of each packet will be 60 and voice packets do not get
> fragmented.
>
This question discussed earlier many more times.
*) priority queue (LLQ) doesn't fragment any packet while sending.
*) try to make voice packet less than 60 byte (ip header compress, rtp
header
compress)
>
>
> configure random detect on an interface for outgoing traffic so that
the
> following policies are met. Default traffic should be flow based.
> Bandwidth http: 1000
> Bandwidth tftp: 500
>
class-map match-all tftp
match protocol tftp
class-map match-all http
match protocol http
!
!
policy-map child
class http
bandwidth 1000
class tftp
bandwidth 500
policy-map main
class class-default
fair-queue
random-detect
service-policy child
R2#sh policy-map
Policy Map main
Class class-default
Flow based Fair Queueing
Bandwidth 0 (kbps)
exponential weight 9
class min-threshold max-threshold mark-probablity
----------------------------------------------------------
0 - - 1/10
1 - - 1/10
2 - - 1/10
3 - - 1/10
4 - - 1/10
5 - - 1/10
6 - - 1/10
7 - - 1/10
rsvp - - 1/10
service-policy child
Policy Map ex
Policy Map child
Class http
Bandwidth 1000 (kbps) Max Threshold 64 (packets)
Class tftp
Bandwidth 500 (kbps) Max Threshold 64 (packets)
>
> Rate limit multicast traffic on interface Fe0/0 inbound to that
multicast
> group 239.5.5.5 gets 100k of bandwidth and all other multicast gets
600k.
>
R2#sh run int s1/0
Building configuration...
Current configuration : 241 bytes
!
interface Serial1/0
ip address 10.23.20.1 255.255.255.0
ip pim sparse-dense-mode
ip multicast rate-limit out group-list 1 100
ip multicast rate-limit out 600
encapsulation ppp
ip igmp join-group 239.5.5.5
serial restart-delay 0
end
R2#sh ip mroute int s1/0
IP Multicast Routing Table
Flags: D - Dense, S - Sparse, B - Bidir Group, s - SSM Group, C -
Connected,
L - Local, P - Pruned, R - RP-bit set, F - Register flag,
T - SPT-bit set, J - Join SPT, M - MSDP created entry,
X - Proxy Join Timer Running, A - Candidate for MSDP
Advertisement,
U - URD, I - Received Source Specific Host Report,
Z - Multicast Tunnel, z - MDT-data group sender,
Y - Joined MDT-data group, y - Sending to MDT-data group
Outgoing interface flags: H - Hardware switched, A - Assert winner
Timers: Uptime/Expires
Interface state: Interface, Next-Hop or VCD, State/Mode
(*, 239.5.5.5), 00:05:09/00:02:29, RP 0.0.0.0, flags: DCL
Incoming interface: Null, RPF nbr 0.0.0.0
Outgoing interface list:
Serial1/0, Forward/Sparse-Dense, 00:05:09/00:00:00, limit 100 kbps
<-----
(*, 224.0.1.40), 00:05:10/00:02:28, RP 0.0.0.0, flags: DCL
Incoming interface: Null, RPF nbr 0.0.0.0
Outgoing interface list:
Serial1/0, Forward/Sparse-Dense, 00:05:10/00:00:00, Int limit 600
kbps <--
>
>
>
>
> Joe
>
>
This archive was generated by hypermail 2.1.4 : Wed Nov 01 2006 - 07:29:07 ART