1#!/bin/bash 2# SPDX-License-Identifier: GPL-2.0 3# 4# A test for strict prioritization of traffic in the switch. Run two streams of 5# traffic, each through a different ingress port, one tagged with PCP of 1, the 6# other with PCP of 2. Both streams converge at one egress port, where they are 7# assigned TC of, respectively, 1 and 2, with strict priority configured between 8# them. In H3, we expect to see (almost) exclusively the high-priority traffic. 9# 10# Please see qos_mc_aware.sh for an explanation of why we use mausezahn and 11# counters instead of just running iperf3. 12# 13# +---------------------------+ +-----------------------------+ 14# | H1 | | H2 | 15# | $h1.111 + | | + $h2.222 | 16# | 192.0.2.33/28 | | | | 192.0.2.65/28 | 17# | e-qos-map 0:1 | | | | e-qos-map 0:2 | 18# | | | | | | 19# | $h1 + | | + $h2 | 20# +-----------------|---------+ +---------|-------------------+ 21# | | 22# +-----------------|-------------------------------------|-------------------+ 23# | $swp1 + + $swp2 | 24# | >1Gbps | | >1Gbps | 25# | +---------------|-----------+ +----------|----------------+ | 26# | | $swp1.111 + | | + $swp2.222 | | 27# | | BR111 | SW | BR222 | | 28# | | $swp3.111 + | | + $swp3.222 | | 29# | +---------------|-----------+ +----------|----------------+ | 30# | \_____________________________________/ | 31# | | | 32# | + $swp3 | 33# | | 1Gbps bottleneck | 34# | | ETS: (up n->tc n for n in 0..7) | 35# | | strict priority | 36# +------------------------------------|--------------------------------------+ 37# | 38# +--------------------|--------------------+ 39# | + $h3 H3 | 40# | / \ | 41# | / \ | 42# | $h3.111 + + $h3.222 | 43# | 192.0.2.34/28 192.0.2.66/28 | 44# +-----------------------------------------+ 45 46ALL_TESTS=" 47 ping_ipv4 48 test_ets_strict 49" 50 51lib_dir=$(dirname $0)/../../../net/forwarding 52 53NUM_NETIFS=6 54source $lib_dir/lib.sh 55source $lib_dir/devlink_lib.sh 56source qos_lib.sh 57 58h1_create() 59{ 60 simple_if_init $h1 61 mtu_set $h1 10000 62 63 vlan_create $h1 111 v$h1 192.0.2.33/28 64 ip link set dev $h1.111 type vlan egress-qos-map 0:1 65} 66 67h1_destroy() 68{ 69 vlan_destroy $h1 111 70 71 mtu_restore $h1 72 simple_if_fini $h1 73} 74 75h2_create() 76{ 77 simple_if_init $h2 78 mtu_set $h2 10000 79 80 vlan_create $h2 222 v$h2 192.0.2.65/28 81 ip link set dev $h2.222 type vlan egress-qos-map 0:2 82} 83 84h2_destroy() 85{ 86 vlan_destroy $h2 222 87 88 mtu_restore $h2 89 simple_if_fini $h2 90} 91 92h3_create() 93{ 94 simple_if_init $h3 95 mtu_set $h3 10000 96 97 vlan_create $h3 111 v$h3 192.0.2.34/28 98 vlan_create $h3 222 v$h3 192.0.2.66/28 99} 100 101h3_destroy() 102{ 103 vlan_destroy $h3 222 104 vlan_destroy $h3 111 105 106 mtu_restore $h3 107 simple_if_fini $h3 108} 109 110switch_create() 111{ 112 ip link set dev $swp1 up 113 mtu_set $swp1 10000 114 115 ip link set dev $swp2 up 116 mtu_set $swp2 10000 117 118 # prio n -> TC n, strict scheduling 119 lldptool -T -i $swp3 -V ETS-CFG up2tc=0:0,1:1,2:2,3:3,4:4,5:5,6:6,7:7 120 lldptool -T -i $swp3 -V ETS-CFG tsa=$( 121 )"0:strict,"$( 122 )"1:strict,"$( 123 )"2:strict,"$( 124 )"3:strict,"$( 125 )"4:strict,"$( 126 )"5:strict,"$( 127 )"6:strict,"$( 128 )"7:strict" 129 sleep 1 130 131 ip link set dev $swp3 up 132 mtu_set $swp3 10000 133 tc qdisc replace dev $swp3 root handle 101: tbf rate 1gbit \ 134 burst 128K limit 1G 135 136 vlan_create $swp1 111 137 vlan_create $swp2 222 138 vlan_create $swp3 111 139 vlan_create $swp3 222 140 141 ip link add name br111 up type bridge vlan_filtering 0 142 ip link set dev $swp1.111 master br111 143 ip link set dev $swp3.111 master br111 144 145 ip link add name br222 up type bridge vlan_filtering 0 146 ip link set dev $swp2.222 master br222 147 ip link set dev $swp3.222 master br222 148 149 # Make sure that ingress quotas are smaller than egress so that there is 150 # room for both streams of traffic to be admitted to shared buffer. 151 devlink_pool_size_thtype_save 0 152 devlink_pool_size_thtype_set 0 dynamic 10000000 153 devlink_pool_size_thtype_save 4 154 devlink_pool_size_thtype_set 4 dynamic 10000000 155 156 devlink_port_pool_th_save $swp1 0 157 devlink_port_pool_th_set $swp1 0 6 158 devlink_tc_bind_pool_th_save $swp1 1 ingress 159 devlink_tc_bind_pool_th_set $swp1 1 ingress 0 6 160 161 devlink_port_pool_th_save $swp2 0 162 devlink_port_pool_th_set $swp2 0 6 163 devlink_tc_bind_pool_th_save $swp2 2 ingress 164 devlink_tc_bind_pool_th_set $swp2 2 ingress 0 6 165 166 devlink_tc_bind_pool_th_save $swp3 1 egress 167 devlink_tc_bind_pool_th_set $swp3 1 egress 4 7 168 devlink_tc_bind_pool_th_save $swp3 2 egress 169 devlink_tc_bind_pool_th_set $swp3 2 egress 4 7 170 devlink_port_pool_th_save $swp3 4 171 devlink_port_pool_th_set $swp3 4 7 172} 173 174switch_destroy() 175{ 176 devlink_port_pool_th_restore $swp3 4 177 devlink_tc_bind_pool_th_restore $swp3 2 egress 178 devlink_tc_bind_pool_th_restore $swp3 1 egress 179 180 devlink_tc_bind_pool_th_restore $swp2 2 ingress 181 devlink_port_pool_th_restore $swp2 0 182 183 devlink_tc_bind_pool_th_restore $swp1 1 ingress 184 devlink_port_pool_th_restore $swp1 0 185 186 devlink_pool_size_thtype_restore 4 187 devlink_pool_size_thtype_restore 0 188 189 ip link del dev br222 190 ip link del dev br111 191 192 vlan_destroy $swp3 222 193 vlan_destroy $swp3 111 194 vlan_destroy $swp2 222 195 vlan_destroy $swp1 111 196 197 tc qdisc del dev $swp3 root handle 101: 198 mtu_restore $swp3 199 ip link set dev $swp3 down 200 lldptool -T -i $swp3 -V ETS-CFG up2tc=0:0,1:0,2:0,3:0,4:0,5:0,6:0,7:0 201 202 mtu_restore $swp2 203 ip link set dev $swp2 down 204 205 mtu_restore $swp1 206 ip link set dev $swp1 down 207} 208 209setup_prepare() 210{ 211 h1=${NETIFS[p1]} 212 swp1=${NETIFS[p2]} 213 214 swp2=${NETIFS[p3]} 215 h2=${NETIFS[p4]} 216 217 swp3=${NETIFS[p5]} 218 h3=${NETIFS[p6]} 219 220 h3mac=$(mac_get $h3) 221 222 vrf_prepare 223 224 h1_create 225 h2_create 226 h3_create 227 switch_create 228} 229 230cleanup() 231{ 232 pre_cleanup 233 234 switch_destroy 235 h3_destroy 236 h2_destroy 237 h1_destroy 238 239 vrf_cleanup 240} 241 242ping_ipv4() 243{ 244 ping_test $h1 192.0.2.34 " from H1" 245 ping_test $h2 192.0.2.66 " from H2" 246} 247 248rel() 249{ 250 local old=$1; shift 251 local new=$1; shift 252 253 bc <<< " 254 scale=2 255 ret = 100 * $new / $old 256 if (ret > 0) { ret } else { 0 } 257 " 258} 259 260test_ets_strict() 261{ 262 RET=0 263 264 # Run high-prio traffic on its own. 265 start_traffic $h2.222 192.0.2.65 192.0.2.66 $h3mac 266 local -a rate_2 267 rate_2=($(measure_rate $swp2 $h3 rx_octets_prio_2 "prio 2")) 268 check_err $? "Could not get high enough prio-2 ingress rate" 269 local rate_2_in=${rate_2[0]} 270 local rate_2_eg=${rate_2[1]} 271 stop_traffic # $h2.222 272 273 # Start low-prio stream. 274 start_traffic $h1.111 192.0.2.33 192.0.2.34 $h3mac 275 276 local -a rate_1 277 rate_1=($(measure_rate $swp1 $h3 rx_octets_prio_1 "prio 1")) 278 check_err $? "Could not get high enough prio-1 ingress rate" 279 local rate_1_in=${rate_1[0]} 280 local rate_1_eg=${rate_1[1]} 281 282 # High-prio and low-prio on their own should have about the same 283 # throughput. 284 local rel21=$(rel $rate_1_eg $rate_2_eg) 285 check_err $(bc <<< "$rel21 < 95") 286 check_err $(bc <<< "$rel21 > 105") 287 288 # Start the high-prio stream--now both streams run. 289 start_traffic $h2.222 192.0.2.65 192.0.2.66 $h3mac 290 rate_3=($(measure_rate $swp2 $h3 rx_octets_prio_2 "prio 2 w/ 1")) 291 check_err $? "Could not get high enough prio-2 ingress rate with prio-1" 292 local rate_3_in=${rate_3[0]} 293 local rate_3_eg=${rate_3[1]} 294 stop_traffic # $h2.222 295 296 stop_traffic # $h1.111 297 298 # High-prio should have about the same throughput whether or not 299 # low-prio is in the system. 300 local rel32=$(rel $rate_2_eg $rate_3_eg) 301 check_err $(bc <<< "$rel32 < 95") 302 303 log_test "strict priority" 304 echo "Ingress to switch:" 305 echo " p1 in rate $(humanize $rate_1_in)" 306 echo " p2 in rate $(humanize $rate_2_in)" 307 echo " p2 in rate w/ p1 $(humanize $rate_3_in)" 308 echo "Egress from switch:" 309 echo " p1 eg rate $(humanize $rate_1_eg)" 310 echo " p2 eg rate $(humanize $rate_2_eg) ($rel21% of p1)" 311 echo " p2 eg rate w/ p1 $(humanize $rate_3_eg) ($rel32% of p2)" 312} 313 314trap cleanup EXIT 315 316setup_prepare 317setup_wait 318 319tests_run 320 321exit $EXIT_STATUS 322