1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
1001
1002
1003
1004
1005
1006
1007
1008
1009
1010
1011
1012
1013
1014
1015
1016
1017
1018
1019
1020
1021
1022
1023
1024
1025
1026
1027
1028
1029
1030
1031
1032
1033
1034
1035
1036
1037
1038
1039
1040
1041
1042
1043
1044
1045
1046
1047
1048
1049
1050
1051
1052
1053
1054
1055
1056
1057
1058
1059
1060
1061
1062
1063
1064
1065
1066
1067
1068
1069
1070
1071
1072
1073
1074
1075
1076
1077
1078
1079
1080
1081
1082
1083
1084
1085
1086
1087
1088
1089
1090
1091
1092
1093
1094
1095
1096
1097
1098
1099
1100
1101
1102
1103
1104
1105
1106
1107
1108
1109
1110
1111
1112
1113
1114
1115
1116
1117
1118
1119
1120
1121
1122
1123
1124
1125
1126
1127
1128
1129
1130
1131
1132
1133
1134
1135
1136
1137
1138
1139
1140
1141
1142
1143
1144
1145
1146
1147
1148
1149
1150
1151
1152
1153
1154
1155
1156
1157
1158
1159
1160
1161
1162
1163
1164
1165
1166
1167
1168
1169
1170
1171
1172
1173
1174
1175
1176
1177
1178
1179
1180
1181
1182
1183
1184
1185
1186
1187
1188
1189
1190
1191
1192
1193
1194
1195
1196
1197
1198
1199
1200
1201
1202
1203
1204
1205
1206
1207
1208
1209
1210
1211
1212
1213
1214
1215
1216
1217
1218
1219
1220
1221
1222
1223
1224
1225
1226
1227
1228
1229
1230
1231
1232
1233
1234
1235
1236
1237
1238
1239
1240
1241
1242
1243
1244
1245
1246
1247
1248
1249
1250
1251
1252
1253
1254
1255
1256
1257
1258
1259
1260
1261
1262
1263
1264
1265
1266
1267
1268
1269
1270
1271
1272
1273
1274
1275
1276
1277
1278
1279
1280
1281
1282
1283
1284
1285
1286
1287
1288
1289
1290
1291
1292
1293
1294
1295
1296
1297
1298
1299
1300
1301
1302
1303
1304
1305
1306
1307
1308
1309
1310
1311
1312
1313
1314
1315
1316
1317
1318
1319
1320
1321
1322
1323
1324
1325
1326
1327
1328
1329
1330
1331
1332
1333
1334
1335
1336
1337
1338
1339
1340
1341
1342
1343
1344
1345
1346
1347
1348
1349
1350
1351
1352
1353
1354
1355
1356
1357
1358
1359
1360
1361
1362
1363
1364
1365
1366
1367
1368
1369
1370
1371
1372
1373
1374
1375
1376
1377
1378
1379
1380
1381
1382
1383
1384
1385
1386
1387
1388
1389
1390
1391
1392
1393
1394
1395
1396
1397
1398
1399
1400
1401
1402
1403
1404
1405
1406
1407
1408
1409
1410
1411
1412
1413
1414
1415
1416
1417
1418
1419
1420
1421
1422
1423
1424
1425
1426
1427
1428
1429
1430
1431
1432
1433
1434
1435
1436
1437
1438
1439
1440
1441
1442
1443
1444
1445
1446
1447
1448
1449
1450
1451
1452
1453
1454
1455
1456
1457
1458
1459
1460
1461
1462
1463
1464
1465
1466
1467
1468
1469
1470
1471
1472
1473
1474
1475
1476
1477
1478
1479
1480
1481
1482
1483
1484
1485
1486
1487
1488
1489
1490
1491
1492
1493
1494
1495
1496
1497
1498
1499
1500
1501
1502
1503
1504
1505
1506
1507
1508
1509
1510
1511
1512
1513
1514
1515
1516
1517
1518
1519
1520
1521
1522
1523
1524
1525
1526
1527
1528
1529
1530
1531
1532
1533
1534
1535
1536
1537
1538
1539
1540
1541
1542
1543
1544
1545
1546
1547
1548
1549
1550
1551
1552
1553
1554
1555
1556
1557
1558
1559
1560
1561
1562
1563
1564
1565
1566
1567
1568
1569
1570
1571
1572
1573
1574
1575
1576
1577
1578
1579
1580
1581
1582
1583
1584
1585
1586
1587
1588
1589
1590
1591
1592
1593
1594
1595
1596
1597
1598
1599
1600
1601
1602
1603
1604
1605
1606
1607
1608
1609
1610
1611
1612
1613
1614
1615
1616
1617
1618
1619
1620
1621
1622
1623
1624
1625
1626
1627
1628
1629
1630
1631
1632
1633
1634
1635
1636
1637
1638
1639
1640
1641
1642
1643
1644
1645
1646
1647
1648
1649
1650
1651
1652
1653
1654
1655
1656
1657
1658
1659
1660
1661
1662
1663
1664
1665
1666
1667
1668
1669
1670
1671
1672
1673
1674
1675
1676
1677
1678
1679
1680
1681
1682
1683
1684
1685
1686
1687
1688
1689
1690
1691
1692
1693
1694
1695
1696
1697
1698
1699
1700
1701
1702
1703
1704
1705
1706
1707
1708
1709
1710
1711
1712
1713
1714
1715
1716
1717
1718
1719
1720
1721
1722
1723
1724
|
Internet Architecture Board (IAB) W. Hardaker
Request for Comments: 9318
Category: Informational O. Shapira
ISSN: 2070-1721 October 2022
IAB Workshop Report: Measuring Network Quality for End-Users
Abstract
The Measuring Network Quality for End-Users workshop was held
virtually by the Internet Architecture Board (IAB) on September
14-16, 2021. This report summarizes the workshop, the topics
discussed, and some preliminary conclusions drawn at the end of the
workshop.
Note that this document is a report on the proceedings of the
workshop. The views and positions documented in this report are
those of the workshop participants and do not necessarily reflect IAB
views and positions.
Status of This Memo
This document is not an Internet Standards Track specification; it is
published for informational purposes.
This document is a product of the Internet Architecture Board (IAB)
and represents information that the IAB has deemed valuable to
provide for permanent record. It represents the consensus of the
Internet Architecture Board (IAB). Documents approved for
publication by the IAB are not candidates for any level of Internet
Standard; see Section 2 of RFC 7841.
Information about the current status of this document, any errata,
and how to provide feedback on it may be obtained at
https://www.rfc-editor.org/info/rfc9318.
Copyright Notice
Copyright (c) 2022 IETF Trust and the persons identified as the
document authors. All rights reserved.
This document is subject to BCP 78 and the IETF Trust's Legal
Provisions Relating to IETF Documents
(https://trustee.ietf.org/license-info) in effect on the date of
publication of this document. Please review these documents
carefully, as they describe your rights and restrictions with respect
to this document.
Table of Contents
1. Introduction
1.1. Problem Space
2. Workshop Agenda
3. Position Papers
4. Workshop Topics and Discussion
4.1. Introduction and Overviews
4.1.1. Key Points from the Keynote by Vint Cerf
4.1.2. Introductory Talks
4.1.3. Introductory Talks - Key Points
4.2. Metrics Considerations
4.2.1. Common Performance Metrics
4.2.2. Availability Metrics
4.2.3. Capacity Metrics
4.2.4. Latency Metrics
4.2.5. Measurement Case Studies
4.2.6. Metrics Key Points
4.3. Cross-Layer Considerations
4.3.1. Separation of Concerns
4.3.2. Security and Privacy Considerations
4.3.3. Metric Measurement Considerations
4.3.4. Towards Improving Future Cross-Layer Observability
4.3.5. Efficient Collaboration between Hardware and Transport
Protocols
4.3.6. Cross-Layer Key Points
4.4. Synthesis
4.4.1. Measurement and Metrics Considerations
4.4.2. End-User Metrics Presentation
4.4.3. Synthesis Key Points
5. Conclusions
5.1. General Statements
5.2. Specific Statements about Detailed Protocols/Techniques
5.3. Problem Statements and Concerns
5.4. No-Consensus-Reached Statements
6. Follow-On Work
7. IANA Considerations
8. Security Considerations
9. Informative References
Appendix A. Program Committee
Appendix B. Workshop Chairs
Appendix C. Workshop Participants
IAB Members at the Time of Approval
Acknowledgments
Contributors
Authors' Addresses
1. Introduction
The Internet Architecture Board (IAB) holds occasional workshops
designed to consider long-term issues and strategies for the
Internet, and to suggest future directions for the Internet
architecture. This long-term planning function of the IAB is
complementary to the ongoing engineering efforts performed by working
groups of the Internet Engineering Task Force (IETF).
The Measuring Network Quality for End-Users workshop [WORKSHOP] was
held virtually by the Internet Architecture Board (IAB) on September
14-16, 2021. This report summarizes the workshop, the topics
discussed, and some preliminary conclusions drawn at the end of the
workshop.
1.1. Problem Space
The Internet in 2021 is quite different from what it was 10 years
ago. Today, it is a crucial part of everyone's daily life. People
use the Internet for their social life, for their daily jobs, for
routine shopping, and for keeping up with major events. An
increasing number of people can access a gigabit connection, which
would be hard to imagine a decade ago. Additionally, thanks to
improvements in security, people trust the Internet for financial
banking transactions, purchasing goods, and everyday bill payments.
At the same time, some aspects of the end-user experience have not
improved as much. Many users have typical connection latencies that
remain at decade-old levels. Despite significant reliability
improvements in data center environments, end users also still often
see interruptions in service. Despite algorithmic advances in the
field of control theory, one still finds that the queuing delays in
the last-mile equipment exceeds the accumulated transit delays.
Transport improvements, such as QUIC, Multipath TCP, and TCP Fast
Open, are still not fully supported in some networks. Likewise,
various advances in the security and privacy of user data are not
widely supported, such as encrypted DNS to the local resolver.
Some of the major factors behind this lack of progress is the popular
perception that throughput is often the sole measure of the quality
of Internet connectivity. With such a narrow focus, the Measuring
Network Quality for End-Users workshop aimed to discuss various
topics:
* What is user latency under typical working conditions?
* How reliable is connectivity across longer time periods?
* Do networks allow the use of a broad range of protocols?
* What services can be run by network clients?
* What kind of IPv4, NAT, or IPv6 connectivity is offered, and are
there firewalls?
* What security mechanisms are available for local services, such as
DNS?
* To what degree are the privacy, confidentiality, integrity, and
authenticity of user communications guarded?
* Improving these aspects of network quality will likely depend on
measuring and exposing metrics in a meaningful way to all involved
parties, including to end users. Such measurement and exposure of
the right metrics will allow service providers and network
operators to concentrate focus on their users' experience and will
simultaneously empower users to choose the Internet Service
Providers (ISPs) that can deliver the best experience based on
their needs.
* What are the fundamental properties of a network that contributes
to a good user experience?
* What metrics quantify these properties, and how can we collect
such metrics in a practical way?
* What are the best practices for interpreting those metrics and
incorporating them in a decision-making process?
* What are the best ways to communicate these properties to service
providers and network operators?
* How can these metrics be displayed to users in a meaningful way?
2. Workshop Agenda
The Measuring Network Quality for End-Users workshop was divided into
the following main topic areas; see further discussion in Sections 4
and 5:
* Introduction overviews and a keynote by Vint Cerf
* Metrics considerations
* Cross-layer considerations
* Synthesis
* Group conclusions
3. Position Papers
The following position papers were received for consideration by the
workshop attendees. The workshop's web page [WORKSHOP] contains
archives of the papers, presentations, and recorded videos.
* Ahmed Aldabbagh. "Regulatory perspective on measuring network
quality for end users" [Aldabbagh2021]
* Al Morton. "Dream-Pipe or Pipe-Dream: What Do Users Want (and how
can we assure it)?" [Morton2021]
* Alexander Kozlov. "The 2021 National Internet Segment Reliability
Research"
* Anna Brunstrom. "Measuring network quality - the MONROE
experience"
* Bob Briscoe, Greg White, Vidhi Goel, and Koen De Schepper. "A
Single Common Metric to Characterize Varying Packet Delay"
[Briscoe2021]
* Brandon Schlinker. "Internet Performance from Facebook's Edge"
[Schlinker2019]
* Christoph Paasch, Kristen McIntyre, Randall Meyer, Stuart
Cheshire, and Omer Shapira. "An end-user approach to the Internet
Score" [McIntyre2021]
* Christoph Paasch, Randall Meyer, Stuart Cheshire, and Omer
Shapira. "Responsiveness under Working Conditions" [Paasch2021]
* Dave Reed and Levi Perigo. "Measuring ISP Performance in
Broadband America: A Study of Latency Under Load" [Reed2021]
* Eve M. Schooler and Rick Taylor. "Non-traditional Network
Metrics"
* Gino Dion. "Focusing on latency, not throughput, to provide
better internet experience and network quality" [Dion2021]
* Gregory Mirsky, Xiao Min, Gyan Mishra, and Liuyan Han. "The error
performance metric in a packet-switched network" [Mirsky2021]
* Jana Iyengar. "The Internet Exists In Its Use" [Iyengar2021]
* Jari Arkko and Mirja Kuehlewind. "Observability is needed to
improve network quality" [Arkko2021]
* Joachim Fabini. "Network Quality from an End User Perspective"
[Fabini2021]
* Jonathan Foulkes. "Metrics helpful in assessing Internet Quality"
[Foulkes2021]
* Kalevi Kilkki and Benajamin Finley. "In Search of Lost QoS"
[Kilkki2021]
* Karthik Sundaresan, Greg White, and Steve Glennon. "Latency
Measurement: What is latency and how do we measure it?"
* Keith Winstein. "Five Observations on Measuring Network Quality
for Users of Real-Time Media Applications"
* Ken Kerpez, Jinous Shafiei, John Cioffi, Pete Chow, and Djamel
Bousaber. "Wi-Fi and Broadband Data" [Kerpez2021]
* Kenjiro Cho. "Access Network Quality as Fitness for Purpose"
* Koen De Schepper, Olivier Tilmans, and Gino Dion. "Challenges and
opportunities of hardware support for Low Queuing Latency without
Packet Loss" [DeSchepper2021]
* Kyle MacMillian and Nick Feamster. "Beyond Speed Test: Measuring
Latency Under Load Across Different Speed Tiers" [MacMillian2021]
* Lucas Pardue and Sreeni Tellakula. "Lower-layer performance not
indicative of upper-layer success" [Pardue2021]
* Matt Mathis. "Preliminary Longitudinal Study of Internet
Responsiveness" [Mathis2021]
* Michael Welzl. "A Case for Long-Term Statistics" [Welzl2021]
* Mikhail Liubogoshchev. "Cross-layer Cooperation for Better
Network Service" [Liubogoshchev2021]
* Mingrui Zhang, Vidhi Goel, and Lisong Xu. "User-Perceived Latency
to Measure CCAs" [Zhang2021]
* Neil Davies and Peter Thompson. "Measuring Network Impact on
Application Outcomes Using Quality Attenuation" [Davies2021]
* Olivier Bonaventure and Francois Michel. "Packet delivery time as
a tie-breaker for assessing Wi-Fi access points" [Michel2021]
* Pedro Casas. "10 Years of Internet-QoE Measurements. Video,
Cloud, Conferencing, Web and Apps. What do we Need from the
Network Side?" [Casas2021]
* Praveen Balasubramanian. "Transport Layer Statistics for Network
Quality" [Balasubramanian2021]
* Rajat Ghai. "Using TCP Connect Latency for measuring CX and
Network Optimization" [Ghai2021]
* Robin Marx and Joris Herbots. "Merge Those Metrics: Towards
Holistic (Protocol) Logging" [Marx2021]
* Sandor Laki, Szilveszter Nadas, Balazs Varga, and Luis M.
Contreras. "Incentive-Based Traffic Management and QoS
Measurements" [Laki2021]
* Satadal Sengupta, Hyojoon Kim, and Jennifer Rexford. "Fine-
Grained RTT Monitoring Inside the Network" [Sengupta2021]
* Stuart Cheshire. "The Internet is a Shared Network"
[Cheshire2021]
* Toerless Eckert and Alex Clemm. "network-quality-eckert-clemm-
00.4"
* Vijay Sivaraman, Sharat Madanapalli, and Himal Kumar. "Measuring
Network Experience Meaningfully, Accurately, and Scalably"
[Sivaraman2021]
* Yaakov (J) Stein. "The Futility of QoS" [Stein2021]
4. Workshop Topics and Discussion
The agenda for the three-day workshop was broken into four separate
sections that each played a role in framing the discussions. The
workshop started with a series of introduction and problem space
presentations (Section 4.1), followed by metrics considerations
(Section 4.2), cross-layer considerations (Section 4.3), and a
synthesis discussion (Section 4.4). After the four subsections
concluded, a follow-on discussion was held to draw conclusions that
could be agreed upon by workshop participants (Section 5).
4.1. Introduction and Overviews
The workshop started with a broad focus on the state of user Quality
of Service (QoS) and Quality of Experience (QoE) on the Internet
today. The goal of the introductory talks was to set the stage for
the workshop by describing both the problem space and the current
solutions in place and their limitations.
The introduction presentations provided views of existing QoS and QoE
measurements and their effectiveness. Also discussed was the
interaction between multiple users within the network, as well as the
interaction between multiple layers of the OSI stack. Vint Cerf
provided a keynote describing the history and importance of the
topic.
4.1.1. Key Points from the Keynote by Vint Cerf
We may be operating in a networking space with dramatically different
parameters compared to 30 years ago. This differentiation justifies
reconsidering not only the importance of one metric over the other
but also reconsidering the entire metaphor.
It is time for the experts to look at not only adjusting TCP but also
exploring other protocols, such as QUIC has done lately. It's
important that we feel free to consider alternatives to TCP. TCP is
not a teddy bear, and one should not be afraid to replace it with a
transport layer with better properties that better benefit its users.
A suggestion: we should consider exercises to identify desirable
properties. As we are looking at the parametric spaces, one can
identify "desirable properties", as opposed to "fundamental
properties", for example, a low-latency property. An example coming
from the Advanced Research Projects Agency (ARPA): you want to know
where the missile is now, not where it was. Understanding drives
particular parameter creation and selection in the design space.
When parameter values are changed in extreme, such as connectiveness,
alternative designs will emerge. One case study of note is the
interplanetary protocol, where "ping" is no longer indicative of
anything useful. While we look at responsiveness, we should not
ignore connectivity.
Unfortunately, maintaining backward compatibility is painful. The
work on designing IPv6 so as to transition from IPv4 could have been
done better if the backward compatibility was considered. It is too
late for IPv6, but it is not too late to consider this issue for
potential future problems.
IPv6 is still not implemented fully everywhere. It's been a long
road to deployment since starting work in 1996, and we are still not
there. In 1996, the thinking was that it was quite easy to implement
IPv6, but that failed to hold true. In 1996, the dot-com boom began,
where a lot of money was spent quickly, and the moment was not caught
in time while the market expanded exponentially. This should serve
as a cautionary tale.
One last point: consider performance across multiple hops in the
Internet. We've not seen many end-to-end metrics, as successfully
developing end-to-end measurements across different network and
business boundaries is quite hard to achieve. A good question to ask
when developing new protocols is "will the new protocol work across
multiple network hops?"
Multi-hop networks are being gradually replaced by humongous, flat
networks with sufficient connectivity between operators so that
systems become 1 hop, or 2 hops at most, away from each other (e.g.,
Google, Facebook, and Amazon). The fundamental architecture of the
Internet is changing.
4.1.2. Introductory Talks
The Internet is a shared network built on IP protocols using packet
switching to interconnect multiple autonomous networks. The
Internet's departure from circuit-switching technologies allowed it
to scale beyond any other known network design. On the other hand,
the lack of in-network regulation made it difficult to ensure the
best experience for every user.
As Internet use cases continue to expand, it becomes increasingly
more difficult to predict which network characteristics correlate
with better user experiences. Different application classes, e.g.,
video streaming and teleconferencing, can affect user experience in
ways that are complex and difficult to measure. Internet utilization
shifts rapidly during the course of each day, week, and year, which
further complicates identifying key metrics capable of predicting a
good user experience.
QoS initiatives attempted to overcome these difficulties by strictly
prioritizing different types of traffic. However, QoS metrics do not
always correlate with user experience. The utility of the QoS metric
is further limited by the difficulties in building solutions with the
desired QoS characteristics.
QoE initiatives attempted to integrate the psychological aspects of
how quality is perceived and create statistical models designed to
optimize the user experience. Despite these high modeling efforts,
the QoE approach proved beneficial in certain application classes.
Unfortunately, generalizing the models proved to be difficult, and
the question of how different applications affect each other when
sharing the same network remains an open problem.
The industry's focus on giving the end user more throughput/bandwidth
led to remarkable advances. In many places around the world, a home
user enjoys gigabit speeds to their ISP. This is so remarkable that
it would have been brushed off as science fiction a decade ago.
However, the focus on increased capacity came at the expense of
neglecting another important core metric: latency. As a result, end
users whose experience is negatively affected by high latency were
advised to upgrade their equipment to get more throughput instead.
[MacMillian2021] showed that sometimes such an upgrade can lead to
latency improvements, due to the economical reasons of overselling
the "value-priced" data plans.
As the industry continued to give end users more throughput, while
mostly neglecting latency concerns, application designs started to
employ various latency and short service disruption hiding
techniques. For example, a user's web browser performance experience
is closely tied to the content in the browser's local cache. While
such techniques can clearly improve the user experience when using
stale data is possible, this development further decouples user
experience from core metrics.
In the most recent 10 years, efforts by Dave Taht and the bufferbloat
society have led to significant progress in updating queuing
algorithms to reduce latencies under load compared to simpler FIFO
queues. Unfortunately, the home router industry has yet to implement
these algorithms, mostly due to marketing and cost concerns. Most
home router manufacturers depend on System on a Chip (SoC)
acceleration to create products with a desired throughput. SoC
manufacturers opt for simpler algorithms and aggressive aggregation,
reasoning that a higher-throughput chip will have guaranteed demand.
Because consumers are offered choices primarily among different high-
throughput devices, the perception that a higher throughput leads to
higher a QoS continues to strengthen.
The home router is not the only place that can benefit from clearer
indications of acceptable performance for users. Since users
perceive the Internet via the lens of applications, it is important
that we call upon application vendors to adopt solutions that stress
lower latencies. Unfortunately, while bandwidth is straightforward
to measure, responsiveness is trickier. Many applications have found
a set of metrics that are helpful to their realm but do not
generalize well and cannot become universally applicable.
Furthermore, due to the highly competitive application space, vendors
may have economic reasons to avoid sharing their most useful metrics.
4.1.3. Introductory Talks - Key Points
1. Measuring bandwidth is necessary but is not alone sufficient.
2. In many cases, Internet users don't need more bandwidth but
rather need "better bandwidth", i.e., they need other
connectivity improvements.
3. Users perceive the quality of their Internet connection based on
the applications they use, which are affected by a combination of
factors. There's little value in exposing a typical user to the
entire spectrum of possible reasons for the poor performance
perceived in their application-centric view.
4. Many factors affecting user experience are outside the users'
sphere of control. It's unclear whether exposing users to these
other factors will help them understand the state of their
network performance. In general, users prefer simple,
categorical choices (e.g., "good", "better", and "best" options).
5. The Internet content market is highly competitive, and many
applications develop their own "secret sauce".
4.2. Metrics Considerations
In the second agenda section, the workshop continued its discussion
about metrics that can be used instead of or in addition to available
bandwidth. Several workshop attendees presented deep-dive studies on
measurement methodology.
4.2.1. Common Performance Metrics
Losing Internet access entirely is, of course, the worst user
experience. Unfortunately, unless rebooting the home router restores
connectivity, there is little a user can do other than contacting
their service provider. Nevertheless, there is value in the
systematic collection of availability metrics on the client side;
these can help the user's ISP localize and resolve issues faster
while enabling users to better choose between ISPs. One can measure
availability directly by simply attempting connections from the
client side to distant locations of interest. For example, Ookla's
[Speedtest] uses a large number of Android devices to measure network
and cellular availability around the globe. Ookla collects hundreds
of millions of data points per day and uses these for accurate
availability reporting. An alternative approach is to derive
availability from the failure rates of other tests. For example,
[FCC_MBA] and [FCC_MBA_methodology] use thousands of off-the-shelf
routers, with measurement software developed by [SamKnows]. These
routers perform an array of network tests and report availability
based on whether test connections were successful or not.
Measuring available capacity can be helpful to end users, but it is
even more valuable for service providers and application developers.
High-definition video streaming requires significantly more capacity
than any other type of traffic. At the time of the workshop, video
traffic constituted 90% of overall Internet traffic and contributed
to 95% of the revenues from monetization (via subscriptions, fees, or
ads). As a result, video streaming services, such as Netflix, need
to continuously cope with rapid changes in available capacity. The
ability to measure available capacity in real time leverages the
different adaptive bitrate (ABR) compression algorithms to ensure the
best possible user experience. Measuring aggregated capacity demand
allows ISPs to be ready for traffic spikes. For example, during the
end-of-year holiday season, the global demand for capacity has been
shown to be 5-7 times higher than during other seasons. For end
users, knowledge of their capacity needs can help them select the
best data plan given their intended usage. In many cases, however,
end users have more than enough capacity, and adding more bandwidth
will not improve their experience -- after a point, it is no longer
the limiting factor in user experience. Finally, the ability to
differentiate between the "throughput" and the "goodput" can be
helpful in identifying when the network is saturated.
In measuring network quality, latency is defined as the time it takes
a packet to traverse a network path from one end to the other. At
the time of this report, users in many places worldwide can enjoy
Internet access that has adequately high capacity and availability
for their current needs. For these users, latency improvements,
rather than bandwidth improvements, can lead to the most significant
improvements in QoE. The established latency metric is a round-trip
time (RTT), commonly measured in milliseconds. However, users often
find RTT values unintuitive since, unlike other performance metrics,
high RTT values indicate poor latency and users typically understand
higher scores to be better. To address this, [Paasch2021] and
[Mathis2021] present an inverse metric, called "Round-trips Per
Minute" (RPM).
There is an important distinction between "idle latency" and "latency
under working conditions". The former is measured when the network
is underused and reflects a best-case scenario. The latter is
measured when the network is under a typical workload. Until
recently, typical tools reported a network's idle latency, which can
be misleading. For example, data presented at the workshop shows
that idle latencies can be up to 25 times lower than the latency
under typical working loads. Because of this, it is essential to
make a clear distinction between the two when presenting latency to
end users.
Data shows that rapid changes in capacity affect latency.
[Foulkes2021] attempts to quantify how often a rapid change in
capacity can cause network connectivity to become "unstable" (i.e.,
having high latency with very little throughput). Such changes in
capacity can be caused by infrastructure failures but are much more
often caused by in-network phenomena, like changing traffic
engineering policies or rapid changes in cross-traffic.
Data presented at the workshop shows that 36% of measured lines have
capacity metrics that vary by more than 10% throughout the day and
across multiple days. These differences are caused by many
variables, including local connectivity methods (Wi-Fi vs. Ethernet),
competing LAN traffic, device load/configuration, time of day, and
local loop/backhaul capacity. These factor variations make measuring
capacity using only an end-user device or other end-network
measurement difficult. A network router seeing aggregated traffic
from multiple devices provides a better vantage point for capacity
measurements. Such a test can account for the totality of local
traffic and perform an independent capacity test. However, various
factors might still limit the accuracy of such a test. Accurate
capacity measurement requires multiple samples.
As users perceive the Internet through the lens of applications, it
may be difficult to correlate changes in capacity and latency with
the quality of the end-user experience. For example, web browsers
rely on cached page versions to shorten page load times and mitigate
connectivity losses. In addition, social networking applications
often rely on prefetching their "feed" items. These techniques make
the core in-network metrics less indicative of the users' experience
and necessitates collecting data from the end-user applications
themselves.
It is helpful to distinguish between applications that operate on a
"fixed latency budget" from those that have more tolerance to latency
variance. Cloud gaming serves as an example application that
requires a "fixed latency budget", as a sudden latency spike can
decide the "win/lose" ratio for a player. Companies that compete in
the lucrative cloud gaming market make significant infrastructure
investments, such as building entire data centers closer to their
users. These data centers highlight the economic benefit that lower
numbers of latency spikes outweigh the associated deployment costs.
On the other hand, applications that are more tolerant to latency
spikes can continue to operate reasonably well through short spikes.
Yet, even those applications can benefit from consistently low
latency depending on usage shifts. For example, Video-on-Demand
(VOD) apps can work reasonably well when the video is consumed
linearly, but once the user tries to "switch a channel" or to "skip
ahead", the user experience suffers unless the latency is
sufficiently low.
Finally, as applications continue to evolve, in-application metrics
are gaining in importance. For example, VOD applications can assess
the QoE by application-specific metrics, such as whether the video
player is able to use the highest possible resolution, identifying
when the video is smooth or freezing, or other similar metrics.
Application developers can then effectively use these metrics to
prioritize future work. All popular video platforms (YouTube,
Instagram, Netflix, and others) have developed frameworks to collect
and analyze VOD metrics at scale. One example is the Scuba framework
used by Meta [Scuba].
Unfortunately, in-application metrics can be challenging to use for
comparative research purposes. First, different applications often
use different metrics to measure the same phenomena. For example,
application A may measure the smoothness of video via "mean time to
rebuffer", while application B may rely on the "probability of
rebuffering per second" for the same purpose. A different challenge
with in-application metrics is that VOD is a significant source of
revenue for companies, such as YouTube, Facebook, and Netflix,
placing a proprietary incentive against exchanging the in-application
data. A final concern centers on the privacy issues resulting from
in-application metrics that accurately describe the activities and
preferences of an individual end user.
4.2.2. Availability Metrics
Availability is simply defined as whether or not a packet can be sent
and then received by its intended recipient. Availability is naively
thought to be the simplest to measure, but it is more complex when
considering that continual, instantaneous measurements would be
needed to detect the smallest of outages. Also difficult is
determining the root cause of infallibility: was the user's line
down, was something in the middle of the network, or was it the
service with which the user was attempting to communicate?
4.2.3. Capacity Metrics
If the network capacity does not meet user demands, the network
quality will be impacted. Once the capacity meets the demands,
increasing capacity won't lead to further quality improvements.
The actual network connection capacity is determined by the equipment
and the lines along the network path, and it varies throughout the
day and across multiple days. Studies involving DSL lines in North
America indicate that over 30% of the DSL lines have capacity metrics
that vary by more than 10% throughout the day and across multiple
days.
Some factors that affect the actual capacity are:
1. Presence of a competing traffic, either in the LAN or in the WAN
environments. In the LAN setting, the competing traffic reflects
the multiple devices that share the Internet connection. In the
WAN setting, the competing traffic often originates from the
unrelated network flows that happen to share the same network
path.
2. Capabilities of the equipment along the path of the network
connection, including the data transfer rate and the amount of
memory used for buffering.
3. Active traffic management measures, such as traffic shapers and
policers that are often used by the network providers.
There are other factors that can negatively affect the actual line
capacities.
The user demands of the traffic follow the usage patterns and
preferences of the particular users. For example, large data
transfers can use any available capacity, while the media streaming
applications require limited capacity to function correctly.
Videoconferencing applications typically need less capacity than
high-definition video streaming.
4.2.4. Latency Metrics
End-to-end latency is the time that a particular packet takes to
traverse the network path from the user to their destination and
back. The end-to-end latency comprises several components:
1. The propagation delay, which reflects the path distance and the
individual link technologies (e.g., fiber vs. satellite). The
propagation doesn't depend on the utilization of the network, to
the extent that the network path remains constant.
2. The buffering delay, which reflects the time segments spent in
the memory of the network equipment that connect the individual
network links, as well as in the memory of the transmitting
endpoint. The buffering delay depends on the network
utilization, as well as on the algorithms that govern the queued
segments.
3. The transport protocol delays, which reflect the time spent in
retransmission and reassembly, as well as the time spent when the
transport is "head-of-line blocked".
4. Some of the workshop submissions that have explicitly called out
the application delay, which reflects the inefficiencies in the
application layer.
Typically, end-to-end latency is measured when the network is idle.
Results of such measurements mostly reflect the propagation delay but
not other kinds of delay. This report uses the term "idle latency"
to refer to results achieved under idle network conditions.
Alternatively, if the latency is measured when the network is under
its typical working conditions, the results reflect multiple types of
delays. This report uses the term "working latency" to refer to such
results. Other sources use the term "latency under load" (LUL) as a
synonym.
Data presented at the workshop reveals a substantial difference
between the idle latency and the working latency. Depending on the
traffic direction and the technology type, the working latency is
between 6 to 25 times higher than the idle latency:
+============+============+========+=========+============+=========+
| Direction | Technology |Working | Idle | Working - |Working /|
| | Type |Latency | Latency | Idle |Idle |
| | | | | Difference |Ratio |
+============+============+========+=========+============+=========+
| Downstream | FTTH |148 | 10 | 138 |15 |
+------------+------------+--------+---------+------------+---------+
| Downstream | Cable |103 | 13 | 90 |8 |
+------------+------------+--------+---------+------------+---------+
| Downstream | DSL |194 | 10 | 184 |19 |
+------------+------------+--------+---------+------------+---------+
| Upstream | FTTH |207 | 12 | 195 |17 |
+------------+------------+--------+---------+------------+---------+
| Upstream | Cable |176 | 27 | 149 |6 |
+------------+------------+--------+---------+------------+---------+
| Upstream | DSL |686 | 27 | 659 |25 |
+------------+------------+--------+---------+------------+---------+
Table 1
While historically the tooling available for measuring latency
focused on measuring the idle latency, there is a trend in the
industry to start measuring the working latency as well, e.g.,
Apple's [NetworkQuality].
4.2.5. Measurement Case Studies
The participants have proposed several concrete methodologies for
measuring the network quality for the end users.
[Paasch2021] introduced a methodology for measuring working latency
from the end-user vantage point. The suggested method incrementally
adds network flows between the user device and a server endpoint
until a bottleneck capacity is reached. From these measurements, a
round-trip latency is measured and reported to the end user. The
authors chose to report results with the RPM metric. The methodology
had been implemented in Apple's macOS Monterey.
[Mathis2021] applied the RPM metric to the results of more than 4
billion download tests that M-Lab performed from 2010-2021. During
this time frame, the M-Lab measurement platform underwent several
upgrades that allowed the research team to compare the effect of
different TCP congestion control algorithms (CCAs) on the measured
end-to-end latency. The study showed that the use of cubic CCA leads
to increased working latency, which is attributed to its use of
larger queues.
[Schlinker2019] presented a large-scale study that aimed to establish
a correlation between goodput and QoE on a large social network. The
authors performed the measurements at multiple data centers from
which video segments of set sizes were streamed to a large number of
end users. The authors used the goodput and throughput metrics to
determine whether particular paths were congested.
[Reed2021] presented the analysis of working latency measurements
collected as part of the Measuring Broadband America (MBA) program by
the Federal Communication Commission (FCC). The FCC does not include
working latency in its yearly report but does offer it in the raw
data files. The authors used a subset of the raw data to identify
important differences in the working latencies across different ISPs.
[MacMillian2021] presented analysis of working latency across
multiple service tiers. They found that, unsurprisingly, "premium"
tier users experienced lower working latency compared to a "value"
tier. The data demonstrated that working latency varies
significantly within each tier; one possible explanation is the
difference in equipment deployed in the homes.
These studies have stressed the importance of measurement of working
latency. At the time of this report, many home router manufacturers
rely on hardware-accelerated routing that uses FIFO queues. Focusing
on measuring the working latency measurements on these devices and
making the consumer aware of the effect of choosing one manufacturer
vs. another can help improve the home router situation. The ideal
test would be able to identify the working latency and pinpoint the
source of the delay (home router, ISP, server side, or some network
node in between).
Another source of high working latency comes from network routers
exposed to cross-traffic. As [Schlinker2019] indicated, these can
become saturated during the peak hours of the day. Systematic
testing of the working latency in routers under load can help improve
both our understanding of latency and the impact of deployed
infrastructure.
4.2.6. Metrics Key Points
The metrics for network quality can be roughly grouped into the
following:
1. Availability metrics, which indicate whether the user can access
the network at all.
2. Capacity metrics, which indicate whether the actual line capacity
is sufficient to meet the user's demands.
3. Latency metrics, which indicate if the user gets the data in a
timely fashion.
4. Higher-order metrics, which include both the network metrics,
such as inter-packet arrival time, and the application metrics,
such as the mean time between rebuffering for video streaming.
The availability metrics can be seen as a derivative of either the
capacity (zero capacity leading to zero availability) or the latency
(infinite latency leading to zero availability).
Key points from the presentations and discussions included the
following:
1. Availability and capacity are "hygienic factors" -- unless an
application is capable of using extra capacity, end users will
see little benefit from using over-provisioned lines.
2. Working latency has a stronger correlation with the user
experience than latency under an idle network load. Working
latency can exceed the idle latency by order of magnitude.
3. The RPM metric is a stable metric, with positive values being
better, that may be more effective when communicating latency to
end users.
4. The relationship between throughput and goodput can be effective
in finding the saturation points, both in client-side
[Paasch2021] and server-side [Schlinker2019] settings.
5. Working latency depends on the algorithm choice for addressing
endpoint congestion control and router queuing.
Finally, it was commonly agreed to that the best metrics are those
that are actionable.
4.3. Cross-Layer Considerations
In the cross-layer segment of the workshop, participants presented
material on and discussed how to accurately measure exactly where
problems occur. Discussion centered especially on the differences
between physically wired and wireless connections and the
difficulties of accurately determining problem spots when multiple
different types of network segments are responsible for the quality.
As an example, [Kerpez2021] showed that a limited bandwidth of 2.4
Ghz Wi-Fi bottlenecks the most frequently. In comparison, the wider
bandwidth of the 5 Ghz Wi-Fi has only bottlenecked in 20% of
observations.
The participants agreed that no single component of a network
connection has all the data required to measure the effects of the
network performance on the quality of the end-user experience.
* Applications that are running on the end-user devices have the
best insight into their respective performance but have limited
visibility into the behavior of the network itself and are unable
to act based on their limited perspective.
* ISPs have good insight into QoS considerations but are not able to
infer the effect of the QoS metrics on the quality of end-user
experiences.
* Content providers have good insight into the aggregated behavior
of the end users but lack the insight on what aspects of network
performance are leading indicators of user behavior.
The workshop had identified the need for a standard and extensible
way to exchange network performance characteristics. Such an
exchange standard should address (at least) the following:
* A scalable way to capture the performance of multiple (potentially
thousands of) endpoints.
* The data exchange format should prevent data manipulation so that
the different participants won't be able to game the mechanisms.
* Preservation of end-user privacy. In particular, federated
learning approaches should be preferred so that no centralized
entity has the access to the whole picture.
* A transparent model for giving the different actors on a network
connection an incentive to share the performance data they
collect.
* An accompanying set of tools to analyze the data.
4.3.1. Separation of Concerns
Commonly, there's a tight coupling between collecting performance
metrics, interpreting those metrics, and acting upon the
interpretation. Unfortunately, such a model is not the best for
successfully exchanging cross-layer data, as:
* actors that are able to collect particular performance metrics
(e.g., the TCP RTT) do not necessarily have the context necessary
for a meaningful interpretation,
* the actors that have the context and the computational/storage
capacity to interpret metrics do not necessarily have the ability
to control the behavior of the network/application, and
* the actors that can control the behavior of networks and/or
applications typically do not have access to complete measurement
data.
The participants agreed that it is important to separate the above
three aspects, so that:
* the different actors that have the data, but not the ability to
interpret and/or act upon it, should publish their measured data
and
* the actors that have the expertise in interpreting and
synthesizing performance data should publish the results of their
interpretations.
4.3.2. Security and Privacy Considerations
Preserving the privacy of Internet end users is a difficult
requirement to meet when addressing this problem space. There is an
intrinsic trade-off between collecting more data about user
activities and infringing on their privacy while doing so.
Participants agreed that observability across multiple layers is
necessary for an accurate measurement of the network quality, but
doing so in a way that minimizes privacy leakage is an open question.
4.3.3. Metric Measurement Considerations
* The following TCP protocol metrics have been found to be effective
and are available for passive measurement:
- TCP connection latency measured using selective acknowledgment
(SACK) or acknowledgment (ACK) timing, as well as the timing
between TCP retransmission events, are good proxies for end-to-
end RTT measurements.
- On the Linux platform, the tcp_info structure is the de facto
standard for an application to inspect the performance of
kernel-space networking. However, there is no equivalent de
facto standard for user-space networking.
* The QUIC and MASQUE protocols make passive performance
measurements more challenging.
- An approach that uses federated measurement/hierarchical
aggregation may be more valuable for these protocols.
- The QLOG format seems to be the most mature candidate for such
an exchange.
4.3.4. Towards Improving Future Cross-Layer Observability
The ownership of the Internet is spread across multiple
administrative domains, making measurement of end-to-end performance
data difficult. Furthermore, the immense scale of the Internet makes
aggregation and analysis of this difficult. [Marx2021] presented a
simple logging format that could potentially be used to collect and
aggregate data from different layers.
Another aspect of the cross-layer collaboration hampering measurement
is that the majority of current algorithms do not explicitly provide
performance data that can be used in cross-layer analysis. The IETF
community could be more diligent in identifying each protocol's key
performance indicators and exposing them as part of the protocol
specification.
Despite all these challenges, it should still be possible to perform
limited-scope studies in order to have a better understanding of how
user quality is affected by the interaction of the different
components that constitute the Internet. Furthermore, recent
development of federated learning algorithms suggests that it might
be possible to perform cross-layer performance measurements while
preserving user privacy.
4.3.5. Efficient Collaboration between Hardware and Transport Protocols
With the advent of the low latency, low loss, and scalable throughput
(L4S) congestion notification and control, there is an even higher
need for the transport protocols and the underlying hardware to work
in unison.
At the time of the workshop, the typical home router uses a single
FIFO queue that is large enough to allow amortizing the lower-layer
header overhead across multiple transport PDUs. These designs worked
well with the cubic congestion control algorithm, yet the newer
generation of algorithms can operate on much smaller queues. To
fully support latencies less than 1 ms, the home router needs to work
efficiently on sequential transmissions of just a few segments vs.
being optimized for large packet bursts.
Another design trait common in home routers is the use of packet
aggregation to further amortize the overhead added by the lower-layer
headers. Specifically, multiple IP datagrams are combined into a
single, large transfer frame. However, this aggregation can add up
to 10 ms to the packet sojourn delay.
Following the famous "you can't improve what you don't measure"
adage, it is important to expose these aggregation delays in a way
that would allow identifying the source of the bottlenecks and making
hardware more suitable for the next generation of transport
protocols.
4.3.6. Cross-Layer Key Points
* Significant differences exist in the characteristics of metrics to
be measured and the required optimizations needed in wireless vs.
wired networks.
* Identification of an issue's root cause is hampered by the
challenges in measuring multi-segment network paths.
* No single component of a network connection has all the data
required to measure the effects of the complete network
performance on the quality of the end-user experience.
* Actionable results require both proper collection and
interpretation.
* Coordination among network providers is important to successfully
improve the measurement of end-user experiences.
* Simultaneously providing accurate measurements while preserving
end-user privacy is challenging.
* Passive measurements from protocol implementations may provide
beneficial data.
4.4. Synthesis
Finally, in the synthesis section of the workshop, the presentations
and discussions concentrated on the next steps likely needed to make
forward progress. Of particular concern is how to bring forward
measurements that can make sense to end users trying to select
between various networking subscription options.
4.4.1. Measurement and Metrics Considerations
One important consideration is how decisions can be made and what
actions can be taken based on collected metrics. Measurements must
be integrated with applications in order to get true application
views of congestion, as measurements over different infrastructure or
via other applications may return incorrect results. Congestion
itself can be a temporary problem, and mitigation strategies may need
to be different depending on whether it is expected to be a short-
term or long-term phenomenon. A significant challenge exists in
measuring short-term problems, driving the need for continuous
measurements to ensure critical moments and long-term trends are
captured. For short-term problems, workshop participants debated
whether an issue that goes away is indeed a problem or is a sign that
a network is properly adapting and self-recovering.
Important consideration must be taken when constructing metrics in
order to understand the results. Measurements can also be affected
by individual packet characteristics -- differently sized packets
typically have a linear relationship with their delay. With this in
mind, measurements can be divided into a delay based on geographical
distances, a packet-size serialization delay, and a variable (noise)
delay. Each of these three sub-component delays can be different and
individually measured across each segment in a multi-hop path.
Variable delay can also be significantly impacted by external
factors, such as bufferbloat, routing changes, network load sharing,
and other local or remote changes in performance. Network
measurements, especially load-specific tests, must also be run long
enough to ensure that any problems associated with buffering,
queuing, etc. are captured. Measurement technologies should also
distinguish between upstream and downstream measurements, as well as
measure the difference between end-to-end paths and sub-path
measurements.
4.4.2. End-User Metrics Presentation
Determining end-user needs requires informative measurements and
metrics. How do we provide the users with the service they need or
want? Is it possible for users to even voice their desires
effectively? Only high-level, simplistic answers like "reliability",
"capacity", and "service bundling" are typical answers given in end-
user surveys. Technical requirements that operators can consume,
like "low-latency" and "congestion avoidance", are not terms known to
and used by end users.
Example metrics useful to end users might include the number of users
supported by a service and the number of applications or streams that
a network can support. An example solution to combat networking
issues include incentive-based traffic management strategies (e.g.,
an application requesting lower latency may also mean accepting lower
bandwidth). User-perceived latency must be considered, not just
network latency -- user experience in-application to in-server
latency and network-to-network measurements may only be studying the
lowest-level latency. Thus, picking the right protocol to use in a
measurement is critical in order to match user experience (for
example, users do not transmit data over ICMP, even though it is a
common measurement tool).
In-application measurements should consider how to measure different
types of applications, such as video streaming, file sharing, multi-
user gaming, and real-time voice communications. It may be that
asking users for what trade-offs they are willing to accept would be
a helpful approach: would they rather have a network with low latency
or a network with higher bandwidth? Gamers may make different
decisions than home office users or content producers, for example.
Furthermore, how can users make these trade-offs in a fair manner
that does not impact other users? There is a tension between
solutions in this space vs. the cost associated with solving these
problems, as well as which customers are willing to front these
improvement costs.
Challenges in providing higher-priority traffic to users centers
around the ability for networks to be willing to listen to client
requests for higher incentives, even though commercial interests may
not flow to them without a cost incentive. Shared mediums in general
are subject to oversubscribing, such that the number of users a
network can support is either accurate on an underutilized network or
may assume an average bandwidth or other usage metric that fails to
be accurate during utilization spikes. Individual metrics are also
affected by in-home devices from cheap routers to microwaves and by
(multi-)user behaviors during tests. Thus, a single metric alone or
a single reading without context may not be useful in assisting a
user or operator to determine where the problem source actually is.
User comprehension of a network remains a challenging problem.
Multiple workshop participants argued for a single number
(potentially calculated with a weighted aggregation formula) or a
small number of measurements per expected usage (e.g., a "gaming"
score vs. a "content producer" score). Many agreed that some users
may instead prefer to consume simplified or color-coded ratings
(e.g., good/better/best, red/yellow/green, or bronze/gold/platinum).
4.4.3. Synthesis Key Points
* Some proposed metrics:
- Round-trips Per Minute (RPM)
- users per network
- latency
- 99% latency and bandwidth
* Median and mean measurements are distractions from the real
problems.
* Shared network usage greatly affects quality.
* Long measurements are needed to capture all facets of potential
network bottlenecks.
* Better-funded research in all these areas is needed for progress.
* End users will best understand a simplified score or ranking
system.
5. Conclusions
During the final hour of the three-day workshop, statements that the
group deemed to be summary statements were gathered. Later, any
statements that were in contention were discarded (listed further
below for completeness). For this document, the authors took the
original list and divided it into rough categories, applied some
suggested edits discussed on the mailing list, and further edited for
clarity and to provide context.
5.1. General Statements
1. Bandwidth is necessary but not alone sufficient.
2. In many cases, Internet users don't need more bandwidth but
rather need "better bandwidth", i.e., they need other
improvements to their connectivity.
3. We need both active and passive measurements -- passive
measurements can provide historical debugging.
4. We need passive measurements to be continuous, archivable, and
queriable, including reliability/connectivity measurements.
5. A really meaningful metric for users is whether their application
will work properly or fail because of a lack of a network with
sufficient characteristics.
6. A useful metric for goodness must actually incentivize goodness
-- good metrics should be actionable to help drive industries
towards improvement.
7. A lower-latency Internet, however achieved, would benefit all end
users.
5.2. Specific Statements about Detailed Protocols/Techniques
1. Round-trips Per Minute (RPM) is a useful, consumable metric.
2. We need a usable tool that fills the current gap between network
reachability, latency, and speed tests.
3. End users that want to be involved in QoS decisions should be
able to voice their needs and desires.
4. Applications are needed that can perform and report good quality
measurements in order to identify insufficient points in network
access.
5. Research done by regulators indicate that users/consumers prefer
a simple metric per application, which frequently resolves to
whether the application will work properly or not.
6. New measurements and QoS or QoE techniques should not rely only
or depend on reading TCP headers.
7. It is clear from developers of interactive applications and from
network operators that lower latency is a strong factor in user
QoE. However, metrics are lacking to support this statement
directly.
5.3. Problem Statements and Concerns
1. Latency mean and medians are distractions from better
measurements.
2. It is frustrating to only measure network services without
simultaneously improving those services.
3. Stakeholder incentives aren't aligned for easy wins in this
space. Incentives are needed to motivate improvements in public
network access. Measurements may be one step towards driving
competitive market incentives.
4. For future-proof networking, it is important to measure the
ecological impact of material and energy usage.
5. We do not have incontrovertible evidence that any one metric
(e.g., latency or speed) is more important than others to
persuade device vendors to concentrate on any one optimization.
5.4. No-Consensus-Reached Statements
Additional statements were discussed and recorded that did not have
consensus of the group at the time, but they are listed here for
completeness:
1. We do not have incontrovertible evidence that bufferbloat is a
prevalent problem.
2. The measurement needs to support reporting localization in order
to find problems. Specifically:
* Detecting a problem is not sufficient if you can't find the
location.
* Need more than just English -- different localization
concerns.
3. Stakeholder incentives aren't aligned for easy wins in this
space.
6. Follow-On Work
There was discussion during the workshop about where future work
should be performed. The group agreed that some work could be done
more immediately within existing IETF working groups (e.g., IPPM,
DetNet, and RAW), while other longer-term research may be needed in
IRTF groups.
7. IANA Considerations
This document has no IANA actions.
8. Security Considerations
A few security-relevant topics were discussed at the workshop,
including but not limited to:
* what prioritization techniques can work without invading the
privacy of the communicating parties and
* how oversubscribed networks can essentially be viewed as a DDoS
attack.
9. Informative References
[Aldabbagh2021]
Aldabbagh, A., "Regulatory perspective on measuring
network quality for end-users", September 2021,
<https://www.iab.org/wp-content/IAB-
uploads/2021/09/2021-09-07-Aldabbagh-Ofcom-presentationt-
to-IAB-1v00-1.pdf>.
[Arkko2021]
Arkko, J. and M. Kühlewind, "Observability is needed to
improve network quality", August 2021,
<https://www.iab.org/wp-content/IAB-uploads/2021/09/iab-
position-paper-observability.pdf>.
[Balasubramanian2021]
Balasubramanian, P., "Transport Layer Statistics for
Network Quality", February 2021, <https://www.iab.org/wp-
content/IAB-uploads/2021/09/transportstatsquality.pdf>.
[Briscoe2021]
Briscoe, B., White, G., Goel, V., and K. De Schepper, "A
Single Common Metric to Characterize Varying Packet
Delay", September 2021, <https://www.iab.org/wp-content/
IAB-uploads/2021/09/single-delay-metric-1.pdf>.
[Casas2021]
Casas, P., "10 Years of Internet-QoE Measurements Video,
Cloud, Conferencing, Web and Apps. What do we need from
the Network Side?", August 2021, <https://www.iab.org/wp-
content/IAB-uploads/2021/09/
net_quality_internet_qoe_CASAS.pdf>.
[Cheshire2021]
Cheshire, S., "The Internet is a Shared Network", August
2021, <https://www.iab.org/wp-content/IAB-uploads/2021/09/
draft-cheshire-internet-is-shared-00b.pdf>.
[Davies2021]
Davies, N. and P. Thompson, "Measuring Network Impact on
Application Outcomes Using Quality Attenuation", September
2021, <https://www.iab.org/wp-content/IAB-uploads/2021/09/
PNSol-et-al-Submission-to-Measuring-Network-Quality-for-
End-Users-1.pdf>.
[DeSchepper2021]
De Schepper, K., Tilmans, O., and G. Dion, "Challenges and
opportunities of hardware support for Low Queuing Latency
without Packet Loss", February 2021, <https://www.iab.org/
wp-content/IAB-uploads/2021/09/Nokia-IAB-Measuring-
Network-Quality-Low-Latency-measurement-workshop-
20210802.pdf>.
[Dion2021] Dion, G., De Schepper, K., and O. Tilmans, "Focusing on
latency, not throughput, to provide a better internet
experience and network quality", August 2021,
<https://www.iab.org/wp-content/IAB-uploads/2021/09/Nokia-
IAB-Measuring-Network-Quality-Improving-and-focusing-on-
latency-.pdf>.
[Fabini2021]
Fabini, J., "Network Quality from an End User
Perspective", February 2021, <https://www.iab.org/wp-
content/IAB-uploads/2021/09/Fabini-IAB-
NetworkQuality.txt>.
[FCC_MBA] FCC, "Measuring Broadband America",
<https://www.fcc.gov/general/measuring-broadband-america>.
[FCC_MBA_methodology]
FCC, "Measuring Broadband America - Open Methodology",
<https://www.fcc.gov/general/measuring-broadband-america-
open-methodology>.
[Foulkes2021]
Foulkes, J., "Metrics helpful in assessing Internet
Quality", September 2021, <https://www.iab.org/wp-content/
IAB-uploads/2021/09/
IAB_Metrics_helpful_in_assessing_Internet_Quality.pdf>.
[Ghai2021] Ghai, R., "Using TCP Connect Latency for measuring CX and
Network Optimization", February 2021,
<https://www.iab.org/wp-content/IAB-uploads/2021/09/
xfinity-wifi-ietf-iab-v2-1.pdf>.
[Iyengar2021]
Iyengar, J., "The Internet Exists In Its Use", August
2021, <https://www.iab.org/wp-content/IAB-uploads/2021/09/
The-Internet-Exists-In-Its-Use.pdf>.
[Kerpez2021]
Shafiei, J., Kerpez, K., Cioffi, J., Chow, P., and D.
Bousaber, "Wi-Fi and Broadband Data", September 2021,
<https://www.iab.org/wp-content/IAB-uploads/2021/09/Wi-Fi-
Report-ASSIA.pdf>.
[Kilkki2021]
Kilkki, K. and B. Finley, "In Search of Lost QoS",
February 2021, <https://www.iab.org/wp-content/IAB-
uploads/2021/09/Kilkki-In-Search-of-Lost-QoS.pdf>.
[Laki2021] Nadas, S., Varga, B., Contreras, L.M., and S. Laki,
"Incentive-Based Traffic Management and QoS Measurements",
February 2021, <https://www.iab.org/wp-content/IAB-
uploads/2021/11/CamRdy-
IAB_user_meas_WS_Nadas_et_al_IncentiveBasedTMwQoS.pdf>.
[Liubogoshchev2021]
Liubogoshchev, M., "Cross-layer Cooperation for Better
Network Service", February 2021, <https://www.iab.org/wp-
content/IAB-uploads/2021/09/Cross-layer-Cooperation-for-
Better-Network-Service-2.pdf>.
[MacMillian2021]
MacMillian, K. and N. Feamster, "Beyond Speed Test:
Measuring Latency Under Load Across Different Speed
Tiers", February 2021, <https://www.iab.org/wp-content/
IAB-uploads/2021/09/2021_nqw_lul.pdf>.
[Marx2021] Marx, R. and J. Herbots, "Merge Those Metrics: Towards
Holistic (Protocol) Logging", February 2021,
<https://www.iab.org/wp-content/IAB-uploads/2021/09/
MergeThoseMetrics_Marx_Jul2021.pdf>.
[Mathis2021]
Mathis, M., "Preliminary Longitudinal Study of Internet
Responsiveness", August 2021, <https://www.iab.org/wp-
content/IAB-uploads/2021/09/Preliminary-Longitudinal-
Study-of-Internet-Responsiveness-1.pdf>.
[McIntyre2021]
Paasch, C., McIntyre, K., Shapira, O., Meyer, R., and S.
Cheshire, "An end-user approach to an Internet Score",
September 2021, <https://www.iab.org/wp-content/IAB-
uploads/2021/09/Internet-Score-2.pdf>.
[Michel2021]
Michel, F. and O. Bonaventure, "Packet delivery time as a
tie-breaker for assessing Wi-Fi access points", February
2021, <https://www.iab.org/wp-content/IAB-uploads/2021/09/
camera_ready_Packet_delivery_time_as_a_tie_breaker_for_ass
essing_Wi_Fi_access_points.pdf>.
[Mirsky2021]
Mirsky, G., Min, X., Mishra, G., and L. Han, "The error
performance metric in a packet-switched network", February
2021, <https://www.iab.org/wp-content/IAB-uploads/2021/09/
IAB-worshop-Error-performance-measurement-in-packet-
switched-networks.pdf>.
[Morton2021]
Morton, A. C., "Dream-Pipe or Pipe-Dream: What Do Users
Want (and how can we assure it)?", Work in Progress,
Internet-Draft, draft-morton-ippm-pipe-dream-01, 6
September 2021, <https://datatracker.ietf.org/doc/html/
draft-morton-ippm-pipe-dream-01>.
[NetworkQuality]
Apple, "Network Quality",
<https://support.apple.com/en-gb/HT212313>.
[Paasch2021]
Paasch, C., Meyer, R., Cheshire, S., and O. Shapira,
"Responsiveness under Working Conditions", Work in
Progress, Internet-Draft, draft-cpaasch-ippm-
responsiveness-01, 25 October 2021,
<https://datatracker.ietf.org/doc/html/draft-cpaasch-ippm-
responsiveness-01>.
[Pardue2021]
Pardue, L. and S. Tellakula, "Lower-layer performance is
not indicative of upper-layer success", February 2021,
<https://www.iab.org/wp-content/IAB-uploads/2021/09/Lower-
layer-performance-is-not-indicative-of-upper-layer-
success-20210906-00-1.pdf>.
[Reed2021] Reed, D.P. and L. Perigo, "Measuring ISP Performance in
Broadband America: A Study of Latency Under Load",
February 2021, <https://www.iab.org/wp-content/IAB-
uploads/2021/09/Camera_Ready_-Measuring-ISP-Performance-
in-Broadband-America.pdf>.
[SamKnows] "SamKnows", <https://www.samknows.com/>.
[Schlinker2019]
Schlinker, B., Cunha, I., Chiu, Y., Sundaresan, S., and E.
Katz-Basset, "Internet Performance from Facebook's Edge",
February 2019, <https://www.iab.org/wp-content/IAB-
uploads/2021/09/Internet-Performance-from-Facebooks-
Edge.pdf>.
[Scuba] Abraham, L. et al., "Scuba: Diving into Data at Facebook",
<https://research.facebook.com/publications/scuba-diving-
into-data-at-facebook/>.
[Sengupta2021]
Sengupta, S., Kim, H., and J. Rexford, "Fine-Grained RTT
Monitoring Inside the Network", February 2021,
<https://www.iab.org/wp-content/IAB-uploads/2021/09/
Camera_Ready__Fine-
Grained_RTT_Monitoring_Inside_the_Network.pdf>.
[Sivaraman2021]
Sivaraman, V., Madanapalli, S., and H. Kumar, "Measuring
Network Experience Meaningfully, Accurately, and
Scalably", February 2021, <https://www.iab.org/wp-content/
IAB-uploads/2021/09/CanopusPositionPaperCameraReady.pdf>.
[Speedtest]
Ookla, "Speedtest", <https://www.speedtest.net>.
[Stein2021]
Stein, Y., "The Futility of QoS", August 2021,
<https://www.iab.org/wp-content/IAB-uploads/2021/09/QoS-
futility.pdf>.
[Welzl2021]
Welzl, M., "A Case for Long-Term Statistics", February
2021, <https://www.iab.org/wp-content/IAB-uploads/2021/09/
iab-longtermstats_cameraready.docx-1.pdf>.
[WORKSHOP] IAB, "IAB Workshop: Measuring Network Quality for End-
Users, 2021", September 2021,
<https://www.iab.org/activities/workshops/network-
quality>.
[Zhang2021]
Zhang, M., Goel, V., and L. Xu, "User-Perceived Latency to
Measure CCAs", September 2021, <https://www.iab.org/wp-
content/IAB-uploads/2021/09/User_Perceived_Latency-1.pdf>.
Appendix A. Program Committee
The program committee consisted of:
Jari Arkko
Olivier Bonaventure
Vint Cerf
Stuart Cheshire
Sam Crowford
Nick Feamster
Jim Gettys
Toke Hoiland-Jorgensen
Geoff Huston
Cullen Jennings
Katarzyna Kosek-Szott
Mirja Kühlewind
Jason Livingood
Matt Mathis
Randall Meyer
Kathleen Nichols
Christoph Paasch
Tommy Pauly
Greg White
Keith Winstein
Appendix B. Workshop Chairs
The workshop chairs consisted of:
Wes Hardaker
Evgeny Khorov
Omer Shapira
Appendix C. Workshop Participants
The following is a list of participants who attended the workshop
over a remote connection:
Ahmed Aldabbagh
Jari Arkko
Praveen Balasubramanian
Olivier Bonaventure
Djamel Bousaber
Bob Briscoe
Rich Brown
Anna Brunstrom
Pedro Casas
Vint Cerf
Stuart Cheshire
Kenjiro Cho
Steve Christianson
John Cioffi
Alexander Clemm
Luis M. Contreras
Sam Crawford
Neil Davies
Gino Dion
Toerless Eckert
Lars Eggert
Joachim Fabini
Gorry Fairhurst
Nick Feamster
Mat Ford
Jonathan Foulkes
Jim Gettys
Rajat Ghai
Vidhi Goel
Wes Hardaker
Joris Herbots
Geoff Huston
Toke Høiland-Jørgensen
Jana Iyengar
Cullen Jennings
Ken Kerpez
Evgeny Khorov
Kalevi Kilkki
Joon Kim
Zhenbin Li
Mikhail Liubogoshchev
Jason Livingood
Kyle MacMillan
Sharat Madanapalli
Vesna Manojlovic
Robin Marx
Matt Mathis
Jared Mauch
Kristen McIntyre
Randall Meyer
François Michel
Greg Mirsky
Cindy Morgan
Al Morton
Szilveszter Nadas
Kathleen Nichols
Lai Yi Ohlsen
Christoph Paasch
Lucas Pardue
Tommy Pauly
Levi Perigo
David Reed
Alvaro Retana
Roberto
Koen De Schepper
David Schinazi
Brandon Schlinker
Eve Schooler
Satadal Sengupta
Jinous Shafiei
Shapelez
Omer Shapira
Dan Siemon
Vijay Sivaraman
Karthik Sundaresan
Dave Taht
Rick Taylor
Bjørn Ivar Teigen
Nicolas Tessares
Peter Thompson
Balazs Varga
Bren Tully Walsh
Michael Welzl
Greg White
Russ White
Keith Winstein
Lisong Xu
Jiankang Yao
Gavin Young
Mingrui Zhang
IAB Members at the Time of Approval
Internet Architecture Board members at the time this document was
approved for publication were:
Jari Arkko
Deborah Brungard
Lars Eggert
Wes Hardaker
Cullen Jennings
Mallory Knodel
Mirja Kühlewind
Zhenbin Li
Tommy Pauly
David Schinazi
Russ White
Qin Wu
Jiankang Yao
Acknowledgments
The authors would like to thank the workshop participants, the
members of the IAB, and the program committee for creating and
participating in many interesting discussions.
Contributors
Thank you to the people that contributed edits to this document:
Erik Auerswald
Simon Leinen
Brian Trammell
Authors' Addresses
Wes Hardaker
Email: ietf@hardakers.net
Omer Shapira
Email: omer_shapira@apple.com
|