1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
1001
1002
1003
1004
1005
1006
1007
1008
1009
1010
1011
1012
1013
1014
1015
1016
1017
1018
1019
1020
1021
1022
1023
1024
1025
1026
1027
1028
1029
1030
1031
1032
1033
1034
1035
1036
1037
1038
1039
1040
1041
1042
1043
1044
1045
1046
1047
1048
1049
1050
1051
1052
1053
1054
1055
1056
1057
1058
1059
1060
1061
1062
1063
1064
1065
1066
1067
1068
1069
1070
1071
1072
1073
1074
1075
1076
1077
1078
1079
1080
1081
1082
1083
1084
1085
1086
1087
1088
1089
1090
1091
1092
1093
1094
1095
1096
1097
1098
1099
1100
1101
1102
1103
1104
1105
1106
1107
1108
1109
1110
1111
1112
1113
1114
1115
1116
1117
1118
1119
1120
1121
1122
1123
1124
1125
1126
1127
1128
1129
1130
1131
1132
1133
1134
1135
1136
1137
1138
1139
1140
1141
1142
1143
1144
1145
1146
1147
1148
1149
1150
1151
1152
1153
1154
1155
1156
1157
1158
1159
1160
1161
1162
1163
1164
1165
1166
1167
1168
1169
1170
1171
1172
1173
1174
1175
1176
1177
1178
1179
1180
1181
1182
1183
1184
1185
1186
1187
1188
1189
1190
1191
1192
1193
1194
1195
1196
1197
1198
1199
1200
1201
1202
1203
1204
1205
1206
1207
1208
1209
1210
1211
1212
1213
1214
1215
1216
1217
1218
1219
1220
1221
1222
1223
1224
1225
1226
1227
1228
1229
1230
1231
1232
1233
1234
1235
1236
1237
1238
1239
1240
1241
1242
1243
1244
1245
1246
1247
1248
1249
1250
1251
1252
1253
1254
1255
1256
1257
1258
1259
1260
1261
1262
1263
1264
1265
1266
1267
1268
1269
1270
1271
1272
1273
1274
1275
1276
1277
1278
1279
1280
1281
1282
1283
1284
1285
1286
1287
1288
1289
1290
1291
1292
1293
1294
1295
1296
1297
1298
1299
1300
1301
1302
1303
1304
1305
1306
1307
1308
1309
1310
1311
1312
1313
1314
1315
1316
1317
1318
1319
1320
1321
1322
1323
1324
1325
1326
1327
1328
1329
1330
1331
1332
1333
1334
1335
1336
1337
1338
1339
1340
1341
1342
1343
1344
1345
1346
1347
1348
1349
1350
1351
1352
1353
1354
1355
1356
1357
1358
1359
1360
1361
1362
1363
1364
1365
1366
1367
1368
1369
1370
1371
1372
1373
1374
1375
1376
1377
1378
1379
1380
1381
1382
1383
1384
1385
1386
1387
1388
1389
1390
1391
1392
1393
1394
1395
1396
1397
1398
1399
1400
1401
1402
1403
1404
1405
1406
1407
1408
1409
1410
1411
1412
1413
1414
1415
1416
1417
1418
1419
1420
1421
1422
1423
1424
1425
1426
1427
1428
1429
1430
1431
1432
1433
1434
1435
1436
1437
1438
1439
1440
1441
1442
1443
1444
1445
1446
1447
1448
1449
1450
1451
1452
1453
1454
1455
1456
1457
1458
1459
1460
1461
1462
1463
1464
1465
1466
1467
1468
1469
1470
1471
1472
1473
1474
1475
1476
1477
1478
1479
1480
1481
1482
1483
1484
1485
1486
1487
1488
1489
1490
1491
1492
1493
1494
1495
1496
1497
1498
1499
1500
1501
1502
1503
1504
1505
1506
1507
1508
1509
1510
1511
1512
1513
1514
1515
1516
1517
1518
1519
1520
1521
1522
1523
1524
1525
1526
1527
1528
1529
1530
1531
1532
1533
1534
1535
1536
1537
1538
1539
1540
1541
1542
1543
1544
1545
1546
1547
1548
1549
1550
1551
1552
1553
1554
1555
1556
1557
1558
1559
1560
1561
1562
1563
1564
1565
1566
1567
1568
1569
1570
1571
1572
1573
1574
1575
1576
1577
1578
1579
1580
1581
1582
1583
1584
1585
1586
1587
1588
1589
1590
1591
1592
1593
1594
1595
1596
1597
1598
1599
1600
1601
1602
1603
1604
1605
1606
1607
1608
1609
1610
1611
1612
1613
1614
1615
1616
1617
1618
1619
1620
1621
1622
1623
1624
1625
1626
1627
1628
1629
1630
1631
1632
1633
1634
1635
1636
1637
1638
1639
1640
1641
1642
1643
1644
1645
1646
1647
1648
1649
1650
1651
1652
1653
1654
1655
1656
1657
1658
1659
1660
1661
1662
1663
1664
1665
1666
1667
1668
1669
1670
1671
1672
1673
1674
1675
1676
1677
1678
1679
1680
1681
1682
1683
1684
1685
1686
1687
1688
1689
1690
1691
1692
1693
1694
1695
1696
1697
1698
1699
1700
1701
1702
1703
1704
1705
1706
1707
1708
1709
1710
1711
1712
1713
1714
1715
1716
1717
1718
1719
1720
1721
1722
1723
1724
1725
1726
1727
1728
1729
1730
1731
1732
1733
1734
1735
1736
1737
1738
1739
1740
1741
1742
1743
1744
1745
1746
1747
1748
1749
1750
1751
1752
1753
1754
1755
1756
1757
1758
1759
1760
1761
1762
1763
1764
1765
1766
1767
1768
1769
1770
1771
1772
1773
1774
1775
1776
1777
1778
1779
1780
1781
1782
1783
1784
1785
1786
1787
1788
1789
1790
1791
1792
1793
1794
1795
1796
1797
1798
1799
1800
1801
1802
1803
1804
1805
1806
1807
1808
1809
1810
1811
1812
1813
1814
1815
1816
1817
1818
1819
1820
1821
1822
1823
1824
1825
1826
1827
1828
1829
1830
1831
1832
1833
1834
1835
1836
1837
1838
1839
1840
1841
1842
1843
1844
1845
1846
1847
1848
1849
1850
1851
1852
1853
1854
1855
1856
1857
1858
1859
1860
1861
1862
1863
1864
1865
1866
1867
1868
1869
1870
1871
1872
1873
1874
1875
1876
1877
1878
1879
1880
1881
1882
1883
1884
1885
1886
1887
1888
1889
1890
1891
1892
1893
1894
1895
1896
1897
1898
1899
1900
1901
1902
1903
1904
1905
1906
1907
1908
1909
1910
1911
1912
1913
1914
1915
1916
1917
1918
1919
1920
1921
1922
1923
1924
1925
1926
1927
1928
1929
1930
1931
1932
1933
1934
1935
1936
1937
1938
1939
1940
1941
1942
1943
1944
1945
1946
1947
1948
1949
1950
1951
1952
1953
1954
1955
1956
1957
1958
1959
1960
1961
1962
1963
1964
1965
1966
1967
1968
1969
1970
1971
1972
1973
1974
1975
1976
1977
1978
1979
1980
1981
1982
1983
1984
1985
1986
1987
1988
1989
1990
1991
1992
1993
1994
1995
1996
1997
1998
1999
2000
2001
2002
2003
2004
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
2025
2026
2027
2028
2029
2030
2031
2032
2033
2034
2035
2036
2037
2038
2039
2040
2041
2042
2043
2044
2045
2046
2047
2048
2049
2050
2051
2052
2053
2054
2055
2056
2057
2058
2059
2060
2061
2062
2063
2064
2065
2066
2067
2068
2069
2070
2071
2072
2073
2074
2075
2076
2077
2078
2079
2080
2081
2082
2083
2084
2085
2086
2087
2088
2089
2090
2091
2092
2093
2094
2095
2096
2097
2098
2099
2100
2101
2102
2103
2104
2105
2106
2107
2108
2109
2110
2111
2112
2113
2114
2115
2116
2117
2118
2119
2120
2121
2122
2123
2124
2125
2126
2127
2128
2129
2130
2131
2132
2133
2134
|
Internet Research Task Force (IRTF) J. L. Hall
Request for Comments: 9505 Internet Society
Category: Informational M. D. Aaron
ISSN: 2070-1721 CU Boulder
A. Andersdotter
B. Jones
N. Feamster
U Chicago
M. Knodel
Center for Democracy & Technology
November 2023
A Survey of Worldwide Censorship Techniques
Abstract
This document describes technical mechanisms employed in network
censorship that regimes around the world use for blocking or
impairing Internet traffic. It aims to make designers, implementers,
and users of Internet protocols aware of the properties exploited and
mechanisms used for censoring end-user access to information. This
document makes no suggestions on individual protocol considerations,
and is purely informational, intended as a reference. This document
is a product of the Privacy Enhancement and Assessment Research Group
(PEARG) in the IRTF.
Status of This Memo
This document is not an Internet Standards Track specification; it is
published for informational purposes.
This document is a product of the Internet Research Task Force
(IRTF). The IRTF publishes the results of Internet-related research
and development activities. These results might not be suitable for
deployment. This RFC represents the consensus of the Privacy
Enhancements and Assessments Research Group of the Internet Research
Task Force (IRTF). Documents approved for publication by the IRSG
are not candidates for any level of Internet Standard; see Section 2
of RFC 7841.
Information about the current status of this document, any errata,
and how to provide feedback on it may be obtained at
https://www.rfc-editor.org/info/rfc9505.
Copyright Notice
Copyright (c) 2023 IETF Trust and the persons identified as the
document authors. All rights reserved.
This document is subject to BCP 78 and the IETF Trust's Legal
Provisions Relating to IETF Documents
(https://trustee.ietf.org/license-info) in effect on the date of
publication of this document. Please review these documents
carefully, as they describe your rights and restrictions with respect
to this document.
Table of Contents
1. Introduction
2. Terminology
3. Technical Prescription
4. Technical Identification
4.1. Points of Control
4.2. Application Layer
4.2.1. HTTP Request Header Identification
4.2.2. HTTP Response Header Identification
4.2.3. Transport Layer Security (TLS)
4.2.4. Instrumenting Content Distributors
4.2.5. DPI Identification
4.3. Transport Layer
4.3.1. Shallow Packet Inspection and Transport Header
Identification
4.3.2. Protocol Identification
4.4. Residual Censorship
5. Technical Interference
5.1. Application Layer
5.1.1. DNS Interference
5.2. Transport Layer
5.2.1. Performance Degradation
5.2.2. Packet Dropping
5.2.3. RST Packet Injection
5.3. Routing Layer
5.3.1. Network Disconnection
5.3.2. Adversarial Route Announcement
5.4. Multi-layer and Non-layer
5.4.1. Distributed Denial of Service (DDoS)
5.4.2. Censorship in Depth
6. Non-technical Interference
6.1. Manual Filtering
6.2. Self-Censorship
6.3. Server Takedown
6.4. Notice and Takedown
6.5. Domain Name Seizures
7. Future Work
8. IANA Considerations
9. Security Considerations
10. Informative References
Acknowledgments
Authors' Addresses
1. Introduction
Censorship is where an entity in a position of power -- such as a
government, organization, or individual -- suppresses communication
that it considers objectionable, harmful, sensitive, or inconvenient
[WP-Def-2020]. Although censors that engage in censorship must do so
through legal, martial, or other means, this document focuses largely
on technical mechanisms used to achieve network censorship.
This document describes technical mechanisms that censorship regimes
around the world use for blocking or impairing Internet traffic. See
[RFC7754] for a discussion of Internet blocking and filtering in
terms of implications for Internet architecture rather than end-user
access to content and services. There is also a growing field of
academic study of censorship circumvention (see the review article of
[Tschantz-2016]), results from which we seek to make relevant here
for protocol designers and implementers.
Censorship circumvention also impacts the cost of implementation of a
censorship measure, and we include mentions of trade-offs in relation
to such costs in conjunction with each technical method identified
below.
This document has seen extensive discussion and review in the IRTF
Privacy Enhancement and Assessment Research Group (PEARG) and
represents the consensus of that group. It is not an IETF product
and is not a standard.
2. Terminology
We describe three elements of Internet censorship: prescription,
identification, and interference. This document contains three major
sections, each corresponding to one of these elements. Prescription
is the process by which censors determine what types of material they
should censor, e.g., classifying pornographic websites as
undesirable. Identification is the process by which censors classify
specific traffic or traffic identifiers to be blocked or impaired,
e.g., deciding that webpages containing "sex" in an HTTP header or
that accept traffic through the URL "www.sex.example" are likely to
be undesirable. Interference is the process by which censors
intercede in communication and prevent access to censored materials
by blocking access or impairing the connection, e.g., implementing a
technical solution capable of identifying HTTP headers or URLs and
ensuring they are rendered wholly or partially inaccessible.
3. Technical Prescription
Prescription is the process of figuring out what censors would like
to block [Glanville-2008]. Generally, censors aggregate information
"to block" in blocklists, databases of image hashes [ekr-2021], or
use real-time heuristic assessment of content [Ding-1999]. Some
national networks are designed to more naturally serve as points of
control [Leyba-2019]. There are also indications that online censors
use probabilistic machine learning techniques [Tang-2016]. Indeed,
web crawling and machine learning techniques are an active research
area in the effort to identify content deemed as morally or
commercially harmful to companies or consumers in some jurisdictions
[SIDN-2020].
There are typically a few types of blocklist elements: keyword,
domain name, protocol, or IP address. Keyword and domain name
blocking take place at the application level, e.g., HTTP; protocol
blocking often occurs using deep packet inspection (DPI) to identify
a forbidden protocol; IP blocking tends to take place using IP
addresses in IPv4/IPv6 headers. Some censors also use the presence
of certain keywords to enable more aggressive blocklists
[Rambert-2021] or to be more permissive with content [Knockel-2021].
The mechanisms for building up these blocklists vary. Censors can
purchase from private industry "content control" software, which lets
censors filter traffic from broad categories they would like to
block, such as gambling or pornography [Knight-2005]. In these
cases, these private services attempt to categorize every semi-
questionable website to allow for meta-tag blocking. Similarly, they
tune real-time content heuristic systems to map their assessments
onto categories of objectionable content.
Countries that are more interested in retaining specific political
control typically have ministries or organizations that maintain
blocklists. Examples include the Ministry of Industry and
Information Technology in China, the Ministry of Culture and Islamic
Guidance in Iran, and the organizations specific to copyright law in
France [HADOPI] and consumer protection law across the EU
[Reda-2017].
Content-layer filtering of images and video requires institutions or
organizations to store hashes of images or videos to be blocked in
databases, which can then be compared, with some degree of tolerance,
to content that is sent, received, or stored using centralized
content applications and services [ekr-2021].
4. Technical Identification
4.1. Points of Control
Internet censorship takes place in all parts of the network topology.
It may be implemented in the network itself (e.g., local loop or
backhaul), on the services side of communication (e.g., web hosts,
cloud providers, or content delivery networks), in the ancillary
services ecosystem (e.g., domain name system (DNS) or certificate
authorities (CAs)), or on the end-client side (e.g., in an end-user
device, such as a smartphone, laptop, or desktop, or software
executed on such devices). An important aspect of pervasive
technical interception is the necessity to rely on software or
hardware to intercept the content the censor is interested in. There
are various logical and physical points of control that censors may
use for interception mechanisms, including, though not limited to,
the following:
Internet Backbone:
If a censor controls elements of Internet network infrastructure,
such as the international gateways into a region or Internet
Exchange Points (IXPs), those choke points can be used to filter
undesirable traffic that is traveling into and out of the region
by packet sniffing and port mirroring. Censorship at gateways is
most effective at controlling the flow of information between a
region and the rest of the Internet, but is ineffective at
identifying content traveling between the users within a region,
which would have to be accomplished at exchange points or other
network aggregation points. Some national network designs
naturally serve as more effective choke points and points of
control [Leyba-2019].
Internet Service Providers (ISPs):
ISPs are frequently exploited points of control. They have the
benefit of being easily enumerable by a censor -- often falling
under the jurisdictional or operational control of a censor in an
indisputable way -- with the additional feature that an ISP can
identify the regional and international traffic of all their
users. The censor's filtration mechanisms can be placed on an ISP
via governmental mandates, ownership, or voluntary/coercive
influence.
Institutions:
Private institutions such as corporations, schools, and Internet
cafes can use filtration mechanisms. These mechanisms are
occasionally at the request of a government censor but can also be
implemented to help achieve institutional goals, such as fostering
a particular moral outlook on life by schoolchildren, independent
of broader society or government goals.
Content Distribution Network (CDN):
CDNs seek to collapse network topology in order to better locate
content closer to the service's users. This reduces content
transmission latency and improves QoS. The CDN service's content
servers, located "close" to the user in a network sense, can be
powerful points of control for censors, especially if the location
of CDN repositories allows for easier interference.
CAs for Public Key Infrastructures (PKIs):
Authorities that issue cryptographically secured resources can be
a significant point of control. CAs that issue certificates to
domain holders for TLS/HTTPS (the Web PKI) or Regional or Local
Internet Registries (RIRs or LIRs) that issue Route Origin
Authorizations (ROAs) to BGP operators can be forced to issue
rogue certificates that may allow compromise, i.e., by allowing
censorship software to engage in identification and interference
where it may not have been possible before. CAs may also be
forced to revoke certificates. This may lead to adversarial
traffic routing, TLS interception being allowed, or an otherwise
rightful origin or destination point of traffic flows being unable
to communicate in a secure way.
Services:
Application service providers can be pressured, coerced, or
legally required to censor specific content or data flows.
Service providers naturally face incentives to maximize their
potential customer base, and potential service shutdowns or legal
liability due to censorship efforts may seem much less attractive
than potentially excluding content, users, or uses of their
service. Services have increasingly become focal points of
censorship discussions as well as discussions of moral imperatives
to use censorship tools.
Content Sites:
On the service side of communications lie many platforms that
publish user-generated content and require terms of service
compliance with all content and user accounts in order to avoid
intermediary liability for the web hosts. In aggregate, these
policies, actions, and remedies are known as content moderation.
Content moderation happens above the services or application
layer, but these mechanisms are built to filter, sort, and block
content and users, thus making them available to censors through
direct pressure on the private entity.
Personal Devices:
Censors can mandate censorship software be installed on the device
level. This has many disadvantages in terms of scalability, ease
of circumvention, and operating system requirements. (Of course,
if a personal device is treated with censorship software before
sale and this software is difficult to reconfigure, this may work
in favor of those seeking to control information, say, for
children, students, customers, or employees.) The emergence of
mobile devices has exacerbated these feasibility problems. This
software can also be mandated by institutional actors acting on
non-governmentally mandated moral imperatives.
At all levels of the network hierarchy, the filtration mechanisms
used to censor undesirable traffic are essentially the same: a censor
either directly identifies undesirable content using the identifiers
described below and then uses a blocking or shaping mechanism (such
as the ones exemplified below to prevent or impair access), or
requests that an actor ancillary to the censor (such as a private
entity) perform these functions. Identification of undesirable
traffic can occur at the application, transport, or network layer of
the IP stack. Censors often focus on web traffic, so the relevant
protocols tend to be filtered in predictable ways (see Sections 4.2.1
and 4.2.2). For example, a subversive image might make it past a
keyword filter. However, if later the image is deemed undesirable, a
censor may then blocklist the provider site's IP address.
4.2. Application Layer
The following subsections describe properties and trade-offs of
common ways in which censors filter using application-layer
information. Each subsection includes empirical examples describing
these common behaviors for further reference.
4.2.1. HTTP Request Header Identification
An HTTP header contains a lot of useful information for traffic
identification. Although "host" is the only required field in an
HTTP request header (for HTTP/1.1 and later), an HTTP method field is
necessary to do anything useful. As such, "method" and "host" are
the two fields used most often for ubiquitous censorship. A censor
can sniff traffic and identify a specific domain name (host) and
usually a page name (for example, GET /page) as well. This
identification technique is usually paired with transport header
identification (see Section 4.3.1) for a more robust method.
Trade-offs: HTTP request header identification is a technically
straightforward identification method that can be easily implemented
at the backbone or ISP level. The hardware needed for this sort of
identification is cheap and easy to acquire, making it desirable when
budget and scope are a concern. HTTPS (Hypertext Transport Protocol
Secure) will encrypt the relevant request and response fields, so
pairing with transport identification (see Section 4.3.1) is
necessary for HTTPS filtering. However, some countermeasures can
trivially defeat simple forms of HTTP request header identification.
For example, two cooperating endpoints -- an instrumented web server
and client -- could encrypt or otherwise obfuscate the "host" header
in a request, potentially thwarting techniques that match against
"host" header values.
Empirical Examples: Studies exploring censorship mechanisms have
found evidence of HTTP header and/or URL filtering in many countries,
including Bangladesh, Bahrain, China, India, Iran, Malaysia,
Pakistan, Russia, Saudi Arabia, South Korea, Thailand, and Turkey
[Verkamp-2012] [Nabi-2013] [Aryan-2013]. Commercial technologies are
often purchased by censors [Dalek-2013]. These commercial
technologies use a combination of HTTP request header identification
and transport header identification to filter specific URLs. Dalek
et al. and Jones et al. identified the use of these products in the
wild [Dalek-2013] [Jones-2014].
4.2.2. HTTP Response Header Identification
While HTTP request header identification relies on the information
contained in the HTTP request from client to server, HTTP response
header identification uses information sent in response by the server
to client to identify undesirable content.
Trade-offs: As with HTTP request header identification, the
techniques used to identify HTTP traffic are well-known, cheap, and
relatively easy to implement. However, they are made useless by
HTTPS because HTTPS encrypts the response and its headers.
The response fields are also less helpful for identifying content
than request fields, as "Server" could easily be identified using
HTTP request header identification, and "Via" is rarely relevant.
HTTP response censorship mechanisms normally let the first n packets
through while the mirrored traffic is being processed; this may allow
some content through, and the user may be able to detect that the
censor is actively interfering with undesirable content.
Empirical Examples: In 2009, Jong Park et al. at the University of
New Mexico demonstrated that the Great Firewall of China (GFW) has
used this technique [Crandall-2010]. However, Jong Park et al. found
that the GFW discontinued this practice during the course of the
study. Due to the overlap in HTTP response filtering and keyword
filtering (see Section 4.2.4), it is likely that most censors rely on
keyword filtering over TCP streams instead of HTTP response
filtering.
4.2.3. Transport Layer Security (TLS)
Similar to HTTP, censors have deployed a variety of techniques
towards censoring TLS (and by extension HTTPS). Most of these
techniques relate to the Server Name Indication (SNI) field,
including censoring SNI, Encrypted SNI (ESNI), or omitted SNI.
Censors can also censor HTTPS content via server certificates. Note
that TLS 1.3 acts as a security component of QUIC.
4.2.3.1. Server Name Indication (SNI)
In encrypted connections using TLS, there may be servers that host
multiple "virtual servers" at a given network address, and the client
will need to specify in the ClientHello message which domain name it
seeks to connect to (so that the server can respond with the
appropriate TLS certificate) using, the SNI TLS extension [RFC6066].
The ClientHello message is unencrypted for TCP-based TLS. When using
QUIC, the ClientHello message is encrypted, but its confidentiality
is not effectively protected because the initial encryption keys are
derived using a value that is visible on the wire. Since SNI is
often sent in the clear (as are the cert fields sent in response),
censors and filtering software can use it (and response cert fields)
as a basis for blocking, filtering, or impairment by dropping
connections to domains that match prohibited content (e.g.,
"bad.foo.example" may be censored while "good.foo.example" is not)
[Shbair-2015]. There are ongoing standardization efforts in the TLS
Working Group to encrypt SNI [RFC8744] [TLS-ESNI], and recent
research shows promising results in the use of ESNI in the face of
SNI-based filtering [Chai-2019] in some countries.
Domain fronting has been one popular way to avoid identification by
censors [Fifield-2015]. To avoid identification by censors,
applications using domain fronting put a different domain name in the
SNI extension than in the "host" header, which is protected by HTTPS.
The visible SNI would indicate an unblocked domain, while the blocked
domain remains hidden in the encrypted application header. Some
encrypted messaging services relied on domain fronting to enable
their provision in countries employing SNI-based filtering. These
services used the cover provided by domains for which blocking at the
domain level would be undesirable to hide their true domain names.
However, the companies holding the most popular domains have since
reconfigured their software to prevent this practice. It may be
possible to achieve similar results using potential future options to
encrypt SNI.
Trade-offs: Some clients do not send the SNI extension (e.g., clients
that only support versions of SSL and not TLS), rendering this method
ineffective (see Section 4.2.3.3). In addition, this technique
requires deep packet inspection (DPI) techniques that can be
expensive in terms of computational complexity and infrastructure,
especially when applied to QUIC where DPI requires key extraction and
decryption of the ClientHello in order to read the SNI. Improper
configuration of an SNI-based block can result in significant over-
blocking, e.g., when a second-level domain like
"populardomain.example" is inadvertently blocked. In the case of
ESNI, pressure to censor may transfer to other points of
intervention, such as content and application providers.
Empirical Examples: There are many examples of security firms that
offer SNI-based filtering products [Trustwave-2015] [Sophos-2023]
[Shbair-2015]. The governments of China, Egypt, Iran, Qatar, South
Korea, Turkey, Turkmenistan, and the United Arab Emirates all do
widespread SNI filtering or blocking [OONI-2018] [OONI-2019]
[NA-SK-2019] [CitizenLab-2018] [Gatlan-2019] [Chai-2019]
[Grover-2019] [Singh-2019]. SNI blocking against QUIC traffic was
first observed in Russia in March 2022 [Elmenhorst-2022].
4.2.3.2. Encrypted SNI (ESNI)
With the data leakage present with the SNI field, a natural response
is to encrypt it, which is forthcoming in TLS 1.3 with Encrypted
Client Hello (ECH). Prior to ECH, the ESNI extension is available to
prevent the data leakage caused by SNI, which encrypts only the SNI
field. Unfortunately, censors can target connections that use the
ESNI extension specifically for censorship. This guarantees over-
blocking for the censor but can be worth the cost if ESNI is not yet
widely deployed within the country. ECH is the emerging standard for
protecting the entire TLS ClientHello, but it is not yet widely
deployed.
Trade-offs: The cost to censoring ESNI is significantly higher than
SNI to a censor, as the censor can no longer target censorship to
specific domains and guarantees over-blocking. In these cases, the
censor uses the over-blocking to discourage the use of ESNI entirely.
Empirical Examples: In 2020, China began censoring all uses of ESNI
[Bock-2020b], even for innocuous connections. The censorship
mechanism for China's ESNI censorship differs from how China censors
SNI-based connections, suggesting that new middleboxes were deployed
specifically to target ESNI connections.
4.2.3.3. Omitted SNI
Researchers have observed that some clients omit the SNI extension
entirely. This omitted-SNI approach limits the information available
to a censor. Like with ESNI, censors can choose to block connections
that omit the SNI, though this too risks over-blocking.
Trade-offs: The approach of censoring all connections that omit the
SNI field is guaranteed to over-block, though connections that omit
the SNI field should be relatively rare in the wild.
Empirical Examples: In the past, researchers have observed censors in
Russia blocking connections that omit the SNI field [Bock-2020b].
4.2.3.4. Server Response Certificate
During the TLS handshake after the TLS ClientHello, the server will
respond with the TLS certificate. This certificate also contains the
domain the client is trying to access, creating another avenue that
censors can use to perform censorship. This technique will not work
in TLS 1.3, as the certificate will be encrypted.
Trade-offs: Censoring based on the server certificate requires DPI
techniques that can be more computationally expensive compared to
other methods. Additionally, the certificate is sent later in the
TLS handshake compared to the SNI field, forcing the censor to track
the connection longer.
Empirical Examples: Researchers have observed the Reliance Jio ISP in
India using certificate response fields to censor connections
[Satija-2021].
4.2.4. Instrumenting Content Distributors
Many governments pressure content providers to censor themselves, or
provide the legal framework, within which content distributors are
incentivized to follow the content restriction preferences of agents
external to the content distributor [Boyle-1997]. Due to the
extensive reach of such censorship, we define "content distributor"
as any service that provides utility to users, including everything
from websites to storage to locally installed programs.
A commonly used method of instrumenting content distributors consists
of keyword identification to detect restricted terms on their
platforms. Governments may provide the terms on such keyword lists.
Alternatively, the content provider may be expected to come up with
their own list.
An increasingly common method of instrumenting content distribution
consists of hash matching to detect and take action against images
and videos known to be restricted either by governments,
institutions, organizations or the distributor themselves [ekr-2021].
A different method of instrumenting content distributors consists of
requiring a distributor to disassociate with some categories of
users. See also Section 6.4.
Trade-offs: By instrumenting content distributors to identify
restricted content or content providers, the censor can gain new
information at the cost of political capital with the companies it
forces or encourages to participate in censorship. For example, the
censor can gain insight about the content of encrypted traffic by
coercing websites to identify restricted content. Coercing content
distributors to regulate users, categories of users, content, and
content providers may encourage users and content providers to
exhibit self-censorship, an additional advantage for censors (see
Section 6.2). The trade-offs for instrumenting content distributors
are highly dependent on the content provider and the requested
assistance. A typical concern is that the targeted keywords or
categories of users are too broad, risk being too broadly applied, or
are not subjected to a sufficiently robust legal process prior to
their mandatory application (see page 8 of [EC-2012]).
Empirical Examples: Researchers discovered keyword identification by
content providers on platforms ranging from instant messaging
applications [Senft-2013] to search engines [Rushe-2014] [Cheng-2010]
[Whittaker-2013] [BBC-2013] [Condliffe-2013]. To demonstrate the
prevalence of this type of keyword identification, we look to search
engine censorship.
Search engine censorship demonstrates keyword identification by
content providers and can be regional or worldwide. Implementation
is occasionally voluntary, but normally it is based on laws and
regulations of the country a search engine is operating in. The
keyword blocklists are most likely maintained by the search engine
provider. China is known to require search engine providers to
"voluntarily" maintain search term blocklists to acquire and keep an
Internet Content Provider (ICP) license [Cheng-2010]. It is clear
these blocklists are maintained by each search engine provider based
on the slight variations in the intercepted searches [Zhu-2011]
[Whittaker-2013]. The United Kingdom has been pushing search engines
to self-censor with the threat of litigation if they do not do it
themselves: Google and Microsoft have agreed to block more than
100,000 queries in the U.K. to help combat abuse [BBC-2013]
[Condliffe-2013]. European Union law, as well as United States law,
requires modification of search engine results in response to either
copyright, trademark, data protection, or defamation concerns
[EC-2012].
Depending on the output, search engine keyword identification may be
difficult or easy to detect. In some cases, specialized or blank
results provide a trivial enumeration mechanism, but more subtle
censorship can be difficult to detect. In February 2015, Microsoft's
search engine, Bing, was accused of censoring Chinese content outside
of China [Rushe-2014] because Bing returned different results for
censored terms in Chinese and English. However, it is possible that
censorship of the largest base of Chinese search users, China, biased
Bing's results so that the more popular results in China (the
uncensored results) were also more popular for Chinese speakers
outside of China.
Disassociation by content distributors from certain categories of
users has happened for instance in Spain, as a result of the conflict
between the Catalan independence movement and the Spanish legal
presumption of a unitary state [Lomas-2019]. E-sport event
organizers have also disassociated themselves from top players who
expressed political opinions in relation to the 2019 Hong Kong
protests [Victor-2019]. See also Section 5.3.1.
4.2.5. DPI Identification
DPI technically is any kind of packet analysis beyond IP address and
port number and has become computationally feasible as a component of
censorship mechanisms in recent years [Wagner-2009]. Unlike other
techniques, DPI reassembles network flows to examine the application
"data" section, as opposed to only headers, and is therefore often
used for keyword identification. DPI also differs from other
identification technologies because it can leverage additional packet
and flow characteristics, e.g., packet sizes and timings, when
identifying content. To prevent substantial QoS impacts, DPI
normally analyzes a copy of data while the original packets continue
to be routed. Typically, the traffic is split using either a mirror
switch or fiber splitter and analyzed on a cluster of machines
running Intrusion Detection Systems (IDSs) configured for censorship.
Trade-offs: DPI is one of the most expensive identification
mechanisms and can have a large QoS impact [Porter-2005]. When used
as a keyword filter for TCP flows, DPI systems can cause also major
over-blocking problems. Like other techniques, DPI is less useful
against encrypted data, though DPI can leverage unencrypted elements
of an encrypted data flow (e.g., the Server Name Indication (SNI)
sent in the clear for TLS) or metadata about an encrypted flow (e.g.,
packet sizes, which differ across video and textual flows) to
identify traffic. See Section 4.2.3.1 for more information about
SNI-based filtration mechanisms.
Other kinds of information can be inferred by comparing certain
unencrypted elements exchanged during TLS handshakes to similar data
points from known sources. This practice, called "TLS
fingerprinting", allows a probabilistic identification of a party's
operating system, browser, or application, based on a comparison of
the specific combinations of TLS version, ciphersuites, compression
options, etc., sent in the ClientHello message to similar signatures
found in unencrypted traffic [Husak-2016].
Despite these problems, DPI is the most powerful identification
method and is widely used in practice. The Great Firewall of China
(GFW), the largest censorship system in the world, uses DPI to
identify restricted content over HTTP and DNS and to inject TCP RSTs
and bad DNS responses, respectively, into connections [Crandall-2010]
[Clayton-2006] [Anonymous-2014].
Empirical Examples: Several studies have found evidence of censors
using DPI for censoring content and tools. Clayton et al., Crandal
et al., Anonymous, and Khattak et al., all explored the GFW
[Crandall-2010] [Clayton-2006] [Anonymous-2014]. Khattak et al. even
probed the firewall to discover implementation details like how much
state it stores [Khattak-2013]. The Tor project claims that China,
Iran, Ethiopia, and others must have used DPI to block the obfs2
protocol [Wilde-2012]. Malaysia has been accused of using targeted
DPI, paired with DDoS, to identify and subsequently attack pro-
opposition material [Wagstaff-2013]. It also seems likely that
organizations that are not so worried about blocking content in real
time could use DPI to sort and categorically search gathered traffic
using technologies such as high-speed packet processing
[Hepting-2011].
4.3. Transport Layer
4.3.1. Shallow Packet Inspection and Transport Header Identification
Of the various shallow packet inspection methods, transport header
identification is the most pervasive, reliable, and predictable type
of identification. Transport headers contain a few invaluable pieces
of information that must be transparent for traffic to be
successfully routed: destination and source IP address and port.
Destination and source IP are doubly useful, as not only do they
allow a censor to block undesirable content via IP blocklisting but
also allow a censor to identify the IP of the user making the request
and the IP address of the destination being visited, which in most
cases can be used to infer the domain being visited [Patil-2019].
Port is useful for allowlisting certain applications.
By combining IP address, port, and protocol information found in the
transport header, shallow packet inspection can be used by a censor
to identify specific TCP or UDP endpoints. UDP endpoint blocking has
been observed in the context of QUIC blocking [Elmenhorst-2021].
Trade-offs: Header identification is popular due to its simplicity,
availability, and robustness.
Header identification is trivial to implement in some routers, but is
difficult to implement in backbone or ISP routers at scale, and is
therefore typically implemented with DPI. Blocklisting an IP is
equivalent to installing a specific route on a router (such as a /32
route for IPv4 addresses and a /128 route for IPv6 addresses).
However, due to limited flow table space, this cannot scale beyond a
few thousand IPs at most. IP blocking is also relatively crude. It
often leads to over-blocking and cannot deal with some services like
Content Distribution Networks (CDNs) that host content at hundreds or
thousands of IP addresses. Despite these limitations, IP blocking is
extremely effective because the user needs to proxy their traffic
through another destination to circumvent this type of
identification. In addition, IP blocking is effective against all
protocols above IP, e.g., TCP and QUIC.
Port blocking is generally not useful because many types of content
share the same port, and it is possible for censored applications to
change their port. For example, most HTTP traffic goes over port 80,
so the censor cannot differentiate between restricted and allowed web
content solely on the basis of port. HTTPS goes over port 443, with
similar consequences for the censor except only partial metadata may
now be available to the censor. Port allowlisting is occasionally
used, where a censor limits communication to approved ports (such as
80 for HTTP traffic), and is most effective when used in conjunction
with other identification mechanisms. For example, a censor could
block the default HTTPS port (port 443), thereby forcing most users
to fall back to HTTP. A counterexample is that port 25 (SMTP) has
long been blocked on residential ISP networks to reduce the risk of
email spam, but doing this also prohibits residential ISP customers
from running their own email servers.
4.3.2. Protocol Identification
Censors sometimes identify entire protocols to be blocked using a
variety of traffic characteristics. For example, Iran impairs the
performance of HTTPS traffic, a protocol that prevents further
analysis, to encourage users to switch to HTTP, a protocol that they
can analyze [Aryan-2013]. A simple protocol identification would be
to recognize all TCP traffic over port 443 as HTTPS, but a more
sophisticated analysis of the statistical properties of payload data
and flow behavior would be more effective, even when port 443 is not
used [Hjelmvik-2010] [Sandvine-2015].
If censors can detect circumvention tools, they can block them.
Therefore, censors like China are extremely interested in identifying
the protocols for censorship circumvention tools. In recent years,
this has devolved into a competition between censors and
circumvention tool developers. As part of this competition, China
developed an extremely effective protocol identification technique
that researchers call "active probing" or "active scanning".
In active probing, the censor determines whether hosts are running a
circumvention protocol by trying to initiate communication using the
circumvention protocol. If the host and the censor successfully
negotiate a connection, then the censor conclusively knows that the
host is running a circumvention tool. China has used active scanning
to great effect to block Tor [Winter-2012].
Trade-offs: Protocol identification only provides insight into the
way information is traveling, and not the information itself.
Protocol identification is useful for detecting and blocking
circumvention tools (like Tor) or traffic that is difficult to
analyze (like Voice over IP (VoIP) or SSL) because the censor can
assume that this traffic should be blocked. However, this can lead
to over-blocking problems when used with popular protocols. These
methods are expensive, both computationally and financially, due to
the use of statistical analysis and can be ineffective due to their
imprecise nature.
Censors have also used protocol identification in the past in an
"allowlist" filtering capacity, such as by only allowing specific,
pre-vetted protocols to be used and blocking any unrecognized
protocols [Bock-2020]. These protocol filtering approaches can also
lead to over-blocking if the allowed lists of protocols are too small
or incomplete but can be cheap to implement, as many standard
"allowed" protocols are simple to identify (such as HTTP).
Empirical Examples: Protocol identification can be easy to detect if
it is conducted in real time and only a particular protocol is
blocked. However, some types of protocol identification, like active
scanning, are much more difficult to detect. Protocol identification
has been used by Iran to identify and throttle Secure Shell (SSH)
protocol traffic to make it unusable [Van-der-Sar-2007] and by China
to identify and block Tor relays [Winter-2012]. Protocol
identification has also been used for traffic management, such as the
2007 case where Comcast in the United States used RST injection
(injection of a TCP RST packet into the stream) to interrupt
BitTorrent traffic [Winter-2012]. In 2020, Iran deployed an
allowlist protocol filter, which only allowed three protocols to be
used (DNS, TLS, and HTTP) on specific ports, and censored any
connection it could not identify [Bock-2020]. In 2022, Russia seemed
to have used protocol identification to block most HTTP/3 connections
[Elmenhorst-2022].
4.4. Residual Censorship
Another feature of some modern censorship systems is residual
censorship, a punitive form of censorship whereby after a censor
disrupts a forbidden connection, the censor continues to target
subsequent connections, even if they are innocuous [Bock-2021].
Residual censorship can take many forms and often relies on the
methods of technical interference described in the next section.
An important facet of residual censorship is precisely what the
censor continues to block after censorship is initially triggered.
There are three common options available to an adversary: 2-tuple
(client IP, server IP), 3-tuple (client IP, server IP, server port),
or 4-tuple (client IP, client port, server IP, server port). Future
connections that match the tuple of information the censor records
will be disrupted [Bock-2021].
Residual censorship can sometimes be difficult to identify and can
often complicate censorship measurement.
Trade-offs: The impact of residual censorship is to provide users
with further discouragement from trying to access forbidden content,
though it is not clear how successful it is at accomplishing this.
Empirical Examples: China has used 3-tuple residual censorship in
conjunction with their HTTP censorship for years, and researchers
have reported seeing similar residual censorship for HTTPS. China
seems to use a mix of 3-tuple and 4-tuple residual censorship for
their censorship of HTTPS with ESNI. Some censors that perform
censorship via packet dropping often accidentally implement 4-tuple
residual censorship, including Iran and Kazakhstan [Bock-2021].
5. Technical Interference
5.1. Application Layer
5.1.1. DNS Interference
There are a variety of mechanisms that censors can use to block or
filter access to content by altering responses from the DNS
[AFNIC-2013] [ICANN-SSAC-2012], including blocking the response,
replying with an error message, or responding with an incorrect
address. Note that there are now encrypted transports for DNS
queries in DNS over HTTPS [RFC8484] and DNS over TLS [RFC7858] that
can mitigate interference with DNS queries between the stub and the
resolver.
Responding to a DNS query with an incorrect address can be achieved
with on-path interception, off-path cache poisoning, or lying by the
name server.
"DNS mangling" is a network-level technique of on-path interception
where an incorrect IP address is returned in response to a DNS query
to a censored destination. Some Chinese networks, for example, do
this. (We are not aware of any other wide-scale uses of mangling.)
On those Chinese networks, each DNS request in transit is examined
(presumably by network inspection technologies such as DPI), and if
it matches a censored domain, a false response is injected. End
users can see this technique in action by simply sending DNS requests
to any unused IP address in China (see example below). If it is not
a censored name, there will be no response. If it is censored, a
forged response will be returned. For example, using the command-
line dig utility to query an unused IP address in China of 192.0.2.2
for the name "www.uncensored.example" compared with
"www.censored.example" (censored at the time of writing), we get a
forged IP address "198.51.100.0" as a response:
% dig +short +nodnssec @192.0.2.2 A www.uncensored.example
;; connection timed out; no servers could be reached
% dig +short +nodnssec @192.0.2.2 A www.censored.example
198.51.100.0
DNS cache poisoning happens off-path and refers to a mechanism where
a censor interferes with the response sent by an authoritative DNS
name server to a recursive resolver by responding more quickly than
the authoritative name server can respond with an alternative IP
address [Halley-2008]. Cache poisoning occurs after the requested
site's name servers resolve the request and attempt to forward the
true IP back to the requesting device. On the return route, the
resolved IP is recursively cached by each DNS server that initially
forwarded the request. During this caching process if an undesirable
keyword is recognized, the resolved IP is "poisoned", and an
alternative IP (or NXDOMAIN error) is returned more quickly than the
upstream resolver can respond, causing a forged IP address to be
cached (and potentially recursively so). The alternative IPs usually
direct to a nonsense domain or a warning page. Alternatively,
Iranian censorship appears to prevent the communication en route,
preventing a response from ever being sent [Aryan-2013].
There are also cases of what is colloquially called "DNS lying",
where a censor mandates that the DNS responses provided -- by an
operator of a recursive resolver such as an Internet Access Provider
-- be different than what an authoritative name server would provide
[Bortzmeyer-2015].
Trade-offs: These forms of DNS interference require the censor to
force a user to traverse a controlled DNS hierarchy (or intervening
network on which the censor serves as an active pervasive attacker
[RFC7624] to rewrite DNS responses) for the mechanism to be
effective. DNS interference can be circumvented by using alternative
DNS resolvers (such as any of the public DNS resolvers) that may fall
outside of the jurisdictional control of the censor or Virtual
Private Network (VPN) technology. DNS mangling and cache poisoning
also imply returning an incorrect IP to those attempting to resolve a
domain name, but in some cases the destination may be technically
accessible. For example, over HTTP, the user may have another method
of obtaining the IP address of the desired site and may be able to
access it if the site is configured to be the default server
listening at this IP address. Target blocking has also been a
problem, as occasionally users outside of the censor's region will be
directed through DNS servers or DNS-rewriting network equipment
controlled by a censor, causing the request to fail. The ease of
circumvention paired with the large risk of content blocking and
target blocking make DNS interference a partial, difficult, and less-
than-ideal censorship mechanism.
Additionally, the above mechanisms rely on DNSSEC not being deployed
or DNSSEC validation not being active on the client or recursive
resolver (neither of which is hard to imagine given limited
deployment of DNSSEC and limited client support for DNSSEC
validation). Note that an adversary seeking to merely block
resolution can serve a DNSSEC record that doesn't validate correctly,
assuming of course that the client or recursive resolver validates.
Previously, techniques were used for censorship that relied on DNS
requests being passed in cleartext over port 53 [SSAC-109-2020].
With the deployment of encrypted DNS (e.g., DNS over HTTPS [RFC8484])
these requests are now increasingly passed on port 443 with other
HTTPS traffic, or in the case of DNS over TLS [RFC7858] no longer
passed in the clear (see also Section 4.3.1).
Empirical Examples: DNS interference, when properly implemented, is
easy to identify based on the shortcomings identified above. Turkey
relied on DNS interference for its country-wide block of websites,
including Twitter and YouTube, for almost a week in March of 2014.
The ease of circumvention resulted in an increase in the popularity
of Twitter until Turkish ISPs implemented an IP blocklist to achieve
the governmental mandate [Zmijewski-2014]. Ultimately, Turkish ISPs
started hijacking all requests to Google and Level 3's international
DNS resolvers [Zmijewski-2014]. DNS interference, when incorrectly
implemented, has resulted in some of the largest censorship
disasters. In January 2014, China started directing all requests
passing through the Great Fire Wall to a single domain
"dongtaiwang.com", due to an improperly configured DNS poisoning
attempt. This incident is thought to be the largest Internet service
outage in history [AFP-2014] [Anon-SIGCOMM12]. Countries such as
China, Turkey, and the United States have discussed blocking entire
Top-Level Domains (TLDs) as well [Albert-2011]. DNS blocking is
commonly deployed in European countries to deal with undesirable
content, such as
* child abuse content (Norway, United Kingdom, Belgium, Denmark,
Finland, France, Germany, Ireland, Italy, Malta, the Netherlands,
Poland, Spain, and Sweden [Wright-2013] [Eneman-2010]),
* online gambling (Belgium, Bulgaria, Czech Republic, Cyprus,
Denmark, Estonia, France, Greece, Hungary, Italy, Latvia,
Lithuania, Poland, Portugal, Romania, Slovakia, Slovenia, and
Spain (see Section 6.3.2 of [EC-gambling-2012],
[EC-gambling-2019])),
* copyright infringement (all European Economic Area countries),
* hate speech and extremism (France [Hertel-2015]), and
* terrorism content (France [Hertel-2015]).
5.2. Transport Layer
5.2.1. Performance Degradation
While other interference techniques outlined in this section mostly
focus on blocking or preventing access to content, it can be an
effective censorship strategy in some cases to not entirely block
access to a given destination or service but instead to degrade the
performance of the relevant network connection. The resulting user
experience for a site or service under performance degradation can be
so bad that users opt to use a different site, service, or method of
communication or may not engage in communication at all if there are
no alternatives. Traffic-shaping techniques that rate-limit the
bandwidth available to certain types of traffic is one example of a
performance degradation.
Trade-offs: While implementing a performance degradation will not
always eliminate the ability of people to access a desire resource,
it may force them to use other means of communication where
censorship (or surveillance) is more easily accomplished.
Empirical Examples: Iran has been known to shape the bandwidth
available to HTTPS traffic to encourage unencrypted HTTP traffic
[Aryan-2013].
5.2.2. Packet Dropping
Packet dropping is a simple mechanism to prevent undesirable traffic.
The censor identifies undesirable traffic and chooses to not properly
forward any packets it sees associated with the traversing
undesirable traffic instead of following a normal routing protocol.
This can be paired with any of the previously described mechanisms so
long as the censor knows the user must route traffic through a
controlled router.
Trade-offs: Packet dropping is most successful when every traversing
packet has transparent information linked to undesirable content,
such as a destination IP. One downside packet dropping suffers from
is the necessity of blocking all content from otherwise allowable IPs
based on a single subversive subdomain; blogging services and GitHub
repositories are good examples. China famously dropped all GitHub
packets for three days based on a single repository hosting
undesirable content [Anonymous-2013]. The need to inspect every
traversing packet in almost real time also makes packet dropping
somewhat challenging from a QoS perspective.
Empirical Examples: Packet dropping is a very common form of
technical interference and lends itself to accurate detection given
the unique nature of the timeout requests it leaves in its wake. The
Great Firewall of China has been observed using packet dropping as
one of its primary technical censorship mechanisms [Ensafi-2013].
Iran has also used packet dropping as the mechanism for throttling
SSH [Aryan-2013]. These are but two examples of a ubiquitous
censorship practice. Notably, packet dropping during the handshake
or working connection is the only interference technique observed for
QUIC traffic to date (e.g., in India, Iran, Russia, and Uganda
[Elmenhorst-2021] [Elmenhorst-2022]).
5.2.3. RST Packet Injection
Packet injection, generally, refers to a machine-in-the-middle (MITM)
network interference technique that spoofs packets in an established
traffic stream. RST packets are normally used to let one side of a
TCP connection know the other side has stopped sending information
and that the receiver should close the connection. RST packet
injection is a specific type of packet injection attack that is used
to interrupt an established stream by sending RST packets to both
sides of a TCP connection; as each receiver thinks the other has
dropped the connection, the session is terminated.
QUIC is not vulnerable to these types of injection attacks once the
connection has been set up. While QUIC implements a stateless reset
mechanism, such a reset is only accepted by a peer if the packet ends
in a previously issued (stateless reset) token, which is difficult to
guess. During the handshake, QUIC only provides effective protection
against off-path attackers but is vulnerable to injection attacks by
attackers that have parsed prior packets. (See [RFC9000] for more
details.)
Trade-offs: Although ineffective against non-TCP protocols (QUIC,
IPsec), RST packet injection has a few advantages that make it
extremely popular as a technique employed for censorship. RST packet
injection is an out-of-band interference mechanism, allowing the
avoidance of the QoS bottleneck that one can encounter with inline
techniques such as packet dropping. This out-of-band property allows
a censor to inspect a copy of the information, usually mirrored by an
optical splitter, making it an ideal pairing for DPI and protocol
identification [Weaver-2009]. (This asynchronous version of a MITM
is often called a machine-on-the-side (MOTS).) RST packet injection
also has the advantage of only requiring one of the two endpoints to
accept the spoofed packet for the connection to be interrupted.
The difficult part of RST packet injection is spoofing "enough"
correct information to ensure one endpoint accepts a RST packet as
legitimate; this generally implies a correct IP, port, and TCP
sequence number. The sequence number is the hardest to get correct,
as [RFC9293] specifies that a RST packet should be in sequence to be
accepted, although that RFC also recommends allowing in-window
packets. This in-window recommendation is important; if it is
implemented, it allows for successful Blind RST Injection attacks
[Netsec-2011]. When in-window sequencing is allowed, it is trivial
to conduct a Blind RST Injection. While the term "blind" injection
implies the censor doesn't know any sensitive sequencing information
about the TCP stream they are injecting into, they can simply
enumerate all ~70000 possible windows. This is particularly useful
for interrupting encrypted/obfuscated protocols such as SSH or Tor
[Gilad]. Some censorship evasion systems work by trying to confuse
the censor into tracking incorrect information, rendering their RST
packet injection useless [Khattak-2013] [Wang-2017] [Li-2017]
[Bock-2019] [Wang-2020].
RST packet injection relies on a stateful network, making it useless
against UDP connections. RST packet injection is among the most
popular censorship techniques used today given its versatile nature
and effectiveness against all types of TCP traffic. Recent research
shows that a TCP RST packet injection attack can even work in the
case of an off-path attacker [Cao-2016].
Empirical Examples: RST packet injection, as mentioned above, is most
often paired with identification techniques that require splitting,
such as DPI or protocol identification. In 2007, Comcast was accused
of using RST packet injection to interrupt traffic it identified as
BitTorrent [Schoen-2007], subsequently leading to a US Federal
Communications Commission ruling against Comcast [VonLohmann-2008].
China has also been known to use RST packet injection for censorship
purposes. This interference is especially evident in the
interruption of encrypted/obfuscated protocols, such as those used by
Tor [Winter-2012].
5.3. Routing Layer
5.3.1. Network Disconnection
While it is perhaps the crudest of all techniques employed for
censorship, there is no more effective way of making sure undesirable
information isn't allowed to propagate on the web than by shutting
off the network. The network can be logically cut off in a region
when a censoring entity withdraws all of the Border Gateway Protocol
(BGP) prefixes routing through the censor's country.
Trade-offs: The impact of a network disconnection in a region is huge
and absolute; the censor pays for absolute control over digital
information by losing the benefits a globally accessible Internet
brings. Network disconnections are also politically expensive as
citizens accustomed to accessing Internet platforms and services see
such disconnections as a loss of civil liberty. Network
disconnection is rarely a long-term solution for any censor and is
normally only used as a last resort in times of substantial civil
unrest in a country.
Empirical Examples: Network disconnections tend to only happen in
times of substantial unrest, largely due to the huge social,
political, and economic impact such a move has. One of the first,
highly covered occurrences was when the junta in Myanmar employed
network disconnection to help junta forces quash a rebellion in 2007
[Dobie-2007]. China disconnected the network in the Xinjiang region
during unrest in 2009 in an effort to prevent the protests from
spreading to other regions [Heacock-2009]. The Arab Spring saw the
most frequent usage of network disconnection, with events in Egypt
and Libya in 2011 [Cowie-2011] and Syria in 2012 [Thomson-2012].
Russia indicated that it would attempt to disconnect all Russian
networks from the global Internet in April 2019 as part of a test of
the nation's network independence. Reports also indicate that, as
part of the test disconnect, Russian telecommunications firms must
now route all traffic to state-operated monitoring points
[Cimpanu-2019]. India saw the largest number of Internet shutdowns
per year in 2016 and 2017 [Dada-2017].
5.3.2. Adversarial Route Announcement
More fine-grained and potentially wide-spread censorship can be
achieved with BGP hijacking, which adversarially re-routes BGP IP
prefixes incorrectly within a region and beyond. This restricts and
effectively censors the correctly known location of information that
flows into or out of a jurisdiction and will similarly prevent people
from outside your jurisdiction from viewing content generated outside
that jurisdiction as the adversarial route announcement propagates.
The first can be achieved by an adversarial BGP announcement of
incorrect routes that are not intended to leak beyond a jurisdiction,
where the latter attacks traffic by deliberately introducing bogus
BGP announcements that reach the global Internet.
Trade-offs: A global leak of a misrouted website can overwhelm an ISP
if the website gets a lot of traffic. It is not a permanent solution
because incorrect BGP routes that leak globally can be fixed, but
leaks within a jurisdiction can only be corrected by an ISP/IXP for
local users.
Empirical Examples: In 2008, Pakistan Telecom censored YouTube at the
request of the Pakistan government by changing its BGP routes for the
website. The new routes were announced to the ISP's upstream
providers and beyond. The entire Internet began directing YouTube
routes to Pakistan Telecom and continued doing so for many hours. In
2018, nearly all Google services and Google Cloud customers, like
Spotify, all lost more than one hour of service after Google lost
control of several million of its IP addresses. Those IP prefixes
were being misdirected to China Telecom, a Chinese government-owned
ISP [Google-2018], in a manner similar to the BGP hijacking of US
government and military websites by China Telecom in 2010. ISPs in
both Russia (2022) and Myanmar (2021) have tried to hijack the same
Twitter prefix more than once [Siddiqui-2022].
5.4. Multi-layer and Non-layer
5.4.1. Distributed Denial of Service (DDoS)
Distributed Denial of Service attacks are a common attack mechanism
used by "hacktivists" and malicious hackers. Censors have also used
DDoS in the past for a variety of reasons. There is a wide variety
of DDoS attacks [Wikip-DoS]. However, at a high level, two possible
impacts from the attack tend to occur: a flood attack results in the
service being unusable while resources are being spent to flood the
service, and a crash attack aims to crash the service so resources
can be reallocated elsewhere without "releasing" the service.
Trade-offs: DDoS is an appealing mechanism when a censor would like
to prevent all access (not just regional access) to undesirable
content for a limited period of time. Temporal impermanence is
really the only uniquely beneficial feature of DDoS as a technique
employed for censorship. The resources required to carry out a
successful DDoS against major targets are computationally expensive,
usually requiring rental or ownership of a malicious distributed
platform such as a botnet, and they are imprecise. DDoS is an
incredibly crude censorship technique and appears to largely be used
as a timely, easy-to-access mechanism for blocking undesirable
content for a limited period of time.
Empirical Examples: In 2012, the U.K.'s signals intelligence
organization, the Government Communications Headquarters (GCHQ), used
DDoS to temporarily shutdown Internet Relay Chat (IRC) chat rooms
frequented by members of Anonymous using the Syn Flood DDoS method;
Syn Flood exploits the handshake used by TCP to overload the victim
server with so many requests that legitimate traffic becomes slow or
impossible [NBC-2014] [CERT-2000]. Dissenting opinion websites are
frequently victims of DDoS around politically sensitive events like
the DDoS in Burma [Villeneuve-2011]. Controlling parties in Russia
[Kravtsova-2012], Zimbabwe [Orion-2013], and Malaysia
[Muncaster-2013] have been accused of using DDoS to interrupt
opposition support and access during elections. In 2015, China
launched a DDoS attack using a true MITM system (dubbed "Great
Cannon"), collocated with the Great Firewall, that was able to inject
JavaScript code into web visits to a Chinese search engine that
commandeered those user agents to send DDoS traffic to various sites
[Marczak-2015].
5.4.2. Censorship in Depth
Often, censors implement multiple techniques in tandem, creating
"censorship in depth". Censorship in depth can take many forms; some
censors block the same content through multiple techniques (such as
blocking a domain by DNS, IP blocking, and HTTP simultaneously), some
deploy parallel systems to improve censorship reliability (such as
deploying multiple different censorship systems to block the same
domain), and others can use complimentary systems to limit evasion
(such as by blocking unwanted protocols entirely, forcing users to
use other filtered protocols).
Trade-offs: Censorship in depth can be attractive for censors to
deploy, as it offers additional guarantees about censorship: even if
someone evades one type of censorship, they may still be blocked by
another. The main drawback to this approach is the cost to initial
deployment, as it requires the system to deploy multiple censorship
systems in tandem.
Empirical Examples: Censorship in depth is present in many large
censoring nation states today. Researchers have observed that China
has deployed significant censorship in depth, often censoring the
same resource across multiple protocols [Chai-2019] [Bock-2020b] or
deploying additional censorship systems to censor the same content
and protocol [Bock-2021b]. Iran also has deployed a complimentary
protocol filter to limit which protocols can be used on certain
ports, forcing users to rely on protocols their censorship system can
filter [Bock-2020].
6. Non-technical Interference
6.1. Manual Filtering
As the name implies, sometimes manual labor is the easiest way to
figure out which content to block. Manual filtering differs from the
common tactic of building up blocklists in that it doesn't
necessarily target a specific IP or DNS but instead removes or flags
content. Given the imprecise nature of automatic filtering, manually
sorting through content and flagging dissenting websites, blogs,
articles, and other media for filtration can be an effective
technique on its own or combined with other automated techniques of
detection that are then followed by an action that would require
manual confirmation. This filtration can occur on the backbone or
ISP level. China's army of monitors is a good example [BBC-2013b],
but more commonly, manual filtering occurs on an institutional level.
ICPs, such as Google or Weibo, require a business license to operate
in China. One of the prerequisites for a business license is an
agreement to sign a "voluntary pledge" known as the "Public Pledge on
Self-discipline for the Chinese Internet Industry". The failure to
"energetically uphold" the pledged values can lead to the ICPs being
held liable for the offending content by the Chinese government
[BBC-2013b].
6.2. Self-Censorship
Self-censorship is difficult to document as it manifests primarily
through a lack of undesirable content. Tools that encourage self-
censorship may lead a prospective speaker to believe that speaking
increases the risk of unfavorable outcomes for the speaker (technical
monitoring, identification requirements, etc.). Reporters Without
Borders exemplify methods of imposing self-censorship in their annual
World Press Freedom Index reports [RWB-2020].
6.3. Server Takedown
As mentioned in passing by [Murdoch-2008], servers must have a
physical location somewhere in the world. If undesirable content is
hosted in the censoring country, the servers can be physically
seized, or -- in cases where a server is virtualized in a cloud
infrastructure where it may not necessarily have a fixed physical
location -- the hosting provider can be required to prevent access.
6.4. Notice and Takedown
In many countries, legal mechanisms exist where an individual or
other content provider can issue a legal request to a content host
that requires the host to take down content. Examples include the
systems employed by companies like Google to comply with "Right to be
Forgotten" policies in the European Union [Google-RTBF], intermediary
liability rules for electronic platform providers [EC-2012], or the
copyright-oriented notice and takedown regime of the United States
Digital Millennium Copyright Act (DMCA) Section 512 [DMLP-512].
6.5. Domain Name Seizures
Domain names are catalogued in name servers operated by legal
entities called registries. These registries can be made to cede
control over a domain name to someone other than the entity that
registered the domain name through a legal procedure grounded in
either private contracts or public law. Domain name seizure is
increasingly used by both public authorities and private entities to
deal with undesired content dissemination [ICANN-2012] [EFF-2017].
7. Future Work
In addition to establishing a thorough resource for describing
censorship techniques, this document implicates critical areas for
future work.
Taken as a whole, the apparent costs of implementation of censorship
techniques indicate a need for better classification of censorship
regimes as they evolve and mature and better specification of
censorship circumvention techniques themselves. Censor maturity
refers to the technical maturity required of the censor to perform
the specific censorship technique. Future work might classify
techniques by essentially how hard a censor must work, including what
infrastructure is required, in order to successfully censor content,
users, or services.
On circumvention, the increase in protocols leveraging encryption is
an effective countermeasure against some forms of censorship
described in this document, but that thorough research on
circumvention and encryption is left for another document. Moreover,
the censorship circumvention community has developed an area of
research on "pluggable transports," which collect, document, and make
agile methods for obfuscating the on-path traffic of censorship
circumvention tools such that it appears indistinguishable from other
kinds of traffic [Tor-2019]. Those methods would benefit from future
work in the Internet standards community, too.
Lastly, the empirical examples demonstrate that censorship techniques
can evolve quickly, and experience shows that this document can only
be a point-in-time statement. Future work might extend this document
with updates and new techniques described using a comparable
methodology.
8. IANA Considerations
This document has no IANA actions.
9. Security Considerations
This document is a survey of existing literature on network
censorship techniques. As such, it does not introduce any new
security considerations to be taken into account beyond what is
already discussed in each paper surveyed.
10. Informative References
[AFNIC-2013]
AFNIC, "Report of the AFNIC Scientific Council:
Consequences of DNS-based Internet filtering", January
2013,
<http://www.afnic.fr/medias/documents/conseilscientifique/
SC-consequences-of-DNS-based-Internet-filtering.pdf>.
[AFP-2014] AFP, "China Has Massive Internet Breakdown Reportedly
Caused By Their Own Censoring Tools", January 2014,
<http://www.businessinsider.com/chinas-internet-breakdown-
reportedly-caused-by-censoring-tools-2014-1>.
[Albert-2011]
Albert, K., "DNS Tampering and the new ICANN gTLD Rules",
June 2011, <https://opennet.net/blog/2011/06/dns-
tampering-and-new-icann-gtld-rules>.
[Anon-SIGCOMM12]
Anonymous, "The Collateral Damage of Internet Censorship
by DNS Injection", July 2012,
<http://www.sigcomm.org/sites/default/files/ccr/
papers/2012/July/2317307-2317311.pdf>.
[Anonymous-2013]
Anonymous, "GitHub blocked in China - how it happened, how
to get around it, and where it will take us", January
2013, <https://en.greatfire.org/blog/2013/jan/github-
blocked-china-how-it-happened-how-get-around-it-and-where-
it-will-take-us>.
[Anonymous-2014]
Anonymous, "Towards a Comprehensive Picture of the Great
Firewall's DNS Censorship", August 2014,
<https://www.usenix.org/system/files/conference/foci14/
foci14-anonymous.pdf>.
[Aryan-2013]
Aryan, S., Aryan, H., and J. A. Halderman, "Internet
Censorship in Iran: A First Look", 2012,
<https://jhalderm.com/pub/papers/iran-foci13.pdf>.
[BBC-2013] BBC News, "Google and Microsoft agree steps to block abuse
images", November 2013,
<http://www.bbc.com/news/uk-24980765>.
[BBC-2013b]
BBC, "China employs two million microblog monitors state
media say", 2013,
<https://www.bbc.com/news/world-asia-china-24396957>.
[Bock-2019]
Bock, K., Hughey, G., Qiang, X., and D. Levin, "Geneva:
Evolving Censorship Evasion Strategies",
DOI 10.1145/3319535.3363189, November 2019,
<https://geneva.cs.umd.edu/papers/geneva_ccs19.pdf>.
[Bock-2020]
Bock, K., Fax, Y., Reese, K., Singh, J., and D. Levin,
"Detecting and Evading Censorship-in-Depth: A Case Study
of Iran's Protocol Filter", January 2020,
<https://geneva.cs.umd.edu/papers/evading-censorship-in-
depth.pdf>.
[Bock-2020b]
Bock, K., iyouport, Anonymous, Merino, L-H., Fifield, D.,
Houmansadr, A., and D. Levin, "Exposing and Circumventing
China's Censorship of ESNI", August 2020,
<https://geneva.cs.umd.edu/posts/china-censors-esni/
esni/>.
[Bock-2021]
Bock, K., Bharadwaj, P., Singh, J., and D. Levin, "Your
Censor is My Censor: Weaponizing Censorship Infrastructure
for Availability Attacks",
DOI 10.1109/SPW53761.2021.00059, May 2021,
<https://geneva.cs.umd.edu/papers/woot21-weaponizing-
availability.pdf>.
[Bock-2021b]
Bock, K., Naval, G., Reese, K., and D. Levin, "Even
Censors Have a Backup: Examining China's Double HTTPS
Censorship Middleboxes", FOCI '21: Proceedings of the ACM
SIGCOMM 2021 Workshop on Free and Open Communications on
the Internet, Pages 1-7, DOI 10.1145/3473604.3474559,
August 2021,
<https://geneva.cs.umd.edu/papers/foci21.pdf>.
[Bortzmeyer-2015]
Bortzmeyer, S., "DNS Censorship (DNS Lies) As Seen By RIPE
Atlas", December 2015,
<https://labs.ripe.net/Members/stephane_bortzmeyer/dns-
censorship-dns-lies-seen-by-atlas-probes>.
[Boyle-1997]
Boyle, J., "Foucault in Cyberspace: Surveillance,
Sovereignty, and Hardwired Censors", 66 University of
Cincinnati Law Review 177-205, 1997,
<https://scholarship.law.duke.edu/
faculty_scholarship/619/>.
[Cao-2016] Cao, Y., Qian, Z., Wang, Z., Dao, T., Krishnamurthy, S.,
and L. Marvel, "Off-Path TCP Exploits: Global Rate Limit
Considered Dangerous", August 2016,
<https://www.usenix.org/system/files/conference/
usenixsecurity16/sec16_paper_cao.pdf>.
[CERT-2000]
CERT, "CERT Advisory CA-1996-21 TCP SYN Flooding and IP
Spoofing Attacks", 2000,
<https://vuls.cert.org/confluence/display/historical/
CERT+Advisory+CA-
1996-21+TCP+SYN+Flooding+and+IP+Spoofing+Attacks>.
[Chai-2019]
Chai, Z., Ghafari, A., and A. Houmansadr, "On the
Importance of Encrypted-SNI (ESNI) to Censorship
Circumvention", 2019,
<https://www.usenix.org/system/files/
foci19-paper_chai_update.pdf>.
[Cheng-2010]
Cheng, J., "Google stops Hong Kong auto-redirect as China
plays hardball", June 2010, <http://arstechnica.com/tech-
policy/2010/06/google-tweaks-china-to-hong-kong-redirect-
same-results/>.
[Cimpanu-2019]
Cimpanu, C., "Russia to disconnect from the internet as
part of a planned test", February 2019,
<https://www.zdnet.com/article/russia-to-disconnect-from-
the-internet-as-part-of-a-planned-test/>.
[CitizenLab-2018]
Marczak, B., Dalek, J., McKune, S., Senft, A., Scott-
Railton, J., and R. Deibert, "Bad Traffic: Sandvine's
PacketLogic Devices Used to Deploy Government Spyware in
Turkey and Redirect Egyptian Users to Affiliate Ads?",
March 2018, <https://citizenlab.ca/2018/03/bad-traffic-
sandvines-packetlogic-devices-deploy-government-spyware-
turkey-syria/>.
[Clayton-2006]
Clayton, R., Murdoch, S.J., and R.N.M. Watson, "Ignoring
the Great Firewall of China", Lecture Notes in Computer
Science, Volume 4258, DOI 10.1007/11957454_2, 2006,
<https://link.springer.com/chapter/10.1007/11957454_2>.
[Condliffe-2013]
Condliffe, J., "Google Announces Massive New Restrictions
on Child Abuse Search Terms", November 2013,
<http://gizmodo.com/google-announces-massive-new-
restrictions-on-child-abus-1466539163>.
[Cowie-2011]
Cowie, J., "Egypt Leaves The Internet", NANOG 51, February
2011,
<https://archive.nanog.org/meetings/nanog51/presentations/
Tuesday/LT-Cowie-Egypt%20Leaves%20The%20Internet.pdf>.
[Crandall-2010]
Park, J.C. and J. Crandall, "Empirical Study of a
National-Scale Distributed Intrusion Detection System:
Backbone-Level Filtering of HTML Responses in China", June
2010, <http://www.cs.unm.edu/~crandall/icdcs2010.pdf>.
[Dada-2017]
Dada, T. and P. Micek, "Launching STOP: the #KeepItOn
internet shutdown tracker", September 2017,
<https://www.accessnow.org/keepiton-shutdown-tracker/>.
[Dalek-2013]
Dalek, J., Haselton, B., Noman, H., Senft, A., Crete-
Nishihata, M., Gill, P., and R. J. Deibert, "A Method for
Identifying and Confirming the Use of URL Filtering
Products for Censorship", IMC '13: Proceedings of the 2013
conference on Internet measurement conference, Pages
23-30, DOI 10.1145/2504730.2504763, October 2013,
<http://conferences.sigcomm.org/imc/2013/papers/imc112s-
dalekA.pdf>.
[Ding-1999]
Ding, C., Chi, C. H., Deng, J., and C. L. Dong,
"Centralized Content-Based Web Filtering and Blocking: How
Far Can It Go?", IEEE SMC'99 Conference Proceedings,
DOI 10.1109/ICSMC.1999.825218, October 1999,
<http://citeseerx.ist.psu.edu/viewdoc/
download?doi=10.1.1.132.3302&rep=rep1&type=pdf>.
[DMLP-512] Digital Media Law Project, "Protecting Yourself Against
Copyright Claims Based on User Content", May 2012,
<https://www.dmlp.org/legal-guide/protecting-yourself-
against-copyright-claims-based-user-content>.
[Dobie-2007]
Dobie, M., "Junta tightens media screw", BBC News,
September 2007,
<http://news.bbc.co.uk/2/hi/asia-pacific/7016238.stm>.
[EC-2012] European Commission, "Summary of the results of the Public
Consultation on the future of electronic commerce in the
Internal Market and the implementation of the Directive on
electronic commerce (2000/31/EC)", January 2012,
<https://ec.europa.eu/information_society/newsroom/image/
document/2017-4/
consultation_summary_report_en_2010_42070.pdf>.
[EC-gambling-2012]
European Commission, "Online gambling in the Internal
Market Accompanying the document Communication from the
Commission to the European Parliament, the Council, the
Economic and Social Committee and the Committee of the
Regions Towards a comprehensive framework for online
gambling", 2012, <https://eur-lex.europa.eu/legal-
content/EN/TXT/?uri=CELEX:52012SC0345>.
[EC-gambling-2019]
European Commission, "Evaluation of regulatory tools for
enforcing online gambling rules and channelling demand
towards controlled offers", January 2019,
<https://ec.europa.eu/growth/content/evaluation-
regulatory-tools-enforcing-online-gambling-rules-and-
channelling-demand-towards-1_en>.
[EFF-2017] Malcom, J., Rossi, G., and M. Stoltz, "Which Internet
registries offer the best protection for domain owners?",
Electronic Frontier Foundation, July 2017,
<https://www.eff.org/files/2017/08/02/
domain_registry_whitepaper.pdf>.
[ekr-2021] Rescorla, E., "Overview of Apple's Client-side CSAM
Scanning", August 2021,
<https://educatedguesswork.org/posts/apple-csam-intro/>.
[Elmenhorst-2021]
Elmenhorst, K., Schuetz, B., Aschenbruck, N., and S.
Basso, "Web Censorship Measurements of HTTP/3 over QUIC",
IMC '21: Proceedings of the 21st ACM Internet Measurement
Conference, Pages 276-282, DOI 10.1145/3487552.3487836,
November 2021,
<https://dl.acm.org/doi/pdf/10.1145/3487552.3487836>.
[Elmenhorst-2022]
Elmenhorst, K., "A Quick Look at QUIC Censorship", April
2022,
<https://www.opentech.fund/news/a-quick-look-at-quic/>.
[Eneman-2010]
Eneman, M., "Internet service provider (ISP) filtering of
child-abusive material: A critical reflection of its
effectiveness", DOI 10.1080/13552601003760014, June 2010,
<https://www.tandfonline.com/doi/
abs/10.1080/13552601003760014>.
[Ensafi-2013]
Ensafi, R., Knockel, J., Alexander, G., and J.R. Crandall,
"Detecting Intentional Packet Drops on the Internet via
TCP/IP Side Channels: Extended Version",
DOI 10.48550/arXiv.1312.5739, December 2013,
<http://arxiv.org/pdf/1312.5739v1.pdf>.
[Fifield-2015]
Fifield, D., Lan, C., Hynes, R., Wegmann, P., and V.
Paxson, "Blocking-resistant communication through domain
fronting", DOI 10.1515/popets-2015-0009, May 2015,
<https://petsymposium.org/2015/papers/03_Fifield.pdf>.
[Gatlan-2019]
Gatlan, S., "South Korea is Censoring the Internet by
Snooping on SNI Traffic", February 2019,
<https://www.bleepingcomputer.com/news/security/south-
korea-is-censoring-the-internet-by-snooping-on-sni-
traffic/>.
[Gilad] Gilad, Y. and A. Herzberg, "Off-Path TCP Injection
Attacks", ACM Transactions on Information and System
Security, Volume 16, Issue 4, Article No.: 13, pp. 1-32,
DOI 10.1145/2597173, April 2014,
<https://doi.org/10.1145/2597173>.
[Glanville-2008]
Glanville, J., "The big business of net censorship", The
Guardian, November 2008,
<http://www.theguardian.com/commentisfree/2008/nov/17/
censorship-internet>.
[Google-2018]
"Google Cloud Networking Incident #18018", November 2018,
<https://status.cloud.google.com/incident/cloud-
networking/18018>.
[Google-RTBF]
Google, Inc., "Search removal request under data
protection law in Europe", 2015,
<https://support.google.com/legal/contact/
lr_eudpa?product=websearch>.
[Grover-2019]
Grover, G., Singh, K., and E. Hickok, Ed., "Reliance Jio
is using SNI inspection to block websites", November 2019,
<https://cis-india.org/internet-governance/blog/reliance-
jio-is-using-sni-inspection-to-block-websites>.
[HADOPI] Hadopi, "Hadopi | Haute Autorité pour la diffusion des
oeuvres et la protection des droits sur internet",
<https://www.hadopi.fr/>.
[Halley-2008]
Halley, B., "How DNS cache poisoning works", October 2008,
<https://www.networkworld.com/article/2277316/tech-
primers/tech-primers-how-dns-cache-poisoning-works.html>.
[Heacock-2009]
Heacock, R., "China shuts down Internet in Xinjiang region
after riots", OpenNet Initiative, July 2009,
<https://opennet.net/blog/2009/07/china-shuts-down-
internet-xinjiang-region-after-riots>.
[Hepting-2011]
Wikipedia, "Hepting v. AT&T", September 2023,
<https://en.wikipedia.org/wiki/
Hepting_v._AT%26T&oldid=1175143505>.
[Hertel-2015]
Hertel, O., "Comment les autorités peuvent bloquer un site
Internet" [How authorities can block a website], March
2015, <https://www.sciencesetavenir.fr/high-tech/comment-
les-autorites-peuvent-bloquer-un-site-internet_35828>.
[Hjelmvik-2010]
Hjelmvik, E. and W. John, "Breaking and Improving Protocol
Obfuscation", Technical Report No. 2010-05, ISSN
1652-926X, July 2010,
<https://www.iis.se/docs/hjelmvik_breaking.pdf>.
[Husak-2016]
Husák, M., Čermák, M., Jirsík, T., and P. Čeleda, "HTTPS
traffic analysis and client identification using passive
SSL/TLS fingerprinting", DOI 10.1186/s13635-016-0030-7,
February 2016, <https://link.springer.com/article/10.1186/
s13635-016-0030-7>.
[ICANN-2012]
ICANN Security and Stability Advisory Committee, "Guidance
for Preparing Domain Name Orders, Seizures & Takedowns",
January 2012,
<https://www.icann.org/en/system/files/files/guidance-
domain-seizures-07mar12-en.pdf>.
[ICANN-SSAC-2012]
ICANN Security and Stability Advisory Committee (SSAC),
"SAC 056: SSAC Advisory on Impacts of Content Blocking via
the Domain Name System", October 2012,
<https://www.icann.org/en/system/files/files/sac-
056-en.pdf>.
[Jones-2014]
Jones, B., Lee, T-W., Feamster, N., and P. Gill,
"Automated Detection and Fingerprinting of Censorship
Block Pages", IMC '14: Proceedings of the 2014 Conference
on Internet Measurement Conference, Pages 299-304,
DOI 10.1145/2663716.2663722, November 2014,
<http://conferences2.sigcomm.org/imc/2014/papers/
p299.pdf>.
[Khattak-2013]
Khattak, S., Javed, M., Anderson, P.D., and V. Paxson,
"Towards Illuminating a Censorship Monitor's Model to
Facilitate Evasion", August 2013, <http://0b4af6cdc2f0c599
8459-c0245c5c937c5dedcca3f1764ecc9b2f.r43.cf2.rackcdn.com/
12389-foci13-khattak.pdf>.
[Knight-2005]
Knight, W., "Iranian net censorship powered by US
technology", June 2005,
<https://www.newscientist.com/article/dn7589-iranian-net-
censorship-powered-by-us-technology/>.
[Knockel-2021]
Knockel, J. and L. Ruan, "Measuring QQMail's automated
email censorship in China", FOCI '21: Proceedings of the
ACM SIGCOMM 2021 Workshop on Free and Open Communications
on the Internet, Pages 8-15, DOI 10.1145/3473604.3474560,
April 2021,
<https://dl.acm.org/doi/10.1145/3473604.3474560>.
[Kravtsova-2012]
Kravtsova, Y., "Cyberattacks Disrupt Opposition's
Election", The Moscow Times, October 2012,
<http://www.themoscowtimes.com/news/article/cyberattacks-
disrupt-oppositions-election/470119.html>.
[Leyba-2019]
Leyba, K., Edwards, B., Freeman, C., Crandall, J., and S.
Forrest, "Borders and gateways: measuring and analyzing
national as chokepoints", COMPASS '19: Proceedings of the
2nd ACM SIGCAS Conference on Computing and Sustainable
Societies, pages 184-194, DOI 10.1145/3314344.3332502,
July 2019, <https://doi.org/10.1145/3314344.3332502>.
[Li-2017] Li, F., Razaghpanah, A., Molavi Kakhki, A., Akhavan Niaki,
A., Choffnes, D., Gill, P., and A. Mislove, "lib•erate,
(n): a library for exposing (traffic-classification) rules
and avoiding them efficiently",
DOI 10.1145/3131365.3131376, November 2017,
<https://david.choffnes.com/pubs/liberate-imc17.pdf>.
[Lomas-2019]
Lomas, N., "Github removes Tsunami Democràtic's APK after
a takedown order from Spain", October 2019,
<https://techcrunch.com/2019/10/30/github-removes-tsunami-
democratics-apk-after-a-takedown-order-from-spain/>.
[Marczak-2015]
Marczak, B., Weaver, N., Dalek, J., Ensafi, R., Fifield,
D., McKune, S., Rey, A., Scott-Railton, J., Deibert, R.,
and V. Paxson, "An Analysis of China's "Great Cannon"",
August 2015,
<https://www.usenix.org/system/files/conference/foci15/
foci15-paper-marczak.pdf>.
[Muncaster-2013]
Muncaster, P., "Malaysian election sparks web blocking/
DDoS claims", The Register, May 2013,
<http://www.theregister.co.uk/2013/05/09/
malaysia_fraud_elections_ddos_web_blocking/>.
[Murdoch-2008]
Murdoch, S. J. and R. Anderson, "Tools and Technology of
Internet Filtering" in "Access Denied: The Practice and
Policy of Global Internet Filtering",
DOI 10.7551/mitpress/7617.003.0006, 2008,
<https://doi.org/10.7551/mitpress/7617.003.0006>.
[NA-SK-2019]
Morgus, R., Sherman, J., and S. Nam, "Analysis: South
Korea's New Tool for Filtering Illegal Internet Content",
March 2019, <https://www.newamerica.org/cybersecurity-
initiative/c2b/c2b-log/analysis-south-koreas-sni-
monitoring/>.
[Nabi-2013]
Nabi, Z., "The Anatomy of Web Censorship in Pakistan",
August 2013, <http://0b4af6cdc2f0c5998459-c0245c5c937c5ded
cca3f1764ecc9b2f.r43.cf2.rackcdn.com/12387-foci13-nabi.pdf
>.
[NBC-2014] NBC News, "Exclusive: Snowden Docs Show UK Spies Attacked
Anonymous, Hackers", February 2014,
<http://www.nbcnews.com/feature/edward-snowden-interview/
exclusive-snowden-docs-show-uk-spies-attacked-anonymous-
hackers-n21361>.
[Netsec-2011]
n3t2.3c, "TCP-RST Injection", October 2011,
<https://nets.ec/TCP-RST_Injection>.
[OONI-2018]
Evdokimov, L., "Iran Protests: DPI blocking of Instagram
(Part 2)", February 2018,
<https://ooni.org/post/2018-iran-protests-pt2/>.
[OONI-2019]
Singh, S., Filastò, A., and M. Xynou, "China is now
blocking all language editions of Wikipedia", May 2019,
<https://ooni.org/post/2019-china-wikipedia-blocking/>.
[Orion-2013]
Orion, E., "Zimbabwe election hit by hacking and DDoS
attacks", Wayback Machine archive, August 2013, <https://w
eb.archive.org/web/20130825010947/http://www.theinquirer.n
et/inquirer/news/2287433/zimbabwe-election-hit-by-hacking-
and-ddos-attacks>.
[Patil-2019]
Patil, S. and N. Borisov, "What can you learn from an
IP?", Proceedings of the Applied Networking Research
Workshop, Pages 45-51, DOI 10.1145/3340301.3341133, July
2019, <https://irtf.org/anrw/2019/
anrw2019-final44-acmpaginated.pdf>.
[Porter-2005]
Porter, T., "The Perils of Deep Packet Inspection", 2010,
<http://www.symantec.com/connect/articles/perils-deep-
packet-inspection>.
[Rambert-2021]
Rampert, R., Weinberg, Z., Barradas, D., and N. Christin,
"Chinese Wall or Swiss Cheese? Keyword filtering in the
Great Firewall of China", DOI 10.1145/3442381.3450076,
April 2021,
<https://www.andrew.cmu.edu/user/nicolasc/publications/
Rambert-WWW21.pdf>.
[Reda-2017]
Reda, F., "New EU law prescribes website blocking in the
name of "consumer protection"", November 2017,
<https://felixreda.eu/2017/11/eu-website-blocking/>.
[RFC6066] Eastlake 3rd, D., "Transport Layer Security (TLS)
Extensions: Extension Definitions", RFC 6066,
DOI 10.17487/RFC6066, January 2011,
<https://www.rfc-editor.org/info/rfc6066>.
[RFC7624] Barnes, R., Schneier, B., Jennings, C., Hardie, T.,
Trammell, B., Huitema, C., and D. Borkmann,
"Confidentiality in the Face of Pervasive Surveillance: A
Threat Model and Problem Statement", RFC 7624,
DOI 10.17487/RFC7624, August 2015,
<https://www.rfc-editor.org/info/rfc7624>.
[RFC7754] Barnes, R., Cooper, A., Kolkman, O., Thaler, D., and E.
Nordmark, "Technical Considerations for Internet Service
Blocking and Filtering", RFC 7754, DOI 10.17487/RFC7754,
March 2016, <https://www.rfc-editor.org/info/rfc7754>.
[RFC7858] Hu, Z., Zhu, L., Heidemann, J., Mankin, A., Wessels, D.,
and P. Hoffman, "Specification for DNS over Transport
Layer Security (TLS)", RFC 7858, DOI 10.17487/RFC7858, May
2016, <https://www.rfc-editor.org/info/rfc7858>.
[RFC8484] Hoffman, P. and P. McManus, "DNS Queries over HTTPS
(DoH)", RFC 8484, DOI 10.17487/RFC8484, October 2018,
<https://www.rfc-editor.org/info/rfc8484>.
[RFC8744] Huitema, C., "Issues and Requirements for Server Name
Identification (SNI) Encryption in TLS", RFC 8744,
DOI 10.17487/RFC8744, July 2020,
<https://www.rfc-editor.org/info/rfc8744>.
[RFC9000] Iyengar, J., Ed. and M. Thomson, Ed., "QUIC: A UDP-Based
Multiplexed and Secure Transport", RFC 9000,
DOI 10.17487/RFC9000, May 2021,
<https://www.rfc-editor.org/info/rfc9000>.
[RFC9293] Eddy, W., Ed., "Transmission Control Protocol (TCP)",
STD 7, RFC 9293, DOI 10.17487/RFC9293, August 2022,
<https://www.rfc-editor.org/info/rfc9293>.
[Rushe-2014]
Rushe, D., "Bing censoring Chinese language search results
for users in the US", The Guardian, February 2014,
<http://www.theguardian.com/technology/2014/feb/11/bing-
censors-chinese-language-search-results>.
[RWB-2020] Reporters Without Borders (RSF), "2020 World Press Freedom
Index: 'Entering a decisive decade for journalism,
exacerbated by coronavirus'", April 2020,
<https://rsf.org/en/2020-world-press-freedom-index-
entering-decisive-decade-journalism-exacerbated-
coronavirus>.
[Sandvine-2015]
Sandvine, "Internet Traffic Classification: A Sandvine
Technology Showcase", 2015,
<https://www.researchgate.net/profile/Nirmala-Svsg/post/
Anybody-working-on-Internet-traffic-
classification/attachment/59d63a5779197b807799782d/
AS%3A405810988503040%401473764287142/download/traffic-
classification-identifying-and-measuring-internet-
traffic.pdf>.
[Satija-2021]
Satija, S. and R. Chatterjee, "BlindTLS: Circumventing
TLS-based HTTPS censorship", FOCI '21: Proceedings of the
ACM SIGCOMM 2021 Workshop on Free and Open Communications
on the Internet, Pages 43-49, DOI 10.1145/3473604.3474564,
August 2021,
<https://sambhav.info/files/blindtls-foci21.pdf>.
[Schoen-2007]
Schoen, S., "EFF tests agree with AP: Comcast is forging
packets to interfere with user traffic", October 2007,
<https://www.eff.org/deeplinks/2007/10/eff-tests-agree-ap-
comcast-forging-packets-to-interfere>.
[Senft-2013]
, Crete-Nishihata, M., Dalek, J., Hardy, S., Hilts, A.,
Kleemola, K., Ng, J., Poetranto, I., Senft, A., Sinpeng,
A., Sonne, B., and G. Wiseman, "Asia Chats: Analyzing
Information Controls and Privacy in Asian Messaging
Applications", November 2013,
<https://citizenlab.org/2013/11/asia-chats-analyzing-
information-controls-privacy-asian-messaging-
applications/>.
[Shbair-2015]
Shbair, W. M., Cholez, T., Goichot, A., and I. Chrisment,
"Efficiently Bypassing SNI-based HTTPS Filtering", May
2015, <https://hal.inria.fr/hal-01202712/document>.
[Siddiqui-2022]
Siddiqui, A., "Lesson Learned: Twitter Shored Up Its
Routing Security", March 2022,
<https://www.manrs.org/2022/03/lesson-learned-twitter-
shored-up-its-routing-security/>.
[SIDN-2020]
Moura, G., "Detecting and Taking Down Fraudulent Webshops
at the .nl ccTLD", February 2020,
<https://labs.ripe.net/Members/giovane_moura/detecting-
and-taking-down-fraudulent-webshops-at-a-cctld>.
[Singh-2019]
Singh, K., Grover, G., and V. Bansal, "How India Censors
the Web", DOI 10.48550/arXiv.1912.08590, December 2019,
<https://arxiv.org/abs/1912.08590>.
[Sophos-2023]
Sophos, "Sophos Firewall: Web filtering basics", 2023,
<https://support.sophos.com/support/s/article/KB-
000036518?language=en_US>.
[SSAC-109-2020]
ICANN Security and Stability Advisory Committee (SSAC),
"SAC109: The Implications of DNS over HTTPS and DNS over
TLS", March 2020,
<https://www.icann.org/en/system/files/files/sac-
109-en.pdf>.
[Tang-2016]
Tang, C., "In-depth analysis of the Great Firewall of
China", December 2016,
<https://www.cs.tufts.edu/comp/116/archive/fall2016/
ctang.pdf>.
[Thomson-2012]
Thomson, I., "Syria cuts off internet and mobile
communication", The Register, November 2012,
<http://www.theregister.co.uk/2012/11/29/
syria_internet_blackout/>.
[TLS-ESNI] Rescorla, E., Oku, K., Sullivan, N., and C. A. Wood, "TLS
Encrypted Client Hello", Work in Progress, Internet-Draft,
draft-ietf-tls-esni-17, 9 October 2023,
<https://datatracker.ietf.org/doc/html/draft-ietf-tls-
esni-17>.
[Tor-2019] Tor, "Tor: Pluggable Transports", 2019,
<https://2019.www.torproject.org/docs/pluggable-
transports.html.en>.
[Trustwave-2015]
Trustwave, "Filter : SNI extension feature and HTTPS
blocking", 2015,
<https://www3.trustwave.com/software/8e6/hlp/r3000/
files/1system_filter.html>.
[Tschantz-2016]
Tschantz, M., Afroz, S., Anonymous, and V. Paxson, "SoK:
Towards Grounding Censorship Circumvention in Empiricism",
DOI 10.1109/SP.2016.59, May 2016,
<https://oaklandsok.github.io/papers/tschantz2016.pdf>.
[Van-der-Sar-2007]
Van der Sar, E., "How To Bypass Comcast's BitTorrent
Throttling", October 2012, <https://torrentfreak.com/how-
to-bypass-comcast-bittorrent-throttling-071021>.
[Verkamp-2012]
Verkamp, J. P. and M. Gupta, "Inferring Mechanics of Web
Censorship Around the World", August 2012,
<https://www.usenix.org/system/files/conference/foci12/
foci12-final1.pdf>.
[Victor-2019]
Victor, D., "Blizzard Sets Off Backlash for Penalizing
Hearthstone Gamer in Hong Kong", The New York Times,
October 2019,
<https://www.nytimes.com/2019/10/09/world/asia/blizzard-
hearthstone-hong-kong.html>.
[Villeneuve-2011]
Villeneuve, N. and M. Crete-Nishihata, "Open Access:
Chapter 8, Control and Resistance, Attacks on Burmese
Opposition Media", January 2011,
<http://access.opennet.net/wp-content/uploads/2011/12/
accesscontested-chapter-08.pdf>.
[VonLohmann-2008]
VonLohmann, F., "FCC Rules Against Comcast for BitTorrent
Blocking", August 2008,
<https://www.eff.org/deeplinks/2008/08/fcc-rules-against-
comcast-bit-torrent-blocking>.
[Wagner-2009]
Wagner, B., "Deep Packet Inspection and Internet
Censorship: International Convergence on an 'Integrated
Technology of Control'", Global Voices Advocacy, 2009,
<http://advocacy.globalvoicesonline.org/wp-
content/uploads/2009/06/deeppacketinspectionandinternet-
censorship2.pdf>.
[Wagstaff-2013]
Wagstaff, J., "In Malaysia, online election battles take a
nasty turn", NBC News, May 2013,
<https://www.nbcnews.com/tech/tech-news/malaysia-online-
election-battles-take-nasty-turn-flna6c9783842>.
[Wang-2017]
Wang, Z., Cao, Y., Qian, Z., Song, C., and S.V.
Krishnamurthy, "Your State is Not Mine: A Closer Look at
Evading Stateful Internet Censorship",
DOI 10.1145/3131365.3131374, November 2017,
<https://www.cs.ucr.edu/~zhiyunq/pub/
imc17_censorship_tcp.pdf>.
[Wang-2020]
Wang, Z., Zhu, S., Cao, Y., Qian, Z., Song, C.,
Krishnamurthy, S.V., Chan, K.S., and T.D. Braun, "SYMTCP:
Eluding Stateful Deep Packet Inspection with Automated
Discrepancy Discovery", DOI 10.14722/ndss.2020.24083,
February 2020,
<https://www.cs.ucr.edu/~zhiyunq/pub/ndss20_symtcp.pdf>.
[Weaver-2009]
Weaver, N., Sommer, R., and V. Paxson, "Detecting Forged
TCP Reset Packets", September 2009,
<http://www.icir.org/vern/papers/reset-
injection.ndss09.pdf>.
[Whittaker-2013]
Whittaker, Z., "1,168 keywords Skype uses to censor,
monitor its Chinese users", March 2013,
<http://www.zdnet.com/1168-keywords-skype-uses-to-censor-
monitor-its-chinese-users-7000012328/>.
[Wikip-DoS]
Wikipedia, "Denial-of-service attack", March 2016,
<https://en.wikipedia.org/w/index.php?title=Denial-of-
service_attack&oldid=710558258>.
[Wilde-2012]
Wilde, T., "Knock Knock Knockin' on Bridges Doors", The
Tor Project, July 2012, <https://blog.torproject.org/blog/
knock-knock-knockin-bridges-doors>.
[Winter-2012]
Winter, P. and S. Lindskog, "How China Is Blocking Tor",
April 2012, <http://arxiv.org/pdf/1204.0447v1.pdf>.
[WP-Def-2020]
Wikipedia, "Censorship", March 2020,
<https://en.wikipedia.org/w/
index.php?title=Censorship&oldid=943938595>.
[Wright-2013]
Wright, J. and Y. Breindl, "Internet filtering trends in
liberal democracies: French and German regulatory
debates", DOI 10.14763/2013.2.122, April 2013,
<https://policyreview.info/articles/analysis/internet-
filtering-trends-liberal-democracies-french-and-german-
regulatory-debates>.
[Zhu-2011] Zhu, T., Bronk, C., and D.S. Wallach, "An Analysis of
Chinese Search Engine Filtering",
DOI 10.48550/arXiv.1107.3794, July 2011,
<http://arxiv.org/ftp/arxiv/papers/1107/1107.3794.pdf>.
[Zmijewski-2014]
Zmijewski, E., "Turkish Internet Censorship Takes a New
Turn", Wayback Machine archive, March 2014,
<http://web.archive.org/web/20200726222723/
https://blogs.oracle.com/internetintelligence/turkish-
internet-censorship-takes-a-new-turn>.
Acknowledgments
This document benefited from discussions with and input from David
Belson, Stéphane Bortzmeyer, Vinicius Fortuna, Gurshabad Grover,
Andrew McConachie, Martin Nilsson, Michael Richardson, Patrick Vacek,
and Chris Wood.
Coauthor Hall performed work on this document before employment at
the Internet Society, and his affiliation listed in this document is
for identification purposes only.
Authors' Addresses
Joseph Lorenzo Hall
Internet Society
Email: hall@isoc.org
Michael D. Aaron
CU Boulder
Email: michael.drew.aaron@gmail.com
Amelia Andersdotter
Email: amelia.ietf@andersdotter.cc
Ben Jones
Email: ben.jones.irtf@gmail.com
Nick Feamster
U Chicago
Email: feamster@uchicago.edu
Mallory Knodel
Center for Democracy & Technology
Email: mknodel@cdt.org
|