aboutsummaryrefslogtreecommitdiffstatshomepage
path: root/apt-panopticon.lua
blob: b91d47cec2db63b71ef243dc8ee77a29e3ce29f5 (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
1001
1002
1003
1004
1005
1006
1007
1008
1009
1010
1011
1012
1013
1014
1015
1016
1017
1018
1019
1020
1021
1022
1023
1024
1025
1026
1027
1028
1029
1030
1031
1032
1033
1034
1035
1036
1037
1038
1039
1040
1041
1042
1043
1044
1045
1046
1047
1048
1049
1050
1051
1052
1053
1054
1055
1056
1057
1058
1059
1060
1061
1062
1063
1064
1065
1066
1067
1068
1069
1070
1071
1072
1073
1074
1075
1076
1077
1078
1079
1080
1081
1082
1083
1084
1085
1086
1087
1088
1089
1090
1091
1092
1093
1094
1095
1096
1097
1098
1099
1100
1101
1102
1103
1104
1105
1106
1107
1108
1109
1110
1111
1112
1113
1114
1115
1116
1117
1118
1119
1120
1121
1122
1123
1124
1125
1126
1127
1128
1129
1130
1131
1132
1133
1134
1135
1136
1137
1138
1139
1140
1141
1142
1143
1144
1145
1146
1147
1148
1149
1150
1151
1152
1153
1154
1155
1156
1157
1158
1159
1160
1161
1162
1163
1164
1165
1166
1167
1168
1169
1170
1171
1172
1173
1174
1175
1176
1177
1178
1179
1180
1181
1182
1183
1184
1185
1186
1187
1188
1189
1190
1191
1192
1193
1194
1195
1196
1197
1198
1199
1200
1201
1202
1203
1204
1205
1206
1207
1208
1209
1210
1211
1212
1213
1214
1215
1216
1217
1218
1219
1220
1221
1222
1223
1224
1225
1226
1227
1228
1229
1230
1231
1232
1233
1234
1235
1236
1237
1238
1239
1240
1241
1242
1243
1244
1245
1246
1247
1248
1249
1250
1251
1252
1253
1254
1255
1256
1257
1258
1259
1260
1261
1262
1263
1264
1265
1266
1267
1268
1269
1270
1271
#!/usr/bin/env luajit

local now = os.time()

local APT = require 'apt-panopticommon'
local D = APT.D
local I = APT.I
local T = APT.T
local W = APT.W
local E = APT.E
local C = APT.C
local arg, sendArgs = APT.parseArgs({...})
APT.html = true


local defaultURL = {scheme = "http"}
local releases = {"jessie", "ascii", "beowulf", "ceres"}
local releaseFiles =
{
    -- Release file.
    "Release",				-- 3.7 MB
    "Release.gpg",			--
--    "InRelease",			-- 3.7 MB
--    "main/binary-all/Packages.xz",	-- 2.6 GB for all that changed recently.
    -- Contents files.			-- 3.3 GB
--    "main/Contents-all.xz",
--    "main/Contents-amd64.xz",
--    "main/Contents-arm64.xz",
--    "-security/main/Contents-all.xz",
--    "-security/main/Contents-amd64.xz",
--    "-security/main/Contents-arm64.xz",
}
local notExist =
{
    "ceres-security"	-- This will never exist, it's our code name for the testing suite.
}
local referenceDebs =
{
    -- Debian package.
    "merged/pool/DEBIAN/main/d/debian-keyring/debian-keyring_2019.02.25_all.deb",
    -- Debian security package.  NOTE this one should always be redirected?
    "merged/pool/DEBIAN-SECURITY/updates/main/a/apt/apt-transport-https_1.4.9_amd64.deb",
}
local referenceDevs =
{
    -- Devuan package.  NOTE this one should not get redirected, but that's more a warning than an error.
    "merged/pool/DEVUAN/main/d/devuan-keyring/devuan-keyring_2017.10.03_all.deb",
}

local curlStatus = 
{
    [1 ] = "Unsupported protocol. This build of curl has no support for this protocol.",
    [2 ] = "Failed to initialize.",
    [3 ] = "URL malformed. The syntax was not correct.",
    [4 ] = "A feature or option that was needed to perform the desired request was not enabled or was explicitly disabled at build-time. To make curl able to do this, you probably need another build of libcurl!",
    [5 ] = "Couldn't resolve proxy. The given proxy host could not be resolved.",
    [6 ] = "Couldn't resolve host. The given remote host was not resolved.",
    [7 ] = "Failed to connect to host.",
    [8 ] = "Weird server reply. The server sent data curl couldn't parse.",
    [9 ] = "FTP access denied. The server denied login or denied access to the particular resource or directory you wanted to reach. Most often you tried to change to a directory that doesn't exist on the server.",
    [10] = "While waiting for the server to connect back when an active FTP session is used, an error code was sent over the control connection or similar.",
    [11] = "FTP weird PASS reply. Curl couldn't parse the reply sent to the PASS request.",
    [12] = "During an active FTP session while waiting for the server to connect, the CURLOPT_ACCEPTTIMEOUT_MS (or the internal default) timeout expired.",
    [13] = "FTP weird PASV reply, Curl couldn't parse the reply sent to the PASV request.",
    [14] = "FTP weird 227 format. Curl couldn't parse the 227-line the server sent.",
    [15] = "FTP can't get host. Couldn't resolve the host IP we got in the 227-line.",
    [16] = "A problem was detected in the HTTP2 framing layer. This is somewhat generic and can be one out of several problems, see the error buffer for details.",
    [17] = "FTP couldn't set binary. Couldn't change transfer method to binary.",
    [18] = "Partial file. Only a part of the file was transferred.",
    [19] = "FTP couldn't download/access the given file, the RETR (or similar) command failed.",

    [21] = "FTP quote error. A quote command returned error from the server.",
    [22] = "HTTP page not retrieved. The requested url was not found or returned another error with the HTTP error code being 400 or above. This return code only appears if -f, --fail is used.",
    [23] = "Write error. Curl couldn't write data to a local filesystem or similar.",

    [25] = "FTP couldn't STOR file. The server denied the STOR operation, used for FTP uploading.",
    [26] = "Read error. Various reading problems.",
    [27] = "Out of memory. A memory allocation request failed.",
    [28] = "Operation timeout. The specified time-out period was reached according to the conditions.",

    [30] = "FTP PORT failed. The PORT command failed. Not all FTP servers support the PORT command, try doing a transfer using PASV instead!",
    [31] = "FTP couldn't use REST. The REST command failed. This command is used for resumed FTP transfers.",

    [33] = "HTTP range error. The range \"command\" didn't work.",
    [34] = "HTTP post error. Internal post-request generation error.",
    [35] = "SSL connect error. The SSL handshaking failed.",
    [36] = "FTP bad download resume. Couldn't continue an earlier aborted download.",
    [37] = "FILE couldn't read file. Failed to open the file. Permissions?",
    [38] = "LDAP cannot bind. LDAP bind operation failed.",
    [39] = "LDAP search failed.",

    [41] = "Function not found. A required LDAP function was not found.",
    [42] = "Aborted by callback. An application told curl to abort the operation.",
    [43] = "Internal error. A function was called with a bad parameter.",

    [45] = "Interface error. A specified outgoing interface could not be used.",

    [47] = "Too many redirects. When following redirects, curl hit the maximum amount.",
    [48] = "Unknown option specified to libcurl. This indicates that you passed a weird option to curl that was passed on to libcurl and rejected. Read up in the manual!",
    [49] = "Malformed telnet option.",

    [51] = "The peer's SSL certificate or SSH MD5 fingerprint was not OK.",
    [52] = "The server didn't reply anything, which here is considered an error.",
    [53] = "SSL crypto engine not found.",
    [54] = "Cannot set SSL crypto engine as default.",
    [55] = "Failed sending network data.",
    [56] = "Failure in receiving network data.",

    [58] = "Problem with the local certificate.",
    [59] = "Couldn't use specified SSL cipher.",
    [60] = "Peer certificate cannot be authenticated with known CA certificates.",
    [61] = "Unrecognized transfer encoding.",
    [62] = "Invalid LDAP URL.",
    [63] = "Maximum file size exceeded.",
    [64] = "Requested FTP SSL level failed.",
    [65] = "Sending the data requires a rewind that failed.",
    [66] = "Failed to initialise SSL Engine.",
    [67] = "The user name, password, or similar was not accepted and curl failed to log in.",
    [68] = "File not found on TFTP server.",
    [69] = "Permission problem on TFTP server.",
    [70] = "Out of disk space on TFTP server.",
    [71] = "Illegal TFTP operation.",
    [72] = "Unknown TFTP transfer ID.",
    [73] = "File already exists (TFTP).",
    [74] = "No such user (TFTP).",
    [75] = "Character conversion failed.",
    [76] = "Character conversion functions required.",
    [77] = "Problem with reading the SSL CA cert (path? access rights?).",
    [78] = "The resource referenced in the URL does not exist.",
    [79] = "An unspecified error occurred during the SSH session.",
    [80] = "Failed to shut down the SSL connection.",
    [81] = "Socket is not ready for send/recv wait till it's ready and try again. This return code is only returned from curl_easy_recv and curl_easy_send.",
    [82] = "Could not load CRL file, missing or wrong format (added in 7.19.0).",
    [83] = "Issuer check failed (added in 7.19.0).",
    [84] = "The FTP PRET command failed",
    [85] = "RTSP: mismatch of CSeq numbers",
    [86] = "RTSP: mismatch of Session Identifiers",
    [87] = "unable to parse FTP file list",
    [88] = "FTP chunk callback reported error",
    [89] = "No connection available, the session will be queued",
    [90] = "SSL public key does not matched pinned public key",
    [91] = "Status returned failure when asked with CURLOPT_SSL_VERIFYSTATUS.",
    [92] = "Stream error in the HTTP/2 framing layer.",
    [93] = "An API function was called from inside a callback.",
    [94] = "An authentication function returned an error.",
    [95] = "A problem was detected in the HTTP/3 layer. This is somewhat generic and can be one out of several problems, see the error buffer for details.",
}


local socket = require 'socket'
local ftp = require 'socket.ftp'
local http = require 'socket.http'
local url = require 'socket.url'


local ip = ""
local cor = nil
local Updating = false

local downloadLock = "flock -n results/curl-"
local arw = ' &nbsp; <font color="magenta"><b>-></b></font> &nbsp; '

local repoExists = function (r)
    r = r:match("([%a-]*)")
    if nil == r then return false end
    for k, v in pairs(notExist) do
	if v == r then return false end
    end
    return true
end

local IP = {}
gatherIPs = function (host)
    if nil == IP[host] then
        local IPs
        -- Takes about 30 seconds to look up the lot.
        -- I tested using dig's -f option, it didn't seem much faster.
        -- The sort -r assumes that deb.devuan.org is the first alphabetically.
	local dig = io.popen('dig +keepopen +noall +nottlid +answer ' .. host .. ' A ' .. host .. ' AAAA ' .. host .. ' CNAME ' .. host .. ' SRV | sort -r | uniq')
	repeat
	    IPs = dig:read("*l")
	    if nil ~= IPs then
		for k, t, v in IPs:gmatch("([%w_%-%.]*)%.%s*IN%s*(%a*)%s*(.*)") do
		    if "." == v:sub(-1, -1) then v = v:sub(1, -2) end
		    if nil == IP[k] then IP[k] = {} end
		    IP[k][v] = t
		    D("  DNS record " .. host .. " == " .. k .. " type " .. t .. " -> " .. v)
		    if t == "CNAME" then
			gatherIPs(v)
			IP[k][v] = IP[v]
		    elseif v == "SRV" then
			print("SVR record found, now what do we do?")
		    end
		end
	    end
	until nil == IPs
    end

    -- If this is the DNS-RR domain name, gather the IPs for the mirrors that mirror_list.txt says should be in it.
    if host == APT.options.roundRobin.value then
	for k, m in pairs(APT.mirrors) do
	    if "yes" == m.DNSRR then
		gatherIPs(m.FQDN)
		IP[host][m.FQDN] = IP[m.FQDN]
	    end
	end
    end
    return IP[host]
end

-- Returns FTP directory listing
local nlst = function (u)
    local t = {}
    local p = url.parse(u)
    p.command = "nlst"
    p.sink = ltn12.sink.table(t)
    local r, e = ftp.get(p)
    return r and table.concat(t), e
end

local timeouts = 0;
local totalTimeouts = 0
local spcd = ' &nbsp; '
checkHEAD = function (host, URL, r, retry, sanity)
    if nil == r then r = 0 end
    if nil == retry then retry = 0 end
    if true == sanity then sanity = 'URLSanity' else sanity = '' end
    local check = "HEAD testing file"
    local PU = url.parse(URL, defaultURL)
    local pu = url.parse(PU.scheme .. "://" .. host, defaultURL)
    local fname = host .. "_" .. PU.host .. "_" .. PU.path:gsub("/", "_") .. ".log.txt"
    local hdr = ""
    local IP = ""
    if pu.host ~= PU.host then
	if "http" == PU.scheme then
	    hdr = '-H "Host: ' .. host .. '"'
	end
	IP = '--connect-to "' .. pu.host .. '::' .. PU.host .. ':"'
	fname = host .. "_" .. pu.host .. '_' .. PU.host .. "_" .. PU.path:gsub("/", "_") .. ".txt"
    end
    os.execute('rm -f results/HEADERS_' .. fname .. ' 2>/dev/null; rm -f results/STATUS_' .. fname .. ' 2>/dev/null')
    if not APT.testing(PU.scheme, host) and APT.redir then I(spcd .. string.upper(PU.scheme) .. " not supported, not tested. &nbsp; " .. URL, host); return end
    if 0 < r then
	check = "Redirecting to"
    end
    if 0 < retry then
	os.execute("sleep " .. math.random(1, 3))
	check = "Retry " .. retry .. " " .. check
    end
    if 2 <= timeouts then
	E(spcd .. spcd .. "too many timeouts!  " .. check .. " " .. host .. arw .. URL, PU.scheme, "", host)
	return
    end
    if APT.options.timeouts.value <= (totalTimeouts) then
	E(spcd .. spcd .. "Way too many timeouts!", PU.scheme, "", host)
	return
    end
    if 20 <= r then
	E(spcd .. spcd .. "too many redirects!  " .. check .. " " .. host .. arw .. URL, PU.scheme, "", host)
	return
    end
    if APT.options.retries.value <= retry then
	E(spcd .. spcd .. "too many retries!  " .. check .. " " .. host .. arw .. URL, PU.scheme, "", host)
	return
    end
    if "https" == PU.scheme and APT.options.roundRobin.value == host then
	I(spcd .. "Not testing " .. APT.lnk(URL) .. " mirrors wont have the correct HTTPS certificate for the round robin.", host)
	return
    else
	I(spcd .. check .. " " .. APT.lnk(URL), host)
    end

    --[[ Using curl command line -
	-I - HEAD
	--connect-to domain:port:IP:port - connect to IP, but use SNI from URL.
	-header "" - add extra headers.
	-L - DO follow redirects.
	--max-redirs n - set maximum redirects, default is 50, -1 = unlimited.
	-4 or -6 - Only use IPv4 or IPv6
	--retry n - maximum retries, default is 0, no retries.
	-o file - write to file instead of stdout.
	--path-as-is - https://curl.haxx.se/libcurl/c/CURLOPT_PATH_AS_IS.html might be useful for URLSanity.
	-s silent - don't output progress or error messages.
	--connect-timeout n - timeout in seconds.
	    Should return with error code 28 on a timeout?
	-D file - write the received headers to a file.  This includes the status code and string.
    ]]
    local status = APT.exe(
	'curl -I --retry 0 -s --path-as-is --connect-timeout ' .. APT.options.timeout.value .. ' --max-redirs 0 ' .. APT.IPv46 .. ' ' ..
	IP .. ' ' .. '-o /dev/null -D results/"HEADERS_' .. fname .. '" ' ..
	hdr .. ' -w "#%{http_code} %{ssl_verify_result} %{url_effective}\\n" ' .. PU.scheme .. '://' .. host .. PU.path .. ' >>results/"STATUS_' .. fname .. '"'
    ):Nice():log():Do().status
    local code = "???"
    local cstr = ""
    local location = nil
    local tmot = 1
    while not APT.checkFile('results/STATUS_' .. fname) do
	D(spcd .. spcd .. 'Waiting for results/STATUS_' .. fname .. ' file.')
	os.execute('sleep ' .. tmot)
	tmot = tmot * 2
	if 8 < tmot then
	    T(spcd .. spcd .. "TIMEOUT " .. timeouts + 1 .. ", retry " .. retry + 1 .. ' ' .. APT.lnk(URL), PU.scheme, sanity, host)
	    timeouts = timeouts + 1
	    checkHEAD(host, URL, r, retry + 1, '' ~= sanity)
	    os.execute('cat results/"HEADERS_' .. fname .. '" >>results/"STATUS_' .. fname .. '" 2>/dev/null; rm -f results/"HEADERS_' .. fname .. '" 2>/dev/null')
	    return
	end
    end
    os.execute('cat results/"HEADERS_' .. fname .. '" >>results/"STATUS_' .. fname .. '" 2>/dev/null; rm -f results/"HEADERS_' .. fname .. '" 2>/dev/null')
    if 0 ~= status then
	local msg = curlStatus[status]
	if nil == msg then msg = "UNKNOWN CURL STATUS CODE!" end
	if (28 == status) or (7 == status) then
	    T(spcd .. spcd .. "TIMEOUT " .. timeouts + 1 .. ", retry " .. retry + 1 .. ' ' .. APT.lnk(URL), PU.scheme, sanity, host)
	    timeouts = timeouts + 1
	else
	    E(spcd .. spcd .. "The curl command return an error code of " .. status .. " - " .. msg .. ' for '.. APT.lnk(URL), PU.scheme, sanity, host)
	end
	if 60 == status then return end	-- Certificate is invalid, don't bother retrying.
	checkHEAD(host, URL, r, retry + 1, '' ~= sanity)
	return
    end
    local rfile, e = io.open("results/STATUS_" .. fname, "r")
    if nil == rfile then W("opening results/STATUS_" .. fname .. " file - " .. e) else
	for line in rfile:lines("*l") do
	    if "#" == line:sub(1, 1) then
		code = line:sub(2, 4)
		if ("https" == PU.scheme) and ("0" ~= line:sub(6, 6)) then
		    os.execute('cp results/STATUS_' .. fname .. ' results/STATUS_' .. fname .. '_SAVED')
		    if '' ~= sanity then
			E(spcd .. spcd .. "The certificate is invalid.", PU.scheme, sanity, host)
		    else
			E(spcd .. spcd .. "The certificate is invalid.", PU.scheme, "https", host)
		    end
		end
	    elseif "http" == line:sub(1, 4):lower() then
		-- -2 coz the headers file gets a \r at the end.
		cstr = line:sub(14, -2)
	    elseif "location" == line:sub(1, 8):lower() then
		location = line:sub(11, -2)
	    end
	end
	if '???' == code then
	    W(spcd .. spcd .. 'Could not find response code. ' .. APT.lnk(URL), PU.scheme, sanity, host)
	end
    end
    os.execute('cat results/STATUS_' .. fname .. ' >> results/curl_HEAD_' .. fname .. '; rm -f results/STATUS_' .. fname .. ' 2>/dev/null')

    if ("4" == tostring(code):sub(1, 1)) or ("5" == tostring(code):sub(1, 1)) then
	E(spcd .. spcd .. code .. " " .. cstr .. ".   " .. check .. " " .. APT.lnk(URL), PU.scheme, sanity, host)
    else
	if not APT.testing(PU.scheme, host) then
	    I(spcd .. spcd .. "Not supported, but works " .. PU.scheme .. " " .. APT.lnk(URL), PU.scheme, "", host)
	end
	I(spcd .. spcd .. code .. " " .. cstr .. ".   " .. check .. " " .. APT.lnk(URL), host)
--	timeouts = timeouts - 1	-- Backoff the timeouts count if we managed to get through.
	if nil ~= location then
	    pu = url.parse(location, defaultURL)
	    if (pu.host == APT.options.roundRobin.value) and (nil ~= PU.path:find('merged/pool/DEVUAN/')) then
		E('DEVUAN packages must not be redirected to ' .. APT.options.roundRobin.value .. ' - ' .. APT.lnk(URL) .. arw .. APT.lnk(location), PU.scheme, 'Redirects', host)
	    end
	    if ('http' == location:sub(1, 4)) and (pu.scheme ~= PU.scheme) then	-- Sometimes a location sans scheme is returned, this is not a protocol change.
		if APT.testing("Protocol") then
		    if APT.options.roundRobin.value == host then  -- Coz HTTPS shouldn't happen via the round robin.
			E(spcd .. spcd .. "Protocol changed during redirect!  " .. check .. " " .. APT.lnk(URL) .. arw .. APT.lnk(location), PU.scheme, "Protocol", host)
		    end
			W(spcd .. spcd .. "Protocol changed during redirect!  " .. check .. " " .. APT.lnk(URL) .. arw .. APT.lnk(location), PU.scheme, "Protocol", host)
		    else
		end
	    end

	    if location == URL then
		E(spcd .. spcd .. "Redirect loop!  " .. check .. " " .. APT.lnk(URL) .. arw .. APT.lnk(location), PU.scheme, "", host)
	    elseif nil == pu.host then
		I(spcd .. spcd .. "Relative redirect.  " .. check .. " " .. APT.lnk(URL) .. arw .. APT.lnk(location), host)
		if 1 <= APT.options.bandwidth.value then checkHEAD(host, PU.scheme .. "://" .. PU.host .. location, r + 1, retry, '' ~= sanity) end
	    elseif (PU.host == pu.host) or (host == pu.host) then
		if PU.host ~= host then
		    local t = pu.host
		    pu.host = PU.host
		    location = url.build(pu)
		    pu.host = t
		end
		I(spcd .. spcd .. "Redirect to same host.  " .. check .. " " .. APT.lnk(URL) .. arw .. APT.lnk(location), host)
		if 1 <= APT.options.bandwidth.value then checkHEAD(host, location, r + 1, retry, '' ~= sanity) end
	    else
		I(spcd .. spcd .. "Redirect to different host.  " .. check .. " " .. APT.lnk(URL) .. arw .. APT.lnk(location), host)
		if 1 <= APT.options.bandwidth.value then 
		    --[[ The hard part here is that we end up throwing ALL of the test files at the redirected location.
		        Not good for deb.debian.org, which we should only be throwing .debs at.
		        What we do is loop through the DNS entries, and only test the specific protocol & file being tested here.
		    ]]
		    local u = pu.host .. "/" .. pu.path
		    local file = pu.path:match(".*/([%w%.%+%-_]*)$")	-- Get the filename.
		    local path = pu.path:sub(2, -1 -(#file))
		    local check = u:gsub("/", "_")
		    local extraArgs = sendArgs .. ' -o -r '
		    if 'https' == pu.scheme then extraArgs = extraArgs .. ' --tests=-http' end
		    if 'http'  == pu.scheme then extraArgs = extraArgs .. ' --tests=-https' end
		    local pth = path:match('^(.*/pool/).*$')
		    if nil ~= pth then table.insert(APT.results[PU.scheme].redirects, pu.host .. "/" .. pth) else E(spcd .. spcd .. 'Odd redirect path ' .. path) end
		    I(spcd .. spcd .. "Now checking redirected host " .. u .. ' &nbsp; for &nbsp; ' .. APT.lnk(URL) .. arw .. APT.lnk(location), host)
		    APT.exe(downloadLock .. "REDIR-" .. check .. ".log.txt" .. " ./apt-panopticon.lua " .. extraArgs .. ' ' .. pu.host .. "/" .. path .. " " .. file):Nice():log():fork()
		    D(spcd .. 'logging to ' .. APT.logName(pu.host, nil, file)[2])
		end
	    end
	elseif nil ~= PU.path:find('merged/pool/DEBIAN-SECURITY/') then
	    W('DEBIAN-SECURITY packages must be redirected to a Debian mirror - ' .. APT.lnk(URL) .. arw .. APT.lnk(location), PU.scheme, 'Redirects', host)
	end
    end
end

local checkTimeouts = function(host, scheme, URL)
    totalTimeouts = totalTimeouts + timeouts; timeouts = 0
    checkHEAD(host, scheme .. "://" .. URL)
    if (1 <= APT.options.bandwidth.value) and APT.testing("URLSanity") then
	URL = URL:gsub("/", "///")
	URL = URL:gsub("///", "/", 1)
	checkHEAD(host, scheme .. "://" .. URL, 0, 0, true)
    end
    if nil ~= cor then
	D('*&gt;* About to resume coroutine after checkHEAD(' .. host .. ' , ' .. scheme .. ' :// ' .. URL .. ')')
	local ok, message = coroutine.resume(cor)
	if not ok then cor = nil; print(message) end
    end
    if APT.options.timeouts.value <= (totalTimeouts) then
	E("Way too many timeouts!", scheme, "URLSanity", host)
	return true
    end
    return false
end

local checkFiles = function (host, ip, path, file)
    timeouts = 0
    if nil == path then path = "" end
    if nil ~= file then
	if "redir" == ip then ip = host end
	if checkTimeouts(host, "http",  ip .. path .. "/" .. file) then return end
	if checkTimeouts(host, "https", ip .. path .. "/" .. file) then return end
    else
	I("  HEAD testing files for " .. host .. arw .. ip .. " " .. path, host)
	if 1 <= APT.options.bandwidth.value then
	    -- Do these first, coz they are likely to fork off a different server.
	    for i, s in pairs(referenceDebs) do
		if checkTimeouts(host, "http",  ip .. path .. "/" .. s) then return end
		if checkTimeouts(host, "https", ip .. path .. "/" .. s) then return end
	    end
	end
	for i, s in pairs(releases) do
	    for j, k in pairs(releaseFiles) do
		if repoExists(s .. k) then
		    if checkTimeouts(host, "http",  ip .. path .. "/merged/dists/" .. s .. '/' .. k) then return end
		    if 1 <= APT.options.bandwidth.value then
			if checkTimeouts(host, "https", ip .. path .. "/merged/dists/" .. s .. '/' .. k) then return end
		    else
			break
		    end
		end
		if 2 >= APT.options.bandwidth.value then break end
	    end
	    if 2 >= APT.options.bandwidth.value then break end
	end
	if 1 <= APT.options.bandwidth.value then
	    for i, s in pairs(referenceDevs) do
		if checkTimeouts(host, "http",  ip .. path .. "/" .. s) then return end
		if checkTimeouts(host, "https", ip .. path .. "/" .. s) then return end
	    end
	end
    end
end

checkHost = function (orig, host, path, ip, file)
    if nil == host then host = orig end
    if nil == path then path = "" end
    if nil == file then file = "" end
    local ph = url.parse("http://" .. host)
    if (nil ~= ip) and ("redir" ~= ip) then
	local po = url.parse("http://" .. orig)
	if "" ~= file then
	    D("checking redirected file " .. po.host .. " " .. file)
	    checkFiles(po.host, ip, path, file)
	else
	    checkFiles(po.host, ip, path)
	end
    else
	if orig == host then
	    I("Testing mirror " .. orig .. "" .. file)
	    APT.exe("./apt-panopticon.lua " .. sendArgs .. " -o " .. orig .. path .. " " .. file):Nice():log():fork()
	    D('logging to ' .. APT.logName(ph.host, nil, file)[2])
	else D("checkHost " .. orig .. arw .. host) end
    end
end


local addDownload = function(host, URL, f, r, k)
    local file = k:match(".*/([%w%.%+%-_]*)$")	-- Get the filename.
    if APT.checkFile("results/" .. host .. "/merged/dists/" .. r .. '/' .. k) then
	-- Curls "check timestamp and overwrite file" stuff sucks.
	-- -R means the destination file gets the timestamp of the remote file.
	-- Can only do ONE timestamp check per command.
	--  This doesn't work either.  All downloads get all these headers.  Pffft
--	local status, ts = APT.execute('TZ="GMT" ls -l --time-style="+%a, %d %b %Y %T %Z"  results/' .. host .. "/merged/dists/" .. r .. '/' .. k .. ' | cut -d " " -f 6-11')
--	f:write('header "If-Modified-Since: ' .. ts:sub(2, -2) .. '"\n')
	-- Curl will DELETE the existing file if the timestamp fails to download a new one, unless we change directory first, 
	-- which wont work with multiple files in multiple directories.  WTF?
--TODO - change tactic, do a HEAD if-modified test first before adding the file to the list to download.
	os.execute(" mv   results/" .. host .. "/merged/dists/" .. r .. '/' .. k .. 
			" results/" .. host .. "/merged/dists/" .. r .. '/' .. k .. ".old")
    end

    D('Downloading http://' .. host .. URL .. '/merged/dists/' .. r .. '/' .. k)
    f:write('url "' .. 'http://' .. host .. URL .. '/merged/dists/' .. r .. '/' .. k .. '"\n')
    f:write('output "results/' .. host .. '/merged/dists/' .. r .. '/' .. k .. '"\n')
end

local postDownload = function(host, r, k)
    local file = k:match(".*/([%w%.%+%-_]*)$")	-- Get the filename.
    if nil == file then file = k end
    os.execute("if [ -f results/" .. host .. "/merged/dists/" .. r .. '/' .. k .. ".old ]" ..
	    " && [ ! -f results/" .. host .. "/merged/dists/" .. r .. '/' .. k .. " ]; then cp -a" ..
		    "   results/" .. host .. "/merged/dists/" .. r .. '/' .. k .. ".old" ..
		    "   results/" .. host .. "/merged/dists/" .. r .. '/' .. k .. "; fi")
    if APT.checkFile('results/' .. host .. '/merged/dists/' .. r .. '/' .. k) then
	if ".gz"  == k:sub(-3, -1) then APT.exe("gzip -dfk results/" .. host .. "/merged/dists/" .. r .. '/' .. k):Nice():noErr():Do() end
	if ".xz"  == k:sub(-3, -1) then APT.exe("xz   -dfk results/" .. host .. "/merged/dists/" .. r .. '/' .. k):Nice():noErr():Do() end
    end
end

local download = "curl" ..
		 " --connect-timeout " .. APT.options.timeout.value ..
		 " --create-dirs -f -L" ..
		 " --fail-early" ..
		 " --max-time " .. APT.options.maxtime.value ..
		 APT.IPv46 .. ' ' ..
		 " --retry " .. APT.options.retries.value ..
		 " -R -v -z 'results/stamp.old'  --stderr results/"
local downloads = function(host, URL, meta, release, list)
    if nil == URL then URL = "" end
    local files = 'curl-' .. meta .. '-' .. host .. '.files.txt'
    local lock = meta .. "-" .. host .. ".log.txt"
    local log = "curl-" .. meta .. "-" .. host .. ".log.txt"
    local cm = downloadLock .. lock .. " " .. download .. log .. " -K results/" .. files
    if      APT.testing("IPv4")  and (not APT.testing("IPv6")) then cm = cm .. ' -4' end
    if (not APT.testing("IPv4")) and      APT.testing("IPv6")  then cm = cm .. ' -6' end
    f, e = io.open("results/curl-" .. meta .. '-' .. host .. ".files.txt", "a+")
    if nil == f then C("opening curl downloads list file - " .. e); return end

    if nil ~= list then
	if "" ~= list then
	    if nil ~= release then
		for l in list:gmatch("\n*([^\n]+)\n*") do
		    addDownload(host, URL, f, release, "/" .. l)
		end
	    else
		I('Downloading ' .. APT.lnk('http://' .. host .. URL .. '/merged/' .. list))
		f:write('url "' .. 'http://' .. host .. URL .. '/merged/' .. list .. '"\n')
		f:write('output "results/' .. host .. '/merged/' .. list .. '"\n')
	    end
	    f:close()
	    return
	end
    else
	for i, s in pairs(releases) do
	    for j, k in pairs(releaseFiles) do
		if repoExists(s .. k) then
		    addDownload(host, URL, f, s, k)
		end
	    end
	end
    end
    f:close()
    APT.exe(cm):Nice():log():fork()
    D('logging to <a href="' .. log .. '">' .. log .. '</a>, with <a href="' .. files .. '">these files</a>')
end


local validateURL = function(m)
    if " " == m.BaseURL:sub(-1, -1) then
	W("space at end of BaseURL in mirror_list.txt!  " .. m.BaseURL, "", "", m.FQDN)
	m.BaseURL = m.BaseURL:sub(1, -2)
    end
    if "/" == m.BaseURL:sub(-1, -1) then
	W("slash at end of BaseURL in mirror_list.txt!  " .. m.BaseURL, "", "", m.FQDN)
	m.BaseURL = m.BaseURL:sub(1, -2)
    end
    local p = url.parse("http://" .. m.BaseURL)
    if nil == p.path then p.path = '' end
    if nil ~= p.port then p.authority = authority .. ':' .. p.port end
    if nil == m.FDQN then W("Something wrong in FDQN from mirror_list.txt!  nil", "", "", p.authority) else
	if m.FQDN ~= p.authority then W("Something wrong in FDQN from mirror_list.txt!  " .. m.FDQN, "", "", p.authority) end
    end
    if nil == m.BaseURL then W("Something wrong in BaseURL from mirror_list.txt!  nil", "", "", p.authority) else
	if m.BaseURL ~= (p.authority .. p.path) then W("Something wrong in BaseURL from mirror_list.txt!  " .. m.BaseURL, "", "", p.authority) end
    end
    if (nil ~= p.query) or (nil ~= p.fragment) or (nil ~= p.params) then W("Something wrong in BaseURL from mirror_list.txt, should be nothing after the path!  " .. m.BaseURL, "", "", p.authority) end
    if (nil ~= p.user) or (nil ~= p.userinfo) or (nil ~= p.password) then W("Something wrong in BaseURL from mirror_list.txt, should be no credentials!  " .. m.BaseURL, "", "", p.authority) end
    m.FQDN = p.authority
    m.BaseURL = p.authority .. p.path
    return m
end

local getMirrors = function ()
    local mirrors = {}
    local host = ""
    local m = {}
    local active = true
    local URL = 'http://' .. APT.options.referenceSite.value .. '/mirror_list.txt'
    I('Downloading and parsing http://' .. APT.options.referenceSite.value .. '/mirror_list.txt')
    local p, c, h = http.request(URL)
    if nil == p then E(c .. " fetching " .. URL) else

	for l in p:gmatch("\n*([^\n]+)\n*") do
	    local t, d = l:match("(%a*):%s*(.*)")
	    d = string.lower(d)
	    if "FQDN" == t then
		if "" ~= host then
		    mirrors[host] = validateURL(m)
		    m = {}
		    active = true
		end
		host = d
		m[t] = d
	    elseif "Protocols" == t then
		local prot = {}
		for w in d:gmatch("(%w+)") do
		    if APT.search(APT.protocols, w:lower()) then prot[w] = true end
		end
		m[t] = prot
	    elseif "Active" == t and nil == d:sub(1, 3):find("yes", 1, true) then
		W("Mirror " .. host .. " is not active - " .. d, "", "", host)
		active = false
		m[t] = d
-- TODO - Should do some more input validation on everything.
	    elseif "Rate" == t then
		local time, unit = d:match('(%d+) *(%a+)')
		time = tonumber(time)
		unit = unit:sub(1, 1)
		m[t] = time .. ' ' .. unit
		if     'm' == unit then
		    time = time * 60
		elseif 'h' == unit then
		    time = time * 60 * 60
		else
		    C('Unknown Rate for mirror ' .. host)
		end
		m['Updated'] = time
	    else
		m[t] = d
	    end
	end
	if "" ~= host --[[and active]] then
	    mirrors[host] = validateURL(m)
	end
    end
    if APT.testing("DNSRR") then
	mirrors[APT.options.roundRobin.value] = { ["Protocols"] = { ["http"] = true; }; ['Updated'] = 300; ['DNSRR'] = false;
						  ["FQDN"] = APT.options.roundRobin.value; ["Active"] = 'yes'; ["BaseURL"] = APT.options.roundRobin.value; }
    end
    local file, e = io.open("results/mirrors.lua", "w+")
    if nil == file then C("opening mirrors file - " .. e) else
	file:write(APT.dumpTable(mirrors, "mirrors") .. "\nreturn mirrors\n")
        file:close()
    end
    if 42 < #mirrors then print(#mirrors .. ' is too many mirrors!') ; os.exit(false) end
    return mirrors
end


local postParse = function(host, list)
    if APT.options.referenceSite.value == host then
	if nil ~= list then
	    local sem = 'results/NEW_' .. list.out .. '_%s.txt'
	    for i, n in pairs(releases) do
		local f = sem:format(n)
		if APT.checkFile(f .. '.tmp') then
		    os.execute('mv ' .. f .. '.tmp ' .. f)
		else
		    os.execute('touch ' .. f)
		end
	    end
	end
    end
end

local parseDebs = function(host)
    for i, n in pairs(releases) do
	local inFile =  'results/NEW_debs_' .. n .. '.txt'
	local nfile, e = io.open(inFile, "r")
	    if nil == nfile then W("opening " .. inFile .. " file - " .. e) else
	    for l in nfile:lines() do
		local v, p, sz, sha = l:match(' | (.+) | (pool/.+%.deb) | (%d.+) | (%x.+) |')
		if nil ~= p then
		    if APT.checkFile('results/' .. host .. "/merged/" .. p) then
			local fsz = APT.exe('ls -l results/' .. host .. "/merged/" .. p .. ' | cut -d " " -f 5-5'):Do().result
			if APT.testing("Integrity") then
			    if sz ~= fsz:sub(2, -2) then	-- The sub bit is to slice off the EOLs at each end.
				E('Package size mismatch - results/' .. host .. "/merged/" .. p .. ' should be ' .. sz .. ', but is ' .. fsz:sub(2, -2) .. '.', 'http', 'Integrity', host)
			    else
				local fsha = APT.exe('sha256sum results/' .. host .. "/merged/" .. p .. ' | cut -d " " -f 1'):log():Do().result
				if sha ~= fsha:sub(2, -2) then E('Package SHA256 sum mismatch - results/' .. host .. "/merged/" .. p, 'http', 'Integrity', host) end
-- TODO - maybe check the PGP key, though packages are mostly not signed.
			    end
			end
			if APT.testing("Updated") then
			    if sz ~= fsz:sub(2, -2) then
				E('Package size mismatch for ' .. host .. "/merged/" .. p, 'http', 'Updated', host)
			    end
			end
			os.execute('rm -f results/' .. host .. "/merged/" .. p)
		    else
			if Updating then
			    W('Not yet able to download, awaiting update for ' .. host .. "/merged/" .. p, 'http', 'Updated', host)
			else
			    E('Failed to download ' .. host .. "/merged/" .. p, 'http', 'Updated', host)
			end
		    end
		end
	    end
	end
    end
    return nil
end

local parsePackages = function(host)
    local list = {inf = 'Packages', parser = parseDebs, out = 'debs', files = {}, nextf = ''}
    for i, n in pairs(releases) do
	local inFile =  'results/NEW_' .. list.inf  .. '_' .. n .. '.txt'
	local outFile = 'results/NEW_' .. list.out .. '_' .. n .. '.txt'
	if APT.options.referenceSite.value == host then
	    outFile = outFile .. '.tmp'
	end
	local dFile, e = io.open(inFile, "r")
	if nil == dFile then W("opening " .. inFile .. " file - " .. e) else
	    for l in dFile:lines() do
		postDownload(host, n, l)

		l = '/' .. l
		local file = l:match(".*/([%w%.%+%-_]*)$")	-- Get the filename.
		local dir  = l:sub(1, 0 - (#file + 1))

		if "Packages." == file:sub(1, 9) then
-- TODO - compare the SHA256 sums in pkgmaster's Release for both the packed and unpacked versions.
--        Also note that this might get only a partial download due to maxtime.
		    if APT.options.referenceSite.value == host then
			local Pp, e = io.open('results/' .. host .. '/merged/dists/'.. n .. dir .. 'Packages.parsed', "w+")
			if nil == Pp then W('opening results/' .. host .. '/merged/dists/'.. n .. dir .. 'Packages.parsed' .. ' file - ' .. e) else
			    local pp = {}
			    for l in io.lines('results/' .. host .. '/merged/dists/'.. n .. dir .. 'Packages') do
				if "Package: "	== l:sub(1, 9) then
				    if 0 ~= #pp then
					for i = 1, 5 do
					    if nil == pp[i] then print(host .. " " .. n .. " " .. dir .. " " ..  i) else Pp:write(pp[i] .. " | ") end
					end
					Pp:write("\n")
				    end
				    pp = {}
				    pp[1] = 	   l:sub(10, -1)
				elseif "Version: "	== l:sub(1, 9) then
				    pp[2] = 	   l:sub(10, -1)
				elseif "Filename: "	== l:sub(1, 10) then
				    pp[3] = 	   l:sub(11, -1)
				elseif "Size: "	== l:sub(1, 6) then
				    pp[4] = 	   l:sub(7, -1)
				elseif "SHA256: "	== l:sub(1, 8) then
				    pp[5] = 	   l:sub(9, -1)
				end
			    end
			    Pp:close()
			    os.execute('sort results/' .. host .. '/merged/dists/'.. n .. dir .. 'Packages.parsed >results/' .. host .. '/merged/dists/'.. n .. dir .. 'Packages_parsed-sorted')
			    if APT.checkFile('Packages/' .. n .. dir .. 'Packages_parsed-sorted') then
				os.execute('diff -U 0    Packages/' .. n .. dir .. 'Packages_parsed-sorted ' ..
							'results/' .. APT.options.referenceSite.value .. '/merged/dists/' .. n .. dir .. 'Packages_parsed-sorted ' ..
							' | grep -E "^-" | grep -Ev "^\\+\\+\\+|^---" >>results/OLD_' .. list.out  .. '_' .. n .. '.txt')
				os.execute('diff -U 0    Packages/' .. n .. dir .. 'Packages_parsed-sorted ' ..
							'results/' .. APT.options.referenceSite.value .. '/merged/dists/' .. n .. dir .. 'Packages_parsed-sorted ' ..
							' | grep -E "^\\+" | grep -Ev "^\\+\\+\\+|^---" >>results/NEW_' .. list.out  .. '_TMP_' .. n .. '.txt')
				for i, s in pairs(referenceDebs) do
				    if 0 == APT.exe('grep -q "' .. s:sub(8, -1) .. '" results/OLD_' .. list.out  .. '_' .. n .. '.txt'):log():Do().status then
					print('Reference package is out of date (' .. n .. ') - ' .. s)
				    end
				end
				for i, s in pairs(referenceDevs) do
				    if 0 == APT.exe('grep -q "' .. s:sub(8, -1) .. '" results/OLD_' .. list.out  .. '_' .. n .. '.txt'):log():Do().status then
					print('Reference package is out of date (' .. n .. ') - ' .. s)
				    end
				end
			    else
				W("Can't find file Packages/" .. n .. dir .. "Packages_parsed-sorted")
			    end
			    os.execute('mkdir -p Packages/' .. n .. dir)
			    os.execute('mv -f results/' .. APT.options.referenceSite.value .. '/merged/dists/' .. n .. dir .. 'Packages_parsed-sorted Packages/' .. n .. dir .. 'Packages_parsed-sorted')
			end
		    else
		    end
		    os.execute('rm -fr results/' .. host .. '/merged/dists/' .. n .. dir .. ' 2>/dev/null')
		end
	    end

	    if APT.checkFile('results/NEW_' .. list.out  .. '_TMP_' .. n .. '.txt') then
		-- Sort by size.
		os.execute('sort -b -k 9,9 -n results/NEW_' .. list.out  .. '_TMP_' .. n .. '.txt >results/NEW_' .. list.out  .. '_' .. n .. '.sorted.txt')
		os.execute('grep -s " | pool/DEVUAN/"          results/NEW_' .. list.out  .. '_' .. n .. '.sorted.txt 2>/dev/null | head -n 1 >>' .. outFile)
		os.execute('grep -s " | pool/DEBIAN-SECURITY/" results/NEW_' .. list.out  .. '_' .. n .. '.sorted.txt 2>/dev/null | head -n 1 >>' .. outFile)
		os.execute('grep -s " | pool/DEBIAN/"          results/NEW_' .. list.out  .. '_' .. n .. '.sorted.txt 2>/dev/null | head -n 1  >' .. outFile)
		os.execute('rm -f results/NEW_' .. list.out  .. '_TMP_' .. n .. '.txt')
	    end

	end

	local nfile, e = io.open(outFile, "r")
	if nil ~= nfile then
--	    for l in nfile:lines() do
		local l = nfile:read('*l')
		if nil ~= l then
		    local p = l:match('(pool/.*%.deb)')
		    if nil ~= p then
			table.insert(list.files, p)
		    end
		end
--	    end
	end
    end

    postParse(host, list)
    return list
end

local parseRelease = function(host)
    local list = {inf = 'Release', parser = parsePackages, out = 'Packages', files = {}, nextf = 'debs'}
    local updated = false
    local now = tonumber(os.date('%s'))
    for i, n in pairs(releases) do
	for l, o in pairs(releaseFiles) do
	    if repoExists(i .. o) then
		postDownload(host, n, o)
		if (".gpg" == o:sub(-4, -1)) and (APT.checkFile('results/' .. host .. '/merged/dists/' .. n .. '/' .. o)) then
		    if APT.testing("Integrity") then
			local status = APT.exe( "gpgv --keyring /usr/share/keyrings/devuan-keyring.gpg results/" .. host .. "/merged/dists/" .. n .. '/' .. o .. 
						" results/" .. host .. "/merged/dists/" .. n .. '/' .. o:sub(1, -5)):Nice():noErr():log():Do().status
			if 0 ~= status then E("GPG check failed for " .. host .. "/merged/dists/" .. n .. '/' .. o, "http", "Integrity", host) end
-- TODO - should check the PGP sig of InRelease as well.
		    end
		    os.execute('rm results/' .. host .. '/merged/dists/' .. n .. '/' .. o)
		end
	    end
	end

	local fR  = 'results/' .. host .. '/merged/dists/' .. n .. '/Release'
	local fRp = APT.options.referenceSite.value .. '/merged/dists/' .. n .. '/Release.SORTED'
	if  APT.checkFile(fR) then
	    os.execute('sort -k 3 ' .. fR .. ' >' .. fR .. '.SORTED')
	    local outFile = 'results/NEW_' .. list.out .. '_' .. n .. '.txt'
	    if APT.checkFile('results_old/' .. fRp) then
		if APT.options.referenceSite.value == host then
		    outFile = outFile .. '.tmp'
		    os.execute('diff -U 0    results_old/' .. fRp .. ' ' ..
					    'results/' .. fRp .. ' ' ..
					    '| grep -v "@@" | grep "^+" | grep "Packages.xz$" |  cut -c 77- >' .. outFile)
-- TODO - Maybe check the date in Release, though since they are updated daily, is there any point?  Perhaps it's for checking amprolla got run?
--	Also check if that date is in the future, apt recently got a check for that, though not sure why.
		    os.execute('rm -f ' .. fR .. ' 2>/dev/null; ')
		else
-- TODO - compare to the pkgmaster copy.
		    if APT.testing('Updated') then
			while not APT.checkFile('results_old/' .. fRp) do
			    D('*&lt;* About to yield coroutine while waiting on - not APT.checkFile(results_old/' .. fRp .. ')')
			    coroutine.yield()
			    D('*&gt;* Resumed coroutine while waiting on - not APT.checkFile(results_old/' .. fRp .. ')')
			end
			local pkt = tonumber(APT.exe([[TZ="GMT" date -d "$(grep '^Date:' results/]] .. fRp .. [[  | cut -d ' ' -f 2-)" '+%s']]):Do().result:sub(2, -2))
			local new = tonumber(APT.exe([[TZ="GMT" date -d "$(grep '^Date:' ]]  ..  fR  .. [[.SORTED | cut -d ' ' -f 2-)" '+%s']]):Do().result:sub(2, -2))
			local upd  = pkt +  APT.mirrors[host].Updated
			local updd = pkt + (APT.mirrors[host].Updated * 1.5)	-- Give the mirror time to actually do the update.
			if pkt > new then
			    D( 'pkt is ' .. os.date('!%F %T', pkt) .. ' new is ' .. os.date('!%F %T', new) ..
			      ' upd is ' .. os.date('!%F %T', upd) .. ' updd is ' .. os.date('!%F %T', updd) ..
			      ' now is ' .. os.date('!%F %T', now) .. ' Updated is ' .. APT.mirrors[host].Updated)
			    if updd >= now then
				W('Release ' .. n .. ' not updated yet, should update @ ' .. os.date('!%F %T', upd) .. ', and was last updated @ ' .. os.date('!%F %T', new), 'http', 'Updated', host)
				Updating = true
			    else
				E('Release ' .. n .. ' not updated, should have updated @ ' .. os.date('!%F %T', upd) .. ', but was last updated @ ' .. os.date('!%F %T', new), 'http', 'Updated', host)
			    end
			    updated = false
			else
			    updated = true
			end
		    end
		end

-- TODO - if it's not Integrity and not reference, then just do a HEAD check and compare file times?
-- TODO - like we do with debs, pick just the smallest Packages.xz that has changed.
--	    Though we are sorting Release by name, so we can do the diff with the one from results_old, so we'll need to sort by size to.
--	    pkgmaster still needs to download all the changed Packages.xz files though.
		if (2 <= APT.options.bandwidth.value) and (updated or APT.testing("Integrity") or (APT.options.referenceSite.value == host)) then
		    local dfile, e = io.open(outFile, "r")
		    if nil == dfile then W("opening " .. outFile .. " file - " .. e) else
			for l in dfile:lines() do
			    table.insert(list.files, 'dists/' .. n .. '/' .. l)
			end
		    end
		end
	    end
	end

    end
    postParse(host, list)
    return list
end

local parseStart = function(host)
    local list = {inf = '', parser = parseRelease, out = 'Release', files = {}, nextf = 'Packages'}
    for i, n in pairs(releases) do
	local outFile = 'results/NEW_' .. list.out .. '_' .. n .. '.txt'
	for l, o in pairs(releaseFiles) do
	    if repoExists(n .. o) then
		if APT.options.referenceSite.value == host then
		    local dfile, e = io.open(outFile .. '.tmp', "a+")
		    if nil == dfile then W("opening " .. outFile .. ".tmp file - " .. e) else
			dfile:write(o .. '\n')
		    end
		end
		table.insert(list.files, 'dists/' .. n .. '/' .. o)
	    end
	end
    end
    postParse(host, list)
    return list
end

local doDownloads = function(host, path, list)
    while nil ~= list do
	if 0 ~= #(list.files) then
	    for j, f in pairs(list.files) do
		downloads(host, path, list.out, nil, f)
	    end
	    downloads(host, path, list.out, nil, '')
--[[ I've seen flock & curl act oddly.  Perhaps flock didn't have time to start up?
    /var/www/html/apt-panopticon/apt-panopticon/results_2019-12-22-15-00

Mon Dec 23 01:02:54 2019 DEBUG : forking

flock -n results/curl-debs-pkgmaster.devuan.org.log curl --connect-timeout 5 --create-dirs -f -L --fail-early --max-time 300 --retry 3 -R -v -z 'results/stamp.old'  --stderr results/curl-debs-pkgmaster.devuan.org.log -K results/curl-debs-pkgmaster.devuan.org.files


Mon Dec 23 01:02:54 2019 DEBUG : 0 flock -n results/curl-debs-pkgmaster.devuan.org.log commands still running.
Mon Dec 23 01:02:54 2019 DEBUG : *>* Resumed coroutine NO LONGER waiting on - 0 < APT.checkExes(flock -n results/curl-debs-pkgmaster.devuan.org.log
Mon Dec 23 01:02:54 2019 DEBUG : *** Doing list.parser() for debs
Mon Dec 23 01:02:54 2019 ERROR (http Updated pkgmaster.devuan.org): Failed to download - results/pkgmaster.devuan.org/merged/pool/DEBIAN/main/a/aptly/aptly_1.3.0+ds1-4_amd64.deb

drwxr-x--- 2 www-data www-data 4096 2019-12-23 01:02:57.000000000 +1000 aptly
-rw-r--r-- 1 www-data www-data  7129 2019-12-23 01:03:54.000000000 +1000 curl-debs-pkgmaster.devuan.org.log

]]
os.execute('sleep 1')	-- Wait for things to start up before checking for them.
	    while 0 < APT.checkExes(downloadLock .. list.out .. "-" .. host .. ".log.txt") do
		D('*&lt;* About to yield coroutine while waiting on - 0 < APT.checkExes(' .. downloadLock .. list.out .. '-' .. host .. '.log.txt')
		coroutine.yield()
		D('*&gt;* Resumed coroutine while waiting on - 0 < APT.checkExes(' .. downloadLock .. list.out .. '-' .. host .. '.log.txt')
	    end
	    D('*&gt;* Resumed coroutine NO LONGER waiting on - 0 < APT.checkExes(' .. downloadLock .. list.out .. '-' .. host .. '.log.txt')

	    local f = 'results/curl-' .. list.out .. "-" .. host .. ".log.txt"

	    local f = 'results/curl-' .. list.out .. "-" .. host .. ".log.txt"
	    -- Should not be needed, but maybe this is why sometimes I don't see the speeds, though the file is there aand valid when I look later.
	    while not APT.checkFile(f) do
		D('*&lt;* About to yield coroutine while waiting on - not APT.checkFile(' .. f .. ')')
		coroutine.yield()
		D('*&gt;* Resumed coroutine while waiting on - not APT.checkFile(' .. f .. ')')
	    end

--[[ TODO - should try to figure out which server the file actually got downloaded from, and attribute the speed and errors to that server.
	    Which means parsing the curl logs, not just a simple match().
	    Watch out for misplaced ^M, they don't all come at the end of the line.
	    Also note curl-Release-mirror.devuan.de.log.txt, timeouts don't always show the "Connected to" string.
	    * Immediate connect fail for 2001:4ca0:4300::1:19: Network is unreachable
	    * Connected to debian.ipacct.com (2a01:9e40::251) port 80 (#1)
	      0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0* Connected to devuan.bio.lmu.de (141.84.43.19) port 80 (#0)
	    curl: (22) The requested URL returned error: 404 Not Found
]]
	    local min, max, spd = 999999999999, 0
	    local num = '[%d%.]+[kM]?'
	    if APT.checkFile(f) then
		for l in io.lines(f) do
		    local speed, crrnt = l:match('^%c *'..num..' +'..num..' +%d+ +'..num..' +%d+ +%d+ +('..num..') +%d+ +[%d%-]+:[%d%-]+:[%d%-]+ +[%d%-]+:[%d%-]+:[%d%-]+ +[%d%-]+:[%d%-]+:[%d%-]+ +('..num..')')
		    if nil ~= speed then
			if 'k' == speed:sub(-1, -1)         then speed = tonumber(speed:sub(1, -2)) * 1024
			    elseif 'M' == speed:sub(-1, -1) then speed = tonumber(speed:sub(1, -2)) * 1024 * 1024
			end
			speed = tonumber(speed)
			if 'k' == crrnt:sub(-1, -1)         then crrnt = tonumber(crrnt:sub(1, -2)) * 1024
			    elseif 'M' == crrnt:sub(-1, -1) then crrnt = tonumber(crrnt:sub(1, -2)) * 1024 * 1024
			end
			crrnt = tonumber(crrnt)
			if speed < min and speed ~= 0 then min = speed end
			if speed > max                then max = speed end
			if crrnt < min and crrnt ~= 0 then min = crrnt end
			if crrnt > max                then max = crrnt end
		    end
		    if l:find('timed out') ~= nil then
			E("  TIMEOUT " .. timeouts + 1 .. ', details in <a href="curl-' .. list.out .. '-' .. host .. '.log.txt">curl-' .. list.out .. '-' .. host .. '.log.txt</a>', 'http', '', host)
			timeouts = timeouts + 1
			APT.results["timeout"] = true
		    end
		end
	    end
	    APT.results["speed"] = {["min"] = min, ["max"] = max}
	end

	if (APT.options.referenceSite.value ~= host) and ('' ~= list.nextf) then
	    local sem = 'results/NEW_' .. list.nextf .. '_%s.txt'
	    for i, n in pairs(releases) do
		local f = sem:format(n)
		while not APT.checkFile(f) do
		    D('*&lt;* About to yield coroutine while waiting on - not APT.checkFile(' .. f .. ')')
		    coroutine.yield()
		    D('*&gt;* Resumed coroutine while waiting on - not APT.checkFile(' .. f .. ')')
		end
	    end
	end
	D('*** Doing list.parser() for ' .. list.out)
	list = list.parser(host)
	if APT.options.timeouts.value <= (totalTimeouts) then break end
    end
    D('*&lt;&lt;* About to end coroutine.')
    cor = nil
end


if 0 < #arg then
    if "/" == arg[1]:sub(-1, -1) then
	W("slash at end of path!  " .. arg[1])
	arg[1] = arg[1]:sub(1, -2)
    end
    if " " == arg[1]:sub(-1, -1) then
	W("space at end of path!  " .. arg[1])
	arg[1] = arg[1]:sub(1, -2)
    end
    local pu = url.parse("http://" .. arg[1])

    if APT.redir and (nil == arg[3])then
	arg[3] = arg[2]
	arg[2] = nil
    end

    if APT.testing("Integrity") or APT.testing("Updated") then
	if APT.origin and APT.options.referenceSite.value == pu.host then
--	    if not APT.keep then os.execute("rm -fr results/" .. pu.host .. " 2>/dev/null") end
	end
    end

    if not APT.logOpen(pu.host, arg[2], arg[3]) then return end
    I("Starting tests for " .. arg[1] .. " with these tests - " .. table.concat(APT.options.tests.value, ", "))
    APT.mirrors = loadfile("results/mirrors.lua")()
    APT.results = APT.padResults(APT.results)
    if APT.origin or APT.redir then APT.results["IPs"] = gatherIPs(pu.host) end
    if nil ~= arg[2] then I(" &nbsp; Using IP " .. arg[2]); ip = arg[2] end
    if nil ~= arg[3] then I(" &nbsp; Using file " .. arg[3]); end

    if APT.origin then
	local file = arg[3]
	if nil == file then file = '' end
	local path = pu.path
	if nil == path then path = '' end
	if APT.origin then
	    local ips = APT.results["IPs"]
	    if nil ~= ips then
		APT.allpairs(ips,
		    function(k, v)
			if v == "A" then
			    if APT.testing("IPv4") then APT.exe('./apt-panopticon.lua ' .. sendArgs .. ' -4 ' .. pu.host .. path .. ' ' .. k .. ' ' .. file):Nice():log():fork() end
			elseif v == "AAAA" then
			    if APT.testing("IPv6") then APT.exe('./apt-panopticon.lua ' .. sendArgs .. ' -6 ' .. APT.IPv46 .. ' ' .. pu.host .. path .. ' ' .. k .. ' ' .. file):Nice():log():fork() end
			end
			D('logging to ' .. APT.logName(pu.host, k, file)[2])
		    end
		)
	    else
		E("no IPs for " .. pu.host)
	    end
	end

	if not APT.redir then
	    if (1 <= APT.options.bandwidth.value) and (APT.testing("Integrity") or APT.testing("Updated")) then
		if APT.origin and (APT.options.roundRobin.value ~= pu.host) then
		    I("Starting file downloads for " .. pu.host)
		    D('*&gt;* About to create coroutine.')
		    cor = coroutine.create(doDownloads)
		    local ok, message = coroutine.resume(cor, pu.host, pu.path, parseStart(pu.host))
		    if not ok then cor = nil; print(message) end
		end
	    end
	    checkFiles(pu.host, pu.host, pu.path);
	else
	    checkFiles(pu.host, pu.host, pu.path:sub(1, -1), file);
	end
    else
	checkHost(pu.host, pu.host, pu.path, arg[2], arg[3])
    end

    while nil ~= cor do
	os.execute('sleep 10')
	D('*&gt;* About to resume coroutine before writing results.')

	local ok, message = coroutine.resume(cor)
	if not ok then cor = nil; print(message); break end
    end

    local f = pu.host
    if "" ~= ip then f = f .. "_" .. ip end
-- TODO - perhaps number them if there's more than one?
    if APT.redir then f = f .. '_' .. 'R' end
    local rfile, e = io.open("results/" .. f .. ".lua", "w+")
    if nil == rfile then C("opening results file - " .. e) else
	rfile:write(APT.dumpTable(APT.results, "results") .. "\nreturn results\n")
        rfile:close()
    end

    if APT.origin and (not APT.redir) and (APT.options.referenceSite.value ~= pu.host) then
os.execute('sleep 1')	-- Wait for things to start up before checking for them.
	while 0 < APT.checkExes(downloadLock .. "Release-"  .. pu.host .. ".log.txt") do os.execute("sleep 10") end
	while 0 < APT.checkExes(downloadLock .. "Packages-" .. pu.host .. ".log.txt") do os.execute("sleep 10") end
	while 0 < APT.checkExes(downloadLock .. "package-"  .. pu.host .. ".log.txt") do os.execute("sleep 10") end
	os.execute("sleep 5")
	if not APT.keep then os.execute("rm -fr results/" .. pu.host .. " 2>/dev/null") end
    end

    APT.logPost()
else
    local adt = APT.exe("ls -dl results_old 2>/dev/null | cut -d '>' -f 2 | cut -d ' ' -f 2"):Do().result:sub(2, -2)
    if (nil ~= adt) and APT.checkFile(adt) then
	APT.exe('mkdir -p ' .. adt:sub(1, 18))
	:And():Nice('tar -c --xz ' .. adt .. ' -f ' .. adt:sub(1, 18) .. '/' .. adt .. '.tar.xz')
	:And('rm -fr ' .. adt):noErr():fork()
    end
    local dt = os.date('!%F-%H-%M')
    local odt = APT.exe('TZ="GMT" date -r results/stamp +%F-%H-%M 2>/dev/null'):Do().result:sub(2, -2)
    if nil ~= odt then os.execute('                               rm -f results_old; ln -s results_' .. odt .. ' results_old 2>/dev/null') end
    if nil ~= dt  then os.execute('mkdir -p results_' .. dt .. '; rm -f results;     ln -s results_' .. dt  .. ' results     2>/dev/null') end
    os.execute('if [ -f results/stamp ]; then mv results/stamp results/stamp.old; else touch results/stamp.old -t 199901010000; fi;  touch results/stamp')
    if not APT.keep then
	os.execute("rm -f results/*.html 2>/dev/null")
	os.execute("rm -f results/*.txt  2>/dev/null")
    end

    if not APT.logOpen('apt-panopticon') then return end
    I("Starting tests " .. table.concat(APT.options.tests.value, ", "))
    os.execute("mkdir -p results")
    APT.mirrors = getMirrors()
    checkHost(APT.options.referenceSite.value)

    for k, m in pairs(APT.mirrors) do
	local pu = url.parse("http://" .. m.BaseURL)
	if APT.options.referenceSite.value ~= pu.host then
	    checkHost(m.BaseURL)
	end
    end

    while not APT.checkFile('results/LOG_' .. APT.options.referenceSite.value .. '.html') do	-- Wait for things to start up before checking for them.
	D('Waiting for results/LOG_' .. APT.options.referenceSite.value .. '.html');
	os.execute("sleep 5")
-- TODO -  count these, and abort if it takes too long.
	-- Try something similar for the other "Wait for things to start up before checking for them.", maybe fold it into APT.exe.
    end
    while 1 <= APT.checkExes("apt-panopticon.lua " .. sendArgs) do os.execute("sleep 5") end

    local APT_args    = APT.args
    local APT_logFile = APT.logFile
    local debians = {}
    local srvs = io.popen('ls -1 results/*.lua')
    for l in srvs:lines() do
	local hst = l:sub(9, -5)
	if nil ~= l:find('_R%.lua') then hst = hst:sub(1, -3) end
	if (hst:find('_') == nil) and (nil == APT.mirrors[hst]) then
	    local ips = loadfile(l)().IPs
	    if nil ~= ips then
		debians[hst] = {Country = '', FDQN = hst, Active = 'yes', Rate = '', BaseURL = hst, Protocols = {http = true, https = true}, Bandwidth = '', IPs = ips}
		local baseFiles = {}
		local IPfiles = {}
		for i, a in pairs(ips) do
		    IPfiles[i] = {}
		    if type(a) == 'table' then
			for j, b in pairs(a) do
			    IPfiles[i][j] = {}
			end
		    end
		end
		local files = io.popen('ls -1 results/LOG_' .. hst .. '_*.html')
		for ll in files:lines() do
		    local dn = false
		    for i, a in pairs(ips) do
			if type(a) == 'table' then
			    for j, b in pairs(a) do
				if nil ~= ll:match('(results/LOG_' .. hst .. '_' .. j .. '_.*)') then
				    table.insert(IPfiles[i][j], ll)
				    dn = true
				end
			    end
			else
			    if nil ~= ll:match('(results/LOG_' .. hst .. '_' .. i .. '_.*)') then
				table.insert(IPfiles[i], ll)
				dn = true
			    end
			end
		    end
		    if not dn then table.insert(baseFiles, ll) end
		end

		local combine = function(ip, a)
		    if not APT.logOpen(hst, ip) then
			print('PROBLEM OPENING LOG FILE ' .. hst .. ' ' .. ip) 
		    else
			APT.logFile:write('<h1>Note log lines will be out of order, this is a bunch of other log files combined.</h1>\n')
			for i, f in pairs(a) do
			    f = f:sub(9, -1)
			    APT.logFile:write('<hr>\n<hr>\n<h2><a href="' .. f .. '">' .. f .. '</a></h2>\n')
			    for ln in io.lines('results/' .. f) do
				if ln:match('^' .. os.date('!%Y%-%m%-%d ') .. '.*$') then APT.logFile:write(ln .. '\n') end	-- %F isn't good enough, coz we have to escape the '-'.
			    end
			end
		    end
		    APT.logPost()
		    APT.args = APT_args
		    APT.logFile = APT_logFile
		end
		combine('', baseFiles)
		for ip, a in pairs(IPfiles) do
		    if nil == a[1] then
			for i, f in pairs(a) do
			    combine(i, f)
			end
		    else
			combine(ip, a)
		    end
		end

	    end
	end
    end
    local file, e = io.open("results/debians.lua", "w+")
    if nil == file then C("opening debians file - " .. e) else
	file:write(APT.dumpTable(debians, "debians") .. "\nreturn debians\n")
        file:close()
    end


    for k, v in pairs(APT.mirrors) do
	local f =  'results/' .. k .. '.lua'
	if APT.checkFile(f) then
	    results = loadfile(f)()
	    APT.mirrors[k]['IPs'] = results.IPs
	end
    end
    local file, e = io.open("results/mirrors.lua", "w+")
    if nil == file then C("opening mirrors file - " .. e) else
	file:write(APT.dumpTable(APT.mirrors, "mirrors") .. "\nreturn mirrors\n")
        file:close()
    end

    -- Create the reports.
    for n, r in pairs(APT.options.reports.value) do
	if APT.checkFile("apt-panopticon-report-" .. r .. ".lua") then
	    I("Creating " .. r .. " report.")
	    APT.exe("./apt-panopticon-report-" .. r .. ".lua " .. sendArgs):log():Do()
	end
    end

    I('Total run time was ' .. (os.time() - now) .. ' seconds.')
    APT.logPost()
end