diff options
Diffstat (limited to '')
-rwxr-xr-x | SuckIt | 69 | ||||
-rwxr-xr-x | SuckItFos | 68 | ||||
-rwxr-xr-x | SuckItPm | 72 |
3 files changed, 69 insertions, 140 deletions
@@ -0,0 +1,69 @@ | |||
1 | #!/bin/bash | ||
2 | |||
3 | TIMEFORMAT=" took %lR using %P%% CPU" | ||
4 | time { | ||
5 | pushd /opt/merged | ||
6 | |||
7 | rm -fr Foswiki/* | ||
8 | cp -r /opt/merged_EMPTY/Foswiki . | ||
9 | rm -fr PmWiki/* | ||
10 | cp -r /opt/merged_EMPTY/PmWiki . | ||
11 | |||
12 | |||
13 | filter=" | ||
14 | -name _default -prune -o \ | ||
15 | -name _empty -prune -o \ | ||
16 | -name System -prune -o \ | ||
17 | -name Trash -prune -o \ | ||
18 | -name TWiki -prune -o \ | ||
19 | " | ||
20 | URL="https://fos.wiki.devuan.org" | ||
21 | time find /opt/Foswiki/data ${filter} \ | ||
22 | -name "*.txt" -type f,l -printf "%P\n" | while read line | ||
23 | do | ||
24 | base=`echo "${line}" | cut -d '/' -f 1` | ||
25 | file=`echo "${line}" | cut -d '/' -f 2- | rev | cut -b 5- | rev` | ||
26 | if [[ ! $file =~ (AdminGroup|AdminUser|AdminUserLeftBar|EditorGroup|GroupTemplate|GroupViewTemplate|NobodyGroup|PatternSkinUserViewTemplate|ProjectContributor|RegistrationAgent|SitePreferences|UnprocessedRegistrations|UnprocessedRegistrationsLog|UserHomepageHeader|UserList|UserListByDateJoined|UserListByLocation|UserList|UserListHeader|WebAtom|WebChanges|WebCreateNewTopic|WebHome|WebIndex|WebLeftBar|WebLeftBarExample|WebNotify|WebPreferences|WebRss|WebSearch|WebSearchAdvanced|WebTopicList|WikiGroups|WikiUsers)$ ]]; then | ||
27 | time=`date --rfc-3339=seconds -ur /opt/Foswiki/data/${base}/${file}.txt | cut -d '+' -f 1` | ||
28 | mkdir -p Foswiki/$base | ||
29 | mkdir -p Foswiki/${base}/`dirname ${file}` | ||
30 | echo -e "ogWiki=Foswiki\nogURL=${URL}\nogBase=${base}\nogFile=${file}\ntimestamp=${time}\n" > Foswiki/${base}/${file}.md.md | ||
31 | echo "downloading ${URL}/${base}/${file}?cover=print" | ||
32 | # Doesn't help with redownloads, coz natch a dynamic site isn't cached. But I can at least comment out the curl command during testing to save time. | ||
33 | curl --silent --no-progress-meter ${URL}/${base}/${file}?cover=print -o Foswiki/${base}/${file}.HTM | ||
34 | fi | ||
35 | done | ||
36 | |||
37 | |||
38 | filter=" | ||
39 | -not -name "*~" -a \ | ||
40 | -not -name ".flock" -a \ | ||
41 | -not -name ".htaccess" -a \ | ||
42 | -not -name ".lastmod" -a \ | ||
43 | -not -name ".pageindex" -a \ | ||
44 | " | ||
45 | URL="https://wiki.devuan.org" | ||
46 | time find /opt/pmwiki/wiki.d ${filter} \ | ||
47 | -name "*.*" -type f,l -printf "%P\n" | while read line | ||
48 | do | ||
49 | base=`echo "${line}" | cut -d '.' -f 1` | ||
50 | if [[ "${base}" != "Site" ]]; then | ||
51 | file=`echo "${line}" | cut -d '.' -f 2` | ||
52 | time=`date --rfc-3339=seconds -ur /opt/pmwiki/wiki.d/${base}.${file} | cut -d '+' -f 1` | ||
53 | mkdir -p PmWiki/$base | ||
54 | echo -e "ogWiki=PmWiki\nogURL=${URL}\nogBase=${base}\nogFile=${file}\ntimestamp=${time}\n" > PmWiki/${base}/${file}.md.md | ||
55 | # echo "downloading ${URL}/?n=${base}.${file}?action=markdown" | ||
56 | # curl --no-progress-meter ${URL}/?n=${base}.${file}?action=markdown -o PmWiki/${base}/${file}.MARKDOWN | ||
57 | echo "downloading ${URL}/?n=${base}.${file}?action=print" | ||
58 | curl --no-progress-meter ${URL}/?n=${base}.${file}?action=print -o PmWiki/${base}/${file}.HTM | ||
59 | |||
60 | # pandoc -f markdown -t commonmark_x --self-contained PmWiki//${base}/${file}.MD >PmWiki/${base}/${file}.md | ||
61 | # pandoc -f html -t commonmark_x --self-contained PmWiki//${base}/${file}.HTM >PmWiki/${base}/${file}.md | ||
62 | fi | ||
63 | done | ||
64 | |||
65 | |||
66 | time notYetAnotherWiki.lua | ||
67 | |||
68 | popd | ||
69 | } | ||
diff --git a/SuckItFos b/SuckItFos deleted file mode 100755 index dc65505..0000000 --- a/SuckItFos +++ /dev/null | |||
@@ -1,68 +0,0 @@ | |||
1 | #!/bin/bash | ||
2 | |||
3 | URL="https://fos.wiki.devuan.org" | ||
4 | |||
5 | filter=" | ||
6 | -name _default -prune -o \ | ||
7 | -name _empty -prune -o \ | ||
8 | -name System -prune -o \ | ||
9 | -name Trash -prune -o \ | ||
10 | -name TWiki -prune -o \ | ||
11 | " | ||
12 | |||
13 | pushd /opt/mergedWork | ||
14 | |||
15 | find /opt/Foswiki/data ${filter} \ | ||
16 | -name "*.txt" -type f,l -printf "%P\n" | while read line | ||
17 | do | ||
18 | base=`echo "${line}" | cut -d '/' -f 1` | ||
19 | file=`echo "${line}" | cut -d '/' -f 2- | rev | cut -b 5- | rev` | ||
20 | time=`date --rfc-3339=seconds -ur /opt/Foswiki/data/${base}/${file}.txt | cut -d '+' -f 1` | ||
21 | mkdir -p Foswiki/$base | ||
22 | mkdir -p Foswiki/${base}/`dirname ${file}` | ||
23 | mkdir -p combined/$base | ||
24 | mkdir -p combined/${base}/`dirname ${file}` | ||
25 | echo "Converting ${URL}/${base}/${file}?cover=print -> Foswiki/${base}/${file}.md" | ||
26 | echo -e "ogWiki=Foswiki\nogURL=${URL}\nogBase=${base}\nogFile=${file}\ntimestamp=${time}\n" > Foswiki/${base}/${file}.md.md | ||
27 | # Doesn't help with redownloads, coz natch a dynamic site isn't cached. But I can at least comment out the curl command during testing to save time. | ||
28 | curl --silent --no-progress-meter ${URL}/${base}/${file}?cover=print -o Foswiki/${base}/${file}.HTM | ||
29 | cp Foswiki/${base}/${file}.HTM Foswiki/${base}/${file}.HTM_ORIGINAL | ||
30 | csplit -ks Foswiki/${base}/${file}.HTM '%<div id="patternMainContents">%' '/<div class="foswikiAttachments foswikiFormStep" style="overflow:auto">/' | ||
31 | if [ -f xx00 ]; then | ||
32 | rm Foswiki/${base}/${file}.HTM | ||
33 | mv xx00 Foswiki/${base}/${file}.HTM | ||
34 | fi | ||
35 | sed -i -E Foswiki/${base}/${file}.HTM \ | ||
36 | -e "s/rel='nofollow'//g" \ | ||
37 | -e 's/rel="nofollow"//g' \ | ||
38 | -e "s/target='_blank'//g" \ | ||
39 | -e "s/class='foswiki[[:alpha:]]*'//g" \ | ||
40 | -e 's/class="foswikiTopic"/class="FoswikiTopic"/g' \ | ||
41 | -e 's/class="foswiki[[:alpha:]]*"//g' \ | ||
42 | -e "s/style='.*;'//g" | ||
43 | |||
44 | pandoc -f html -t commonmark_x --self-contained Foswiki//${base}/${file}.HTM >Foswiki/${base}/${file}.md | ||
45 | cp Foswiki/${base}/${file}.md Foswiki/${base}/${file}.md_ORIGINAL | ||
46 | |||
47 | csplit -ks Foswiki/${base}/${file}.md '%::: {.FoswikiTopic}%' '/::: {.patternInfo}/' | ||
48 | if [ -f xx00 ]; then | ||
49 | rm Foswiki/${base}/${file}.md | ||
50 | mv xx00 Foswiki/${base}/${file}.md | ||
51 | fi | ||
52 | if [ -f xx01 ]; then | ||
53 | rm xx01 | ||
54 | fi | ||
55 | |||
56 | # Attempt to clean things up, badly. | ||
57 | sed -i -E Foswiki/${base}/${file}.md \ | ||
58 | -e 's/\$/\$dlr\$/g' \ | ||
59 | -e 's/\{#.*\}//g' \ | ||
60 | -e '/^:::/d' \ | ||
61 | -e '/^<!-- -->/d' \ | ||
62 | # -e 's/\{\.pattern.*\}//g' \ | ||
63 | # -e 's/\{\.pattern.*//g' \ | ||
64 | |||
65 | ln -frs Foswiki/${base}/${file}.md combined/${base}/${file}.md | ||
66 | done | ||
67 | |||
68 | popd | ||
diff --git a/SuckItPm b/SuckItPm deleted file mode 100755 index 246f485..0000000 --- a/SuckItPm +++ /dev/null | |||
@@ -1,72 +0,0 @@ | |||
1 | #!/bin/bash | ||
2 | |||
3 | URL="https://wiki.devuan.org" | ||
4 | |||
5 | filter=" | ||
6 | -not -name "*~" -a \ | ||
7 | -not -name ".flock" -a \ | ||
8 | -not -name ".htaccess" -a \ | ||
9 | -not -name ".lastmod" -a \ | ||
10 | -not -name ".pageindex" -a \ | ||
11 | " | ||
12 | |||
13 | pushd /opt/mergedWork | ||
14 | |||
15 | find /opt/pmwiki/wiki.d ${filter} \ | ||
16 | -name "*.*" -type f,l -printf "%P\n" | while read line | ||
17 | do | ||
18 | base=`echo "${line}" | cut -d '.' -f 1` | ||
19 | file=`echo "${line}" | cut -d '.' -f 2` | ||
20 | time=`date --rfc-3339=seconds -ur /opt/pmwiki/wiki.d/${base}.${file} | cut -d '+' -f 1` | ||
21 | mkdir -p PmWiki/$base | ||
22 | mkdir -p combined/$base | ||
23 | echo "Converting ${URL}/?n=${base}.${file}?action=print -> PmWiki/${base}/${file}.md" | ||
24 | echo -e "ogWiki=PmWiki\nogURL=${URL}\nogBase=${base}\nogFile=${file}\ntimestamp=${time}\n" > PmWiki/${base}/${file}.md.md | ||
25 | # Doesn't help with redownloads, coz natch a dynamic site isn't cached. But I can at least comment out the curl command during testing to save time. | ||
26 | # curl --no-progress-meter ${URL}/?n=${base}.${file}?action=markdown -o PmWiki/${base}/${file}.MD | ||
27 | curl --no-progress-meter ${URL}/?n=${base}.${file}?action=print -o PmWiki/${base}/${file}.HTM | ||
28 | cp PmWiki/${base}/${file}.HTM PmWiki/${base}/${file}.HTM_ORIGINAL | ||
29 | csplit -ks PmWiki/${base}/${file}.HTM '%<!--PageText-->%' '/<!--HTMLFooter-->/' | ||
30 | if [ -f xx00 ]; then | ||
31 | rm PmWiki/${base}/${file}.HTM | ||
32 | mv xx00 PmWiki/${base}/${file}.HTM | ||
33 | fi | ||
34 | if [ -f xx01 ]; then | ||
35 | rm xx01 | ||
36 | fi | ||
37 | sed -i -E PmWiki/${base}/${file}.HTM \ | ||
38 | -e "s/rel='nofollow'//g" \ | ||
39 | -e "s/target='_blank'//g" \ | ||
40 | -e "s/class='categorylink'//g" \ | ||
41 | -e "s/class='createlink'//g" \ | ||
42 | -e "s/class='createlinktext'//g" \ | ||
43 | -e "s/class='escaped'//g" \ | ||
44 | -e "s/class='diffmarkup'//g" \ | ||
45 | -e "s/class='selflink'//g" \ | ||
46 | -e "s/class='urllink'//g" \ | ||
47 | -e "s/class='vspace'//g" \ | ||
48 | -e "s/class='wikilink'//g" \ | ||
49 | -e "s/style='.*;'//g" | ||
50 | |||
51 | # pandoc -f markdown -t commonmark_x --self-contained PmWiki//${base}/${file}.MD >PmWiki/${base}/${file}.md | ||
52 | pandoc -f html -t commonmark_x --self-contained PmWiki//${base}/${file}.HTM >PmWiki/${base}/${file}.md | ||
53 | cp PmWiki/${base}/${file}.md PmWiki/${base}/${file}.md_ORIGINAL | ||
54 | |||
55 | # Attempt to clean things up, badly. | ||
56 | sed -i -E PmWiki/${base}/${file}.md \ | ||
57 | -e 's/\$/\$dlr\$/g' \ | ||
58 | -e 's/\{#.*\}//g' \ | ||
59 | -e '/^:::/d' \ | ||
60 | # -e '/\[Site$/d' \ | ||
61 | # -e '/^\[\]/d' \ | ||
62 | # -e "s/\`<a id='trailstart'>\`\{=html\}\`<\/a>\`\{=html\}//g" \ | ||
63 | # -e "s/^\`<img /<img /g" \ | ||
64 | # -e "s/^\`\`\`\{=html\}//g" \ | ||
65 | # -e "s/^\`\`\`//g" \ | ||
66 | # -e "s/\`\{=html\}//g" | ||
67 | |||
68 | ln -frs PmWiki/${base}/${file}.md combined/${base}/${file}.md | ||
69 | ln -frs PmWiki/${base}/${file}.md.md combined/${base}/${file}.md.md | ||
70 | done | ||
71 | |||
72 | popd | ||