aboutsummaryrefslogtreecommitdiffstatshomepage
diff options
context:
space:
mode:
-rwxr-xr-xSuckItFos52
-rwxr-xr-xSuckItPm57
2 files changed, 109 insertions, 0 deletions
diff --git a/SuckItFos b/SuckItFos
new file mode 100755
index 0000000..4f824be
--- /dev/null
+++ b/SuckItFos
@@ -0,0 +1,52 @@
1#!/bin/bash
2
3URL="https://devuan.daffodil.linuxoz.net"
4
5filter="
6 -name _default -prune -o \
7 -name _empty -prune -o \
8 -name System -prune -o \
9 -name Trash -prune -o \
10 -name TWiki -prune -o \
11"
12
13find ../Foswiki/data ${filter} \
14-name "*.txt" -type f,l -printf "%P\n" | while read line
15do
16 base=`echo "${line}" | cut -d '/' -f 1`
17 file=`echo "${line}" | cut -d '/' -f 2- | rev | cut -b 5- | rev`
18 mkdir -p Foswiki/$base
19 echo "Converting ${URL}/${base}/${file} -> Foswiki/${base}/${file}.md"
20# pandoc -f html -t markdown --self-contained ${URL}/${base}/${file} >Foswiki/${base}/${file}.md
21 # TODO - try curl, to see what is actually downloaded, and maybe not download unchanged pages. curl to .HTM
22 # Doesn't help with redownloads, coz natch a dynamic site isn't cached. But I can at least comment out the curl command during testing to save time.
23 curl --silent --no-progress-meter ${URL}/${base}/${file} -o Foswiki/${base}/${file}.HTM
24 pandoc -f html -t commonmark_x --self-contained Foswiki//${base}/${file}.HTM >Foswiki/${base}/${file}.md
25 cp Foswiki/${base}/${file}.md Foswiki/${base}/${file}.md_ORIGINAL
26
27# csplit -ks Foswiki/${base}/${file}.md '%::: foswikiTopic%' '/::: foswikiContentFooter/'
28# if [ -f xx00 ]; then
29# rm Foswiki/${base}/${file}.md
30# mv xx00 Foswiki/${base}/${file}.md
31# fi
32
33 # Attempt to clean things up, badly.
34 sed -i -E Foswiki/${base}/${file}.md \
35 -e 's/\$/\$dlr\$/g' \
36 -e 's/\{#.*\}//g' \
37 -e 's/\{\.foswiki.*\}//g' \
38 -e 's/\{\.foswiki.*//g' \
39 -e 's/\{\.foswikiNewLink rel=“nofollow”\}//g' \
40 -e 's/\{\.foswikiNewLink$//g' \
41 -e 's/^\.foswiki.*\}//g' \
42 -e 's/\{\.pattern.*\}//g' \
43 -e 's/\{\.pattern.*//g' \
44 -e 's/\{rel="nofollow"\}//g' \
45 -e 's/^rel="nofollow"\}//g' \
46 -e 's/rel=“nofollow”\}$//g' \
47 -e '/^:::/d'
48
49 echo "<hr/><p><a href=\"${URL}/${base}/${file}\">Original page</a> where you can edit it.</p>" >> Foswiki/${base}/${file}.md
50done
51
52notYetAnotherWiki.lua
diff --git a/SuckItPm b/SuckItPm
new file mode 100755
index 0000000..6232cdc
--- /dev/null
+++ b/SuckItPm
@@ -0,0 +1,57 @@
1#!/bin/bash
2
3URL="https://wiki.devuan.org"
4
5filter="
6 -not -name "*~" -a \
7 -not -name ".flock" -a \
8 -not -name ".htaccess" -a \
9 -not -name ".lastmod" -a \
10 -not -name ".pageindex" -a \
11"
12
13find ../pmwiki/wiki.d ${filter} \
14-name "*.*" -type f,l -printf "%P\n" | while read line
15do
16 base=`echo "${line}" | cut -d '.' -f 1`
17 file=`echo "${line}" | cut -d '.' -f 2`
18 mkdir -p PmWiki/$base
19 echo "Converting ${URL}/?n=${base}.${file} -> PmWiki/${base}/${file}.md"
20# pandoc -f html -t markdown --self-contained ${URL}/?n=${base}.${file} >PmWiki/${base}/${file}.md
21 # TODO - try curl, to see what is actually downloaded, and maybe not download unchanged pages. curl to .HTM
22 # Doesn't help with redownloads, coz natch a dynamic site isn't cached. But I can at least comment out the curl command during testing to save time.
23 curl --no-progress-meter ${URL}/?n=${base}.${file} -o PmWiki/${base}/${file}.HTM
24 pandoc -f html -t commonmark_x --self-contained PmWiki//${base}/${file}.HTM >PmWiki/${base}/${file}.md
25 cp PmWiki/${base}/${file}.md PmWiki/${base}/${file}.md_ORIGINAL
26
27# csplit -ks PmWiki/${base}/${file}.md '/trailstart/' '/trailend/'
28 csplit -ks PmWiki/${base}/${file}.md '%::: {#wikitext}%'
29 if [ -f xx00 ]; then
30 rm PmWiki/${base}/${file}.md
31 mv xx00 PmWiki/${base}/${file}.md
32 fi
33
34 # Attempt to clean things up, badly.
35 sed -i -E PmWiki/${base}/${file}.md \
36 -e 's/\$/\$dlr\$/g' \
37 -e 's/\{#.*\}//g' \
38 -e '/\{\.wikilink\}/d' \
39 -e '/\[Site$/d' \
40 -e '/^:::/d' \
41 -e '/^Page last modified on /d' \
42 -e '/^\[\]/d' \
43 -e 's/\{rel=".*\}//g' \
44 -e 's/\{rel="nofollow"$//g' \
45 -e 's/^rel="nofollow"\}//g' \
46 -e 's/^target="_blank"\}//g' \
47 -e 's/\{\.createlinktext.*\}//g' \
48 -e 's/\{\.createlinktext$//g' \
49 -e 's/\{\.createlink.*\}//g' \
50 -e 's/\{\.createlink$//g' \
51 -e 's/\{\.urllink.*\}//g' \
52 -e 's/\{\.urllink$//g'
53
54echo "<hr/><p><a href=\"${URL}/?n=${base}.${file}\">Original page</a> where you can edit it.</p>" >> PmWiki/${base}/${file}.md
55done
56
57notYetAnotherWiki.lua