diff options
Diffstat (limited to '')
-rwxr-xr-x | SuckItFos | 9 | ||||
-rwxr-xr-x | SuckItPm | 12 |
2 files changed, 9 insertions, 12 deletions
@@ -22,6 +22,7 @@ do | |||
22 | mkdir -p combined/$base | 22 | mkdir -p combined/$base |
23 | mkdir -p combined/${base}/`dirname ${file}` | 23 | mkdir -p combined/${base}/`dirname ${file}` |
24 | echo "Converting ${URL}/${base}/${file}?cover=print -> Foswiki/${base}/${file}.md" | 24 | echo "Converting ${URL}/${base}/${file}?cover=print -> Foswiki/${base}/${file}.md" |
25 | echo -e "ogWiki=Foswiki\nogURL=${URL}/${base}/${file}\n" > Foswiki/${base}/${file}.md.md | ||
25 | # Doesn't help with redownloads, coz natch a dynamic site isn't cached. But I can at least comment out the curl command during testing to save time. | 26 | # Doesn't help with redownloads, coz natch a dynamic site isn't cached. But I can at least comment out the curl command during testing to save time. |
26 | curl --silent --no-progress-meter ${URL}/${base}/${file}?cover=print -o Foswiki/${base}/${file}.HTM | 27 | curl --silent --no-progress-meter ${URL}/${base}/${file}?cover=print -o Foswiki/${base}/${file}.HTM |
27 | cp Foswiki/${base}/${file}.HTM Foswiki/${base}/${file}.HTM_ORIGINAL | 28 | cp Foswiki/${base}/${file}.HTM Foswiki/${base}/${file}.HTM_ORIGINAL |
@@ -47,6 +48,9 @@ do | |||
47 | rm Foswiki/${base}/${file}.md | 48 | rm Foswiki/${base}/${file}.md |
48 | mv xx00 Foswiki/${base}/${file}.md | 49 | mv xx00 Foswiki/${base}/${file}.md |
49 | fi | 50 | fi |
51 | if [ -f xx01 ]; then | ||
52 | rm xx01 | ||
53 | fi | ||
50 | 54 | ||
51 | # Attempt to clean things up, badly. | 55 | # Attempt to clean things up, badly. |
52 | sed -i -E Foswiki/${base}/${file}.md \ | 56 | sed -i -E Foswiki/${base}/${file}.md \ |
@@ -57,12 +61,7 @@ do | |||
57 | # -e 's/\{\.pattern.*\}//g' \ | 61 | # -e 's/\{\.pattern.*\}//g' \ |
58 | # -e 's/\{\.pattern.*//g' \ | 62 | # -e 's/\{\.pattern.*//g' \ |
59 | 63 | ||
60 | echo -e "****\n[Original page](${URL}/${base}/${file}) where maybe you can edit it." >> Foswiki/${base}/${file}.md | ||
61 | ln -frs Foswiki/${base}/${file}.md combined/${base}/${file}.md | 64 | ln -frs Foswiki/${base}/${file}.md combined/${base}/${file}.md |
62 | |||
63 | if [ -f xx01 ]; then | ||
64 | rm xx01 | ||
65 | fi | ||
66 | done | 65 | done |
67 | 66 | ||
68 | popd | 67 | popd |
@@ -20,6 +20,7 @@ do | |||
20 | mkdir -p PmWiki/$base | 20 | mkdir -p PmWiki/$base |
21 | mkdir -p combined/$base | 21 | mkdir -p combined/$base |
22 | echo "Converting ${URL}/?n=${base}.${file}?action=print -> PmWiki/${base}/${file}.md" | 22 | echo "Converting ${URL}/?n=${base}.${file}?action=print -> PmWiki/${base}/${file}.md" |
23 | echo -e "ogWiki=PmWiki\nogURL=${URL}/?n=${base}.${file}\n" > PmWiki/${base}/${file}.md.md | ||
23 | # Doesn't help with redownloads, coz natch a dynamic site isn't cached. But I can at least comment out the curl command during testing to save time. | 24 | # Doesn't help with redownloads, coz natch a dynamic site isn't cached. But I can at least comment out the curl command during testing to save time. |
24 | # curl --no-progress-meter ${URL}/?n=${base}.${file}?action=markdown -o PmWiki/${base}/${file}.MD | 25 | # curl --no-progress-meter ${URL}/?n=${base}.${file}?action=markdown -o PmWiki/${base}/${file}.MD |
25 | curl --no-progress-meter ${URL}/?n=${base}.${file}?action=print -o PmWiki/${base}/${file}.HTM | 26 | curl --no-progress-meter ${URL}/?n=${base}.${file}?action=print -o PmWiki/${base}/${file}.HTM |
@@ -29,6 +30,9 @@ do | |||
29 | rm PmWiki/${base}/${file}.HTM | 30 | rm PmWiki/${base}/${file}.HTM |
30 | mv xx00 PmWiki/${base}/${file}.HTM | 31 | mv xx00 PmWiki/${base}/${file}.HTM |
31 | fi | 32 | fi |
33 | if [ -f xx01 ]; then | ||
34 | rm xx01 | ||
35 | fi | ||
32 | sed -i -E PmWiki/${base}/${file}.HTM \ | 36 | sed -i -E PmWiki/${base}/${file}.HTM \ |
33 | -e "s/rel='nofollow'//g" \ | 37 | -e "s/rel='nofollow'//g" \ |
34 | -e "s/target='_blank'//g" \ | 38 | -e "s/target='_blank'//g" \ |
@@ -53,7 +57,6 @@ do | |||
53 | -e 's/\{#.*\}//g' \ | 57 | -e 's/\{#.*\}//g' \ |
54 | -e '/^:::/d' \ | 58 | -e '/^:::/d' \ |
55 | # -e '/\[Site$/d' \ | 59 | # -e '/\[Site$/d' \ |
56 | # -e '/^Page last modified on /d' \ | ||
57 | # -e '/^\[\]/d' \ | 60 | # -e '/^\[\]/d' \ |
58 | # -e "s/\`<a id='trailstart'>\`\{=html\}\`<\/a>\`\{=html\}//g" \ | 61 | # -e "s/\`<a id='trailstart'>\`\{=html\}\`<\/a>\`\{=html\}//g" \ |
59 | # -e "s/^\`<img /<img /g" \ | 62 | # -e "s/^\`<img /<img /g" \ |
@@ -61,13 +64,8 @@ do | |||
61 | # -e "s/^\`\`\`//g" \ | 64 | # -e "s/^\`\`\`//g" \ |
62 | # -e "s/\`\{=html\}//g" | 65 | # -e "s/\`\{=html\}//g" |
63 | 66 | ||
64 | # Don't need this, the parts we are grabbing already include that link at the bottom. | ||
65 | # echo -e "****\n[Original page](${URL}/${base}/${page}) where maybe you can edit it." >> PmWiki/${base}/${file}.md | ||
66 | ln -frs PmWiki/${base}/${file}.md combined/${base}/${file}.md | 67 | ln -frs PmWiki/${base}/${file}.md combined/${base}/${file}.md |
67 | 68 | ln -frs PmWiki/${base}/${file}.md.md combined/${base}/${file}.md.md | |
68 | if [ -f xx01 ]; then | ||
69 | rm xx01 | ||
70 | fi | ||
71 | done | 69 | done |
72 | 70 | ||
73 | popd | 71 | popd |