#!/bin/bash URL="https://fos.wiki.devuan.org" filter=" -name _default -prune -o \ -name _empty -prune -o \ -name System -prune -o \ -name Trash -prune -o \ -name TWiki -prune -o \ " pushd /opt/mergedWork find /opt/Foswiki/data ${filter} \ -name "*.txt" -type f,l -printf "%P\n" | while read line do base=`echo "${line}" | cut -d '/' -f 1` file=`echo "${line}" | cut -d '/' -f 2- | rev | cut -b 5- | rev` time=`date --rfc-3339=seconds -ur /opt/Foswiki/data/${base}/${file}.txt | cut -d '+' -f 1` mkdir -p Foswiki/$base mkdir -p Foswiki/${base}/`dirname ${file}` mkdir -p combined/$base mkdir -p combined/${base}/`dirname ${file}` echo "Converting ${URL}/${base}/${file}?cover=print -> Foswiki/${base}/${file}.md" echo -e "ogWiki=Foswiki\nogURL=${URL}\nogBase=${base}\nogFile=${file}\ntimestamp=${time}\n" > Foswiki/${base}/${file}.md.md # Doesn't help with redownloads, coz natch a dynamic site isn't cached. But I can at least comment out the curl command during testing to save time. curl --silent --no-progress-meter ${URL}/${base}/${file}?cover=print -o Foswiki/${base}/${file}.HTM cp Foswiki/${base}/${file}.HTM Foswiki/${base}/${file}.HTM_ORIGINAL csplit -ks Foswiki/${base}/${file}.HTM '%