summaryrefslogtreecommitdiff
path: root/ikiwiki
blob: 771590718dea2a84963fab833d318f124290c9c1 (plain)
  1. #!/usr/bin/perl -T
  2. use warnings;
  3. use strict;
  4. use File::Find;
  5. use Memoize;
  6. use File::Spec;
  7. $ENV{PATH}="/usr/local/bin:/usr/bin:/bin";
  8. BEGIN {
  9. $blosxom::version="is a proper perl module too much to ask?";
  10. do "/usr/bin/markdown";
  11. }
  12. my ($srcdir, $destdir, %links, %oldlinks, %oldpagemtime, %renderedfiles,
  13. %pagesources);
  14. my $link=qr/\[\[([^\s]+)\]\]/;
  15. my $verbose=0;
  16. my $wikiname="wiki";
  17. sub usage {
  18. die "usage: ikiwiki [options] source dest\n";
  19. }
  20. sub error ($) {
  21. die @_;
  22. }
  23. sub debug ($) {
  24. print "@_\n" if $verbose;
  25. }
  26. sub mtime ($) {
  27. my $page=shift;
  28. return (stat($page))[9];
  29. }
  30. sub possibly_foolish_untaint ($) {
  31. my $tainted=shift;
  32. my ($untainted)=$tainted=~/(.*)/;
  33. return $untainted;
  34. }
  35. sub basename {
  36. my $file=shift;
  37. $file=~s!.*/!!;
  38. return $file;
  39. }
  40. sub dirname {
  41. my $file=shift;
  42. $file=~s!/?[^/]+$!!;
  43. return $file;
  44. }
  45. sub pagetype ($) {
  46. my $page=shift;
  47. if ($page =~ /\.mdwn$/) {
  48. return ".mdwn";
  49. }
  50. else {
  51. return "unknown";
  52. }
  53. }
  54. sub pagename ($) {
  55. my $file=shift;
  56. my $type=pagetype($file);
  57. my $page=$file;
  58. $page=~s/\Q$type\E*$// unless $type eq 'unknown';
  59. return $page;
  60. }
  61. sub htmlpage ($) {
  62. my $page=shift;
  63. return $page.".html";
  64. }
  65. sub readpage ($) {
  66. my $page=shift;
  67. local $/=undef;
  68. open (PAGE, "$srcdir/$page") || error("failed to read $page: $!");
  69. my $ret=<PAGE>;
  70. close PAGE;
  71. return $ret;
  72. }
  73. sub writepage ($$) {
  74. my $page=shift;
  75. my $content=shift;
  76. my $dir=dirname("$destdir/$page");
  77. if (! -d $dir) {
  78. my $d="";
  79. foreach my $s (split(m!/+!, $dir)) {
  80. $d.="$s/";
  81. if (! -d $d) {
  82. mkdir($d) || error("failed to create directory $d: $!");
  83. }
  84. }
  85. }
  86. open (PAGE, ">$destdir/$page") || error("failed to write $page: $!");
  87. print PAGE $content;
  88. close PAGE;
  89. }
  90. sub findlinks {
  91. my $content=shift;
  92. my @links;
  93. while ($content =~ /$link/g) {
  94. push @links, lc($1);
  95. }
  96. return @links;
  97. }
  98. # Given a page and the text of a link on the page, determine which existing
  99. # page that link best points to. Prefers pages under a subdirectory with
  100. # the same name as the source page, failing that goes down the directory tree
  101. # to the base looking for matching pages.
  102. sub bestlink ($$) {
  103. my $page=shift;
  104. my $link=lc(shift);
  105. my $cwd=$page;
  106. do {
  107. my $l=$cwd;
  108. $l.="/" if length $l;
  109. $l.=$link;
  110. if (exists $links{$l}) {
  111. #debug("for $page, \"$link\", use $l");
  112. return $l;
  113. }
  114. } while $cwd=~s!/?[^/]+$!!;
  115. #print STDERR "warning: page $page, broken link: $link\n";
  116. return "";
  117. }
  118. sub isinlinableimage ($) {
  119. my $file=shift;
  120. $file=~/\.(png|gif|jpg|jpeg)$/;
  121. }
  122. sub htmllink ($$) {
  123. my $page=shift;
  124. my $link=shift;
  125. my $bestlink=bestlink($page, $link);
  126. return $link if $page eq $bestlink;
  127. # TODO BUG: %renderedfiles may not have it, if the linked to page
  128. # was also added and isn't yet rendered! Note that this bug is
  129. # masked by the bug mentioned below that makes all new files
  130. # be rendered twice.
  131. if (! grep { $_ eq $bestlink } values %renderedfiles) {
  132. $bestlink=htmlpage($bestlink);
  133. }
  134. if (! grep { $_ eq $bestlink } values %renderedfiles) {
  135. return "<a href=\"?\">?</a>$link"
  136. }
  137. $bestlink=File::Spec->abs2rel($bestlink, dirname($page));
  138. if (isinlinableimage($bestlink)) {
  139. return "<img src=\"$bestlink\">";
  140. }
  141. return "<a href=\"$bestlink\">$link</a>";
  142. }
  143. sub linkify ($$) {
  144. my $content=shift;
  145. my $file=shift;
  146. $content =~ s/$link/htmllink(pagename($file), $1)/eg;
  147. return $content;
  148. }
  149. sub htmlize ($$) {
  150. my $type=shift;
  151. my $content=shift;
  152. if ($type eq '.mdwn') {
  153. return Markdown::Markdown($content);
  154. }
  155. else {
  156. error("htmlization of $type not supported");
  157. }
  158. }
  159. sub linkbacks ($$) {
  160. my $content=shift;
  161. my $page=shift;
  162. my @links;
  163. foreach my $p (keys %links) {
  164. next if bestlink($page, $p) eq $page;
  165. if (grep { length $_ && bestlink($p, $_) eq $page } @{$links{$p}}) {
  166. my $href=File::Spec->abs2rel(htmlpage($p), dirname($page));
  167. # Trim common dir prefixes from both pages.
  168. my $p_trimmed=$p;
  169. my $page_trimmed=$page;
  170. my $dir;
  171. 1 while (($dir)=$page_trimmed=~m!^([^/]+/)!) &&
  172. defined $dir &&
  173. $p_trimmed=~s/^\Q$dir\E// &&
  174. $page_trimmed=~s/^\Q$dir\E//;
  175. push @links, "<a href=\"$href\">$p_trimmed</a>";
  176. }
  177. }
  178. $content.="<hr><p>Links: ".join(" ", sort @links)."</p>\n" if @links;
  179. return $content;
  180. }
  181. sub finalize ($$) {
  182. my $content=shift;
  183. my $page=shift;
  184. my $title=basename($page);
  185. $title=~s/_/ /g;
  186. my $pagelink="";
  187. my $path="";
  188. foreach my $dir (reverse split("/", $page)) {
  189. if (length($pagelink)) {
  190. $pagelink="<a href=\"$path$dir.html\">$dir</a>/ $pagelink";
  191. }
  192. else {
  193. $pagelink=$dir;
  194. }
  195. $path.="../";
  196. }
  197. $path=~s/\.\.\/$/index.html/;
  198. $pagelink="<a href=\"$path\">$wikiname</a>/ $pagelink";
  199. $content="<html>\n<head><title>$title</title></head>\n<body>\n".
  200. "<h1>$pagelink</h1>\n".
  201. $content.
  202. "</body>\n</html>\n";
  203. return $content;
  204. }
  205. sub render ($) {
  206. my $file=shift;
  207. my $type=pagetype($file);
  208. my $content=readpage($file);
  209. if ($type ne 'unknown') {
  210. my $page=pagename($file);
  211. $links{$page}=[findlinks($content)];
  212. $content=linkify($content, $file);
  213. $content=htmlize($type, $content);
  214. $content=linkbacks($content, $page);
  215. $content=finalize($content, $page);
  216. writepage(htmlpage($page), $content);
  217. $oldpagemtime{$page}=time;
  218. $renderedfiles{$page}=htmlpage($page);
  219. }
  220. else {
  221. $links{$file}=[];
  222. writepage($file, $content);
  223. $oldpagemtime{$file}=time;
  224. $renderedfiles{$file}=$file;
  225. }
  226. }
  227. sub loadindex () {
  228. open (IN, "$srcdir/.index") || return;
  229. while (<IN>) {
  230. $_=possibly_foolish_untaint($_);
  231. chomp;
  232. my ($mtime, $file, $rendered, @links)=split(' ', $_);
  233. my $page=pagename($file);
  234. $pagesources{$page}=$file;
  235. $oldpagemtime{$page}=$mtime;
  236. $oldlinks{$page}=[@links];
  237. $links{$page}=[@links];
  238. $renderedfiles{$page}=$rendered;
  239. }
  240. close IN;
  241. }
  242. sub saveindex () {
  243. open (OUT, ">$srcdir/.index") || error("cannot write to .index: $!");
  244. foreach my $page (keys %oldpagemtime) {
  245. print OUT "$oldpagemtime{$page} $pagesources{$page} $renderedfiles{$page} ".
  246. join(" ", @{$links{$page}})."\n"
  247. if $oldpagemtime{$page};
  248. }
  249. close OUT;
  250. }
  251. sub prune ($) {
  252. my $file=shift;
  253. unlink($file);
  254. my $dir=dirname($file);
  255. while (rmdir($dir)) {
  256. $dir=dirname($dir);
  257. }
  258. }
  259. sub refresh () {
  260. # Find existing pages.
  261. my %exists;
  262. my @files;
  263. find({
  264. no_chdir => 1,
  265. wanted => sub {
  266. if (/\/\.svn\//) {
  267. $File::Find::prune=1;
  268. }
  269. elsif (! -d $_ && ! /\.html$/ && ! /\/\./) {
  270. my ($f)=/(^[-A-Za-z0-9_.:\/+]+$)/; # untaint
  271. if (! defined $f) {
  272. warn("skipping bad filename $_\n");
  273. }
  274. else {
  275. $f=~s/^\Q$srcdir\E\/?//;
  276. push @files, $f;
  277. $exists{pagename($f)}=1;
  278. }
  279. }
  280. },
  281. }, $srcdir);
  282. my %rendered;
  283. # check for added or removed pages
  284. my @add;
  285. foreach my $file (@files) {
  286. my $page=pagename($file);
  287. if (! $oldpagemtime{$page}) {
  288. debug("new page $page");
  289. push @add, $file;
  290. $links{$page}=[];
  291. $pagesources{$page}=$file;
  292. }
  293. }
  294. my @del;
  295. foreach my $page (keys %oldpagemtime) {
  296. if (! $exists{$page}) {
  297. debug("removing old page $page");
  298. push @del, $renderedfiles{$page};
  299. prune($destdir."/".$renderedfiles{$page});
  300. delete $renderedfiles{$page};
  301. $oldpagemtime{$page}=0;
  302. delete $pagesources{$page};
  303. }
  304. }
  305. # render any updated files
  306. foreach my $file (@files) {
  307. my $page=pagename($file);
  308. if (! exists $oldpagemtime{$page} ||
  309. mtime("$srcdir/$file") > $oldpagemtime{$page}) {
  310. debug("rendering changed file $file");
  311. render($file);
  312. $rendered{$file}=1;
  313. }
  314. }
  315. # if any files were added or removed, check to see if each page
  316. # needs an update due to linking to them
  317. # TODO: inefficient; pages may get rendered above and again here;
  318. # problem is the bestlink may have changed and we won't know until
  319. # now
  320. if (@add || @del) {
  321. FILE: foreach my $file (@files) {
  322. my $page=pagename($file);
  323. foreach my $f (@add, @del) {
  324. my $p=pagename($f);
  325. foreach my $link (@{$links{$page}}) {
  326. if (bestlink($page, $link) eq $p) {
  327. debug("rendering $file, which links to $p");
  328. render($file);
  329. $rendered{$file}=1;
  330. next FILE;
  331. }
  332. }
  333. }
  334. }
  335. }
  336. # handle linkbacks; if a page has added/removed links, update the
  337. # pages it links to
  338. # TODO: inefficient; pages may get rendered above and again here;
  339. # problem is the linkbacks could be wrong in the first pass render
  340. # above
  341. if (%rendered) {
  342. my %linkchanged;
  343. foreach my $file (keys %rendered, @del) {
  344. my $page=pagename($file);
  345. if (exists $links{$page}) {
  346. foreach my $link (@{$links{$page}}) {
  347. $link=bestlink($page, $link);
  348. if (length $link &&
  349. ! exists $oldlinks{$page} ||
  350. ! grep { $_ eq $link } @{$oldlinks{$page}}) {
  351. $linkchanged{$link}=1;
  352. }
  353. }
  354. }
  355. if (exists $oldlinks{$page}) {
  356. foreach my $link (@{$oldlinks{$page}}) {
  357. $link=bestlink($page, $link);
  358. if (length $link &&
  359. ! exists $links{$page} ||
  360. ! grep { $_ eq $link } @{$links{$page}}) {
  361. $linkchanged{$link}=1;
  362. }
  363. }
  364. }
  365. }
  366. foreach my $link (keys %linkchanged) {
  367. my $linkfile=$pagesources{$link};
  368. if (defined $linkfile) {
  369. debug("rendering $linkfile, to update its linkbacks");
  370. render($linkfile);
  371. }
  372. }
  373. }
  374. }
  375. # Generates a C wrapper program for running ikiwiki in a specific way.
  376. # The wrapper may be safely made suid.
  377. sub gen_wrapper ($$) {
  378. my ($offline, $rebuild)=@_;
  379. eval {use Cwd 'abs_path'};
  380. $srcdir=abs_path($srcdir);
  381. $destdir=abs_path($destdir);
  382. my $this=abs_path($0);
  383. if (! -x $this) {
  384. error("$this doesn't seem to be executable");
  385. }
  386. my $call=qq{"$this", "$this", "$srcdir", "$destdir", "--wikiname=$wikiname"};
  387. $call.=', "--verbose"' if $verbose;
  388. $call.=', "--rebuild"' if $rebuild;
  389. $call.=', "--offline"' if $offline;
  390. open(OUT, ">ikiwiki-wrap.c") || error("failed to write ikiwiki-wrap.c: $!");;
  391. print OUT <<"EOF";
  392. /* A suid wraper for ikiwiki */
  393. #include <stdio.h>
  394. #include <unistd.h>
  395. #include <stdlib.h>
  396. int main (void) {
  397. clearenv();
  398. execl($call, NULL);
  399. perror("failed to run $this");
  400. exit(1);
  401. }
  402. EOF
  403. close OUT;
  404. if (system("gcc", "ikiwiki-wrap.c", "-o", "ikiwiki-wrap") != 0) {
  405. error("failed to compile ikiwiki-wrap.c");
  406. }
  407. unlink("ikiwiki-wrap.c");
  408. print "successfully generated ikiwiki-wrap\n";
  409. exit 0;
  410. }
  411. sub update () {
  412. if (-d "$srcdir/.svn") {
  413. if (system("svn", "update", "--quiet", $srcdir) != 0) {
  414. warn("svn update failed\n");
  415. }
  416. }
  417. }
  418. my $rebuild=0;
  419. my $offline=0;
  420. my $gen_wrapper=0;
  421. if (grep /^-/, @ARGV) {
  422. eval {use Getopt::Long};
  423. GetOptions(
  424. "wikiname=s" => \$wikiname,
  425. "verbose|v" => \$verbose,
  426. "rebuild" => \$rebuild,
  427. "gen-wrapper" => \$gen_wrapper,
  428. "offline" => \$offline,
  429. ) || usage();
  430. }
  431. usage() unless @ARGV == 2;
  432. ($srcdir) = possibly_foolish_untaint(shift);
  433. ($destdir) = possibly_foolish_untaint(shift);
  434. gen_wrapper($offline, $rebuild) if $gen_wrapper;
  435. memoize('pagename');
  436. memoize('bestlink');
  437. update() unless $offline;
  438. loadindex() unless $rebuild;
  439. refresh();
  440. saveindex();