summaryrefslogtreecommitdiff
path: root/doc
diff options
context:
space:
mode:
Diffstat (limited to 'doc')
-rw-r--r--doc/bugs/CGI__44___formbuilder__44___non-existent_field_address.mdwn59
-rw-r--r--doc/bugs/Dupe_entry_in_Bundle::IkiWiki::Extras.pm.mdwn5
-rw-r--r--doc/bugs/Inline_doesn__39__t_wikilink_to_pages.mdwn100
-rw-r--r--doc/bugs/user_links_on_recentchanges_pages_problem.mdwn12
-rw-r--r--doc/forum/Darcs_as_the_RCS___63__.mdwn2
-rw-r--r--doc/forum/How_does_ikiwiki_remember_times__63__.mdwn89
-rw-r--r--doc/forum/How_to_fix___34__does_not_map_to_Unicode__34___errors__63__.mdwn20
-rw-r--r--doc/forum/managing_todo_lists.mdwn3
-rw-r--r--doc/git.mdwn1
-rw-r--r--doc/ikiwiki-makerepo.mdwn5
-rw-r--r--doc/ikiwiki/directive/inline.mdwn3
-rw-r--r--doc/ikiwikiusers.mdwn2
-rw-r--r--doc/index/openid/discussion.mdwn50
-rw-r--r--doc/news/version_3.08.mdwn6
-rw-r--r--doc/news/version_3.10.mdwn15
-rw-r--r--doc/plugins/anonok.mdwn4
-rw-r--r--doc/plugins/comments/discussion.mdwn2
-rw-r--r--doc/plugins/contrib/po.mdwn51
-rw-r--r--doc/plugins/contrib/unixauth.mdwn20
-rw-r--r--doc/rcs/darcs.mdwn15
-rw-r--r--doc/rcs/details.mdwn100
-rw-r--r--doc/rcs/mercurial.mdwn6
-rw-r--r--doc/sandbox.mdwn4
-rw-r--r--doc/sandbox/I_hate_making_new_blog_entries_-_sometimes.wiki25
-rw-r--r--doc/style.css6
-rw-r--r--doc/tips/Importing_posts_from_Wordpress.mdwn10
-rw-r--r--doc/tips/Importing_posts_from_Wordpress/discussion.mdwn40
-rw-r--r--doc/tips/add_chatterbox_to_blog.mdwn21
-rw-r--r--doc/tips/add_chatterbox_to_blog/discussion.mdwn25
-rw-r--r--doc/tips/embedding_content.mdwn2
-rw-r--r--doc/tips/github.mdwn55
-rw-r--r--doc/tips/importing_posts_from_wordpress/ikiwiki-wordpress-import.mdwn113
-rw-r--r--doc/todo/darcs.mdwn534
-rw-r--r--doc/todo/hidden_links__47__tags.mdwn13
-rw-r--r--doc/todo/inline_postform_autotitles.mdwn (renamed from doc/todo/inline_autotitles.mdwn)5
-rw-r--r--doc/todo/need_global_renamepage_hook.mdwn2
-rw-r--r--doc/todo/rewrite_ikiwiki_in_haskell.mdwn68
-rw-r--r--doc/todo/rewrite_ikiwiki_in_haskell/discussion.mdwn14
-rw-r--r--doc/usage.mdwn4
-rw-r--r--doc/users/jon.mdwn8
-rw-r--r--doc/users/seanh.mdwn1
-rw-r--r--doc/wikitemplates.mdwn1
42 files changed, 840 insertions, 681 deletions
diff --git a/doc/bugs/CGI__44___formbuilder__44___non-existent_field_address.mdwn b/doc/bugs/CGI__44___formbuilder__44___non-existent_field_address.mdwn
new file mode 100644
index 000000000..ef74deb91
--- /dev/null
+++ b/doc/bugs/CGI__44___formbuilder__44___non-existent_field_address.mdwn
@@ -0,0 +1,59 @@
+Error received when clicking on the "edit" link:
+
+> `Error: [CGI::FormBuilder::AUTOLOAD] Fatal: Attempt to address
+> non-existent field 'text' by name at
+> /home/tealart/bin/share/perl/5.8.4/IkiWiki/CGI.pm line 112`
+
+Error received when following a "Create New Page" (eg. ?) link:
+
+> `Error: [CGI::FormBuilder::AUTOLOAD] Fatal: Attempt to address
+> non-existent field 'param' by name at
+> /home/tealart/bin/share/perl/5.8.4/IkiWiki/Plugin/editpage.pm line 122`
+
+I could probably find several other flavors of this error if I went
+looking, but I trust you get the idea.
+
+The CGI starts to render (this isn't the "you forgot to set the
+permissions/turn on the CGI" error) and then fails.
+
+Further details:
+
+- Running on shared hosting (dreamhost; but everything compiles,
+ dependencies installed, the site generates perfectly, other CGIs
+ work, the file permissions work).
+
+- It's running perl 5.8.4, but I did upgrade gettext to 0.17
+
+- the server is running gcc v3.3.5 (at this point, this is the main
+ difference between the working system and my box.)
+
+- I've removed the locale declarations from both the config file and
+ the environment variable.
+
+- I've also modified the page template and have my templates in a non
+ standard location. The wiki compiles fine, with the template, but
+ might this be an issue? The CGI script doesn't (seem) to load under
+ the new template, but I'm not sure how to address this issue.
+
+- All of the required/suggested module dependencies are installed
+ (finally) to the latest version including (relevantly)
+ CGI::FormBuilder 3.0501.
+
+- I'm running ikiwiki v3.08. Did I mention that it works perfectly in
+ nearly every other way that I've managed to test thusfar?
+
+----
+
+> I suspect that your perl is too old and is incompatible with the version of CGI::FormBuilder you have installed.
+>
+> Is so, it seems likely that the same error message can be reproduced by running a simple command like this at the command line:
+>
+> perl -e 'use warnings; use strict; use CGI::FormBuilder; my $form=CGI::FormBuilder->new; $form->text("boo")'
+>
+> --[[Joey]]
+
+> > nope, that command produces no output. :/
+> >
+> > I considered downgrading CGI::FormBuilder but I saw evidence of previous versions being incompatible with ikiwiki so I decided against that.
+> >
+> > -- [[tychoish]]
diff --git a/doc/bugs/Dupe_entry_in_Bundle::IkiWiki::Extras.pm.mdwn b/doc/bugs/Dupe_entry_in_Bundle::IkiWiki::Extras.pm.mdwn
new file mode 100644
index 000000000..236112786
--- /dev/null
+++ b/doc/bugs/Dupe_entry_in_Bundle::IkiWiki::Extras.pm.mdwn
@@ -0,0 +1,5 @@
+Authen::Passphrase
+
+is entered twice in the .pm file.
+
+[[done]]
diff --git a/doc/bugs/Inline_doesn__39__t_wikilink_to_pages.mdwn b/doc/bugs/Inline_doesn__39__t_wikilink_to_pages.mdwn
new file mode 100644
index 000000000..32f9f1245
--- /dev/null
+++ b/doc/bugs/Inline_doesn__39__t_wikilink_to_pages.mdwn
@@ -0,0 +1,100 @@
+It seems that the [[ikiwiki/directive/inline]] directive doesn't generate wikilinks to the pages it includes. For example I would expect the following to inline all bugs inside this bug report:
+
+\[[!inline pages="bugs/* and !*/discussion and backlink(bugs)" feeds=no postform=no archive=yes show="10"]]
+
+But here it is:
+
+[[!inline pages="bugs/* and !*/discussion and backlink(bugs)" feeds=no postform=no archive=yes show="10"]]
+
+and note that it only included the 'normal' wikilinks (and also note that this page is not marked done even though the done page is inlined).
+One might also wonder if inline would make this page link to any internal links on those inlined pages too, but I think
+that would be overkill.
+
+I'm not even really sure if this is a bug or defined behaviour, but I thought it might work and it didn't. Regardless,
+the correct behaviour, whichever is decided, should be documented. -- [[Will]]
+
+It appears that [[ikiwiki/directive/map]] also doesn't wikilink to the pages it links. Perhaps in each of these
+cases there should be another parameter to the directive that allows linking to switched on. Just switching
+it on universally at this point might break a number of people's pagespecs. -- [[Will]]
+
+> There's a simple reason why these directives don't generate a record of a
+> wikilink between them and the pages they include: Semantically, inlining
+> a page is not the same as writing a link to it. Nor is generating a map that
+> lists a page the same as linking to it. I don't think this is a bug.
+> --[[Joey]]
+
+>> Fair enough. I guess we can mark this as [[done]] then.
+>>
+>> Just a bit of background on where I was going here... I was looking for
+>> a simpler way of attacking [[todo/tracking_bugs_with_dependencies]].
+>> In particular, rather than introducing changes to the pagespec definition,
+>> I wondered if you could use wiki pages as the defined pagespec and
+>> introduce a 'match_mutual' function which matches whenever two pages
+>> link to the same third page, then you don't need to alter the pagespec
+>> handling code.
+>>
+>> But that requires being able use use a pagespec to decide what pages
+>> are linked to. e.g. I want to make an 'openbugs' page that links to all
+>> open bugs. Then I could make a 'readybugs' page that links to
+>> `backlink(openbugs) and !mutualLink(openbugs)`. That is, all bugs
+>> that are open and do not themselves link to an open bug.
+>>
+>> The problem with all this is that it introduces an ordering dependency,
+>> as I noted below. I think the original proposal is better, because it
+>> handles that ordering dependency in the definition of the pagespecs.
+>> --[[Will]]
+
+Here is a patch to make map link to its linked pages (when passed `link="yes"`). It is a bit problematic in that it uses a pagespec
+to decide what to link to (which is why I wanted it). However, at the time the pagespec is used the links
+for each page haven't finished being calculated (we're using the pagespec to figure out those links,
+remember). This means that some pagespec match functions may not work correctly. Sigh.
+It would be nice to find a topological ordering of the pages and scan them in that order
+so that everything we need is found before we need it, but this patch doesn't do that (it would be
+complex).
+
+If you just use simple pagespecs you'll be fine. Unfortunately I really wanted this for more complex
+pagespecs. -- [[Will]]
+
+ diff --git a/IkiWiki/Plugin/map.pm b/IkiWiki/Plugin/map.pm
+ index 3284931..57c0a7a 100644
+ --- a/IkiWiki/Plugin/map.pm
+ +++ b/IkiWiki/Plugin/map.pm
+ @@ -13,7 +13,7 @@ use IkiWiki 3.00;
+
+ sub import {
+ hook(type => "getsetup", id => "map", call => \&getsetup);
+ - hook(type => "preprocess", id => "map", call => \&preprocess);
+ + hook(type => "preprocess", id => "map", call => \&preprocess, scan => 1);
+ }
+
+ sub getsetup () {
+ @@ -27,7 +27,9 @@ sub getsetup () {
+ sub preprocess (@) {
+ my %params=@_;
+ $params{pages}="*" unless defined $params{pages};
+ -
+ +
+ + return if (!defined wantarray && !IkiWiki::yesno($params{link}));
+ +
+ my $common_prefix;
+
+ # Get all the items to map.
+ @@ -42,6 +44,9 @@ sub preprocess (@) {
+ else {
+ $mapitems{$page}='';
+ }
+ + if (!defined wantarray && IkiWiki::yesno($params{link})) {
+ + push @{$links{$params{page}}}, $page;
+ + }
+ # Check for a common prefix.
+ if (! defined $common_prefix) {
+ $common_prefix=$page;
+ @@ -62,6 +67,8 @@ sub preprocess (@) {
+ }
+ }
+
+ + return if ! defined wantarray;
+ +
+ # Common prefix should not be a page in the map.
+ while (defined $common_prefix && length $common_prefix &&
+ exists $mapitems{$common_prefix}) {
diff --git a/doc/bugs/user_links_on_recentchanges_pages_problem.mdwn b/doc/bugs/user_links_on_recentchanges_pages_problem.mdwn
new file mode 100644
index 000000000..d00f6815b
--- /dev/null
+++ b/doc/bugs/user_links_on_recentchanges_pages_problem.mdwn
@@ -0,0 +1,12 @@
+When I click on a linked username for commits on the recentchanges page (on
+the live ikiwiki.info) I get a link such as
+<http://ikiwiki.info/ikiwiki.cgi?page=users%2Fjoey&do=recentchanges_link>
+which returns something like
+
+ <a href="http://ikiwiki.info">ikiwiki</a>/ Error
+ <p class="error">Error: unknown do parameter
+
+-- [[Jon]]
+
+> That was fixed in 3.04 but ikiwiki.info had not been upgraded to it yet.
+> [[done]] --[[Joey]]
diff --git a/doc/forum/Darcs_as_the_RCS___63__.mdwn b/doc/forum/Darcs_as_the_RCS___63__.mdwn
index 2635690f7..9664240ee 100644
--- a/doc/forum/Darcs_as_the_RCS___63__.mdwn
+++ b/doc/forum/Darcs_as_the_RCS___63__.mdwn
@@ -9,3 +9,5 @@ What should I put in the configuration file to use darcs ?
> Darcs is not yet supported. It's being [[worked_on|todo/darcs]].
> > That's good news for me then ! Thank you.
+
+>>> Better news: It will be in version 2.10. --[[Joey]]
diff --git a/doc/forum/How_does_ikiwiki_remember_times__63__.mdwn b/doc/forum/How_does_ikiwiki_remember_times__63__.mdwn
new file mode 100644
index 000000000..5522cbf45
--- /dev/null
+++ b/doc/forum/How_does_ikiwiki_remember_times__63__.mdwn
@@ -0,0 +1,89 @@
+This is similar to the last post in this forum. I want to know exactly how ikiwiki remembers the times associated with pages, especially when using it for blogging, so I know whether I can trust it or not. From that last thread, I think what ikiwiki does is this:
+
+* The created time of a file is when that file was first committed into the versioning repository (in my case git)
+
+ > If `--getctime` it used, yes. In normal operation, when new files
+ > are added, ikiwiki sets the creation time to the ctime of the file
+ > on disk, rather than bothering to ask the VCS. --[[Joey]]
+
+* The modified time of a file is what that file was last updated in the repository
+
+ > Almost right, the modified time is actually taken from the
+ > modification time of the file in disk. --[[Joey]]
+
+And with a blog, by default, the posts are ordered by creation time, although an option can order them by modified time.
+
+Okay. So this should mean that the times are safe if, for example, I delete my working copy and then clone another one from the bare git repository, or otherwise mess up the creation times and mtimes stored as file metadata on the filesystem.
+
+Do I have it right?
+
+> Some VCS, like git, set the file mtimes to the current time
+> when making a new checkout, so they will be lost if you do that.
+> The creation times can be retrived using the `--getctime` option.
+> I suppose it might be nice if there were a `--getmtime` that pulled
+> true modification times out of the VCS, but I haven't found it a big
+> deal in practice for the last modification times to be updated to the
+> current time when rebuilding a wiki like this. --[[Joey]]
+>
+> > Thanks for the clarification. I ran some tests of my own to make sure I understand it right, and I'm satisfied
+> > that the order of posts in my blog can be retrieved from the VCS using the `--getctime` option, at least if I
+> > choose to order my posts by creation time rather than modification time. But I now know that I can't rely on
+> > page modification times in ikiwiki as these can be lost permanently.
+> >
+> > I would suggest that there should at least be a `--getmtime` option like you describe, and perhaps that
+> > `--getctime` and `--getmtime` be _on by default_. In my opinion the creation times and modification times of
+> > pages in ikiwiki are part of the user's content and are important to protect, because the user may be relying
+> > on them, especially if they use blogging or lists of recently modified pages, etc. Right now the modification
+> > times can be lost permanently.
+> >
+> > Is there a typo in the description of `--getctime` in the man page?
+> >
+> > > --getctime
+> > > Pull **last changed time** for each new page out of the revision
+> > > control system. This rarely used option provides a way to get
+> > > the real creation times of items in weblogs, such as when build‐
+> > > ing a wiki from a new Subversion checkout. It is unoptimised and
+> > > quite slow. It is best used with --rebuild, to force ikiwiki to
+> > > get the ctime for all pages.
+> >
+> > Surely it is not the _last changed time_ but the _first seen time_ of each page that is pulled out of the VCS?
+> > If the aim is to get the real creation times of items in weblogs, then the last times that the items were
+> > changed in the VCS is not going to help. -- [[seanh]]
+>>> Typo, fixed. --[[Joey]]
+
+> > > If you want to preserve the date of a page, the best way to do it is to
+> > > use [[ikiwiki/directive/meta]] date="foo". This will survive checkouts,
+> > > VCS migrations, etc. -- [[Jon]]
+> > >
+> > > > That's a good tip Jon. That would also survive renaming a page by renaming its mdwn file, which would
+> > > > normally lose the times also. (And in that case I think both times are irretrievable, even by
+> > > > `--getctime`). I might start using a simple script to make blog posts that creates a file for
+> > > > me, puts today's date in the file as a meta, and opens the file in my editor. -- [[seanh]]
+
+>>>>> I use a script that does that and also sets up templates and tags
+>>>>> for a new item:
+
+ #!/bin/sh
+ set -u
+ set -e
+
+ if [ $# -ne 1 ]; then
+ echo usage: $0 pagename >&2
+ exit 1
+ fi
+
+ pagename="$1"
+
+ if [ -e "$pagename" ]; then
+ echo error: "$pagename" exists >&2
+ exit 1
+ fi
+
+ date=$(date)
+ echo '\[[!template id=draft]]' >> "$pagename"
+ echo "\[[!meta date=\"$date\"]]" >> "$pagename"
+ echo "\[[!tag draft]]" >> "$pagename"
+ git add "$pagename"
+ $EDITOR "$pagename"
+
+>>>>> -- [[Jon]]
diff --git a/doc/forum/How_to_fix___34__does_not_map_to_Unicode__34___errors__63__.mdwn b/doc/forum/How_to_fix___34__does_not_map_to_Unicode__34___errors__63__.mdwn
new file mode 100644
index 000000000..0b3895357
--- /dev/null
+++ b/doc/forum/How_to_fix___34__does_not_map_to_Unicode__34___errors__63__.mdwn
@@ -0,0 +1,20 @@
+I'm getting a number of errors like this when running ikiwiki:
+
+ utf8 "\xA2" does not map to Unicode at /usr/local/share/perl/5.10.0/IkiWiki.pm line 739, <$in> chunk 1.
+
+I think it's because some of my files contain non-utf8, non-unicode, or somehow bad characters in them, probably fancy quotes and the like that have been copy-and-pasted from my web browser. The problem is that I have hundreds of files, I transferred them all over from pyblosxom to ikiwiki at once, and the error message doesn't tell me which file the error comes from. How can I fix this?
+
+Thanks
+-- seanh
+
+> Unfortunatly, these messages are logged by perl so there's no way to add
+> a filename to them.
+>
+> If you run the build in --verbose mode, you should see which page ikiwiki
+> is working on, and unless it inlines some other page, you can be pretty
+> sure that page contains invalid utf-8 if the message is then printed.
+>
+> Another option is to use the `isutf8` program from
+> moreutils](http://kitenet.net/~joey/code/moreutils/),
+> and run it on each file, it will tell you the line number
+> and character position that is invalid. --[[Joey]]
diff --git a/doc/forum/managing_todo_lists.mdwn b/doc/forum/managing_todo_lists.mdwn
index b4bbac255..0a69af805 100644
--- a/doc/forum/managing_todo_lists.mdwn
+++ b/doc/forum/managing_todo_lists.mdwn
@@ -39,3 +39,6 @@ sure how to handle embeds or challenges from the CGI such as a login challenge
>> thanks - I'll give it a look. I spent a few hours writing some javascript to manipulate a ul/li DOM tree in an outliner-fashion the other day. I might be able to join the puzzle pieces together sometime. [[Jon]]
a solution for this could be similar to a solution for [[todo/structured page data]], as todo lists are definitely a form of structured data. (in both cases, the page's current content is rendered into a html form, whose result is then saved as the page's new contents) --[[chrysn]]
+
+> Thanks for the link: yup, there's definitely some common ground there.
+> -- [[Jon]]
diff --git a/doc/git.mdwn b/doc/git.mdwn
index 8a89546cf..87b1aaae6 100644
--- a/doc/git.mdwn
+++ b/doc/git.mdwn
@@ -50,7 +50,6 @@ Some of the branches included in the main repository include:
instead of xhtml.
* `wikiwyg` adds [[todo/wikiwyg]] support. It is unmerged pending some
changes.
-* `darcs` is being used to add darcs support.
* `debian-stable` is used for updates to the old version included in
Debian's stable release, and `debian-testing` is used for updates to
Debian's testing release.
diff --git a/doc/ikiwiki-makerepo.mdwn b/doc/ikiwiki-makerepo.mdwn
index 9c532f201..13f88dc27 100644
--- a/doc/ikiwiki-makerepo.mdwn
+++ b/doc/ikiwiki-makerepo.mdwn
@@ -4,7 +4,7 @@ ikiwiki-makerepo - check an ikiwiki srcdir into revision control
# SYNOPSIS
-ikiwiki-makerepo svn|git|monotone srcdir repository
+ikiwiki-makerepo svn|git|monotone|darcs srcdir repository
ikiwiki-makerepo bzr|mercurial srcdir
@@ -18,6 +18,9 @@ Note that for mercurial and bzr, the srcdir is converted into a
repository. There is no need to have a separate repository with mercurial
or bzr.
+For darcs, the master repo's apply hook will be preconfigured to call a
+ikiwiki wrapper.
+
Note that for monotone, you are assumed to already have run "mtn genkey"
to generate a key.
diff --git a/doc/ikiwiki/directive/inline.mdwn b/doc/ikiwiki/directive/inline.mdwn
index f69d55de3..8afd65a05 100644
--- a/doc/ikiwiki/directive/inline.mdwn
+++ b/doc/ikiwiki/directive/inline.mdwn
@@ -79,7 +79,8 @@ Here are some less often needed parameters:
page. By default the `inlinepage` template is used, while
the `archivepage` template is used for archives. Set this parameter to
use some other, custom template, such as the `titlepage` template that
- only shows post titles. Note that you should still set `archive=yes` if
+ only shows post titles or the `microblog` template, optimised for
+ microblogging. Note that you should still set `archive=yes` if
your custom template does not include the page content.
* `raw` - Rather than the default behavior of creating a blog,
if raw is set to "yes", the page will be included raw, without additional
diff --git a/doc/ikiwikiusers.mdwn b/doc/ikiwikiusers.mdwn
index 1a24ff322..94ae75046 100644
--- a/doc/ikiwikiusers.mdwn
+++ b/doc/ikiwikiusers.mdwn
@@ -41,7 +41,7 @@ Projects & Organizations
* The [Fortran Wiki](http://fortranwiki.org/)
* [Monotone](http://monotone.ca/wiki/FrontPage/)
* The support pages for [Trinity Centre for High Performance Computing](http://www.tchpc.tcd.ie/support/)
-* [St Hugh of Lincoln Primary School in Surrey](http://hugh.vm.bytemark.co.uk/)
+* [St Hugh of Lincoln Catholic Primary School in Surrey](http://www.sthugh-of-lincoln.surrey.sch.uk/)
* [Pigro Network](http://www.pigro.net) is running a hg based ikiwiki. (And provides ikiwiki hosting for $10/m.)
* [Cosin Homepage](http://cosin.ch) uses an Ikiwiki with a subversion repository.
* [Bosco Free Orienteering Software](http://bosco.durcheinandertal.ch)
diff --git a/doc/index/openid/discussion.mdwn b/doc/index/openid/discussion.mdwn
index 43575d4be..4a50fd9dd 100644
--- a/doc/index/openid/discussion.mdwn
+++ b/doc/index/openid/discussion.mdwn
@@ -10,3 +10,53 @@ Hi, there's no return_to from a designated OpenID server page, specs requires (I
> * How can one reproduce the bug?
>
> PS, please file bugs under [[bugs]] in future. --[[Joey]]
+
+>> Oops, my bad, didn't know that existed at the time I wrote this.
+>>
+>> What happened is that the process wouldn't complete, therefore I couldn't login with my OpenID.
+>>
+>> reproducibility: every time
+>>
+>> Should probably move this page, eh? ;)
+>> I'd do that, but I dunno know other than using the SCM backend in question....
+
+Here's some actual output (with my OpenID URL stripped out):
+
+do=postsignin&oic.time=1238224497-1450566d93097caa707f&openid.assoc_handle=%7BHMAC-SHA1%7D%7B49cdce76%7D%7BBhuXXw%3D%3D%7D&openid.identity=|<==== MY OPENID URL GOES HERE ====>|&openid.mode=id_res&openid.op_endpoint=http%3A%2F%2Fwww.myopenid.com%2Fserver&openid.response_nonce=2009-03-28T07%3A15%3A02ZDUFmG3&openid.return_to=http%3A%2F%2Fsimonraven.kisikew.org%2Fbin%2Fikiwiki.cgi%3Fdo%3Dpostsignin%26oic.time%3D1238224497-1450566d93097caa707f&openid.sig=E51Xh6Gnjku%2B0se57qCyhHbT5QY%3D&openid.signed=assoc_handle%2Cidentity%2Cmode%2Cop_endpoint%2Cresponse_nonce%2Creturn_to%2Csigned
+
+The `return_to` arg should NOT be `signed`, it should be the originating URL where you initially logged in.
+
+>> Yes, exactly. That's also my understanding of the spec.
+
+> I think you're confusing 'openid.return_to' with 'return_to'. The
+> former is present above, and is, indeed, the originating url, the latter
+> is part of the *value* of the 'openid.signed' parameter generated by myopenid.com. --[[Joey]]
+
+Also, I dunno what the assoc_handle is doing spitting out an arg like `{HMAC-SHA1}{49cdce76}{BhuXXw%3D%3D}` it should be processed further. I have the needed perl packages installed (latest for Lenny). Hrm, would endianness matter?
+
+> Again, this value is created by the openid server, not by ikiwiki.
+> I see the same HMAC-SHA1 when using myopenid, and completly different
+> things for other openid servers. (Ie, when using livejournal as an openid server,
+> `openid.assoc_handle=1239305290:STLS.QiU6zTZ6w2bM3ttRkdaa:e68f91b751`)
+
+>> OK, I wasn't too sure about that, seemed bogus or somehow wrong or in error, like it wasn't actually being `completed`.
+
+> I'm fairly sure that is all a red herring.
+>
+> So, when I was talking about reproducing the bug, I was thinking perhaps you could tell me what openid server you're using,
+> etc, so I can actually see the bug with my own eyes.
+
+>> myopenid.com, with the CNAME option turned on.
+
+> The sanitised url parameters you've provided are not generated by ikiwiki at all.
+> They don't even seem to be generated by the underlying [[!cpan Net::OpenID]] library.
+
+>> That was a server log entry with date/host/time stripped, and my URL also stripped. Everything else is as was in the log. I installed the Debian packages in Lenny, both server and consumer OpenID Perl packages.
+
+> I'm pretty sure that what you're showing me is the url myopenid redirects
+> the browser to after successfully signing in. At that point, ikiwiki
+> should complete the signin. What fails at this point? How can I reproduce this failure? --[[Joey]]
+
+I'll try it again myself. I had tried it oh probably 6 times before I finally gave up on it. Maybe I'm getting rusty and I'm just PEBKACing all over the place. :P
+
+Also, to address the point about this discussion being in the wrong area (not under bugs), should I move it, or will you? I don't mind doing it, if you can't.
diff --git a/doc/news/version_3.08.mdwn b/doc/news/version_3.08.mdwn
deleted file mode 100644
index 06d795c23..000000000
--- a/doc/news/version_3.08.mdwn
+++ /dev/null
@@ -1,6 +0,0 @@
-ikiwiki 3.08 released with [[!toggle text="these changes"]]
-[[!toggleable text="""
- * git: Fix utf-8 encoding of author names.
- * git: Manually decode git output from utf-8, avoids
- warning messages on invalidly encoded output.
- * Fix bug that caused weird things to appear as page types."""]] \ No newline at end of file
diff --git a/doc/news/version_3.10.mdwn b/doc/news/version_3.10.mdwn
new file mode 100644
index 000000000..9d9c33c58
--- /dev/null
+++ b/doc/news/version_3.10.mdwn
@@ -0,0 +1,15 @@
+ikiwiki 3.10 released with [[!toggle text="these changes"]]
+[[!toggleable text="""
+ * darcs: Finally added support for this VCS, thanks to many
+ contributors:
+ - Thomas Schwinge wrote the original file, implementing only rcs\_commit.
+ - Benjamin A'Lee contributed an alternative implementation.
+ - Tuomo Valkonen contributed rcs\_getctime and stub rcs\_recentchanges.
+ - Simon Michael contributed multiple changes.
+ - Petr Ročkai fixed rcs\_recentchanges.
+ - Sven M. Hallberg merged the above and added missing features.
+ * Add missing newline to Confirm Password prompt.
+ * Add missing permalink support to archivepage and titlepage templates.
+ * debian/control: Wrap fields.
+ * inline: Add author info to archive display.
+ * Add a microblog template that is useful for inlining microblogging posts."""]] \ No newline at end of file
diff --git a/doc/plugins/anonok.mdwn b/doc/plugins/anonok.mdwn
index ab2f744e2..a3fec4d89 100644
--- a/doc/plugins/anonok.mdwn
+++ b/doc/plugins/anonok.mdwn
@@ -5,10 +5,10 @@ By default, anonymous users cannot edit the wiki. This plugin allows
anonymous web users, who have not signed in, to edit any page in the wiki
by default.
-The plugin also has a configuration setting, `anonok_pages`. This
+The plugin also has a configuration setting, `anonok_pagespec`. This
[[PageSpec]] can be used to allow anonymous editing of matching pages.
If you're using the [[comments]] plugin, you can allow anonymous comments
to be posted by setting:
- anonok_pages => "postcomment(*)"
+ anonok_pagespec => "postcomment(*)"
diff --git a/doc/plugins/comments/discussion.mdwn b/doc/plugins/comments/discussion.mdwn
index 2a87a3d93..3d7452b9a 100644
--- a/doc/plugins/comments/discussion.mdwn
+++ b/doc/plugins/comments/discussion.mdwn
@@ -60,7 +60,7 @@ spam problems. So, use `check_canedit` as at least a first-level check?
> have postcomment(blog/*) or something. (Perhaps instead of taking a glob, postcomment
> should take a pagespec, so you can have postcomment(link(tags/commentable))?)
>
-> This is why `anonok_pages => 'postcomment(*)'` and `locked_pages => '!postcomment(*)'`
+> This is why `anonok_pagespec => 'postcomment(*)'` and `locked_pages => '!postcomment(*)'`
> are necessary to allow anonymous and logged-in editing (respectively).
>
>> I changed that to move the flag out of the page name, and into a variable that the `match_postcomment`
diff --git a/doc/plugins/contrib/po.mdwn b/doc/plugins/contrib/po.mdwn
index a5e3375ce..61ec53ea8 100644
--- a/doc/plugins/contrib/po.mdwn
+++ b/doc/plugins/contrib/po.mdwn
@@ -330,12 +330,57 @@ daring a timid "please pull"... or rather, please review again :)
--[[intrigeri]]
> Ok, I've reviewed and merged into my own po branch. It's looking very
-> mergeable. I would still like to go over the `po.pm` code in detail and
-> review it, but it's very complex, and I'm happy with all the changes
-> outside `po.pm`.
+> mergeable.
>
> * Is it worth trying to fix compatability with `indexpages`?
+>>
+>> Supporting `usedirs` being enabled or disabled was already quite
+>> hard IIRC, so supporting all four combinations of `usedirs` and
+>> `indexpages` settings will probably be painful. I propose we forget
+>> about it until someone reports he/she badly needs it, and then
+>> we'll see what can be done.
+>>
> * Would it make sense to go ahead and modify `page.tmpl` to use
> OTHERLANGUAGES and PERCENTTRANSLATED, instead of documenting how to modify it?
+>>
+>> Done in my branch.
+>>
+> * Would it be better to disable po support for pages that use unsupported
+> or poorly-supported markup languages?
+>
+>> I prefer keeping it enabled, as:
+>>
+>> * most wiki markups "almost work"
+>> * when someone needs one of these to be fully supported, it's not
+>> that hard to add dedicated support for it to po4a; if it were
+>> disabled, I fear the ones who could do this would maybe think
+>> it's blandly impossible and give up.
+>>
+
+> * What's the reasoning behind checking that the link plugin
+> is enabled? AFAICS, the same code in the scan hook should
+> also work when other link plugins like camelcase are used.
+> * In `pagetemplate` there is a comment that claims the code
+> relies on `genpage`, but I don't see how it does; it seems
+> to always add a discussion link?
+> * Is there any real reason not to allow removing a translation?
+> I'm imagining a spammy translation, which an admin might not
+> be able to fix, but could remove.
+> * Re the meta title escaping issue worked around by `change`.
+> I suppose this does not only affect meta, but other things
+> at scan time too. Also, handling it only on rebuild feels
+> suspicious -- a refresh could involve changes to multiple
+> pages and trigger the same problem, I think. Also, exposing
+> this rebuild to the user seems really ugly, not confidence inducing.
+>
+> So I wonder if there's a better way. Such as making po, at scan time,
+> re-run the scan hooks, passing them modified content (either converted
+> from po to mdwn or with the escaped stuff cheaply de-escaped). (Of
+> course the scan hook would need to avoid calling itself!)
>
+> (This doesn't need to block the merge, but I hope it can be addressed
+> eventually..)
+>
> --[[Joey]]
+>>
+>> --[[intrigeri]]
diff --git a/doc/plugins/contrib/unixauth.mdwn b/doc/plugins/contrib/unixauth.mdwn
index 76a847744..6108ebfae 100644
--- a/doc/plugins/contrib/unixauth.mdwn
+++ b/doc/plugins/contrib/unixauth.mdwn
@@ -14,25 +14,9 @@ Config variables that affect the behavior of `unixauth`:
__Security__: [As with passwordauth](/security/#index14h2), be wary of sending usernames and passwords in cleartext. Unlike passwordauth, sniffing `unixauth` credentials can get an attacker much further than mere wiki access. Therefore, this plugin defaults to not even _displaying_ the login form fields unless we're running under SSL. Nobody should be able to do anything remotely dumb until the admin has done at least a little thinking. After that, dumb things are always possible. ;-)
-`unixauth` tests for the presence of the `HTTPS` environment variable. `Wrapper.pm` needs to be tweaked to pass it through; without that, the plugin fails closed.
+`unixauth` needs the `HTTPS` environment variable, available in ikiwiki 2.67 or later (fixed in #[502047](http://bugs.debian.org/502047)), without which it fails closed.
-[[!toggle id="diff" text="Wrapper.pm.diff"]]
-
-[[!toggleable id="diff" text="""
-
- --- Wrapper.pm.orig 2008-07-29 00:09:10.000000000 -0400
- +++ Wrapper.pm
- @@ -28,7 +28,7 @@ sub gen_wrapper () {
- my @envsave;
- push @envsave, qw{REMOTE_ADDR QUERY_STRING REQUEST_METHOD REQUEST_URI
- CONTENT_TYPE CONTENT_LENGTH GATEWAY_INTERFACE
- - HTTP_COOKIE REMOTE_USER} if $config{cgi};
- + HTTP_COOKIE REMOTE_USER HTTPS} if $config{cgi};
- my $envsave="";
- foreach my $var (@envsave) {
- $envsave.=<<"EOF"
-
-"""]]
+The plugin has not been tested with newer versions of ikiwiki. [[schmonz]] hopes to have time to polish this plugin soon.
[[!toggle id="code" text="unixauth.pm"]]
diff --git a/doc/rcs/darcs.mdwn b/doc/rcs/darcs.mdwn
new file mode 100644
index 000000000..fbb9bcede
--- /dev/null
+++ b/doc/rcs/darcs.mdwn
@@ -0,0 +1,15 @@
+[Darcs](http://darcs.new) is a distributed revison control
+system. Ikiwiki supports storing a wiki in a
+Darcs repository.
+
+An Ikiwiki wrapper is run by the `posthook` to update a wiki whenever commits
+or remote pushes come in. When running as a [[cgi]] with Darcs, ikiwiki
+automatically commits edited pages, and uses the Darcs history to generate the
+[[RecentChanges]] page.
+
+Example for a `_darcs/prefs/defaults` file in `$SRCDIR`:
+
+ apply posthook /path/to/repository/_darcs/ikiwiki-wrapper
+ apply run-posthook
+
+See also [[todo/darcs|todo/darcs]]
diff --git a/doc/rcs/details.mdwn b/doc/rcs/details.mdwn
index 089221cab..6492cf38c 100644
--- a/doc/rcs/details.mdwn
+++ b/doc/rcs/details.mdwn
@@ -32,98 +32,20 @@ You browse and web-edit the wiki on W.
W "belongs" to ikiwiki and should not be edited directly.
-## [darcs](http://darcs.net/) (not yet included)
+## [[darcs]]
-Support for using darcs as a backend is being worked on by [Thomas
-Schwinge](mailto:tschwinge@gnu.org), although development is on hold curretly.
-There is a patch in [[todo/darcs]].
+Regarding the repository layout: There are two darcs repositories. One is the `srcdir`, the other we'll call `master`.
-### How will it work internally?
+* HTML is generated from `srcdir`.
+* CGI edits happen in `srcdir`.
+* The backend pulls updates from `master` into `srcdir`, i.e. darcs commits should happen to `master`.
+* `master` calls ikiwiki (through a wrapper) in its apply posthook, i.e. `master/_darcs/prefs/defaults` should look like this:
-``Master'' repository R1.
-
-RCS commits from the outside are installed into R1.
-
-HTML is generated from R1. HTML is automatically generated (by using a
-``post-hook'') each time a new change is installed into R1. It follows
-that rcs_update() is not needed.
-
-There is a working copy of R1: R2.
-
-CGI operates on R2. rcs_commit() will push from R2 to R1.
-
-You browse the wiki on R1 and web-edit it on R2. This means for example
-that R2 needs to be updated from R1 if you are going to web-edit a page,
-as the user otherwise might be irritated otherwise...
-
-How do changes get from R1 to R2? Currently only internally in
-rcs\_commit(). Is rcs\_prepedit() suitable?
-
-It follows that the HTML rendering and the CGI handling can be completely
-separated parts in ikiwiki.
-
-What repository should [[RecentChanges]] and History work on? R1?
-
-#### Rationale for doing it differently than in the Subversion case
-
-darcs is a distributed RCS, which means that every checkout of a
-repository is equal to the repository it was checked-out from. There is
-no forced hierarchy.
-
-R1 is nevertheless called the master repository. It's used for
-collecting all the changes and publishing them: on the one hand via the
-rendered HTML and on the other via the standard darcs RCS interface.
-
-R2, the repository the CGI operates on, is just a checkout of R1 and
-doesn't really differ from the other checkouts that people will branch
-off from R1.
-
-(To be continued.)
-
-#### Another possible approach
-
-Here's what I (tuomov) think, would be a “cleaner” approach:
-
- 1. Upon starting to edit, Ikiwiki gets a copy of the page, and `darcs changes --context`.
- This context _and_ the present version of the page are stored in as the “version” of the
- page in a hidden control of the HTML.
- Thus the HTML includes all that is needed to generate a patch wrt. to the state of the
- repository at the time the edit was started. This is of course all that darcs needs.
- 2. Once the user is done with editing, _Ikiwiki generates a patch bundle_ for darcs.
- This should be easy with existing `Text::Diff` or somesuch modules, as the Web edits
- only concern single files. The reason why the old version of the page is stored in
- the HTML (possibly compressed) is that the diff can be generated.
- 3. Now this patch bundle is applied with `darcs apply`, or sent by email for moderation…
- there are many possibilities.
-
-This approach avoids some of the problems of concurrent edits that the previous one may have,
-although there may be conflicts, which may or may not propagate to the displayed web page.
-(Unfortunately there is not an option to `darcs apply` to generate some sort of ‘confliction resolution
-bundle’.) Also, only one repository is needed, as it is never directly modified
-by Ikiwiki.
-
-This approach might be applicable to other distributed VCSs as well, although they're not as oriented
-towards transmitting changes with standalone patch bundles (often by email) as darcs is.
-
-> The mercurial plugin seems to just use one repo and edit it directly - is
-> there some reason that's okay there but not for darcs? I agree with tuomov
-> that having just the one repo would be preferable; the point of a dvcs is
-> that there's no difference between one repo and another. I've got a
-> darcs.pm based on mercurial.pm, that's almost usable... --bma
-
->> IMHO it comes down to whatever works well for a given RCS. Seems like
->> the darcs approach _could_ be done with most any distributed system, but
->> it might be overkill for some (or all?) While there is the incomplete darcs
->> plugin in [[todo/darcs]], if you submit one that's complete, I will
->> probably accept it into ikiwiki.. --[[Joey]]
-
->>> I'd like to help make a robust darcs (2) backend. I also think ikiwiki should use
->>> exactly one darcs repo. I think we can simplify and say conflicting web
->>> edits are not allowed, like most current wiki engines. I don't see that
->>> saving (so much) context in the html is necessary, then.
->>> bma, I would like to see your code. --[[Simon_Michael]]
->>> PS ah, there it is. Let's continue on the [[todo/darcs]] page.
+ apply posthook ikiwrap
+ apply run-posthook
+* The backend pushes CGI edits from `srcdir` back into `master` (triggering the apply hook).
+* The working copies in `srcdir` and `master` should *not* be touched by the user, only by the CGI or darcs, respectively.
## [[Git]]
@@ -319,6 +241,8 @@ please refer to [Emanuele](http://nerd.ocracy.org/em/)
## [[tla]]
+Nobody really understands how tla works. ;-)
+
## rcs
There is a patch that needs a bit of work linked to from [[todo/rcs]].
diff --git a/doc/rcs/mercurial.mdwn b/doc/rcs/mercurial.mdwn
index b4baf07f4..ebfc35202 100644
--- a/doc/rcs/mercurial.mdwn
+++ b/doc/rcs/mercurial.mdwn
@@ -10,9 +10,9 @@ commits edited pages, and uses the Mercurial history to generate the
Example for a `.hg/hgrc` file in `$SRCDIR`:
[hooks]
- post-commit = /home/abe/bin/rebuildwiki
- incoming = /home/abe/bin/rebuildwiki
+ post-commit = ikiwiki --setup /path/to/ikiwiki.setup --post-commit
+ incoming = ikiwiki --setup /path/to/ikiwiki.setup --post-commit
-Do not use `commit` or `precommit` hooks or ikiwiki will run into a dead lock when committing in `$SRCDIR`
+Do not use `commit` or `precommit` hooks or ikiwiki will run into a dead lock when committing in `$SRCDIR`. Also note that `--post-commit` and not `--refresh` must be used to avoid dead locking when editing the pages via CGI interface.
See also [[todo/mercurial|todo/mercurial]]
diff --git a/doc/sandbox.mdwn b/doc/sandbox.mdwn
index 1e5fba304..4b6b9f270 100644
--- a/doc/sandbox.mdwn
+++ b/doc/sandbox.mdwn
@@ -1,7 +1,6 @@
This is the [[SandBox]], a page anyone can edit to try out ikiwiki (version [[!version ]]).
----
-misc test
test more test
[[中文显示]]
@@ -97,3 +96,6 @@ Let's see what happens... ~~
testing
+
+--
+[[!toc levels=2]]
diff --git a/doc/sandbox/I_hate_making_new_blog_entries_-_sometimes.wiki b/doc/sandbox/I_hate_making_new_blog_entries_-_sometimes.wiki
new file mode 100644
index 000000000..54914bb0b
--- /dev/null
+++ b/doc/sandbox/I_hate_making_new_blog_entries_-_sometimes.wiki
@@ -0,0 +1,25 @@
+!!!Heading Three
+
+Itemized
+* list
+* of
+* things
+
+or not
+1. mom
+1. hi
+1. there
+
+
+
+!!!Heading Three
+
+Itemized
+* list
+* of
+* things
+
+or not
+1. mom
+1. hi
+1. there
diff --git a/doc/style.css b/doc/style.css
index 98a28f347..74d968ddf 100644
--- a/doc/style.css
+++ b/doc/style.css
@@ -373,11 +373,13 @@ span.color {
padding: 2px;
}
-.comment-header {
+.comment-header,
+.microblog-header {
font-style: italic;
margin-top: .3em;
}
-.comment .author {
+.comment .author,
+.microblog .author {
font-weight: bold;
}
.comment-subject {
diff --git a/doc/tips/Importing_posts_from_Wordpress.mdwn b/doc/tips/Importing_posts_from_Wordpress.mdwn
index 87ef12079..59330caa4 100644
--- a/doc/tips/Importing_posts_from_Wordpress.mdwn
+++ b/doc/tips/Importing_posts_from_Wordpress.mdwn
@@ -2,4 +2,12 @@ Use case: You want to move away from Wordpress to Ikiwiki as your blogging/websi
[This](http://git.chris-lamb.co.uk/?p=ikiwiki-wordpress-import.git) is a simple tool that generates [git-fast-import](http://www.kernel.org/pub/software/scm/git/docs/git-fast-import.html)-compatible data from a WordPress export XML file. It retains creation time of each post, so you can use Ikiwiki's <tt>--getctime</tt> to get the preserve creation times on checkout.
-WordPress categories are mapped onto Ikiwiki tags. The ability to import comments is planned. \ No newline at end of file
+WordPress categories are mapped onto Ikiwiki tags. The ability to import comments is planned.
+
+-----
+
+I include a modified version of this script. This version includes the ability to write \[[!tag foo]] directives, which the original intended, but didn't actually do.
+
+-- [[users/simonraven]]
+
+[[ikiwiki-wordpress-import]]
diff --git a/doc/tips/Importing_posts_from_Wordpress/discussion.mdwn b/doc/tips/Importing_posts_from_Wordpress/discussion.mdwn
index 3b328649e..55e04d9cb 100644
--- a/doc/tips/Importing_posts_from_Wordpress/discussion.mdwn
+++ b/doc/tips/Importing_posts_from_Wordpress/discussion.mdwn
@@ -2,3 +2,43 @@ When I attempt to use this script, I get the following error:
warning: Not updating refs/heads/master (new tip 26b1787fca04f2f9772b6854843fe99fe06e6088 does not contain fc0ad65d14d88fd27a6cee74c7cef3176f6900ec). I have git 1.5.6.5, any ideas?
Thanks!!
+
+-----
+
+### KeyError: 146
+
+I also get this error, here's the output (it seems to stem from an error in the python script):
+
+<pre>
+Traceback (most recent call last):
+ File "../ikiwiki-wordpress-import.py", line 74, in <module>
+ main(*sys.argv[1:])
+ File "../ikiwiki-wordpress-import.py", line 54, in main
+ data = content.encode('ascii', 'html_replace')
+ File "../ikiwiki-wordpress-import.py", line 30, in <lambda>
+ % htmlentitydefs.codepoint2name[ord(c)] for c in x.object[x.start:x.end]]), x.end))
+KeyError: 146
+warning: Not updating refs/heads/master (new tip 6dca6ac939e12966bd64ce8a822ef14fe60622b2 does not contain 60b798dbf92ec5ae92f18acac3075c4304aca120)
+git-fast-import statistics:
+</pre>
+
+etc.
+
+
+> Well, if this really is a script error, it's not really the script, but the wordpress XML dump, referring to a
+> possible malformed or invalid unicode character in the dump file. This is what I can gather from other scripts.
+> I'll be checking my dump file shortly.
+
+>> This is only part of the problem... I'm not exactly sure what's going on, and it's get late/early for me....
+
+>>> I used --force for fast-import, but then everything seems deleted, so you end up doing a reset, checkout, add, *then* commit.
+>>> Seems really odd. I edited the script however, maybe this is why... this is my changes:
+
+ -print "data %d" % len(data)
+ +print "data %d merge refs/heads/%s" % (len(data), branch)
+
+>>> That control character is a ^q^0 in emacs, see git fast-import --help for more info.
+>>> I'll be trying an import *without* that change, to see what happens.
+
+>>>> I still have to do the above to preserve the changes done by this script... (removed previous note).
+
diff --git a/doc/tips/add_chatterbox_to_blog.mdwn b/doc/tips/add_chatterbox_to_blog.mdwn
new file mode 100644
index 000000000..aa35b9331
--- /dev/null
+++ b/doc/tips/add_chatterbox_to_blog.mdwn
@@ -0,0 +1,21 @@
+If you use twitter or identi.ca, here's how to make a box
+on the side of your blog that holds your recent status updates
+from there, like I have on [my blog](http://kitenet.net/~joey/blog/)
+--[[Joey]]
+
+* Enable the [[plugins/aggregate]] plugin, and set up a cron
+ job for it.
+* At the top of your blog's page, add something like the following.
+ You'll want to change the urls of course. Be sure to also change
+ the inline directive's [[PageSpec]] to link to the location the
+ feed is aggregated to, which will be a subpage of the page
+ you put this on (blog in this example):
+
+ \[[!template id=note text="""
+ \[[!aggregate expirecount=5 name="dents" url="http://identi.ca/joeyh"
+ feedurl="http://identi.ca/api/statuses/user_timeline/joeyh.atom"]]
+ \[[!inline pages="internal(./blog/dents/*)" template=microblog
+ show=5 feeds=no]]
+ """]]
+
+Note: Works best with ikiwiki 3.10 or better.
diff --git a/doc/tips/add_chatterbox_to_blog/discussion.mdwn b/doc/tips/add_chatterbox_to_blog/discussion.mdwn
new file mode 100644
index 000000000..bf7c9b1c3
--- /dev/null
+++ b/doc/tips/add_chatterbox_to_blog/discussion.mdwn
@@ -0,0 +1,25 @@
+The example you gave looks a bit odd.
+
+This is what I did from your example (still trying to learn the more complex things ;).
+
+<pre>
+\[[!template id=note text="""
+\[[!aggregate expirecount=5 name=kijkaqawej url=http://identi.ca/kjikaqawej
+feedurl=http://identi.ca/api/statuses/user_timeline/kjikaqawej.atom]]
+\[[!inline pages="internal(kijkaqawej/*)" template=microblog show=5 feeds=no]] """]]
+</pre>
+
+mine, live, here: <http://simonraven.kisikew.org/blog/meta/microblog-feed/>
+
+I expected something like: sidebar, with a number, and displaying them in the sidebar, but they don't display (similar to what you have on your blog).
+
+On the [[/ikiwiki/pagespec]] page, it says "internal" pages aren't "first-class" wiki pages, so it's best not to directly display them, so how do you manage to display them? I'd like to display their name, and what they link to in the sidebar, or otherwise in the main body.
+
+> That's what the inline does, displays the internal pages.
+>
+> You need to fix your pagespec to refer to where the pages are aggregated
+> to, under the page that contains the aggregate directive. In your example,
+> it should be `internal(./blog/meta/microblog-feed/kijkaqawej/*)` --[[Joey]]
+
+>> Oooh, I see, it's referring to an absolute path (relative to the site), right?
+>> Thanks :).
diff --git a/doc/tips/embedding_content.mdwn b/doc/tips/embedding_content.mdwn
index 142acd16e..bfe458a84 100644
--- a/doc/tips/embedding_content.mdwn
+++ b/doc/tips/embedding_content.mdwn
@@ -15,7 +15,7 @@ you'd better trust that site. And if ikiwiki lets you enter such html, it
needs to trust you.)
The [[plugins/htmlscrubber]] offers a different way around this problem.
-You can configure it to skip scrubbing certian pages, so that content from
+You can configure it to skip scrubbing certain pages, so that content from
elsewhere can be embedded on those pages. Then use [[plugins/lockedit]]
to limit who can edit those unscrubbed pages.
diff --git a/doc/tips/github.mdwn b/doc/tips/github.mdwn
index 974d572f1..c3fdab734 100644
--- a/doc/tips/github.mdwn
+++ b/doc/tips/github.mdwn
@@ -11,51 +11,48 @@ two copies of the website in git (source and the compiled site), and all
historical versions too. So it could happen. If it does, you can pay github
for more space, or you can migrate your site elsewhere.
-## github setup
+## Github Setup
-* Go to [github](http://github.com/) and sign up for an account, if you
- haven't already.
-* Be sure to add your laptop's ssh key to it so you can push
- to github.
-* Create a repository on githib named `$YOU.github.com`, substituting your
- username. This repository will be used to publish your compiled website.
+* Go to [github](http://github.com/) and sign up for an account, if you haven't already.
+* Be sure to add your laptop's ssh key to it so you can push to github.
+* Create a repository on github named `$YOU.github.com`, substituting your
+ *username*. This repository will be used to publish your compiled website.
* Create a repository on github named `$YOU` (or anything else you like).
This repository will be used to publish the source of your website.
This is actually optional.
-## local setup
+## Local Setup
-* On your laptop, create two empty git repositories to correspond to the
- github repositories:
- YOU=# your github username here
- mkdir ~/$YOU.github.com
- cd ~/$YOU.github.com
- git init
- git remote add origin git@github.com:$YOU/$YOU.github.com.git
- mkdir ~/$YOU
- cd ~/$YOU
- git init
- git remote add origin git@github.com:$YOU/$YOU.git
+* On your laptop, create two empty git repositories to correspond to the github repositories: <br />
+ `YOU = your github username here` <br />
+ `mkdir ~/$YOU.github.com` <br />
+ `cd ~/$YOU.github.com` <br />
+ `git init` <br />
+ `git remote add origin git@github.com:$YOU/$YOU.github.com.git` <br />
+ `mkdir ~/$YOU` <br />
+ `cd ~/$YOU` <br />
+ `git init` <br />
+ `git remote add origin git@github.com:$YOU/$YOU.git` <br />
* Add some wiki pages, such as an `index.mdwn`, to `~/$YOU`, and check them
in and commit them to git. You need something to push to github. Run
`git push origin master` to push the source pages to github.
-## publishing to github
+## Publishing to Github
-* Now build your wiki with a command such as:
- ikiwiki ~/$YOU ~/$YOU.github.com --refresh
+* Now build your wiki with a command such as: <br />
+ `ikiwiki ~/$YOU ~/$YOU.github.com --refresh`
* Each time you build the wiki you will need to commit the changes
- to git, and push the compiled pages to github:
- cd ~/YOU.github.com
- git add .
- git commit -a -m update
- git push origin master
+ to git, and push the compiled pages to github: <br />
+ `cd ~/YOU.github.com` <br />
+ `git add .` <br />
+ `git commit -a -m update` <br />
+ `git push origin master` <br />
Your wiki will show up at `http://$YOU.github.com/` within ten
minutes after the first push, and changes you push to it from then on
-should show up immediatly.
+should show up immediately.
-## enhancements
+## Enhancements
You can follow the instructions in [[laptop_wiki_with_git]] to set up an
editable version of your wiki on your laptop. Then you can use the web
diff --git a/doc/tips/importing_posts_from_wordpress/ikiwiki-wordpress-import.mdwn b/doc/tips/importing_posts_from_wordpress/ikiwiki-wordpress-import.mdwn
new file mode 100644
index 000000000..5d7a266ec
--- /dev/null
+++ b/doc/tips/importing_posts_from_wordpress/ikiwiki-wordpress-import.mdwn
@@ -0,0 +1,113 @@
+[[!meta title="ikiwiki-wordpress-import"]]
+
+I modified the script a bit so categories and tags would actually show up in the output file.
+
+
+<pre>
+#!/usr/bin/env python
+
+"""
+ Purpose:
+ Wordpress-to-Ikiwiki import tool
+
+ Copyright:
+ Copyright (C) 2007 Chris Lamb <chris@chris-lamb.co.uk>
+
+ This program is free software: you can redistribute it and/or modify
+ it under the terms of the GNU General Public License as published by
+ the Free Software Foundation, either version 3 of the License, or
+ (at your option) any later version.
+
+ This program is distributed in the hope that it will be useful,
+ but WITHOUT ANY WARRANTY; without even the implied warranty of
+ MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
+ GNU General Public License for more details.
+
+ You should have received a copy of the GNU General Public License
+ along with this program. If not, see <http://www.gnu.org/licenses/>.
+
+ Usage: run --help as an argument with this script.
+
+ Notes:
+ I added some extra bits to include the [[!tag foo]] stuff in the post,
+ as it wasn't before, at all. I'll diff the versions out so you can see
+ the mess I made :).
+
+"""
+
+import os, sys
+import time
+import re
+
+from BeautifulSoup import BeautifulSoup
+
+import codecs, htmlentitydefs
+
+codecs.register_error('html_replace', lambda x: (''.join([u'&%s;' \
+ % htmlentitydefs.codepoint2name[ord(c)] for c in x.object[x.start:x.end]]), x.end))
+
+def main(name, email, subdir, branch='master'):
+ soup = BeautifulSoup(sys.stdin.read())
+
+ # Regular expression to match stub in URL.
+ stub_pattern = re.compile(r'.*\/(.+)\/$')
+
+ for x in soup.findAll('item'):
+ # Ignore draft posts
+ if x.find('wp:status').string != 'publish': continue
+
+ match = stub_pattern.match(x.guid.string)
+ if match:
+ stub = match.groups()[0]
+ else:
+ # Fall back to our own stubs
+ stub = re.sub(r'[^a-zA-Z0-9_]', '-', x.title.string).lower()
+
+ commit_msg = """Importing WordPress post "%s" [%s]""" % (x.title.string, x.guid.string)
+ timestamp = time.mktime(time.strptime(x.find('wp:post_date_gmt').string, "%Y-%m-%d %H:%M:%S"))
+
+ content = '[[!meta title="%s"]]\n\n' % (x.title.string.replace('"', r'\"'))
+ content += x.find('content:encoded').string.replace('\r\n', '\n')
+
+ # categories = x.findAll('category')
+ # categories = x.findAll({'category':True}, attrs={'domain':re.compile(('category|tag'))})
+ # categories = x.findAll({'category':True}, domain=["category", "tag"])
+ # categories = x.findAll({'category':True}, nicename=True)
+ """
+ We do it differently here because we have duplicates otherwise.
+ Take a look:
+ <category><![CDATA[Health]]></category>
+ <category domain="category" nicename="health"><![CDATA[Health]]></category>
+
+ If we do the what original did, we end up with all tags and cats doubled.
+ Therefore we only pick out nicename="foo". Our 'True' below is our 'foo'.
+ I'd much rather have the value of 'nicename', and tried, but my
+ python skillz are extremely limited....
+ """
+ categories = x.findAll('category', nicename=True)
+ if categories:
+ content += "\n"
+ for cat in categories:
+ # remove 'tags/' because we have a 'tagbase' set.
+ # your choice: 'tag', or 'taglink'
+ # content += "\n[[!tag %s]]" % (cat.string.replace(' ', '-'))
+ content += "\n[[!taglink %s]]" % (cat.string.replace(' ', '-'))
+ # print >>sys.stderr, cat.string.replace(' ', '-')
+
+ # moved this thing down
+ data = content.encode('ascii', 'html_replace')
+ print "commit refs/heads/%s" % branch
+ print "committer %s <%s> %d +0000" % (name, email, timestamp)
+ print "data %d" % len(commit_msg)
+ print commit_msg
+ print "M 644 inline %s" % os.path.join(subdir, "%s.mdwn" % stub)
+ print "data %d" % len(data)
+ print data
+
+if __name__ == "__main__":
+ if len(sys.argv) not in (4, 5):
+ print >>sys.stderr, "%s: usage: %s name email subdir [branch] < wordpress-export.xml | git-fast-import " % (sys.argv[0], sys.argv[0])
+ else:
+ main(*sys.argv[1:])
+
+</pre>
diff --git a/doc/todo/darcs.mdwn b/doc/todo/darcs.mdwn
index 882a41379..985ae5f8b 100644
--- a/doc/todo/darcs.mdwn
+++ b/doc/todo/darcs.mdwn
@@ -1,529 +1,21 @@
-Here's Thomas Schwinge unfinished darcs support for ikiwiki.
-
-(Finishing this has been suggested as a [[soc]] project.)
-
-> I haven't been working on this for months and also won't in the near
-> future. Feel free to use what I have done so
-> far and bring it into an usable state! Also, feel free to contact me
-> if there are questions.
-
--- [Thomas Schwinge](mailto:tschwinge@gnu.org)
-
-[[!toggle text="show"]]
-[[!toggleable text="""
- # Support for the darcs rcs, <URL:http://darcs.net/>.
- # Copyright (C) 2006 Thomas Schwinge <tschwinge@gnu.org>
- #
- # This program is free software; you can redistribute it and/or modify it
- # under the terms of the GNU General Public License as published by the
- # Free Software Foundation; either version 2 of the License, or (at your
- # option) any later version.
- #
- # This program is distributed in the hope that it will be useful, but
- # WITHOUT ANY WARRANTY; without even the implied warranty of
- # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
- # General Public License for more details.
- #
- # You should have received a copy of the GNU General Public License along
- # with this program; if not, write to the Free Software Foundation, Inc.,
- # 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA.
-
-
- # We're guaranteed to be the only instance of ikiwiki running at a given
- # time. It is essential that only ikiwiki is working on a particular
- # repository. That means one instance of ikiwiki and it also means that
- # you must not `darcs push' into this repository, as this might create
- # race conditions, as I understand it.
-
-
- use warnings;
- use strict;
- use IkiWiki;
-
- package IkiWiki;
-
-
- # Which darcs executable to use.
- my $darcs = ($ENV{DARCS} or 'darcs');
-
-
- # Internal functions.
-
- sub darcs_info ($$$) {
- my $field = shift;
- my $repodir = shift;
- my $file = shift; # Relative to the repodir.
-
- my $child = open(DARCS_CHANGES, "-|");
- if (! $child) {
- exec($darcs, 'changes', '--repo=' . $repodir, '--xml-output', $file) or
- error('failed to run `darcs changes\'');
- }
-
- # Brute force for now. :-/
- while (<DARCS_CHANGES>) {
- last if /^<\/created_as>$/;
- }
- ($_) = <DARCS_CHANGES> =~ /$field=\'([^\']+)/;
- $field eq 'hash' and s/\.gz//; # Strip away the `.gz' from `hash'es.
-
- close(DARCS_CHANGES) or error('`darcs changes\' exited ' . $?);
-
- return $_;
- }
-
-
- # Exported functions.
-
- sub rcs_update () {
- # Not needed.
- }
-
- sub rcs_prepedit ($) {
- # Prepares to edit a file under revision control. Returns a token that
- # must be passed to rcs_commit() when the file is to be commited. For us,
- # this token the hash value of the latest patch that modifies the file,
- # i.e. something like its current revision. If the file is not yet added
- # to the repository, we return TODO: the empty string.
-
- my $file = shift; # Relative to the repodir.
-
- my $hash = darcs_info('hash', $config{srcdir}, $file);
- return defined $hash ? $hash : "";
- }
-
- sub rcs_commit ($$$) {
- # Commit the page. Returns `undef' on success and a version of the page
- # with conflict markers on failure.
-
- my $file = shift; # Relative to the repodir.
- my $message = shift;
- my $rcstoken = shift;
-
- # Compute if the ``revision'' of $file changed.
- my $changed = darcs_info('hash', $config{srcdir}, $file) ne $rcstoken;
-
- # Yes, the following is a bit convoluted.
- if ($changed) {
- # TODO. Invent a better, non-conflicting name.
- rename("$config{srcdir}/$file", "$config{srcdir}/$file.save") or
- error("failed to rename $file to $file.save: $!");
-
- # Roll the repository back to $rcstoken.
-
- # TODO. Can we be sure that no changes are lost? I think that
- # we can, if we make sure that the `darcs push' below will always
- # succeed.
-
- # We need to revert everything as `darcs obliterate' might choke
- # otherwise.
- # TODO: `yes | ...' needed? Doesn't seem so.
- system($darcs, "revert", "--repodir=" . $config{srcdir}, "--all") and
- error("`darcs revert' failed");
- # Remove all patches starting at $rcstoken.
- # TODO. Something like `yes | darcs obliterate ...' seems to be needed.
- system($darcs, "obliterate", "--quiet", "--repodir" . $config{srcdir},
- "--match", "hash " . $rcstoken) and
- error("`darcs obliterate' failed");
- # Restore the $rcstoken one.
- system($darcs, "pull", "--quiet", "--repodir=" . $config{srcdir},
- "--match", "hash " . $rcstoken, "--all") and
- error("`darcs pull' failed");
-
- # We're back at $rcstoken. Re-install the modified file.
- rename("$config{srcdir}/$file.save", "$config{srcdir}/$file") or
- error("failed to rename $file.save to $file: $!");
- }
-
- # Record the changes.
- # TODO: What if $message is empty?
- writefile("$file.log", $config{srcdir}, $message);
- system($darcs, 'record', '--repodir=' . $config{srcdir}, '--all',
- '--logfile=' . "$config{srcdir}/$file.log",
- '--author=' . 'web commit <web-hurd@gnu.org>', $file) and
- error('`darcs record\' failed');
-
- # Update the repository by pulling from the default repository, which is
- # master repository.
- system($darcs, "pull", "--quiet", "--repodir=" . $config{srcdir},
- "--all") and error("`darcs pull' failed\n");
-
- # If this updating yields any conflicts, we'll record them now to resolve
- # them. If nothing is recorded, there are no conflicts.
- $rcstoken = darcs_info('hash', $config{srcdir}, $file);
- # TODO: Use only the first line here, i.e. only the patch name?
- writefile("$file.log", $config{srcdir}, 'resolve conflicts: ' . $message);
- system($darcs, 'record', '--repodir=' . $config{srcdir}, '--all',
- '--logfile=' . "$config{srcdir}/$file.log",
- '--author=' . 'web commit <web-hurd@gnu.org>', $file) and
- error('`darcs record\' failed');
- my $conflicts = darcs_info('hash', $config{srcdir}, $file) ne $rcstoken;
- unlink("$config{srcdir}/$file.log") or
- error("failed to remove `$file.log'");
-
- # Push the changes to the main repository.
- system($darcs, 'push', '--quiet', '--repodir=' . $config{srcdir}, '--all')
- and error('`darcs push\' failed');
- # TODO: darcs send?
-
- if ($conflicts) {
- my $document = readfile("$config{srcdir}/$file");
- # Try to leave everything in a consistent state.
- # TODO: `yes | ...' needed? Doesn't seem so.
- system($darcs, "revert", "--repodir=" . $config{srcdir}, "--all") and
- warn("`darcs revert' failed.\n");
- return $document;
- } else {
- return undef;
- }
- }
-
- sub rcs_add ($) {
- my $file = shift; # Relative to the repodir.
-
- # Intermediate directories will be added automagically.
- system($darcs, 'add', '--quiet', '--repodir=' . $config{srcdir},
- '--boring', $file) and error('`darcs add\' failed');
- }
-
- sub rcs_recentchanges ($) {
- warn('rcs_recentchanges() is not implemented');
- return 'rcs_recentchanges() is not implemented';
- }
-
- sub rcs_notify () {
- warn('rcs_notify() is not implemented');
- }
-
- sub rcs_getctime () {
- warn('rcs_getctime() is not implemented');
- }
-
- 1
-"""]]
-
-This is my ([bma](bma@bmalee.eu)) darcs.pm - it's messy (my Perl isn't up to much) but seems to work. It uses just one repo, like the mercurial plugin (unlike the above version, which AIUI uses two).
-
-`rcs_commit()` uses backticks instead of `system()`, to prevent darcs' output being sent to the browser and mucking with the HTTP headers (`darcs record` has no --quiet option). And `rcs_recentchanges()` uses regexes rather than parsing darcs' XML output.
-
-[[!toggle text="show" id="bma"]]
-[[!toggleable id="bma" text="""
-
- #!/usr/bin/perl
-
- use warnings;
- use strict;
- use IkiWiki;
- use Date::Parse;
- use open qw{:utf8 :std};
-
- package IkiWiki;
-
- sub rcs_update () {
- # Do nothing - there's nowhere to update *from*.
- }
-
- sub rcs_prepedit ($) {
- }
-
- sub rcs_commit ($$$;$$) {
- my ($file, $message, $rcstoken, $user, $ipaddr) = @_;
-
- # $user should probably be a name and an email address, by darcs
- # convention.
- if (defined $user) {
- $user = possibly_foolish_untaint($user);
- }
- elsif (defined $ipaddr) {
- $user = "Anonymous from $ipaddr";
- }
- else {
- $user = "Anonymous";
- }
-
- $message = possibly_foolish_untaint($message);
-
- # BUG: this outputs one line of text, and there's not a -q or --quiet
- # option. Redirecting output to /dev/null works, but I still get the
- # HTTP status and location headers displayed in the browser - is that
- # darcs' fault or ikiwiki's?
- # Doing it in backticks *works*, but I'm sure it could be done better.
- my @cmdline = ("darcs", "record", "--repodir", "$config{srcdir}",
- "-a", "-m", "$message", "--author", "$user", $file);
- `darcs record --repodir "$config{srcdir}" -a -m "$message" --author "$user" $file`; # Return value? Output? Who needs 'em?
- #if (system(@cmdline) != 0) {
- # warn "'@cmdline' failed: $!";
- #}
-
- return undef; # success
-
- sub rcs_add ($) {
- my ($file) = @_;
-
- my @cmdline = ("darcs", "add", "--repodir", "$config{srcdir}", "-a", "-q", "$file");
- if (system(@cmdline) != 0) {
- warn "'@cmdline' failed: $!";
- }
- }
-
- sub rcs_recentchanges ($) {
- # TODO: This is horrible code. It doesn't work perfectly, and uses regexes
- # rather than parsing Darcs' XML output.
- my $num=shift;
- my @ret;
-
- return unless -d "$config{srcdir}/_darcs";
-
- my $changelog = `darcs changes --xml --summary --repodir "$config{srcdir}"`;
- $changelog = join("", split(/\s*\n\s*/, $changelog));
- my @changes = split(/<\/patch>.*?<patch/m, $changelog);
-
-
- foreach my $change (@changes) {
- $change =~ m/hash='(.*?)'/;
- my $rev = $1;
- $change =~ m/author='(.*?)'/;
- my $user = $1."\n";
- my $committype = "web";
- if($user =~ m/&lt;/) {
- # Author fields generated by darcs include an email address: look for the "<".
- $committype = "darcs";
- use HTML::Entities;
- $user = decode_entities $user;
- }
- $change =~ m/local_date='(.*?)'/;
- my $when = $1;
- $when=time - str2time($when, 'UTC');
- $change =~ m/<name>(.*?)<\/name>/g;
- my @message = {line => $1};
- foreach my $match ($change =~ m/<comment>(.*?)<\/comment>/gm) {
- push @message, {line => $1};
- }
-
- my @pages;
- foreach my $match ($change =~ m/<.*?_(file|directory)>(.*?)(<(added|removed)_lines.*\/>)*<\/.*?_(file|directory)>/g) {
- # My perl-fu is weak. I'm probably going about this all wrong, anyway.
- push @pages, {page => pagename($match)} if ( -f $config{srcdir}."/".$match || -d $config{srcdir}."/".$match) and not $match =~ m/^$/;
- }
- push @ret, { rev => $rev,
- user => $user,
- committype => $committype,
- when => $when,
- message => [@message],
- pages => [@pages],
- }
- }
- return @ret;
- }
-
- sub rcs_notify () {
- # TODO
- }
-
- sub rcs_getctime ($) {
- error gettext("getctime not implemented");
- }
-
- 1
-
-
-
-"""]]
-
----
-
-Well, here's my version too. It only does getctime -- using a real XML parser, instead of regexp ugliness -- and maybe recentchanges, but that may be bitrotted, or maybe I never finished it, as I only need the getctime. As for actual commits, I have previously voiced my opinion, that this should be done by the plugin generating a patch bundle, and forwarding it to darcs in some way (`darcs apply` or even email to another host, possibly moderated), instead of the hacky direct modification of a working copy. It could also be faster to getctime in a batch. Just reading in all the changes the first time they're needed, might not be a big improvement in many cases, but if we got a batch request from ikiwiki, we could keep reaing the changes until all the files in this batch request have been met. --[[tuomov]]
-
-[[!toggle text="show" id="tuomov"]]
-[[!toggleable id="tuomov" text="""
-<pre>
-#!/usr/bin/perl
-# Stubs for no revision control.
-
-use warnings;
-use strict;
-use IkiWiki;
-
-package IkiWiki;
-
-sub rcs_update () {
-}
-
-sub rcs_prepedit ($) {
- return ""
-}
-
-sub rcs_commit ($$$) {
- return undef # success
-}
-
-sub rcs_add ($) {
-}
-
-sub rcs_recentchanges ($) {
- my $num=shift;
- my @ret;
-
- eval q{use Date::Parse};
- eval q{use XML::Simple};
-
- my $repodir=$config{srcdir};
-
- if (-d "$config{srcdir}/_darcs") {
- my $child = open(LOG, "-|");
- if (! $child) {
- exec("darcs", "changes", "--xml",
- "--repodir", "$repodir",
- "--last", "$num")
- || error("darcs changes failed to run");
- }
- my $data=<LOG>;
- close LOG;
-
- my $log = XMLin($data, ForceArray => 1);
-
- foreach my $patch ($log->{patch}) {
- my $date=$patch->{local_date};
- my $hash=$patch->{hash};
- my $when=concise(ago(time - str2time($date)));
- my @pages;
-
- my $child = open(SUMMARY, "-|");
- if (! $child) {
- exec("darcs", "annotate", "-s", "--xml",
- "--match", "hash: $hash",
- "--repodir", "$repodir")
- || error("darcs annotate failed to run");
- }
- my $data=<SUMMARY>;
- close SUMMARY;
-
- my $summary = XMLin("<lame>$data</lame>", ForceArray => 1);
-
- # TODO: find @pages
-
- push @ret, {
- #rev => $rev,
- user => $patch->{author},
- #committype => $committype,
- when => $when,
- #message => [@message],
- pages => [@pages],
- }; # if @pages;
- return @ret if @ret >= $num;
- }
- }
-
- return @ret;
-}
-
-sub rcs_notify () {
-}
-
-sub rcs_getctime ($) {
- my $file=shift;
-
- eval q{use Date::Parse};
- eval q{use XML::Simple};
- local $/=undef;
-
- # Sigh... doing things the hard way again
- my $repodir=$config{srcdir};
-
- my $filer=substr($file, length($repodir));
- $filer =~ s:^[/]+::;
-
- my $child = open(LOG, "-|");
- if (! $child) {
- exec("darcs", "changes", "--xml", "--reverse",
- "--repodir", "$repodir", "$filer")
- || error("darcs changes $filer failed to run");
- }
-
- my $data=<LOG>;
- close LOG;
-
- my $log = XMLin($data, ForceArray => 1);
-
- my $datestr=$log->{patch}[0]->{local_date};
-
- if (! defined $datestr) {
- warn "failed to get ctime for $filer";
- return 0;
- }
-
- my $date=str2time($datestr);
-
- debug("found ctime ".localtime($date)." for $file");
-
- return $date;
-}
-
-1
-</pre>
-"""]]
-
----
-
-I merged the two versions above and made some fixes; it is recording my web edits in darcs and showing a recent changes page.
-It is in a [darcs repository](http://joyful.com/darcsweb/darcsweb.cgi?r=ikiwiki-darcs), please send patches. --[[Simon_Michael]]
-
-> I'd like to see at least the following fixed before I commit this: --[[Joey]]
-> * Running `darcs record $filename` in backticks is not good (security)
-> The thing to do is to open stdout to /dev/null before execing darcs.
-> * Get `rcs_recentchanges_xml` working, parsing xml with regexps does
-> not seem like a maintenance win.
-> * `rcs_notify` should be removed, it's no longer used.
-> * Some form of conflict handling. Using darcs to attempt to merge
-> the changes is I gusss optional (although every other rcs backend,
-> including svn manages to do this), but it needs to at *least* detect
-> conflicts and return a page with conflict markers for the user to fix
-> the conflict.
-
-I have addressed the recentchanges bit, you can find my hacked up darcs.pm at <http://web.mornfall.net/stuff/web-root/IkiWiki/Rcs/darcs.pm>.
-
-It's got couple of FIXMEs, and a very site-specific filter for recentchanges. Not sure how to do that better though. I will eventually add web commits, probably of my own (and mention it here).
-
----
-
-And here's yet another one, including an updated `ikiwiki-makerepo`. :)
-
<http://khjk.org/~pesco/ikiwiki-darcs/> (now a darcs repo)
-> Note that there's a 'darcs' branch in git that I'm keeping a copy of your
-> code in. Just in case. :-)
-
-I've taken all the good stuff from the above and added the missing hooks. The code hasn't seen a lot of testing, so some bugs are likely yet to surface. Also, I'm not experienced with perl and don't know where I should have used the function `possibly_foolish_untaint`.
-
-Regarding the repository layout: There are two darcs repositories. One is the `srcdir`, the other we'll call `master`.
-
- * HTML is generated from `srcdir`.
- * CGI edits happen in `srcdir`.
- * The backend pulls updates from `master` into `srcdir`, i.e. darcs commits should happen to `master`.
- * `master` calls ikiwiki (through a wrapper) in its apply posthook, i.e. `master/_darcs/prefs/defaults` should look like this:
-
- apply posthook ikiwrap
- apply run-posthook
-
- (I'm not sure, should/could it be `ikiwrap --refresh` above?)
- * The backend pushes CGI edits from `srcdir` back into `master` (triggering the apply hook).
- * The working copies in `srcdir` and `master` should *not* be touched by the user, only by the CGI or darcs, respectively.
+I've taken all the good stuff from the above (now deleted --[[Joey]]) and added the missing hooks. The code hasn't seen a lot of testing, so some bugs are likely yet to surface. Also, I'm not experienced with perl and don't know where I should have used the function `possibly_foolish_untaint`.
> Review of this one:
>
-> * Should use tab indentation.
+> * Should use tab indentation. (fixed)
> * `rcs_getctime` should not need to use a ctime cache (such a cache should
> also not be named `.ikiwiki.ctimes`). `rcs_getctime` is run exactly
-> once per page, ever, and the data is cached in ikiwiki's index.
+> once per page, ever, and the data is cached in ikiwiki's index. (fixed)
> * I doubt that ENV{DARCS} will be available, since the wrapper clobbers> the entire
-> environment. I'd say remove that.
+> environment. I'd say remove that. (fixed)
> * I don't understand what `darcs_info` is doing, but it seems to be
> parsing xml with a regexp?
> * Looks like `rcs_commit` needs a few improvements, as marked TODO
-> * `rcs_remove` just calls "rm"? Does darcs record notice the file was removed
-> and automatically commit the removal? (And why `system("rm")` and not
-> `unlink`?)
-> * Is the the darcs info in [[details]] still up-to-date re this version?
+> * `rcs_remove` just calls unlink? Does darcs record notice the file was removed
+> and automatically commit the removal?
+> * Is the the darcs info in [[rcs/details]] still up-to-date re this version? (fixed)
> --[[Joey]]
> Update:
@@ -537,6 +29,8 @@ Regarding the repository layout: There are two darcs repositories. One is the `s
> this version works. It's similar, but the details differ slightly.
> You could copy my description above to replace it.
>
+>> done --[[Joey]]
+>
> There is still some ironing to do, for instance the current version doesn't allow for
> modifying attachments by re-uploading them via CGI ("darcs add failed"). Am I assuming
> correctly that "adding" a file that's already in the repo should just be a no-op?
@@ -548,4 +42,12 @@ Regarding the repository layout: There are two darcs repositories. One is the `s
>>> Done. --pesco
-[[!tag patch]]
+----
+
+I've finally merged this into ikiwiki master. The plugin looks quite
+complete, with only the new `rcs_receive` hook missing, and I
+hope it works as good as it looks. :) If anyone wants to work on improving
+it, there are some TODOs as mentioned above that could still be improved.
+--[[Joey]]
+
+[[!tag patch done]]
diff --git a/doc/todo/hidden_links__47__tags.mdwn b/doc/todo/hidden_links__47__tags.mdwn
new file mode 100644
index 000000000..43d7a8797
--- /dev/null
+++ b/doc/todo/hidden_links__47__tags.mdwn
@@ -0,0 +1,13 @@
+[[!tag wishlist]]
+
+I would like to have the possibility for hidden tags or links.
+Using the tag functionality I could group some news items for including them into other subpages. But I don't want the links or tags to show (and I don't want Tag lists like "Tags: ?mytag").
+The tagged items should not differ from the items, that are not tagged.
+I didn't find any way to hide the tag list or links and I don't want to have to create a "hidden" page containing links to the pages and then using the backlink functionality, because this is more prone to errors. It's easier to forget adding a link on a second page than forgetting to add a needed tag to a new newsitem.
+
+> I found out, that using the meta plugin it is possible to create the hidden link, that I wanted.
+-- [[users/Enno]]
+
+>> Yes, meta link will not show up as a visible link on the page, while
+>> also not showing up in the list of tags of a page, so it seems what you
+>> want. [[done]] --[[Joey]]
diff --git a/doc/todo/inline_autotitles.mdwn b/doc/todo/inline_postform_autotitles.mdwn
index 8bf71deae..5005208be 100644
--- a/doc/todo/inline_autotitles.mdwn
+++ b/doc/todo/inline_postform_autotitles.mdwn
@@ -1,8 +1,9 @@
[[!tag wishlist]]
[[!tag patch]]
-for inlines of pages which follow a certain scheme, it might not be required to
-set the title for each individual post, but to automatically set the title.
+for postforms in inlines of pages which follow a certain scheme, it might not
+be required to set the title for each individual post, but to automatically set
+the title and show no input box prompting for it.
this can either be based on timestamp formatting, or use the already existing
munging mechanism, which appends numbers to page titles in case that page
already exists.
diff --git a/doc/todo/need_global_renamepage_hook.mdwn b/doc/todo/need_global_renamepage_hook.mdwn
index 874460b29..e3cec4a9b 100644
--- a/doc/todo/need_global_renamepage_hook.mdwn
+++ b/doc/todo/need_global_renamepage_hook.mdwn
@@ -111,3 +111,5 @@ hashes is desired, it could return the full set of hashes.
>>>> It looks good. I made some small changes to it in my own po branch.
>>>> Nothing significant really. If this were not tied up in the po branch,
>>>> I've have merged it to master already. --[[Joey]]
+
+>>>> Thanks, this is great :) --[[intrigeri]]
diff --git a/doc/todo/rewrite_ikiwiki_in_haskell.mdwn b/doc/todo/rewrite_ikiwiki_in_haskell.mdwn
new file mode 100644
index 000000000..204c48cd7
--- /dev/null
+++ b/doc/todo/rewrite_ikiwiki_in_haskell.mdwn
@@ -0,0 +1,68 @@
+[[!tag wishlist blue-sky]]
+
+In the long term, I have been considering rewriting ikiwiki in haskell.
+It's appealing for a lot of reasons, including:
+
+* No need to depend on a C compiler and have wrappers. Instead, ikiwiki
+ binaries could be built on demand to do the things wrappers are used for
+ now (cgi, post-commit, etc).
+* Potentially much faster. One problem with the now very modular ikiwiki is
+ that it has to load up dozens of perl modules each time it runs, which
+ means both opening lots of files and evaluating them. A haskell version
+ could run from one pre-compiled file. Other speed efficienies are also
+ likely with haskell. For example, pandoc is apparently an order of
+ magnitude faster than perl markdown implementations.
+* Many plugins could be written in pure functional code, with no side
+ effects. Not all of them, of course.
+* It should be much easier to get ikiwiki to support parallel compilation
+ on multi-core systems using haskell.
+* A rewrite would be an opportunity to utterly break compatability and
+ redo things based on experience. Since the haskell libraries used for
+ markdown, templates, etc, are unlikely to be very compatable with the perl
+ versions, and since perl plugins obviously wouldn't work, and perl setup
+ files wouldn't be practical to keep, a lot of things would unavoidably
+ change, and at that point changinge everything else I can think of
+ probably wouldn't hurt (much).
+
+ - Re templates, it would be nice to have a template library that
+ doesn't use html-ish templating tags, since those are hard for users to
+ edit in html editors currently.
+ - This would be a chance to make WikiLinks with link texts read
+ "the right way round" (ie, vaguely wiki creole compatably).
+ - The data structures would probably be quite different.
+ - I might want to drop a lot of the command-line flags, either
+ requiring a setup file be used for those things, or leaving the
+ general-purpose `--set var=value` flag.
+ - Sometimes the current behavior of `--setup` seems confusing; it might
+ only cause a setup file to be read, and not force rebuild mode.
+ - Hard to say how the very high level plugin interface design would change,
+ but at the least some of the names of hooks could stand a rename, and
+ their parameter passing cleaned up.
+
+We know that a big, break-the-world rewrite like this can be a very
+bad thing for a project to attempt. It would be possible to support
+external plugins written in haskell today, without any rewrite; and a few
+of the benefits could be obtained by, eg, making the mdwn plugin be a
+haskell program that uses pandoc. I doubt that wouod be a good first step
+to converting ikiwiki to haskell, because such a program would have very
+different data structures and intercommuniucation than a pure haskell
+version.
+
+Some other things to be scared about:
+
+* By picking perl, I made a lot of people annoyed (and probably turned
+ several people away from using ikiwiki). But over time there turned out
+ to be a lot of folks who knew perl already (even if rustily), and made
+ some *very* useful contributions. I doubt there's as large a pool of haskell
+ programmers, and it's probably harder for a python user to learn haskell
+ than perl if they want to contribute to ikiwiki.
+* It might be harder for users of hosting services to install a haskell based
+ ikiwiki than the perl version. Such systems probably don't have ghc and
+ a bunch of haskell libraries. OTOH, it might be possible to build a
+ static binary at home and upload it, thus avoiding a messy installation
+ procedure entirely.
+* I can barely code in haskell yet. I'm probably about 100x faster at
+ programming in perl. I need to get some more practical experience before
+ I´m fast and seasoned enough in haskell to attempt such a project.
+ (And so far, progress at learning has been slow and I have not managed
+ to write anything serious in haskell.) --[[Joey]]
diff --git a/doc/todo/rewrite_ikiwiki_in_haskell/discussion.mdwn b/doc/todo/rewrite_ikiwiki_in_haskell/discussion.mdwn
new file mode 100644
index 000000000..1edebe4e8
--- /dev/null
+++ b/doc/todo/rewrite_ikiwiki_in_haskell/discussion.mdwn
@@ -0,0 +1,14 @@
+Ok, I have to admit, I have no idea if this is an April fool's joke or not.
+Congratulations for demonstrating that April fools jokes can still be subtle
+(whether intentionally or not!) -- [[Jon]]
+
+> Having said all that, have you looked at erlang? Have you heard of couchdb?
+> I'd strongly recommend looking at that. -- [[Jon]]
+
+>> I've glanced at couchdb, but don't see how it would tie in with ikiwiki.
+>> --[[Joey]]
+
+
+>>> It doesn't really. I recently (re-)read about couchdb and thought that
+>>> what it was trying to do had some comparisons with the thinking going on
+>>> in [[todo/structured_page_data]]. -- [[Jon]]
diff --git a/doc/usage.mdwn b/doc/usage.mdwn
index e2fe85ff6..0c618de5c 100644
--- a/doc/usage.mdwn
+++ b/doc/usage.mdwn
@@ -308,10 +308,10 @@ also be configured using a setup file.
* --getctime
- Pull last changed time for each new page out of the revision control
+ Pull creation time for each new page out of the revision control
system. This rarely used option provides a way to get the real creation
times of items in weblogs, such as when building a wiki from a new
- Subversion checkout. It is unoptimised and quite slow. It is best used
+ VCS checkout. It is unoptimised and quite slow. It is best used
with --rebuild, to force ikiwiki to get the ctime for all pages.
* --set var=value
diff --git a/doc/users/jon.mdwn b/doc/users/jon.mdwn
index 3e22ded1d..1cda23999 100644
--- a/doc/users/jon.mdwn
+++ b/doc/users/jon.mdwn
@@ -4,8 +4,9 @@ team-documentation management system for a small-sized group of UNIX
sysadmins.
* my edits should appear either as 'Jon' (if I've used
- [[tips/untrusted_git_push]]) or 'alcopop.org/me/openid/'.
-* My [homepage](http://jmtd.net/) is powered by ikiwiki (replacing my [older homepage](http://alcopop.org/), which was a mess of scripts)
+ [[tips/untrusted_git_push]]) or 'jmtd.net' (or once upon a time
+ 'alcopop.org/me/openid/' or 'jondowland').
+* My [homepage](http://jmtd.net/) is powered by ikiwiki
I gave a talk at the [UK UNIX User's Group](http://www.ukuug.org/) annual
[Linux conference](http://www.ukuug.org/events/linux2008/) about organising
@@ -14,3 +15,6 @@ discussing IkiWiki in some technical detail and suggesting it as a good piece
of software for this task.
* slides at <http://www.staff.ncl.ac.uk/jon.dowland/unix/docs/>.
+
+I am also working on some ikiwiki hacks: an alternative approach to
+[[plugins/comments]]; a system for [[forum/managing_todo_lists]].
diff --git a/doc/users/seanh.mdwn b/doc/users/seanh.mdwn
new file mode 100644
index 000000000..d093c2f32
--- /dev/null
+++ b/doc/users/seanh.mdwn
@@ -0,0 +1 @@
+seanh is an ikiwiki user.
diff --git a/doc/wikitemplates.mdwn b/doc/wikitemplates.mdwn
index fc5893677..f365cd5aa 100644
--- a/doc/wikitemplates.mdwn
+++ b/doc/wikitemplates.mdwn
@@ -21,6 +21,7 @@ located in /usr/share/ikiwiki/templates by default.
* `inlinepage.tmpl` - Used for adding a page inline in a blog
page.
* `archivepage.tmpl` - Used for listing a page in a blog archive page.
+* `microblog.tmpl` - Used for showing a microblogging post inline.
* `blogpost.tmpl` - Used for a form to add a post to a blog (and a rss/atom links)
* `feedlink.tmpl` - Used to add rss/atom links if blogpost.tmpl is not used.
* `aggregatepost.tmpl` - Used by the [[plugins/aggregate]] plugin to create