Hiya.
It occurs to me that there's very little point in search engines crawling certain pages of an OpenGuide - anything with "action=edit" or "action=delete" in the URL, at the very least, has no real value to someone searching for information. Unfortunately, unless I'm missing something, robots.txt syntax doesn't allow for matching on anything other than the start of the path component of a URL, which doesn't help us here.
It's been suggested to me that ``<meta robots="noindex">'' tags in the <head> of a page are effective here, but I'm not sure of the best way to implement this for edit pages without implementing it for *all* pages, which would be careless.
I'm thinking of something in header.tt conditional upon the requested URI containing "action=edit" or "action=delete", but before I wander off learning how to talk template::toolkit, I'd be interested to hear better suggestions -- and other values of action we might care about, I guess.
The other, more complex but possibly "better" alternative, would be for someone to run with the idea mooted in this thread: http://openguides.org/mail/openguides-dev/2004-April/000258.html
Then edit pages could be /edit/Node_Name, deletes /delete/Node_Name, and so on. This makes setting up a suitable robots.txt very simple indeed, though it makes setting up the Apache rewrite rules a) a requirement, instead of just a nice thing, and b) more complex than at present.
Thoughts, comments, suggestions and the like all invited.
Cheers, James.