User Tools

Site Tools


Sidebar

DokuWiki plugins

There are DokuWiki plugins for research & education which I developed. All plugins are under GPL2 license, so feel free to use them in your DokuWiki.

Dokuwiki Official

Scientific plugins

DokuWiki plugins for bioscience and biotechnology. Some plugins require cache plugin or bio plugin.

Adora Bella compatible

Before Angua

General plugins

Adora Bella compatible

Before Angua

Links

1) , 2) , 3) interwiki does
en:robots.txt_for_dokuwiki

robots.txt for crawlers in the DokuWiki

If your DokuWiki is configured to change URLs using .htaccess, the following URL

http://www.yourdomain.net/robots.txt

will be changed to like this.

http://www.yourdomain.net/doku.php?id=robots.txt

The page of at this URL only prompts visitors to create a new page…


:-D The latest version allowed crawlers to read special files such as robots.txt, sitemap and so on. You don't have to modify .htaccess if you use the latest version. Take a look 'doku.php' script and search a word 'robots.txt'.


.htaceess modification

In order to pass the robots.txt to the crawlers, insert a line in the .htaccess file that just command to rewrite from ^robots.txt$ to robots.txt.

Before insertion

(part of .htaccess file)

RewriteBase / 
#
RewriteRule ^_media/(.*) lib/exe/fetch.php?media=$1 [QSA,L]
RewriteRule ^_detail/(.*) lib/exe/detail.php?media=$1 [QSA,L]

After insertion

(part of .htaccess file)

RewriteBase / 
#
RewriteRule ^robots.txt$ robots.txt

RewriteRule ^_media/(.*) lib/exe/fetch.php?media=$1 [QSA,L]
RewriteRule ^_detail/(.*) lib/exe/detail.php?media=$1 [QSA,L]

Google sitemap.xml and account file

.htaccess also pass invalid URLs of a 'sitemap.xml' for Google sitemap and a site ID (empty html file). So, set rewrite rules for these files.

RewriteRule ^sitemap.xml$ sitemap.xml.gz
RewriteRule ^googleaFFFFFFFFFFFFFFF.html$ googlea0FFFFFFFFFFFFFFF.html
Translations of this page:
en/robots.txt_for_dokuwiki.txt · Last modified: 2013/01/27 11:19 (external edit)