[17:38:26] How is the interface page MediaWiki:Robots.txt is turned into the actual robots.txt file for wikipedia.org? Extension, script? I've looked at the talk page, and the various wgConfig options but haven't found the process. [17:45:32] I'd like to replicate it if I can. Seems like a good thing if you can have admins edit the page, and that turns into actual webserver configuration [17:45:45] Given the nature of robots.txt [18:11:22] freephile: https://phabricator.wikimedia.org/source/operations-puppet/browse/production/modules/mediawiki/files/apache/sites/public-wiki-rewrites.incl -> redirect from '/robots\.txt' to '/w/robots.php' first. [18:12:40] freephile: https://phabricator.wikimedia.org/source/mediawiki-config/browse/master/w/robots.php ; execute this file. (I'm not sure) [18:12:57] rxy: excellent [18:13:02] thanks [18:13:06] np [18:49:02] I need to add .xmind to application/zip, configured from mime.types. I'm not seeing a secondary source for that information, though. I'd rather not edit the file because it's managed by my distro's package manager. Am I SOL or is there another place to configur it? [19:19:24] anomie: I'm not sure about your edit https://www.mediawiki.org/w/?diff=2558345, test( A > B ) always should give the same result as !test( B > A ) [19:23:58] oh, forget that... I've *now* understood what did you mean by the special case "a[i] == a[i+1]" [19:29:18] * anomie sees that Vulpix answered their own question [23:36:33] Hi. I am trying to parse an XML dump of Simple Wiki. I want the title of the article and its summary. However, I am running into a lot of edge cases when using regex to clean the text. Is there a utility I can use that will simplify this?