Last modified: 2014-10-25 13:28:32 UTC

Wikimedia Bugzilla is closed!

Wikimedia migrated from Bugzilla to Phabricator. Bug reports are handled in Wikimedia Phabricator.
This static website is read-only and for historical purposes. It is not possible to log in and except for displaying bug reports and their history, links might be broken. See T74507, the corresponding Phabricator task for complete and up-to-date bug report information.
Bug 72507 - Allow crawling of bugzilla.wikimedia.org select content
Allow crawling of bugzilla.wikimedia.org select content
Status: NEW
Product: Wikimedia
Classification: Unclassified
Bugzilla (Other open bugs)
wmf-deployment
All All
: Low enhancement (vote)
: ---
Assigned To: Nobody - You can work on this!
http://web.archive.org/save/https://b...
:
Depends on:
Blocks:
  Show dependency treegraph
 
Reported: 2014-10-25 10:16 UTC by Nemo
Modified: 2014-10-25 13:28 UTC (History)
5 users (show)

See Also:
Web browser: ---
Mobile Platform: ---
Assignee Huggle Beta Tester: ---


Attachments

Description Nemo 2014-10-25 10:16:04 UTC
The robots.txt rules are unnecessarily restrictive. As bugzilla is being deprecated, and only a portion of its content migrated to phabricator, it's essential that we allow third parties to do their job. All crawlers, or at least ia_archiver (wayback machine), should be allowed to crawl:
1) any content which
2) doesn't specifically cause load issues and
3) is not being semantically migrated to phabricator.
Ideally we'd drop requirement (3) but let's start somewhere.

Example URLs which shouldn't be blacklisted:
* /page.cgi?id=voting/bug.html*
* /duplicates.cgi*
* /report.cgi* (unless load)
* /weekly-bug-summary.cgi*
* /describecomponents.cgi*

In fact, is there any reason not to allow everything, minus:
* /show_bug.cgi
* /showdependencytree.cgi
* /query.cgi
?
Comment 1 Andre Klapper 2014-10-25 13:28:32 UTC
Duplicate of bug 13881?

Note You need to log in before you can comment on or make changes to this bug.


Navigation
Links