Last modified: 2013-03-25 15:02:58 UTC
When using generate Sitemap all pages in one namespace get the same priority. I think it would be usefull to change this because google is complaining about this and it's not really good for other bots using the sitemap. So I think the priority should depend on the last time it was touched. Pages that was changed short time ago should have a higher priority than pages that wasn't edited for a long time.
*** Bug 17019 has been marked as a duplicate of this bug. ***
agreed. maybe include the factor of hoih page views. although generating the sitemap shouldn't use too much db-load. i run the sitemap script every 2 hours to feed google :) well, i have ~20 new articles per day though. For me it only happens in the main ns and I fear google ignores the complete xml if it a warning is printed. I submitted the following namespaces to google: MAIN 21051 (with warnings) CATEGORIES 6833 IMAGES 2040 HELP 24 PROJECT 17 Maybe give it 1.0 if there was much editing action lately independently from teh page views. But to get rid of the warnings first of all it might be good to set high priorities for recently changed articles. http://www.google.com/support/webmasters/bin/answer.py?hl=en&answer=71936
Well I'm making an update to the sitemap only once a day at night. Google does not scan that often I think. From my point of view every solution with diffenrent priorities would be good. But I think when there is one factor that is used to generate different priorities it's not that different so implement others.
I made some changes to generate Sitemap and added a random priority generator, the best solution I could create myself http://www.mediawiki.org/wiki/User:DaSch/generateSitemap.php
This works fine for now. Warnings are gone after first download of the new sitemap.
Thx :)