Hello,
We have a problem with Google not indexing the subpages of all our DNN websites.
The homepage URLs www.mywebsite.com have pageranks (2/10 or 3/10 for instance)
None of the website subpages have Google pageranks.
Friendly URLs have been enable on the DNN websites. (we are using DNN 4+)
I have used the Google Webmaster Tools to analyze the crawling.
Under "Diagnostic - Crawl Errors" I see all the links to the subpages as "Unreachable URLs" (in one example 67 Unreachable URLs)
This is the error description: "500 (Internal server error) The server encountered an error and cannot fulfil the request."
BUT! -> when I look at these "error 500" URLs, and click them, they all work just fine.
There are no other Crawl errors (no HTTP Errors, Not found errors, not allowed errors, timed outs, restricted by robots.txt, not
followeds...)
We aren't using robots.txt files nor have we submitted sitemaps to Google.
(Because Googlebot should (has to) find the subpages without submitting sitemaps)
I have set the "Preferred domain" in Google Webmaster Tools to www.mywebsite.com (www.mywebsite.com or mywebsite.com)
We are using the dnnMENU, the menu is using <menuitem id="53"..> tags so I added [LINKS] on the page,
which adds the complete Menu a second time with <a href..> tags. We did this to ensure the Googlebot follows the links.
We sometimes find "old" non-friendly URLs in the Google cache from when "Friendly URLs" wasn't enabled on DNN yet, but these are
disappearing gradually.
The problem is when we enter site:www.mywebsite.com in google, the results only show pages followed by links in the content on the front
page, (eg: www.mywebsite.com/LinkClick.aspx?link=123&tabid=12&mid=123), but no results of any subpages.
The timespan on this is several weeks (the websites/URLs have exists for years), not just a few days, I know it takes time for the
Googlebot to index.
So why isn't site:www.mywebsite.com showing my subpages?
- Is the reason Google isn't indexing the subpages because it finds they are not important enough?
This would be weird because they do have a lot of hits, and they all have keywords and all hold relevant textual data.
- These websites used to run on a custom asp engine, and since we transfered to dotnetnuke (summer 2006), we get these results.
We used to have ALL of these subpaged cached/indexed by Google.
Any hints, solutions, tips or comments are greatly appriciated.