Products

Solutions

Resources

Partners

Community

Blog

About

QA

Ideas Test

New Community Website

Ordinarily, you'd be at the right spot, but we've recently launched a brand new community website... For the community, by the community.

Yay... Take Me to the Community!

Welcome to the DNN Community Forums, your preferred source of online community support for all things related to DNN.
In order to participate you must be a registered DNNizen

HomeHomeUsing DNN Platf...Using DNN Platf...Administration ...Administration ...Search Engine OptimimizationSearch Engine Optimimization
Previous
 
Next
New Post
12/1/2006 6:42 AM
 
Hi Folks,

I hope all is well.

I'd like to share what I know about search engine optimization and see if I'm on the right track and if other people have some ideas.

Basically I do these things:
- Add a browserCaps section to my web.config file so that known spiders will be recognized as such. (listed below)
- Add a robots.txt file to the root of my dnn install (listed below)
- Provide Keywords and Descriptions and "Submit to Google" the pages I want indexed
- Use Friendly URLs

Also, I can check my site to make sure the SolPart menu it properly displaying as a site map if I set my FireFox browser to spoof as a spider.

So I'm wondering...
- Do I really need to provide the browserCaps section in my web.config?
- Other than the steps I'm taking, do other folks have any recommendations?

Thanks,
Wylie

Here’s how to set you FireFox browser to spoof a GoogleBot.

*************************************************************

  • Open a FireFox browser and enter about:config as a URL address.
  • You will see all the configuration settings for your browser
  • Now click the right mouse button to get the context menu
  • Select "String" > "New".
  • Enter the preference name "general.useragent.override", without the quotes.
  • Then enter the string value “Googlebot/2.1 (+http://www.googlebot.com/bot.html)”

 When you are finished browsing the web as a googlebot, you may re-open the config screen, find the general.useragent.override preference, right click on it and “reset”.


BrowserCaps - place within <system.web> tags in your web.config
*********************************************************
<browserCaps>
          <filter>
              <!-- SEARCH ENGINES GROUP -->
              <!-- check Google (Yahoo uses this as well) -->
              <case match="^Googlebot(\-Image)?/(?'version'(?'major'\d+)(?'minor'\.\d+)).*">
                  browser=Google
                  version=${version}
                  majorversion=${major}
                  minorversion=${minor}
                  crawler=true
              </case>

              <!-- check Alta Vista (Scooter) -->
              <case match="^Scooter(/|-)(?'version'(?'major'\d+)(?'minor'\.\d+)).*">
                  browser=AltaVista
                  version=${version}
                  majorversion=${major}
                  minorversion=${minor}
                  crawler=true
              </case>

              <!-- check Alta Vista (Mercator) -->
              <case match="Mercator">
                  browser=AltaVista
                  crawler=true
              </case>

              <!-- check Slurp (Yahoo uses this as well) -->
              <case match="Slurp">
                  browser=Slurp
                  crawler=true
              </case>

              <!-- check MSN -->
              <case match="MSNBOT">
                  browser=MSN
                  crawler=true
              </case>

              <!-- check Northern Light -->
              <case match="^Gulliver/(?'version'(?'major'\d+)(?'minor'\.\d+)).*">
                  browser=NorthernLight
                  version=${version}
                  majorversion=${major}
                  minorversion=${minor}
                  crawler=true
              </case>

              <!-- check Excite -->
              <case match="ArchitextSpider">
                  browser=Excite
                  crawler=true
              </case>

              <!-- Lycos -->
              <case match="Lycos_Spider">
                  browser=Lycos
                  crawler=true
              </case>

              <!-- Ask Jeeves -->
              <case match="Ask Jeeves">
                  browser=AskJeaves
                  crawler=true
              </case>

              <!-- check Fast -->
              <case match="^FAST-WebCrawler/(?'version'(?'major'\d+)(?'minor'\.\d+)).*">
                  browser=Fast
                  version=${version}
                  majorversion=${major}
                  minorversion=${minor}
                  crawler=true
              </case>

              <!-- IBM Research Web Crawler -->
              <case match="http\:\/\/www\.almaden.ibm.com\/cs\/crawler">
                  browser=IBMResearchWebCrawler
                  crawler=true
              </case>

          </filter>

      </browserCaps>

Robots.txt
*********************************************
User-agent: *
Disallow: /admin/
Disallow: /app_globalresources/
Disallow: /bin/
Disallow: /Components/
Disallow: /config/
Disallow: /controls/
Disallow: /DesktopModules/
Disallow: /images/
Disallow: /install/
Disallow: /js/
Disallow: /Portals/
Disallow: /Providers/




Chris Wylie
Chief Executive Officer
www.AccordLMS.com
1775 W. State Street Suite 371, Boise ID 83702 USA
 
 
New Post
12/1/2006 6:59 AM
 

- Add a browserCaps section to my web.config file so that known spiders will be recognized as such. (listed below)
- Add a robots.txt file to the root of my dnn install (listed below)
- Provide Keywords and Descriptions and "Submit to Google" the pages I want indexed
- Use Friendly URLs

AFAIK files that cannot be reached from a link, will not be spidered anyway, so robots.txt seems rather pointless to me. ASP.NET never puts direct references to controls in the generated html....

Keywords and Descriptions are hardly ever used anymore by Google. (and luckily so... its more often used to spam search engines than adding real value to the info available to visitors....). On the other hand: text between <H1> tags gets extra attention, also in my experience, google just likes to see meaningfull text in general. Submitting pages to Google is something you should be extremely carefull with: google will find the page anyway if it can be reached from your home page. Pages that cannot be reached by any link could be submitted to google.. but then again.. what would be the purpose of a page that cannot be reached by any link pointing to it...?

Also: it helps if you add google a bit with the structure of your site, for that you can use Google Sitemap. I am usig the free DNN google sitemap module of BiteTheBullit, you can find that here: http://www.bitethebullet.co.uk/tabid/59/Default.aspx

Also: i have the feeling adding google ads to your site helps in getting a better page rank.. but i'm not entirely sure about that

Also: add your site to the dmoz.org index, this is often used by search engines as a starting point for spidering. If your site is there, it will be visited automagically by even the weirdest spiders :)

Also: the most important thing you have to keep in mind: make sure your site is worthwile visiting, and if it is: make sure others link to your site.

Cheers,

Erik


Erik van Ballegoij, Former DNN Corp. Employee and DNN Expert

DNN Blog | Twitter: @erikvb | LinkedIn: Erik van Ballegoij on LinkedIn

 
New Post
12/1/2006 7:18 AM
 
Thanks for the additions Erik. 

Concerning the robots.txt file, you are right in that all the disallows are unnecessary, but I have several experiences with trying to get Google to index my sites.  Their UI reported to me that my site was not getting indexed because no robots.txt file was found.  Once I got this file uploaded then the indexing proceeded.

I will also add that SolPart does present a full site map - just set FF to spoof a GoogleBot and then visit a DNN site with a SolPart menu to see this in action.

Have a good one,
Wylie



Chris Wylie
Chief Executive Officer
www.AccordLMS.com
1775 W. State Street Suite 371, Boise ID 83702 USA
 
 
New Post
12/1/2006 7:24 AM
 

you are right that solpart renders downlevel for browsers that do not have javascript (and google is no more than just a browser with no js support). However.. hidden pages will not show up with this. The google sitemap module also shows hidden pages to google.Also, a "normal" sitemap module will help in this sense, see my site for a sample of that (http://www.apollo-software.nl/Over/SiteMap/tabid/228/Default.aspx)

btw.. i forgot to mention the FF user agent switcher: https://addons.mozilla.org/firefox/59/ this allows you to easily switch user agents in firefox.


Erik van Ballegoij, Former DNN Corp. Employee and DNN Expert

DNN Blog | Twitter: @erikvb | LinkedIn: Erik van Ballegoij on LinkedIn

 
Previous
 
Next
HomeHomeUsing DNN Platf...Using DNN Platf...Administration ...Administration ...Search Engine OptimimizationSearch Engine Optimimization


These Forums are dedicated to discussion of DNN Platform and Evoq Solutions.

For the benefit of the community and to protect the integrity of the ecosystem, please observe the following posting guidelines:

  1. No Advertising. This includes promotion of commercial and non-commercial products or services which are not directly related to DNN.
  2. No vendor trolling / poaching. If someone posts about a vendor issue, allow the vendor or other customers to respond. Any post that looks like trolling / poaching will be removed.
  3. Discussion or promotion of DNN Platform product releases under a different brand name are strictly prohibited.
  4. No Flaming or Trolling.
  5. No Profanity, Racism, or Prejudice.
  6. Site Moderators have the final word on approving / removing a thread or post or comment.
  7. English language posting only, please.
What is Liquid Content?
Find Out
What is Liquid Content?
Find Out
What is Liquid Content?
Find Out