Products

Solutions

Resources

Partners

Community

Blog

About

QA

Ideas Test

New Community Website

Ordinarily, you'd be at the right spot, but we've recently launched a brand new community website... For the community, by the community.

Yay... Take Me to the Community!

Welcome to the DNN Community Forums, your preferred source of online community support for all things related to DNN.
In order to participate you must be a registered DNNizen

HomeHomeOur CommunityOur CommunityGeneral Discuss...General Discuss...Google search results - login & privacy pagesGoogle search results - login & privacy pages
Previous
 
Next
New Post
12/13/2006 10:10 AM
 

I recently created a dot net nule site for a friend, we created a basic site, applied a skin , everything seemed fine. We even applied a google sitemap generator that we could submit to google. When google spidered the site it seemed to get the first few links correct then it seemed to focus on the login and privacy links that appear off our main content pages ... it did this many times ... without actually finishing spidering all the actual main content pages themselves. We also spent quite a lot of time creating relevant meta tag descriptions and keywords for each content page.

What I would like to know, is there any way of controlling the meta tag directives on a per page basis that would allow me to discourage a spider from following that page ... ie the robots/follow/nofollow directive ... in particular I would want the spider to be discouraged from spidering all the various login and privacy "virtual" pages that are created from each main page I actually create ... I dont want to make these pages hidden to normal browsing.

Any help appreciated :)

 
New Post
12/13/2006 6:55 PM
 
I use a robots.txt file to steer the robots away from the login, register, terms and privacy pages.
 
New Post
1/14/2007 9:20 AM
 

Hi Sam,

I was thinking of doing this, but I was unsure if I had to create an entry for each page I didnt want the spider to visit , thus making a very manual process, or if there was some kind of wildcard entry I could use to catch every occurance of a privacy / login version of a page ? How do you do this ?

rgds,
Cameron.

 
New Post
1/31/2007 7:22 AM
 

The answer :

look at www.dotnetnuke.com/robots.txt

within their robots file is an entry for the googlebot :

User-agent: Googlebot
Disallow: /*/ctl/
Disallow: /admin/
Disallow: /App_GlobalResources/
Disallow: /bin/
Disallow: /Components/
Disallow: /contest/
Disallow: /controls/
Disallow: /DesktopModules/
Disallow: /HttpModules/
Disallow: /images/
Disallow: /Install/
Disallow: /js/
Disallow: /Portals/25/Transcripts/
Disallow: /Providers/

the interesting one is :

/*/ctrl

this entry blocks the extra "duplicated" virtual pages e.g. privacy / logon / terms

This is only used for googlebot, not other spiders, I think google has extened the protocol to allow this notation.

 
Previous
 
Next
HomeHomeOur CommunityOur CommunityGeneral Discuss...General Discuss...Google search results - login & privacy pagesGoogle search results - login & privacy pages


These Forums are dedicated to discussion of DNN Platform and Evoq Solutions.

For the benefit of the community and to protect the integrity of the ecosystem, please observe the following posting guidelines:

  1. No Advertising. This includes promotion of commercial and non-commercial products or services which are not directly related to DNN.
  2. No vendor trolling / poaching. If someone posts about a vendor issue, allow the vendor or other customers to respond. Any post that looks like trolling / poaching will be removed.
  3. Discussion or promotion of DNN Platform product releases under a different brand name are strictly prohibited.
  4. No Flaming or Trolling.
  5. No Profanity, Racism, or Prejudice.
  6. Site Moderators have the final word on approving / removing a thread or post or comment.
  7. English language posting only, please.
What is Liquid Content?
Find Out
What is Liquid Content?
Find Out
What is Liquid Content?
Find Out