I would like to maximize the PR on my forum display and thread display pages, without sacrificing PR to less important pages. (for instance, the newthread.php page)

Now suppose there are 5 links on page A. Normally, page A's PR will be spread out to the 5 pages. But now if I added a robots.txt file which blocked indexing of two of the links. Would page A's PR now be spread less thin among 3 pages? Or would it be spread just as thin, but 2 of the pages would be entitled to a share of PR but just wouldn't use it.

I hope this sorta made a bit of sense.

I sent you a PM RE: this topic. :)

And I started this thread RE: your PM :)

Ok I am a knuckelhead. I guess this was a chicken before the egg thing. ;-)

The robots.txt file I'm currently using is in my root directory (not my forum root) and looks like this:

User-agent: googlebot
Disallow: /techtalkforums/announcement.php
Disallow: /techtalkforums/faq.php
Disallow: /techtalkforums/forumdisplay.php
Disallow: /techtalkforums/login.php
Disallow: /techtalkforums/member.php
Disallow: /techtalkforums/newreply.php
Disallow: /techtalkforums/newthread.php
Disallow: /techtalkforums/online.php
Disallow: /techtalkforums/printthread.php
Disallow: /techtalkforums/search.php
Disallow: /techtalkforums/showthread.php

I am disallowing access to showthread.php and forumdisplay.php because I would rather Google only spider the .html mod_rewrite versions of the forums and threads, and therefore not get duplicate content. Was this done correctly? Am I excluding the correct things?

without going too far in, it looks decent to me. I am not sure however, if the robots.txt blocks the weakening link pop from all the links. The less links on a page the more potent the links are. A page with tons of links is spreading the pop thin. That would be a good question to ask SEO-Guy.

It would be very nice if the robots.txt would block the weakening spread of PR. However, even if it doesn't do this, it would still be valuable because it would eliminate spidering duplicate content (i.e. showthread.php?t=10 and thread10.html)

It would be very nice if the robots.txt would block the weakening spread of PR. However, even if it doesn't do this, it would still be valuable because it would eliminate spidering duplicate content (i.e. showthread.php?t=10 and thread10.html)

*nods* for sure.

I don't see how this will help at all.

Google frowns upon multiple pages with the same content. For example, if two different URLs have the exact same content on them, google considers it spamming their search engine. This forum uses Apache's mod_rewrite to rewrite URLs to have a .html extension for search engine purposes. Therefore, the webpage showthread.php?t=100 is the exact same thing as thread100.html - if google spiders see this duplicate contact, they will think that daniweb.com is trying to inflate its page count in google by having multiple URLs with the same content. However, by using robots.txt to block google from spidering the showthread.php pages, google only spiders the pages ending in .html - and therefore doesn't penalize us for duplicate content.

may i aks how you changed it to thread6988.html instead of showthread?

may i aks how you changed it to thread6988.html instead of showthread?

It's done using a technique called url re-writing.

On this server, the page thread6988.htm does not physically exist. Instead, the web server monitors incoming url requests and looks for the word thread in that request..... if so, it grabs the numbers from that and passes it along to showthread.php easy enough.

Hope this helps.

Hi csgal,
can you post the new robots.txt here, please?

Thank you.

Would it make sense to use the robots no follow tag in your particular case?

Can you send me the robots.txt to Matzefn1@web.de?
Thank you very much.

Matzefn1

Can I have the robots.txt, please.

Post #5 shows the robots.txt file that I used to use. I no longer use a robots.txt file.

Post #5 shows the robots.txt file that I used to use. I no longer use a robots.txt file.

Why do you no longer use a robots.txt? Google frowns upon multiple pages with the same content...!?

My robots.txt file: http://www.schachfeld.de/robots.txt

We had a problem where pages that had a no-crawl code at the root directory still were being crawled (they were PDFs that had valuable IP in them).

We discovered that the bots were getting in through links on other pages of ours (the PDFs are "samples" of products that we use as marketing tools), so we put "no follow" codes -- <meta name="robots" content="index,nofollow" /> -- on those pages.

This let's the spider index the page but not follow the links on the page.

But if someone includes a link to the non-HTML thread in a page that you don't control, do you think it will bypass your html rewrite?

Sometimes, even after adding the no index tag to the pages, it will take some weeks before search engines know exactly what you mean.

Aren't robot files bad for tracking purposes, like a no follow? I guess I'm new to this kind of discussion..lol.

Be a part of the DaniWeb community

We're a friendly, industry-focused community of developers, IT pros, digital marketers, and technology enthusiasts meeting, networking, learning, and sharing knowledge.