<?xml version="1.0" encoding="UTF-8" standalone="yes" ?>
<!DOCTYPE bugzilla SYSTEM "https://www.w3.org/Bugs/Public/page.cgi?id=bugzilla.dtd">

<bugzilla version="5.0.4"
          urlbase="https://www.w3.org/Bugs/Public/"
          
          maintainer="sysbot+bugzilla@w3.org"
>

    <bug>
          <bug_id>893</bug_id>
          
          <creation_ts>2004-09-27 17:05:01 +0000</creation_ts>
          <short_desc>cache (non) existence</short_desc>
          <delta_ts>2004-11-07 09:48:05 +0000</delta_ts>
          <reporter_accessible>1</reporter_accessible>
          <cclist_accessible>1</cclist_accessible>
          <classification_id>1</classification_id>
          <classification>Unclassified</classification>
          <product>LinkChecker</product>
          <component>checklink</component>
          <version>4.0</version>
          <rep_platform>All</rep_platform>
          <op_sys>All</op_sys>
          <bug_status>RESOLVED</bug_status>
          <resolution>FIXED</resolution>
          
          
          <bug_file_loc></bug_file_loc>
          <status_whiteboard></status_whiteboard>
          <keywords></keywords>
          <priority>P1</priority>
          <bug_severity>normal</bug_severity>
          <target_milestone>4.1</target_milestone>
          
          
          <everconfirmed>1</everconfirmed>
          <reporter name="Mark Ludwig">uldmjl</reporter>
          <assigned_to name="Ville Skyttä">ville.skytta</assigned_to>
          
          
          <qa_contact name="qa-dev tracking">www-validator-cvs</qa_contact>

      

      

      

          <comment_sort_order>oldest_to_newest</comment_sort_order>  
          <long_desc isprivate="0" >
    <commentid>2373</commentid>
    <comment_count>0</comment_count>
    <who name="Mark Ludwig">uldmjl</who>
    <bug_when>2004-09-27 17:05:01 +0000</bug_when>
    <thetext>128.30.52.13 - - [27/Sep/2004:11:45:30 -0400] &quot;GET /robots.txt HTTP/1.1&quot; 200 26

Why did your site probe the robots.txt file on my server
ublib.buffalo.edu 120 times this morning between 8:55 and 11:45?

It has been doing this since last week and it keeps on probing
whether or not the robots.txt file exists.</thetext>
  </long_desc><long_desc isprivate="0" >
    <commentid>2470</commentid>
    <comment_count>1</comment_count>
    <who name="Olivier Thereaux">ot</who>
    <bug_when>2004-10-06 02:00:00 +0000</bug_when>
    <thetext>Most likely this is the Link Checker doing this, not the CSS validator.
Our site is not &quot;probing&quot; yours. 

Someone (possibly someone local to you, or someone with a site linking to yours) is certainly checking 
links to your site, and the link checker is following the robots exclusion protocol and doing your server 
a favor in doing so.

That said, a possible enhancement would be that the link checker cache the existence/lack of robots.txt 
for a given site instead of querying for it again and again.

Reassigning to proper product and owner.</thetext>
  </long_desc><long_desc isprivate="0" >
    <commentid>2499</commentid>
    <comment_count>2</comment_count>
    <who name="Ville Skyttä">ville.skytta</who>
    <bug_when>2004-10-11 22:13:49 +0000</bug_when>
    <thetext>Right, the /robots.txt fetches should be cached, and actually as far as the low
level implementation (LWP::RobotUA) is concerned, they _are_ cached.

But in the current version of the current link checker codebase, we&apos;re
instantiating several W3C::UserAgent (a superclass of LWP::RobotUA) objects per
link checker run, and the /robots.txt information cache is not shared between
these instances by default; instead, every one of them maintains its own small
cache, practically resulting in very little caching, if at all :(

The real fix would be to instantiate exactly one W3C::UserAgent per link checker
run and use that for fetching all links (unless we want to do parallel fetching
sometime), but that is a very intrusive change and will most likely have to wait
until the next major link checker version.

However, I believe it is possible to come up with an interim solution by
managing a &quot;global&quot; WWW::RobotRules object ourselves and passing that to all
instantiated UserAgents.  I&apos;ll look into it.</thetext>
  </long_desc><long_desc isprivate="0" >
    <commentid>2507</commentid>
    <comment_count>3</comment_count>
    <who name="Ville Skyttä">ville.skytta</who>
    <bug_when>2004-10-12 10:36:04 +0000</bug_when>
    <thetext>Turns out to be that the most trivial one of the workarounds is not possible due
to a bug in upstream WWW::RobotRules.  Fix for that already sent to libwww-perl
mailing list, no comments yet; will think about other workaround alternatives in
the meantime.</thetext>
  </long_desc><long_desc isprivate="0" >
    <commentid>2594</commentid>
    <comment_count>4</comment_count>
    <who name="Ville Skyttä">ville.skytta</who>
    <bug_when>2004-11-07 09:48:05 +0000</bug_when>
    <thetext>Fixed in CVS by using the same W3C::UserAgent instance for all retrievals.
It ain&apos;t pretty, but it works...</thetext>
  </long_desc>
      
      

    </bug>

</bugzilla>