Hello and welcome to our community! Is this your first visit?
Register
Enjoy an ad free experience by logging in. Not a member yet? Register.
Results 1 to 7 of 7
  1. #1
    New Coder
    Join Date
    Oct 2002
    Location
    Mississisppi
    Posts
    14
    Thanks
    0
    Thanked 0 Times in 0 Posts

    Is it possible to tell when Google

    Cache's a website. I thought I saw someone post that google cache their site on such and such date. Is it possible to tell when they visit your website?

  • #2
    Regular Coder
    Join Date
    Jun 2002
    Location
    NL
    Posts
    421
    Thanks
    0
    Thanked 0 Times in 0 Posts
    Sorry, no. You can't tell any SE spider when to visit your site or not.
    You can block them, try to prevent them from caching your pages but you can't tell them when they're welcome or not. They just come and go when they want.

    A link: http://www.google.com/webmasters/

  • #3
    New Coder
    Join Date
    Oct 2002
    Location
    Mississisppi
    Posts
    14
    Thanks
    0
    Thanked 0 Times in 0 Posts
    Well thanks for that info MCookie

  • #4
    Regular Coder
    Join Date
    Jun 2002
    Location
    Dallas, Texas
    Posts
    188
    Thanks
    0
    Thanked 0 Times in 0 Posts
    sure! look at the bottom of a search result, and you'll see the "cache" button. that's there cache of your site.

  • #5
    Regular Coder
    Join Date
    Jun 2002
    Location
    NL
    Posts
    421
    Thanks
    0
    Thanked 0 Times in 0 Posts
    Getting your site spidered by Google and/or letting Google cache your pages are two different things.

    As far as Google concerns, you can
    A. Keep the bot away with a robots.txt or meta tag.
    Pages won't get indexed nor cached and won't be found in a search.

    In your robots.txt:
    User-agent: Googlebot
    Disallow: /

    Or meta tag:
    <META NAME="ROBOTS" CONTENT="NOINDEX, NOFOLLOW">

    B. Get your site spidered but not cached.
    Your pages will get indexed and show up in de search results without a link to cached versions.

    All robots:
    <META NAME="ROBOTS" CONTENT="NOARCHIVE">

    Google only:
    <META NAME="GOOGLEBOT" CONTENT="NOARCHIVE">

  • #6
    jkd
    jkd is offline
    Senior Coder jkd's Avatar
    Join Date
    May 2002
    Location
    metro DC
    Posts
    3,163
    Thanks
    1
    Thanked 18 Times in 18 Posts
    I'd look for their spider in your logs. They might specify something special in the user-agent if it is caching, and then all you need is a server-side script to look for that, and save the time somewhere.

  • #7
    Regular Coder
    Join Date
    Jun 2002
    Location
    NL
    Posts
    421
    Thanks
    0
    Thanked 0 Times in 0 Posts
    No, you can't tell from the user-agent if it's going to cache a page or not, but as far as I know, Google is the only SE with links to cached versions of pages in the search results. I wouldn't block their spiders but if you don't want older versions of your pages in their cache, the noarchive meta tag may be of use.

    Although I know for sure very, very few people actually use the 'cached' link.


  •  

    Posting Permissions

    • You may not post new threads
    • You may not post replies
    • You may not post attachments
    • You may not edit your posts
    •