Google Search Engine Optimisation Starter Guide Manuel d'utilisateur Page 17

  • Télécharger
  • Ajouter à mon manuel
  • Imprimer
  • Page
    / 22
  • Table des matières
  • MARQUE LIVRES
  • Noté. / 5. Basé sur avis des utilisateurs
Vue de la page 16
Google's Search Engine Optimisation Starter Guide, Version 1.1, 13 Nov 2008, latest version at Google Webmaster Central
Use commonly supported file types - Most browsers support JPEG, GIF, PNG and BMP
image formats. It's also a good idea to have the extension of your file name match the file
type.
Make effective use of robots.txt
A "robots.txt" file tells search engines whether they can access and therefore crawl parts of your site.
This file, which must be named "robots.txt", is placed in the root directory of your site.
The address of our robots.txt file
All compliant search engine bots (denoted by the wildcard * symbol) shouldn't access and crawl
the content under /images/ or any URL whose path begins with /search
You may not want certain pages of your site crawled because they might not be useful to users if found
in a search engine's search results. If you do want to prevent search engines from crawling your pages,
Google Webmaster Tools has a friendly robots.txt generator to help you create this file. Note that if your
site uses subdomains and you wish to have certain pages not crawled on a particular subdomain, you'll
have to create a separate robots.txt file for that subdomain.For more information on robots.txt, we
suggest this Webmaster Help Centre guide on using robots.txt files.
There are a handful of other ways to prevent content appearing in search results, such as adding
"NOINDEX" to your robots meta tag, using .htaccess to password-protect directories and using Google
Webmaster Tools to remove content that has already been crawled. Google engineer, Matt Cutts, walks
through the caveats of each URL-blocking method in a helpful video.
Good practices for robots.txt
Use more secure methods for sensitive content - You shouldn't feel comfortable using
robots.txt to block sensitive or confidential material. One reason is that search engines could
still reference the URLs that you block (showing just the URL, no title or snippet) if there
happen to be links to those URLs somewhere on the Internet (e.g. referrer logs). Also, non-
compliant or rogue search engines that don't acknowledge the Robots Exclusion Standard
could disobey the instructions of your robots.txt. Finally, a curious user could examine the
Vue de la page 16
1 2 ... 12 13 14 15 16 17 18 19 20 21 22

Commentaires sur ces manuels

Pas de commentaire