![]() ![]() So why should I care? For me, it's a quick way to find pages or branches of the site that might have been removed from navigation, but might still be on disk. Looking at the file for, we see a number of entries that don't have lastmod or frequency or priority, just the url's (at least at the head of the file anyway). Note that it's created via a script, so as noted it should be updated automagically, but look at the date in the metadata - it's almost 2 years old! (this is NOT unusual) So what do these look like? Let's look at the SANS.ORG file. Either way, normally these are in a cron job or at least in the deployment workflow, so that they get updated either daily or whenever one of your dev's pushes something new to prod. On a more complex site, you'll often see a sitemap_index.xml file that in turn points to a number of other xml files that each cover one aspect or area of your site. The sitemap.xml is basically a set of "signposts" in a file to tell search engines which pages on your site are most important to index. But what about sitemap.xml, what's that for? ![]() Everyone seems to be familiar with robots.txt - the contents of that file are normally used to tell search engines what branches of your site to NOT index, or in some misguided cases folks think that this can be used to "secure" your pages for some reason. ![]()
0 Comments
Leave a Reply. |
Details
AuthorWrite something about yourself. No need to be fancy, just an overview. ArchivesCategories |