30 Mar '13, 5am
800,000 pages blocked by robots... #seo #seomoz
I have experienced something similar after a site redesign the test version was put live with the robots.txt disallowing all. My site was deindexed quickly as when you block pages with a robots.txt their page content wont be indexed so won't appear in the search results. Google may index urls that are disallowed if they are linked from another page online however the rank will be lower due to the page content being ignored. Remove your robots.txt above as it is disallowing all It would appear although that command should allow all but there is no point in having robot file allowing all as this happens without. Though you would usually have disallow: / - to stop all!! Then I would resubmit an updated site map in Google Webmaster Tools and you should see your pages start to be indexed again. If you don't have a site map you can just wait for Google to start to re crawl your ...