My website hasn't been cached for over a month. Can anyone tell me why?
-
I have been working on an eCommerce site www.fuchia.co.uk.
I have asked an earlier question about how to get it working and ranking and I took on board what people said (such as optimising product pages etc...) and I think i'm getting there.
The problem I have now is that Google hasn't indexed my site in over a month and the homepage cache is 404'ing when I check it on Google. At the moment there is a problem with the site being live for both WWW and non-WWW versions, i have told google in Webmaster what preferred domain to use and will also be getting developers to do 301 to the preferred domain. Would this be the problem stopping Google properly indexing me? also I'm only having around 30 pages of 137 indexed from the last crawl.
Can anyone tell me or suggest why my site hasn't been indexed in such a long time?
Thanks
-
Fair point about the Sitemap. Thanks a lot, I'll take these on board and see what happens from there.
Thanks,
-
Cache won't be built or updated overnight so sometimes the first few caches are a waiting game. How long has this site been live? If it's fairly new, what you're experiencing is common. If it's an older site and you recently started changing a lot of the technical stuff - redirecting, canonicals, etc. it may just take a little while to settle in.
The other major recommendation I would give you is to change your sitemap "change frequency" to be slightly more accurate. Does this page http://www.fuchia.co.uk/products/clothing/dresses/dog-tooth-print-dress.aspx really change "daily"? By having daily on every page you aren't helping Google prioritize their crawl, which means you may get a cache for your dog tooth print dress before you get a new cache for your main page.
So I would fix that, resubmit sitemap and then it's a waiting game. Could be a week, could be two, I've seen it go almost a month but not if you use G+.
-
Hi Matt,
I used ping device and it's pinging fine.
I will work on the Google+ suggestion.
I have resubumitted a Sitemap for both fuchia.co.uk and www.fuchia.co.uk as I verified ownership of both to allow me set preferred domain. I submitted one this morning, so maybe that will help. But we will see.
It seems like the main priority at the moment is getting everything redirected and canonicalised and see if that helps anything.
-
Hi Sanket,
The site has been live for around 3 months I would say.
-
I've found that if you manually ping Google, they often update their cache at the same time.
Google doesn't have a cache for either cache: www.fuchia.co.uk. or cache: fuchia.co.uk. so I don't think it's a canonical issue.
I would suggest a few things:
-
Use PingDevice http://www.pingdevice.com/
-
Put your main domain in a Google Plus post every now and then.
-
Resubmit a sitemap. Usually this gets you crawled fairly quickly and possibly updates your cache.
-
-
Hi,
Your site is open with or without WWW so it is major problem you have to do proper 301 redirect in .htaccess file. Need to implement rel=canonical into your site i did not find that code. I see 243 pages are indexed of your site by google. can i know about the domain edge of your site?? when you have live this site?
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Does product environment have impact on main website's SEO
We have two environments - product, where login is necessary and where the customers are working. We also have there our help desk, Q&A and knowledge base. Pretty sophisticated page regarding information on a specific topic. We also have our main page where we promote our products, company and events, etc. Main page is www.example.com, where product environment is login.example.com . Does this product environment have an impact on my main page's SEO?
Intermediate & Advanced SEO | | NeringaA0 -
Would you redirect Website A to Website B, when Website B is in the middle of a HTTP=>HTTPS migration?
Hey guys, I'm curious on your thoughts around this scenario... Website A: 35,000 monthly pageviews 1,000 pages 375 root linking domains currently HTTPS focused on one topic weak rankings for competitive keywords Website B: 3M monthly pageviews 32,500 pages 3,500 root linking domains started HTTP to HTTPS migration 1 week ago. 1/3 of pages indexed as HTTPS. focused on many topics strong rankings for competitive keywords Requirement: I want to have a reliable read on how Website A's keyword rankings change after redirecting it's pages to Website A. This post-migration analysis will be used as a basis to assess the risk of redirecting another website we own that is similar to Website A into Website B. My question: Would you wait until most of the pages on Website B are indexed as HTTPS before doing a 301 of Website A to Website B? Please back up your answer with reasons why or why not 🙂
Intermediate & Advanced SEO | | jeremycabral0 -
Do 404s really 'lose' link juice?
It doesn't make sense to me that a 404 causes a loss in link juice, although that is what I've read. What if you have a page that is legitimate -- think of a merchant oriented page where you sell an item for a given merchant --, and then the merchant closes his doors. It makes little sense 5 years later to still have their merchant page so why would removing them from your site in any way hurt your site? I could redirect forever but that makes little sense. What makes sense to me is keeping the page for a while with an explanation and options for 'similar' products, and then eventually putting in a 404. I would think the eventual dropping out of the index actually REDUCES the overall link juice (ie less pages), so there is no harm in using a 404 in this way. It also is a way to avoid the site just getting bigger and bigger and having more and more 'bad' user experiences over time. Am I looking at it wrong? ps I've included this in 'link building' because it is related in a sense -- link 'paring'.
Intermediate & Advanced SEO | | friendoffood0 -
Can anyone see any issues with the canonical tags on this web site?
The main domain is: http://www.eumom.ie/ And these would be some of the core pages: http://www.eumom.ie/pregnancy/ http://www.eumom.ie/getting-pregnant/ Any help from the Moz community is much appreciated!
Intermediate & Advanced SEO | | IcanAgency0 -
After Receiving a "Googlebot can't access your site" would this stop your site from being crawled?
Hi Everyone,
Intermediate & Advanced SEO | | AMA-DataSet
A few weeks ago now I received a "Googlebot can't access your site..... connection failure rate is 7.8%" message from the webmaster tools, I have since fixed the majority of these issues but iv noticed that all page except the main home page now have a page rank of N/A while the home page has a page rank of 5 still. Has this connectivity issues reduced the page ranks to N/A? or is it something else I'm missing? Thanks in advance.0 -
Preferred domain can't set in Web master Tool
I have put my domain name as xxxxxtours.com without www in web master tool. i have redirect to www version using htaccess file .So I wanna put Preferred domain "Display urls as www.xxxxtours.com .When trying it give error as attached image.but i have verified site the .waiting for expert help . Ar5qx.png
Intermediate & Advanced SEO | | innofidelity0 -
Googlebot Can't Access My Sites After I Repair My Robots File
Hello Mozzers, A colleague and I have been collectively managing about 12 brands for the past several months and we have recently received a number of messages in the sites' webmaster tools instructing us that 'Googlebot was not able to access our site due to some errors with our robots.txt file' My colleague and I, in turn, created new robots.txt files with the intention of preventing the spider from crawling our 'cgi-bin' directory as follows: User-agent: * Disallow: /cgi-bin/ After creating the robots and manually re-submitting it in Webmaster Tools (and receiving the green checkbox), I received the same message about Googlebot not being able to access the site, only difference being that this time it was for a different site that I manage. I repeated the process and everything, aesthetically looked correct, however, I continued receiving these messages for each of the other sites I manage on a daily-basis for roughly a 10-day period. Do any of you know why I may be receiving this error? is it not possible for me to block the Googlebot from crawling the 'cgi-bin'? Any and all advice/insight is very much welcome, I hope I'm being descriptive enough!
Intermediate & Advanced SEO | | NiallSmith1