Jump to content

How to tell Google (or Prestashop) to index English pages only?


Recommended Posts

Hello,

 

My target customers are english speakers for the most part, and I did SEO on english pages of my website only. However, I noticed that Google indexed the other languages that Prestashop in natively providing. As a result, it may lower my Google ranking because pages appear as duplicated contents. I don't want to remove other languages from Prestashop as I don't like to exclude other potential customers who may require a different language setting. I just would like Google to see my english pages. How can I do that?

 

Thanks

 

P.S: It seems Prestashop is creating two sets of pages in English though. One looks like www.example.com/stuff, and the other one is www.example.com/en/stuff. I guess there's one that is not needed there too!

 

 

Link to comment
Share on other sites

if you are only 'concerned about google', then log in (or create) to google webmaster tools and ask google to request removal of URL(s) from search results.

 

10.11.2013-11.15.png
http://screencast.com/t/vU1KvrATp

 

here is what google says on this matter

https://support.google.com/webmasters/answer/59819?hl=en

 

please note: good bots are supposed to follow robots.txt rules, but don't have to.

Link to comment
Share on other sites

now, I can ask this question, why are you doing this exactly? 

 

Are the translations for the other languages complete?

 

Are you worried about duplicate content?

 

If there is an issue with your sitemap, please post in seo section and post link here so we can review it.

Link to comment
Share on other sites

I found something interesting that pertains to no indexing.

Header.tpl

        <meta name="robots" content="{if isset($nobots)}no{/if}index,{if isset($nofollow) && $nofollow}no{/if}follow" />

 

but I don't see where PrestaShop set's nobots.

 

Vekia?
 

Link to comment
Share on other sites

User-agent: *

Disallow: /dir1/

Disallow: /dir2/

Disallow: /dir3/ 

 

where dir1 dir2 and dir3 are language codes that you want to disallow for crawling purposes

 

Vekia, i didn't find the module for that. Could you tell me how you disallow languages?

Link to comment
Share on other sites

In the weblink you shared, Google mentions specific patterns only designed for googlebot. The instance to block all subdirectories is "/private*/".

 

However, in my robots.txt files, it appears directories are disallowed as followed: "/*mails/".

 

That's the reason, I am asking the question in my previous post.

Link to comment
Share on other sites

in the future when you are asked particular questions, like why are you doing this?, you should answer because it might be in your best interest.  adding a disallow to robots.txt is very simple..much like the links explained.

 

here is what you would do, if you already have the User-agent: *, you will not need to add that.

User-agent: *Disallow: /folder1/
Link to comment
Share on other sites

  • 1 year later...
×
×
  • Create New...