No announcement yet.

two domains one site

  • Filter
  • Time
  • Show
Clear All
new posts

  • two domains one site

    I currently have two separate sites using two different cart systems. Each one has it's own domain. Domain A is using 3dCart, I would like to have domain B redirected to domain A when someone clicks on a link or types in the domain. So both domains will point to the same site using 3dCart. Is this possible, if so would it affect Google rankings with two domains pointing to the same site?

    Currently both sites have very similar products and I am finding it hard to keep up with both sites.
    Kay Speranza
    Dish Girl LLC
    Fiesta Dinnerware

  • #2
    I'm thinking you might have to get a separate hosting account for Domain B (after you ditch the shopping cart hosting for that domain), just something like a cheap GoDaddy hosting account, but one that allows you to do redirects. Then redirect each page in Domain B to the corresponding page in Domain A. I think the only way to keep from being penalized by the search engines is to do 301 (permanent) redirects, but I could be wrong. I'm not completely sure this is the correct solution, but it should work.


    • #3
      301 redirects are safer.

      Having a separate 3dcart site for your other domain will keep you SEO Juice. And may be increase your overall sales.

      The goal is not to have duplicate content, google looks down on that. And also anything they feel is spammy...


      • #4
        Regarding duplicate content, I have heard if you own the domains that have duplicate content you will not be penalized. Anyone know if this is true?


        • #5
          read this here>>> Article

          Duplicate content generally refers to substantive blocks of content within or across domains that either completely match other content or are appreciably similar. Mostly, this is not deceptive in origin. Examples of non-malicious duplicate content could include:

          * Discussion forums that can generate both regular and stripped-down pages targeted at mobile devices
          * Store items shown or linked via multiple distinct URLs
          * Printer-only versions of web pages

          If your site contains multiple pages with largely identical content, there are a number of ways you can indicate your preferred URL to Google. (This is called "canonicalization".) More information about canonicalization.

          However, in some cases, content is deliberately duplicated across domains in an attempt to manipulate search engine rankings or win more traffic. Deceptive practices like this can result in a poor user experience, when a visitor sees substantially the same content repeated within a set of search results.

          Google tries hard to index and show pages with distinct information. This filtering means, for instance, that if your site has a "regular" and "printer" version of each article, and neither of these is blocked with a noindex meta tag, we'll choose one of them to list. In the rare cases in which Google perceives that duplicate content may be shown with intent to manipulate our rankings and deceive our users, we'll also make appropriate adjustments in the indexing and ranking of the sites involved. As a result, the ranking of the site may suffer, or the site might be removed entirely from the Google index, in which case it will no longer appear in search results.

          There are some steps you can take to proactively address duplicate content issues, and ensure that visitors see the content you want them to.

          * Use 301s: If you've restructured your site, use 301 redirects ("RedirectPermanent") in your .htaccess file to smartly redirect users, Googlebot, and other spiders. (In Apache, you can do this with an .htaccess file; in IIS, you can do this through the administrative console.)

          * Be consistent: Try to keep your internal linking consistent. For example, don't link to and and

          * Use top-level domains: To help us serve the most appropriate version of a document, use top-level domains whenever possible to handle country-specific content. We're more likely to know that contains Germany-focused content, for instance, than or

          * Syndicate carefully: If you syndicate your content on other sites, Google will always show the version we think is most appropriate for users in each given search, which may or may not be the version you'd prefer. However, it is helpful to ensure that each site on which your content is syndicated includes a link back to your original article. You can also ask those who use your syndicated material to use the noindex meta tag to prevent search engines from indexing their version of the content.

          * Use Webmaster Tools to tell us how you prefer your site to be indexed: You can tell Google your preferred domain (for example, or

          * Minimize boilerplate repetition: For instance, instead of including lengthy copyright text on the bottom of every page, include a very brief summary and then link to a page with more details. In addition, you can use the Parameter Handling tool to specify how you would like Google to treat URL parameters.

          * Avoid publishing stubs: Users don't like seeing "empty" pages, so avoid placeholders where possible. For example, don't publish pages for which you don't yet have real content. If you do create placeholder pages, use the noindex meta tag to block these pages from being indexed.

          * Understand your content management system: Make sure you're familiar with how content is displayed on your web site. Blogs, forums, and related systems often show the same content in multiple formats. For example, a blog entry may appear on the home page of a blog, in an archive page, and in a page of other entries with the same label.

          * Minimize similar content: If you have many pages that are similar, consider expanding each page or consolidating the pages into one. For instance, if you have a travel site with separate pages for two cities, but the same information on both pages, you could either merge the pages into one page about both cities or you could expand each page to contain unique content about each city.

          Google no longer recommends blocking crawler access to duplicate content on your website, whether with a robots.txt file or other methods. If search engines can't crawl pages with duplicate content, they can't automatically detect that these URLs point to the same content and will therefore effectively have to treat them as separate, unique pages. A better solution is to allow search engines to crawl these URLs, but mark them as duplicates by using the rel="canonical" link element, the URL parameter handling tool, or 301 redirects. In cases where duplicate content leads to us crawling too much of your website, you can also adjust the crawl rate setting in Webmaster Tools.

          Duplicate content on a site is not grounds for action on that site unless it appears that the intent of the duplicate content is to be deceptive and manipulate search engine results. If your site suffers from duplicate content issues, and you don't follow the advice listed above, we do a good job of choosing a version of the content to show in our search results.

          However, if our review indicated that you engaged in deceptive practices and your site has been removed from our search results, review your site carefully. If your site has been removed from our search results, review our Webmaster Guidelines for more information. Once you've made your changes and are confident that your site no longer violates our guidelines, submit your site for reconsideration.

          If you find that another site is duplicating your content by scraping (misappropriating and republishing) it, it's unlikely that this will negatively impact your site's ranking in Google search results pages. If you do spot a case that's particularly frustrating, you are welcome to file a DMCA request to claim ownership of the content and request removal of the other site from Google's index.
          Last edited by TurnkeyDropshipBusiness; 03-28-2011, 10:02 PM. Reason: url wrong
          Want 1000 Facebook Likes or Shares? G+1? Tweets? 3Dcart Design and Tweaking - Product Data Solutions - Custom Excel Upload Sheets - SEO Services - FREE Blog Hosting.


          • #6
            How to fix IP Canonicalization?

            I thought by having it checked out if my 3dcart amin setting to enable Enable Canonical URLs , that this issue of IP Canonicalization would not exist. However, after checking my site in another webtool its stating it's not "ON" and states the following: " Be careful you IP is not forwarding your website's domain name. Search engines like Goolge are now able to index your website with its IP instead of its domain name. This will result in duplicate content. You should do an htaccess 301 redirect to make sure the IP doesn't get indexed."- Can this be fixed within the 3dcart setting or will a web developer/designer need to access my account via ftp file and set up a 301 redirect?