Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Using a third party server to host site elements
-
Hi guys -
I have a client who are recently experiencing a great deal of more traffic to their site. As a result, their web development agency have given them a server upgrade to cope with the new demand.
One thing they have also done is put all website scripts, CSS files, images, downloadable content (such as PDFs) - onto a 3rd party server (Amazon S3). Apparently this was done so that my clients server just handles the page requests now - and all other elements are then grabbed from the Amazon s3 server. So basically, this means any HTML content and web pages are still hosted through my clients domain - but all other content is accessible through an Amazon s3 server URL.
I'm wondering what SEO implications this will have for my clients domain? While all pages and HTML content is still accessible thorugh their domain name, each page is of course now making many server calls to the Amazon s3 server through external URLs (s3.amazonaws.com).
I imagine this will mean any elements sitting on the Amazon S3 server can no longer contribute value to the clients SEO profile - because that actual content is not physically part of their domain anymore. However what I am more concerned about is whether all of these external server calls are going to have a negative effect on the web pages value overall. Should I be advising my client to ensure all site elements are hosted on their own server, and therefore all elements are accessible through their domain?
Hope this makes sense (I'm not the best at explaining things!)
-
Hello Zeal Digital,
I use a CDN (Content Delivery Network) for images, CSS and javascript.
Doing that adds only about $10 to cost per month for a site that had around 800,000 pageviews per month.
You have complete control over the images. If there is a problem, you can force the CDN to flush a file and reload it from the source. You add code to your .htaccess file that tells the CDN how long to store images before fluching them and getting a new copy. It is all automated, there is generally no work for you to do. I host with softlayer.com and this is part of their service.
The change from self-sourced images, css and scripts had a massive improvement on the server.
- it is a 16-processor linux box with twin 15,000rpm SCSI drives and 12Gb RAM - it is quite fast!
Page delivery times improved by 1-2 seconds.
The server now is so lightly loaded that it could be downgraded to save more money.
It has zero effect on SEO. The CDN is accessed using a CNAME.
- static.domain.com - so don't worry about it looking like components are from other places.
The CDN has servers all over the world, so no matter where the visitors are, it is only a few hops for them to get most of the content, making it much faster for someone in Australia who would normally pull images from a server in the USA.
Your only problem with Amazon S3 is that they have crashed it a few times, but other than that, it is a good thing to do.
I wouldn't advise them to self-host, unless you want to increase their costs, server loading and page delivery times.
-
Great advice, cheers Jeffery!
-
I work with a number of high traffic sites (TB's of data each day, 10's millions page views/month). With many of these sites, we have offloaded static content to either dedicated static content servers (typically cloud based so we can scale up and down) or to content deliver networks. I've not had anyone report any SEO impact.
In contrast, they often see user engagement (page views/user), repeat visitors, and other traffic metrics improve. Users like fast sites. Also, Google apparently likes fast sites too, so while I've not seen it, you could actually get a boost in your SERPs due to faster loading pages.
If you break down a modern web page, you will find numerous elements required. Dozens of images, CSS, javascript as well as the page itself. All of these items require a request to the web server.
With some graphic intensive sites, I've seen as much as 95% of all web server requests (HTTP requests) be attributable to static content. By moving these HTTP requests to other systems, you free your primary server to handle the application. This provides a better user experience and improves scalability.
Content Delivery Networks
I do not use Amazon's Web Services so I do not know specifically what they offer. But here are two CDN's Ihave used with good success:
Internap:
http://www.internap.com/cdn-services-content-delivery-network/
Edgecast:
One method I look for is called "origin pull." With this method, you do not have to upload files to the CDN. The CDN will fetch them automatically from your site as needed. I found this is much easier to manage on sites that have frequent content updates.
-
Hosting images externally never had any impact on cases I had a chance to observe. The only problem I can think of is that you lose control over loading times or if somebody takes an image and links (credits) the image hosting domain instead of your domain.
-
Couple of notes for you
- There isn't any SEO impact on WHERE the data is loaded from. Look at any major website (especially one that ranks well) and they're openly using content delivery (like Akamai, Amazon S3/Cloudfront, etc) for static content. This is good business practice because it takes that load off your web server and often places the content closer to where the client is. Faster content delivery can help SEO if you have a slow server.
- If they're using the raw S3 buckets I would HIGHLY suggest signing up for Cloudfront. There's two benefits to doing this. First, you put the content into Amazon's cloud, where it is more readily available. Second, you can use domain aliasing to help obscure the source. For instance, let's say you have an images bucket. You could add a CNAME DNS record for images.yourdomain.com and then put that into your source code. You can still see where the DNS takes you, but it's not obvious to the general public. The cost difference between raw S3 delivery and Cloudfront is negligible.
Oh, and I use Amazon Cloudfront for my delivery. Never had any SEO issues with doing so.
-
I don't recomend to have the resources and database to other server than files, it makes some flood traffic between servers, the resources are harder to load and the site optimum speed is decreased. Also you can't compress this content so they are downloaded independently.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
In writing the url, it is better to use the language used by the people of my country or English?
We speak Persian and all people search in Persian on Google. But I read in some sources that the url should be in English. Please tell me which language to use for url writing?
Technical SEO | | ghesta
For example, I brought down two models: 1fb0e134-10dc-4737-904f-bfdf07143a98-image.png https://ghesta.ir/blog/how-to-become-rich/
2)https://ghesta.ir/blog/چگونه-پولدار-شویم/0 -
Sitemap use for very large forum-based community site
I work on a very large site with two main types of content, static landing pages for products, and a forum & blogs (user created) under each product. Site has maybe 500k - 1 million pages. We do not have a sitemap at this time.
Technical SEO | | CommManager
Currently our SEO discoverability in general is good, Google is indexing new forum threads within 1-5 days roughly. Some of the "static" landing pages for our smaller, less visited products however do not have great SEO.
Question is, could our SEO be improved by creating a sitemap, and if so, how could it be implemented? I see a few ways to go about it: Sitemap includes "static" product category landing pages only - i.e., the product home pages, the forum landing pages, and blog list pages. This would probably end up being 100-200 URLs. Sitemap contains the above but is also dynamically updated with new threads & blog posts. Option 2 seems like it would mean the sitemap is unmanageably long (hundreds of thousands of forum URLs). Would a crawler even parse something that size? Or with Option 1, could it cause our organically ranked pages to change ranking due to Google re-prioritizing the pages within the sitemap?
Not a lot of information out there on this topic, appreciate any input. Thanks in advance.0 -
Duplicate content on job sites
Hi, I have a question regarding job boards. Many job advertisers will upload the same job description to multiple websites e.g. monster, gumtree, etc. This would therefore be viewed as duplicate content. What is the best way to handle this if we want to ensure our particular site ranks well? Thanks in advance for the help. H
Technical SEO | | HiteshP0 -
Staging site and "live" site have both been indexed by Google
While creating a site we forgot to password protect the staging site while it was being built. Now that the site has been moved to the new domain, it has come to my attention that both the staging site (site.staging.com) and the "live" site (site.com) are both being indexed. What is the best way to solve this problem? I was thinking about adding a 301 redirect from the staging site to the live site via HTACCESS. Any recommendations?
Technical SEO | | melen0 -
How to create site map for large site (ecommerce type) that has 1000's if not 100,000 of pages.
I know this is kind of a newbie question but I am having an amazing amount of trouble creating a sitemap for our site Bestride.com. We just did a complete redesign (look and feel, functionality, the works) and now I am trying to create a site map. Most of the generators I have used "break" after reaching some number of pages. I am at a loss as to how to create the sitemap. Any help would be greatly appreciated! Thanks
Technical SEO | | BestRide0 -
What is the best way to find missing alt tags on my site (site wide - not page by page)?
I am looking to find all the missing alt tags on my site at once. I have a FF extension that use to do it page by page, but my site is huge and that will take forever. Thanks!!
Technical SEO | | franchisesolutions1 -
Hosting sitemap on another server
I was looking into XML sitemap generators and one that seems to be recommended quite a bit on the forums is the xml-sitemaps.com They have a few versions though. I'll need more than 500 pages indexed, so it is just a case of whether I go for their paid for version and install on our server or go for their pro-sitemaps.com offering. For the pro-sitemaps.com they say: "We host your sitemap files on our server and ping search engines automatically" My question is will this be less effective than my installing it on our server from an SEO perspective because it is no longer on our root domain?
Technical SEO | | design_man0 -
What are the pros and cons of moving one site onto a subdomain of another site?
Two sites. One has weaker sales. What would the benefits and problems for SEO of moving the weak site from its own domain to a subdomain of the stronger site?
Technical SEO | | GriffinHansen0