Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Staging website got indexed by google
-
Our staging website got indexed by google and now MOZ is showing all inbound links from staging site, how should i remove those links and make it no index.
Note- we already added Meta NOINDEX in head tag
-
Hi Dera Moz My Domain Is 18 Years Old But Da is don't increased i don't know why can you please help me and check my url cigars please check sir
#mozda
-
Its good that you already put the Meta NOINDEX.
Now, you can ask to remove the url of website from google index. Visit the google search console and request the url removal.
You can use the URL Removal Tool in Google Search Console to request the removal of specific URLs from Google's index.
To use the URL Removal Tool, you can:
- Open the Removals tool.
- Select the Temporary Removals tab.
- Click New Request.
- Select Next to complete the process.
Warm Regards
Rahul Gupta
Suvidit Academy -
Sydney's Best Chauffeur Car Service | A1 Corporate Cars Au
Sydney's Best Chauffeur Car Service is a premier provider of corporate chauffeured cars in Sydney, Australia. We offer top-of [url=https://a1corporatecars.com.au/]corporate cars Australia[/url] transportation solutions for business professionals, executives, and VIP clients who demand the highest service and comfort. With a fleet of luxury vehicles and experienced professional chauffeurs, we ensure a seamless and luxurious travel experience for our esteemed customers.
-
If your staging website has been indexed by Google, it means that Google's web crawlers have discovered and added your staging site's pages to their search index. This is typically not desirable because staging websites are meant for testing and development purposes and often contain incomplete or confidential content.
To address this issue, you can take several steps. Firstly, ensure that your staging website has a "robots.txt" file configured properly. This file tells search engines which parts of your website to crawl and index. In the case of a staging site, you can disallow all web crawlers from indexing it by using a "robots.txt" file.
Another effective measure is to include a "noindex" meta tag in the HTML of your staging website's pages. This tag instructs search engines not to index the page, adding an extra layer of protection.
Consider password-protecting your staging website using HTTP authentication. This adds an additional layer of security and ensures that only authorized users can access the site.
To further mitigate indexing issues, you can set up your staging website on a subdomain or a subdirectory instead of a separate domain. Google is less likely to index staging content if it's located in a subdomain or subdirectory.
If your staging site is already indexed, you can request the removal of specific URLs from Google's index using the Google Search Console's URL Removal Tool. This is a more proactive approach to remove already indexed content.
Lastly, regularly monitor your staging website to ensure it remains hidden from search engines and that any changes to the robots.txt file or meta tags are being followed. It's a good practice to implement these measures before you create or launch a staging website to prevent it from being indexed in the first place.
Remember that it may take some time for Google to update its index and remove your staging site's pages. Be patient and continue to monitor the situation closely to ensure the desired results are achieved.
-
If a staging website (a non-production or testing version) gets indexed by Google, it can lead to privacy, user experience, and SEO issues. To address this, use methods like robots.txt, "noindex" meta tags, or password protection to prevent indexing. If already indexed, request removal through Google Search Console to ensure only the production site is visible in search results.
-
If your staging website has been indexed by Google, it means that Google's search engine has discovered and included your staging site in its search results. This is not an ideal situation since staging websites are usually intended for testing and development purposes, and you may not want them publicly accessible.
To address this issue, you can take a few steps:
Use a robots.txt file: Create a robots.txt file on your staging website and instruct search engines not to index it. This file specifies which areas of your site search engines should or should not crawl.
Add a noindex meta tag: Insert a "noindex" meta tag in the head section of your staging website's HTML. This tag tells search engines not to index that specific page.
Password protect your staging website: Implement password protection on your staging environment to ensure that only authorized users can access it. This can be done through various authentication methods, depending on your setup.
Remember that these steps can help prevent further indexing, but they may not immediately remove your staging site from the search results. It might take some time for search engines to re-crawl your site and recognize the changes you made.
-
If your staging website gets indexed by Google, you should take these steps:
( Atlantic Immigration Pilot Program application form)
Use a robots.txt file to disallow indexing.
Request removal of indexed pages via Google Search Console.
Canada PR
Add a "noindex, nofollow" meta tag to staging pages.
Consider password protecting the staging site.
Ensure canonical URLs point to the production site.
These actions will help prevent your incomplete or sensitive staging content from appearing in Google search results.
Best digital marketing agency -
If your staging website has been indexed by Google, it means that Google's search engine has crawled and added your staging site's pages to its search index. This is typically not desired because staging websites are not meant for public access and may contain incomplete or sensitive content.
To address this issue, you should take the following steps:
Disallow indexing: Use a robots.txt file to instruct search engines not to crawl and index your staging website. You can add the following lines to your robots.txt file to disallow all search engines:
makefile
Copy code
User-agent: *
Disallow: /
Place this robots.txt file in the root directory of your staging website.Remove indexed pages: You can request Google to remove indexed pages from its search results by using the Google Search Console's "Remove URLs" tool. Log in to your Google Search Console account, select your property, go to the "Index" section, and choose "Removals." From there, you can temporarily hide specific URLs from Google search results.
Use noindex meta tags: On your staging website's pages, you can add a meta tag to indicate that the page should not be indexed. Add the following meta tag within the HTML <head> section of each page you want to exclude:
html
Copy code
<meta name="robots" content="noindex, nofollow">
This tag tells search engines not to index the page or follow any links on it.Password protection: Consider adding password protection to your staging website, so only authorized users can access it. This adds an additional layer of security and privacy.
Update canonical URLs: Ensure that your staging website's canonical URLs (if used) point to the production website, not the staging one. This helps search engines understand the preferred version of your content.
After taking these steps, monitor your staging website to ensure it's no longer being indexed by Google. Keep in mind that it may take some time for changes to take effect and for Google to de-index your staging content.
-
@Asmi-Ta said in Staging website got indexed by google:
Our staging website got indexed by google and now MOZ is showing all inbound links from staging site, how should i remove those links and make it no index.
Note- we already added Meta NOINDEX in head tagTo remove indexed staging site links and prevent further indexing, take these steps: Add a "Disallow" rule for the staging site in your
robots.txt
file, use 301 redirects for indexed staging URLs to point to production, update all internal links to production URLs, request URL removals through Google Search Console's "Fetch as Google" and URL Removal Tool, submit an updated production sitemap, and monitor Google Search Console for updates. Be patient, as it may take time for search engines to de-index staging URLs and re-crawl your site. Ensure the staging site has a "noindex" tag in its<head>
section.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Why MOZ just index some of the links?
hello everyone i've been using moz pro for a while and found a lot of backlink oppertunites as checking my competitor's backlink profile.
Link Building | | seogod123234
i'm doing the same way as my competitors but moz does not see and index lots of them, maybe just index 10% of them. though my backlinks are commenly from sites with +80 and +90 DA like Github, Pinterest, Tripadvisor and .... and the strange point is that 10% are almost from EDU sites with high DA. i go to EDU sites and place a comment and in lots of case, MOZ index them in just 2-3 days!! with maybe just 10 links like this, my DA is incresead from 15 to 19 in less than one month! so, how does this "SEO TOOL" work?? is there anyway to force it to crawl a page?0 -
Advise on the right way to block country specific users but not block Googlebot - and not be seen to be cloaking. Help please!
Hi, I am working on the SEO of an online gaming platform - a platform that can only be accessed by people in certain countries, where the games and content are legally allowed.
International SEO | | MarkCanning
Example: The games are not allowed in the USA, but they are allowed in Canada. Present Situation:
Presently when a user from the USA visits the site they get directed to a restricted location page with the following message: RESTRICTED LOCATION
Due to licensing restrictions, we can't currently offer our services in your location. We're working hard to expand our reach, so stay tuned for updates! Because USA visitors are blocked Google which primarily (but not always) crawls from the USA is also blocked, so the company webpages are not being crawled and indexed. Objective / What we want to achieve: The website will have multiple region and language locations. Some of these will exist as standalone websites and others will exist as folders on the domain. Examples below:
domain.com/en-ca [English Canada]
domain.com/fr-ca [french Canada]
domain.com/es-mx [spanish mexico]
domain.com/pt-br [portugese brazil]
domain.co.in/hi [hindi India] If a user from USA or another restricted location tries to access our site they should not have access but should get a restricted access message.
However we still want google to be able to access, crawl and index our pages. Can i suggest how do we do this without getting done for cloaking etc? Would this approach be ok? (please see below) We continue to work as the present situation is presently doing, showing visitors from the USA a restricted message.
However rather than redirecting these visitors to a restricted location page, we just black out the page and show them a floating message as if it were a model window.
While Googlebot would be allowed to visit and crawl the website. I have also read that it would be good to put paywall schema on each webpage to let Google know that we are not cloaking and its a restricted paid page. All public pages are accessible but only if the visitor is from a location that is not restricted Any feedback and direction that can be given would be greatly appreciated as i am new to this angle of SEO. Sincere thanks,0 -
Who is correct - please help!
I have a website with a lot of product pages - often thousands of pages. As each of these pages is for a specific lease car they are often only fractionally different from other pages. The urls are too long, the H1 is often too long and the Title is often too long for "SEO best practice". And they do create duplication issues according to MOZ. Some people tell me to change them to noindex/nofollow whilst others tell me to leave them as they are as best not to hide from google crawler. Any advice will be gratefully received. Thanks for listening.
Technical SEO | | jlhitch0 -
Unsolved Using NoIndex Tag instead of 410 Gone Code on Discontinued products?
Hello everyone, I am very new to SEO and I wanted to get some input & second opinions on a workaround I am planning to implement on our Shopify store. Any suggestions, thoughts, or insight you have are welcome & appreciated! For those who aren't aware, Shopify as a platform doesn't allow us to send a 410 Gone Code/Error under any circumstance. When you delete or archive a product/page, it becomes unavailable on the storefront. Unfortunately, the only thing Shopify natively allows me to do is set up a 301 redirect. So when we are forced to discontinue a product, customers currently get a 404 error when trying to go to that old URL. My planned workaround is to automatically detect when a product has been discontinued and add the NoIndex meta tag to the product page. The product page will stay up but be unavailable for purchase. I am also adjusting the LD+JSON to list the products availability as Discontinued instead of InStock/OutOfStock.
Technical SEO | | BakeryTech
Then I let the page sit for a few months so that crawlers have a chance to recrawl and remove the page from their indexes. I think that is how that works?
Once 3 or 6 months have passed, I plan on archiving the product followed by setting up a 301 redirect pointing to our internal search results page. The redirect will send the to search with a query aimed towards similar products. That should prevent people with open tabs, bookmarks and direct links to that page from receiving a 404 error. I do have Google Search Console setup and integrated with our site, but manually telling google to remove a page obviously only impacts their index. Will this work the way I think it will?
Will search engines remove the page from their indexes if I add the NoIndex meta tag after they have already been index?
Is there a better way I should implement this? P.S. For those wondering why I am not disallowing the page URL to the Robots.txt, Shopify won't allow me to call collection or product data from within the template that assembles the Robots.txt. So I can't automatically add product URLs to the list.0 -
How to de-index old URLs after redesigning the website?
Thank you for reading. After redesigning my website (5 months ago) in my crawl reports (Moz, Search Console) I still get tons of 404 pages which all seems to be the URLs from my previous website (same root domain). It would be nonsense to 301 redirect them as there are to many URLs. (or would it be nonsense?) What is the best way to deal with this issue?
Intermediate & Advanced SEO | | Chemometec0 -
Best way to remove full demo (staging server) website from Google index
I've recently taken over an in-house role at a property auction company, they have a main site on the top-level domain (TLD) and 400+ agency sub domains! company.com agency1.company.com agency2.company.com... I recently found that the web development team have a demo domain per site, which is found on a subdomain of the original domain - mirroring the site. The problem is that they have all been found and indexed by Google: demo.company.com demo.agency1.company.com demo.agency2.company.com... Obviously this is a problem as it is duplicate content and so on, so my question is... what is the best way to remove the demo domain / sub domains from Google's index? We are taking action to add a noindex tag into the header (of all pages) on the individual domains but this isn't going to get it removed any time soon! Or is it? I was also going to add a robots.txt file into the root of each domain, just as a precaution! Within this file I had intended to disallow all. The final course of action (which I'm holding off in the hope someone comes up with a better solution) is to add each demo domain / sub domain into Google Webmaster and remove the URLs individually. Or would it be better to go down the canonical route?
Intermediate & Advanced SEO | | iam-sold0 -
How to get google to categorize a website in search results?
Hello everyone and thanks in advance for your time. I have a good understanding about SEO, backlinks etc but nowhere near to professional! A good friend of mine has an online store made with opencart e commerce platform he would like to have have category view when his company name is searched on google. Does anyone has any idea how can this be achieved?
Intermediate & Advanced SEO | | superofelia0 -
Best way to de-index content from Google and not Bing?
We have a large quantity of URLs that we would like to de-index from Google (we are affected b Panda), but not Bing. What is the best way to go about doing this?
Intermediate & Advanced SEO | | nicole.healthline0