Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
4XX Errors - Adding %5c%5c to Links
-
Hi all
Hope someone can help me with this.
The internal links on my hubby's business site occasionally break and add %5c%5c%5c endlessly to the end of the url - like this:
site.com/about/hours-of-operation/\\\\\\\\%
I cannot for the life of me figure out why it is doing this and while it has happened to me from time to time, I can't recreate it.
My crawl diagnostics here in my SEOMox campaign show 19-20 urls doing this - it's nuts.
Any insight?
Thank you!!
Jennifer
~PotPieGirl
-
Both Shane and I looked and neither of saw any / vs \ issues.
Then, I just took a peek at my source code and look what I saw:
http://screencast.com/t/y02R4RS2L
Think that is it?
Thanks for replying!
Jennifer
-
Sent you a PM, Shane - thanks so much for the offer!
By "occasionally break", I mean that every now and again, any link on the site will freak out and add the %5c jibberish to the end.
Jennifer
-
It would really have to be....
The only reason for a %5C is the use of backslash - as that is actually what it means in code.
**business site occasionally break **
How do you mean "break" ?
My crawl diagnostics here in my SEOMox campaign show 19-20 urls doing this - it's nuts.
Is the only issue from this in SEOMOZ reports?
If you would like to PM me the site I can attempt to profile it
-
Thanks for replying so quickly, Shane!
I don't believe it's a / vs \ issue. It's a wordpress site. Key pages are in the top nav bar (all urls correct) and all the side bar links are 'widgets' and created by Wordpress.
For some reason I am suspecting a theme issue, but if I can't recreate the error, I'll have no way of knowing if changing the theme solves the problem.
Site has been online since 2010 with no issues...this is a new issue (past couple months according to crawl diagnostics).
Thanks!!!
-
somewhere in your url coding you added backslash instead of slash which is not valid.
-
It appears from research it is actually \ (backslash) not / (slash)
So possibly somewhere in your site you have used \ instead of / but of course this is just a possibility, as it appears that %5C is the delimiter or interpretation of \ in original Unix.
Hope this helps
PS a quick check at http://www.w3schools.com/tags/ref_urlencode.asp verified %5c is URL encode, for backslash - as it is treated as a special character, so somewhere in your CMS or code, you have used backslash instead of slash.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Subdomain 403 error
Hi Everyone, A crawler from our SEO tool detects a 403 error from a link from our main domain to a a couple of subdomains. However, these subdomains are perfect accessibly. What could be the problem? Is this error caused by the server, the crawlbot or something else? I would love to hear your thoughts.
Technical SEO | | WeAreDigital_BE
Jens0 -
Schema Markup Errors - Priority or Not?
Greetings All... I've been digging through the search console on a few of my sites and I've been noticing quite a few structured data errors. Most of the errors are related to: hcard, hentry and hatom. Most of them are missing author & entry-title, while the other one is missing: fn. I recently saw an article on SEL about Google's focus on spammy mark-up. The sites I use are built and managed by vendors, so I would have to impress upon them the impact of these errors and have them prioritize, then fix them. My question is whether or not this should be prioritized? Should I have them correct these errors sooner than later or can I take a phased approach? I haven't noticed any loss in traffic or anything like that, I'm more focused on what negative impact a "phased approach" could have. Any thoughts?
Technical SEO | | AfroSEO0 -
:443 - 404 error
I get strange :443 errors in my 404 monitor on Wordpress https://www.compleetverkleed.nl:443/hoed-al-capone-panter-8713647758068-2/
Technical SEO | | Happy-SEO
https://www.compleetverkleed.nl:443/cart/www.compleetverkleed.nl/feestkleding
https://www.compleetverkleed.nl:443/maskers/ I have no idea where these come from :S2 -
Updating inbound links vs. 301 redirecting the page they link to
Hi everyone, I'm preparing myself for a website redesign and finding conflicting information about inbound links and 301 redirects. If I have a URL (we'll say website.com/website) that is linked to by outside sources, should I get those outside sources to update their links when I change the URL to website.com/webpage? Or is it just as effective from a link juice perspective to simply 301 redirect the old page to the new page? Are there any other implications to this choice that I may want to consider? Thanks!
Technical SEO | | Liggins0 -
Are 404 Errors a bad thing?
Good Morning... I am trying to clean up my e-commerce site and i created a lot of new categories for my parts... I've made the old category pages (which have had their content removed) "hidden" to anyone who visits the site and starts browsing. The only way you could get to those "hidden" pages is either by knowing the URLS that I used to use or if for some reason one of them is spidering in Google. Since I'm trying to clean up the site and get rid of any duplicate content issues, would i be better served by adding those "hidden" pages that don't have much or any content to the Robots.txt file or should i just De-activate them so now even if you type the old URL you will get a 404 page... In this case, are 404 pages bad? You're typically not going to find those pages in the SERPS so the only way you'd land on these 404 pages is to know the old url i was using that has been disabled. Please let me know if you guys think i should be 404'ing them or adding them to Robots.txt Thanks
Technical SEO | | Prime850 -
Self-referencing links
I personally think that self-referencing links are silly. It's blatantly easy for Google to tell and my instinct says that the link juice for this would simply evaporate rather than passing back to itself. Does anyone have information backing me up from an authoritative source? I can't find any info about this linked to Matt Cutts, Rand or any of those I look up to.
Technical SEO | | IPROdigital0 -
404 crawl errors from "tel:" link?
I am seeing thousands of 404 errors. Each of the urls is like this: abc.com/abc123/tel:1231231234 Everything is normal about that url except the "/tel:1231231234" these urls are bad with the tel: extension, they are good without it. The only place I can find this character string is on each page we have this code which is used for Iphones and such. What are we doing wrong? Code: Phone: <a href="[tel:1231231234](tel:7858411943)"> (123) 123-1234a>
Technical SEO | | EugeneF0 -
Does Yelp pass link juice?
This is probably a profoundly obvious question, but I can't seem to find an explicit answer on the internet, so I'll ask it here: Yelp's links out to local business websites are not nofollow'd, but they go through a javascript-based redirect. My understanding is that javascript redirected links do not pass link juice, so a link from a yelp profile will not directly impact my page authority; however, it looks like yelp does use nofollow judiciously for internal links, so I don't understand why they would allow follow for these "useless" outbound links. Do yelp's javascript-redirected links pass link juice?
Technical SEO | | tvkiley0