Welcome to WebmasterWorld Guest from 54.159.179.132

Message Too Old, No Replies

Google WMT reports crawl errors on my PDFs

   
8:46 pm on Oct 6, 2011 (gmt 0)



In the last 3 months I have continually ran into a problem where many of the PDFs on my site keep coming up in the not found area of webmaster tools. These are valid links and I can load them every time I click them. I was wondering does Google bot have a problem with PDFs? Like is there a way I need to host or optimize them to get them to not error. This doesn't happen for all of my pdfs but a good amount.

Thanks
8:57 pm on Oct 6, 2011 (gmt 0)

WebmasterWorld Senior Member tedster is a WebmasterWorld Top Contributor of All Time 10+ Year Member



Do the URLs for those PDF errors contain any odd characters - things that need to be escaped (turned into coded entities) before the browser can send the request?
9:02 pm on Oct 6, 2011 (gmt 0)

WebmasterWorld Senior Member g1smd is a WebmasterWorld Top Contributor of All Time 10+ Year Member Top Contributors Of The Month



What format are the links pointing to those files?

Are they relative, root relative or absolute?

Are there any
../
constructs in those links?
9:40 pm on Oct 6, 2011 (gmt 0)



Thanks for the quick responses. Their are no special characters just alpha numeric upper and lower case and dashes. I wouldnt think these would be a problem right?

The links are relative (they dont include the domain name, i believe that means relative) and are like this example <a href="/pdf/file-name.pdf">
 

Featured Threads

Hot Threads This Week

Hot Threads This Month