
Published: Sep 23, 2008 - 11:55 am
Story Found By: paisley 4543 Days ago
Category: SEO

Search Engine Land produces SMX, the Search Marketing Expo conference series. SMX events deliver the most comprehensive educational and networking experiences - whether you're just starting in search marketing or you're a seasoned expert.
Join us at an upcoming SMX event:
Learn more about search marketing with our free online webcasts and webinars from our sister site, Digital Marketing Depot. Upcoming online events include:
Comments
See also: Google says Stay Away From URL Rewriting! Seriously? started yesterday.
As long as Google isnt going to apply a penalty for continuing to use mod_rewrite Id consider continuing to use static looking pages, as from a user perspective, clean url names are just easier to remember...eg; blog/p2.php vs blog/some-descriptive-title/...and also pre-informs or reinforces what the page is about.
The tone of the original article seems to suggest that the issue here is data management, not SEO. Many large sites from respected organizations have trouble with multipe URLs for each page, sometimes linked to 3 different ways on the same page. Yes, Google can "ignore" 2 versions of the page, but if all other things are equal, then figuring out which one should be considered authorative could be a little like splitting hairs. Making it easy for these larger orgs to feel comfortable using the "naked" dynamic URL could reduce the net number of URLs googlebot has to crawl. I personally wouldnt get too upset about this.
Sure, lets all make it easier for Google, Ill remove rewrites if they give every site in the world +1 toolbar PR :)
So instead of Google making their search product work the way we do, they instead shift the goal posts and expect everyone to play their game? I think these guys need to get their collective act together -- Google != Web
I would definitely not get upset about this. G basically says: hey, a lot of people think we cant follow urls with parameters, but we can. So dont think you MUST rewrite URLs for better rankings. Heck, so many people do it wrong, so please dont do it if you dont know how.The rest of the post seems to me like a (somewhat clumsy) explanation of that with examples that generate more questions than answers
Yup, monchito. WayneSmallman, in my opinion what this post says is "We do a solid job on sites with dynamic parameters, and lots of people make mistakes when they try to rewrite their urls to look static, so you might want to try the dynamic parameter route because that can work quite well."<div></div><div>In essence, its Google saying "Well come to webmasters and the natural way to write dynamic parameters rather than asking you to rewrite everything as static if you dont want to." So were trying to come closer to webmasters, not wanting webmasters to necessarily move toward us. If you already have a site and its doing well the way that it currently is--great. In that case, you probably dont need to change anything. But if youre starting a new site, its worth considering staying with dynamic parameters instead of doing large amounts of rewrites (which some webmasters do in unusual ways that dont always work well in search engines). Thats my take, at least.</div><div></div><div>Its not like either choice would get you penalized in Google; all of this is just advice to give more information to webmasters when theyre making their choice of site architecture.</div>
@MattCutts - Thank you for the clarification sir.
Matt, I am having trouble wrapping my head around this because this for 1 simple reason, which this should be all about, the user. Rewriting URLs and creating a simple clean url structure is not just a tactic for webmasters, but a tactic used to create a simple url structure for users. e.g. Diggs URL Structure of http://digg.com/tech_news/upcoming/most/ and trimming back to http://digg.com/tech_news/upcoming/ or even further http://digg.com/tech_news/ for the popular queries.That is what confuses me about this post the most in that it basically is saying, lets focus on search engines and really in essense, we are hurting the user experience (or even forgetting about the user all together)I could be wrong in my conclusions gathered from the post, but, that is what I came to the conclusion of by reading the post.
Thanks MattWe run large website focused on SEO and Joomla ... this post really put the wind up quite a lot of our users, probably because its a "somewhat clumsy explanation" as mentioned earlier.Your explanation makes a lot more sense.
@tonyadam -- Diggs structure is good, but try "paring back the folders" on most sites that are using rewrites and the new URL will either send a 404 or more likely will give you some junk error pages generated from the CMS with an erroneous "200 OK" status. In a few circumstances, it might end up sending you unexpected random Duplicate Content.There is also serious comment and discussion of this topic over at: http://www.webmasterworld.com/google/3750106-2-20.htm
Im glad Matt clarified, so as not to alarm people. I still think you should rewrite URLs, just be sure you do it correctly, for 2 big reasons: 1) User-Friendliness (which should be more important than Search Engine Friendliness) and 2) there are other search engines out there than Google.I think Matt makes a good point, that when you mod-rewrite make sure the site still behaves as it should. If a page doesnt exist, it should return a 404 and 1 page should only ever have 1 URL. All can be done real easily by including a conditional call in your code right up front before a single stitch of HTML is written. Basically, if page exists, continue, else 404.
This kind of thing keeps SEOs in business. Everything changes, lots of confusion to clear up... :-/
"if youre starting a new site, its worth considering staying with dynamic parameters instead of doing large amounts of rewrites (which some webmasters do in unusual ways that dont always work well in search engines). Thats my take, at least. Its not like either choice would get you penalized in Google; all of this is just advice to give more information to webmasters when theyre making their choice of site architecture."The more Google speaks, the more I see them mudddying the waters with their PC statements.What I find alarmining is how they consider otimizinag a dynamic URL - a rewrite.Rewrite? I dont see any rewrites anywhere. I believe that the correct word is redirect. Optimizining the urls of a new site is certainly neither a redirect nor a rewrite, but rather a first-write.
This kind of thing keeps SEOs in business. Everything changes, lots of confusion to clear up... :-/Thats the problem. Nothing has actually changed, and there shouldnt be any confusion to clear up. Google and their silly propaganda strike again.
The way the article was written, might have caused some confusion. The reason could be thatJuliane Stiller and Kaspar Szymanski of Search Quality Team (who wrote the current article) arent used to communicate with webmasters and SEO communities. Just to be fair.I see Matt Cutts has clearified the "confusion" in few lines. But lets keep in mind, Matt has been communicating with webmasters/SEO communities for ages and therefore he knows exactly how to explain things in understandable and clear way to the said communities.May be next time GOOG wish to post about complex issues, Matt Cutts (or John Mueller @JohnMu) would be doing the writing :-)
It is hard to change it back, you may lost your indexes in google.Our website is real static HTML, I think dynamic URL and static URL should be equal.
Matt - thanks for the clarification. You could really help the guys who wrote that post by giving them some tips on clear, concise writing - I came away thoroughly confused so can only imagine what a new webmaster would have thought.
Its a difficult subject to write clearly about. Even static vs. dynamic there are nuances that would have made the post even longer if you wanted to dot every i. At http://www.mattcutts.com/blog/seo-glossary-url-definitions/ for example (hah: control-shift-V in Chrome makes this comment box less annoying!), I said<div><div><div>"What is a static url vs. a dynamic url? Technically, we consider a static url to be a document that can be returned by a webserver without the webserver doing any computation. A dynamic url is a document that requires the webserver to do some computation before returning the web document.</div><div>Some people simplify static vs. dynamic urls to an easier question: “Does the url have a question mark?” If the url has a question mark, it’s usually considered dynamic; no question mark in the url often implies a static url. That’s not a hard and fast rule though. For example, urls that look static like http://news.google.com/ may require some computation by the web server. Most people just refer to urls as static or dynamic based on whether it has a question mark though."</div><div></div><div><div> </div><div>So even getting to exactly what is static vs. dynamic urls can take some work. I like to see examples, so Im glad that the post included examples. I know its an advanced topic, but Id rather that we post about it and then follow up with answers than not remind people that often times dynamic urls work great in Google.</div></div></div></div>
I guess the primary reason we rewrite our URLs is so that we dont have any link juice flow or indexing problems down the road. Often times we start out with a site using extensions in their URLs like .asp, .cfm, .html, etc. and we prefer .php at the moment. Well implement static URLs (without extensions) and use the existing pages to get started and show our customers that we can perform. After that, we usually end up with a redesign project where we build a new site in PHP. Because we had implemented the static URLs without the .old extensions, we are now able to seamlessly convert the site to PHP without worrying about confusion on SEs part as to which pages should be indexed - the need for .old and the .new extensions was mitigated by converting the dynamic URLs to static ones. The links that we built are all still pointing to the same URLs after the redesign, and the SEs still see the same URLs so they dont have to figure out which ones to index, they just index the same ones they did before.
I say "folder-based URLs" for URLs that look like folders, even when the website internals may be based on a CMS, and being rewritten. I find that "Parameter-based URLs" is a good description for stuff with a ? and with more stuff following it, but I sometimes also end up calling those "dynamic URLs". I say "static page" or "static website" only when the URLs exactly match the internal folder-based structure on the server. Thats my "best effort" on the matter. Its a jungle of incorrectly used terms out there.
For Google, the focus is that the URLs chosen provide the best user-experience. If URL rewriting creates more problems than it solves, this is bad, even though they are seach engine friendly.For SEOs, this is basically thesame, because pretty URLs that are short and descriptive provide a way better user experience. They usually also contain keywords and make sure you get a higher CTR. High CTRs are important for G too, BTW.Its nice to know that we dont ALWAYS need to rewrite parameters in URLs. I will usually however, continue to do so. Because I know how it works, how to avoid duplicate content issues, because it provides a better user experience and because it gives my clients websites more clicks.I also think that a web developer who does not know how to rewrite URLs, misses something essential in his/hers portfolio (that is a nice way of saying they suck at it)
@mattcutts thanks for the twitter ping - I think Rand has pretty much covered my view in his post from yesterday
And now the client emails begin...SIGH...
Hmm, the NYTimes URLs have some parts that look like folders and some parameters on the end after a ?.Strip off the parameters and you get the exact same page of content.Start meddling with the stuff that looks like folders and you get an error message.
While Google has been able to quite easily index dynamic URLs, those with the question marks, for quite some time I would still recommend rewriting the URL to something that makes sense to a human. Why, because you should design with people in mind and not a spider... Or do I have it very wrong?
I always recommend that clients rewrite from dynamic to user friendly, keyword rich static URLs because they:are easier for people to shareare easier for users to understandare easier to type inensure that if the URL is pasted into a webpage, the keywords in the URL provide context about what the link isPrevent multiple versions of a page being created due to site architecture (where a product page is accessible from multiple categories in a retail website for example)of course, if the URL rewriting is implemented badly, it does caus emore of a problem than otherwise, but there are plenty of good resources avaialble, and we should always test what we change to make sure it works properly.
I am still not sure what is causing the confusion ? just stick to static urls (if you have the option to chose between static and dynamic). If i have vbseo mod then I have no reason not to use it (ok, it may not be a good idea in few cases like for established sites with lot of content)Also I hope webmasters still keep usability in mind! I love digg and wiki links, as the url itself convey what the article is about.
I would guess that the problem in Googles eyes is that many people who use URL rewriting do it very badly, and that makes their job of crawling and indexing the web very much harder to do.One major issue is in forgetting to stop site access through the dynamic URLs, either by blocking them, or by adding a redirect that will send the user to the correct URL. So, if both work, youre now serving duplicates for all of your content.Another issue occurs when rewriting is used to try to stuff keywords into a rewritten URL. While doing this, many sites have inadvertently also allowed parts of the rewritten URL to respond to wildcard requests. That is, you internally link to example.com/our-great-products-25374.html but exactly the same page of content would come up if a competitor linked to example.com/this-company-is-fraudulent-25374.html and the issue arises because of the poor design of the CMS, and the rewriting of the URL not being directly tied to the actual title of the page as stored in the content database.Another major issue is poor error handling... where the site will return the templated page for any and every URL that is requested, and with either duplicate content from another page or a blank templete with no content. Many such sites attempt to have a 404 error page, but again, due to poor understanding and even worse implementation, the user sees the message "Page Not Found" while the bot sees "200 OK" in the HTTP header. Some sites even inadvertently rewrite requests for /robots.txt so that the CMS attempts to output an article page, but fails to some error message. Google never gets to see the real robots.txt file, so is that why they are into stuff you thought you were blocking on your site? Even the most well-known of forum, blog, social and CMS software, from the free to the most expensive, is riddled with these issues.