[Seaside] Static sites and spider handling

Colin Putney cputney at wiresong.ca
Tue Aug 26 10:15:18 CEST 2003


Cees and Nevin,

One thing I learned while working at Whistler.com is that search engine 
robots are incredibly paranoid and cynical. It would seem that the 
primary design goal of a search engine is to detect and defeat 
spamdexing in order to present an accurate view of the web to its 
users. As a result, the bots absolutely *hate* two-faced web sites, and 
will penalize them in the rankings accordingly.

IMO the best search engine strategy is to keep as much content as 
possible static, and only use dynamic pages when necessary. Obviously 
that's not an option here,
but it would make a good rule of thumb when designing the robot special 
case: to a robot, the site should appear as much as possible like a 
collection of html files served up by name.

Another data point, and maybe you guys already know this, is that links 
from other sites trump just about every other criterion for determining 
a page's rank. So ideally, you want to provide ensure that all links to 
your site from elsewhere point into the robot index. From there you 
want to get them into a session in way that doesn't annoy the robots.

I figure this will be a difficult project, particularly since the 
search engines are in an arms race with the spamdexers, and the rules 
change constantly. Luckily the optimization community obsesses about 
this stuff, and so the information you need to do it effectively is out 
there, if you can find it.

Good luck,

Colin



More information about the Seaside mailing list