« Postgresql on high availability websites? | Main | Wikimedia architecture »

Profiling WEB applications


Some of the articles of the site claims profiling is essential.
Is there any established approach to profiling WEB apps?

Or it too much depends on technologies used?

Reader Comments (1)

For interactive profiling there are tools like http://developer.yahoo.com/yslow/">Yslow and Load Time Analyzer.

For batch profiling I've seen it built into the infrastructure. You have to log page load times, what links are clicked, browser type, link speed, that sort of thing. Then you can pull all the data from the logs and compare it against norms to see if your pages are slowing down. A red flag during a regression test or too a high percentage of red flags during normal operation would trigger an audit of the page.

November 29, 1990 | Unregistered CommenterTodd Hoff

PostPost a New Comment

Enter your information below to add a new comment.
Author Email (optional):
Author URL (optional):
Some HTML allowed: <a href="" title=""> <abbr title=""> <acronym title=""> <b> <blockquote cite=""> <code> <em> <i> <strike> <strong>