Stress for sale: Page 2

(Page 2 of 2)

"If you use a remote-hosted stress service, you must commit to keeping the scripts current as your site changes. "
Stress for sale (continued) Back

Enter the Web. Whereas before companies could reasonably predict the demands on their systems based on the number of users, Internet access means that the whole world can visit a site without warning. There is no longer a way to predict or control demand and, to make matters worse, performance is now paramount. Internal users might gripe about response times, but they really have no other options. External users, however, are only a click away from a competitor. If your site slows down, customers might not wait--they might just go elsewhere.

So, if your company is a hot Internet startup and your business model says your site is going to attract millions of daily hits, how do you make sure your infrastructure is really up to it?

Stress service

Using a remote-hosted stress service, you (or a consultant) develop a set of scripts that execute the most likely pathways through your Web pages. Then you schedule a date and time for the service to turn up the volume of users and transactions on your site while you monitor the supporting infrastructure to determine where the stress points are.

Of course, you have to do the proper groundwork, making sure that the data being used is valid and that the service can access your test site or staging area so it doesn't bring down the real site.

What this means is that you don't have to make the investment in the software and servers to generate the potentially huge volumes you are hoping for. It also means, of course, that you have to make an ongoing investment in keeping the scripts current as your site changes, as well as an investment in executing the scripts.

For many companies, site changes can be a daily occurrence, although changes to the scripts only matter when they are executed. So the next question is, how often must you stress your site?

Stress testing

The most obvious times to verify performance are when you change the underlying site infrastructure, such as by adding a new server, updating the database, or making significant software changes. Basically, anything that can impact your site's performance should be verified before it goes live.

The other implication, of course, is that it is now possible for the traffic over the entire Internet infrastructure to be escalated. Instead of the usual and constantly expanding population of users going about their daily cyber-travels, we will now have virtual users, whose activities are simulated by software, in the mix. Since there is only one Internet, and no test or staging Internet, all traffic is traveling over the real thing.

From a test integrity point of view, the lack of a "test Internet" is both good and bad. It's good because your site is receiving its simulated users from the same source as the real ones. It's bad because at some point you may see performance degradation resulting from trying to be sure there's no performance degradation. For this reason, many sites schedule their performance tests during off hours, such as the late night or early morning; so long as your customers are not global--thereby spanning the time zones--this may reduce the risk of impacting performance during the test.

Where will it all end? No doubt, the bandwidth and reliability of both Web sites and the Internet itself will continue to expand as more users come online and more capacity is added. Web technology is amazingly young for its pervasiveness, and as it matures the infrastructure for managing high volumes will stabilize. Until then, hackers will make the news by attacking sites with something as simple as too much traffic. Are you prepared? //

Linda Hayes is CEO of WorkSoft Inc. She was one of the founders of AutoTester. She can be reached at linda@worksoft.com.


Page 2 of 2

Previous Page
1 2
 





Comment and Contribute

 


(Maximum characters: 1200). You have characters left.