A watershed event for the testing industry occurred in December 1999, and I’m not talking about Y2K. For the first time–that I am aware of, anyway–the sorry state of test practices was openly discussed in a national forum.
The cover story of a mainstream newspaper was actually focused on the commonplace failure of computer systems and the reasons why they occur. (See www.usatoday.com/life/cyber/tech/ctg838.htm.) This article was revealing for several reasons, and all of them are important to the present and future of software quality.
Primarily, the article pointed out that not only are software failures a daily reality for companies in all industries, but that they result in staggering costs. Furthermore, it shifted focus away from the glitz of the Internet and back to pedestrian applications that truly still operate most corporations–manufacturing, payroll, shipping, and the like.
But the ultimate victory was that in most of the examples the real reason for the failure was admitted to be inadequate attention to testing. Hallelujah.
Forget Y2K
Now that we’re well into the new century, it’s easy to put Y2K fears behind us. Frankly, I was worried that after New Year’s Eve 1999, all of the attention being paid to testing would dissipate and we would return to the old days, when software quality issues were practically invisible.
But this newspaper article, amazingly enough, announced “Y2K is dead” in December–before we knew that Y2K problems were really behind us. The author had the guts to make this statement because the relative cost of what was deemed “everyday failures” is a staggering billion a year, which is, coincidentally, the sum total that was spent on Y2K in this country for the entire preparation effort! In other words, there was a global hue and cry for years about the risks of a once-in-a-millennium event, when we are wasting as much money each and every year on problems that occur each and every day.
One hundred billion dollars a year. Think about it. By anyone’s measure, that’s a whole lot of money. It’s incredible when you think about how many test departments are struggling to eke out the most meager of budgets and protect the stingiest of schedules, while all along the stakes are this enormous.
How could an elephant of this size be lumbering through our economy and just now be noticed? Well, it has a whole lot to do with denial.
Surprising news…not
For industry insiders, especially those who have made a career in testing, the news that systems are regularly slapped together and rushed into production prematurely, without adequate testing, is not news. It’s standard operating procedure. For the rest of the world, however, this comes as a shock.
Yet if you follow the media exposure about software failures and the money being spent accordingly, the only thing that appears to be worth testing is the volume capability of your Internet site. Boring old functional errors in the tiresome applications, such as payroll, inventory, and shipping, that actually operate most corporate giants just don’t seem to rate any real drama–or time or money.
Granted, in most cases, companies don’t rush out a press release when they bungle their own operations, but sometimes they must. Hersey Food Corp., for instance, was forced to explain to stockholders why they had a 12% decrease in 1999 third-quarter sales and a 19% drop in net income from third-quarter 1998. It seems that a complex shipping system that was not thoroughly tested crippled deliveries during the critical Halloween season. But even a case like this, which cost millions in revenues and more in shareholder value, was not picked up as a hot story by either wire services or the networks.
It’s show time
What will it take before companies stop wasting billions of dollars rushing systems into production without adequate testing? My theory is that this will only happen when these previously hidden systems are finally exposed to the outside world. Just as companies can’t deny failures that escape to the Internet, neither will they be able to ignore those that affect other companies.
The fast-approaching business-to-business integration across companies that is sweeping the economy will likely provide the catalyst that reveals the shaky foundations of the most mundane but critical applications that do things like manufacture, inventory, and ship goods. Once the partners, shareholders, and other stakeholders understand that these errors are preventable, tolerance for deficient testing will evaporate and accountability will set in.
Not to be a ghoul about it or to wish public humiliation on anyone, but I actually look forward to the day when testing is no longer treated as an option or an orphan. A day when people don’t glaze over and move away when you tell them what you do for a living, when inexcusable blunders are truly no longer excused. Until then, watch this space for news that might otherwise be buried. //
Linda Hayes is CEO of WorkSoft Inc. She was one of the founders of AutoTester. She can be reached at linda@worksoft.com.
Huawei’s AI Update: Things Are Moving Faster Than We Think
FEATURE | By Rob Enderle,
December 04, 2020
Keeping Machine Learning Algorithms Honest in the ‘Ethics-First’ Era
ARTIFICIAL INTELLIGENCE | By Guest Author,
November 18, 2020
Key Trends in Chatbots and RPA
FEATURE | By Guest Author,
November 10, 2020
FEATURE | By Samuel Greengard,
November 05, 2020
ARTIFICIAL INTELLIGENCE | By Guest Author,
November 02, 2020
How Intel’s Work With Autonomous Cars Could Redefine General Purpose AI
ARTIFICIAL INTELLIGENCE | By Rob Enderle,
October 29, 2020
Dell Technologies World: Weaving Together Human And Machine Interaction For AI And Robotics
ARTIFICIAL INTELLIGENCE | By Rob Enderle,
October 23, 2020
The Super Moderator, or How IBM Project Debater Could Save Social Media
FEATURE | By Rob Enderle,
October 16, 2020
FEATURE | By Cynthia Harvey,
October 07, 2020
ARTIFICIAL INTELLIGENCE | By Guest Author,
October 05, 2020
CIOs Discuss the Promise of AI and Data Science
FEATURE | By Guest Author,
September 25, 2020
Microsoft Is Building An AI Product That Could Predict The Future
FEATURE | By Rob Enderle,
September 25, 2020
Top 10 Machine Learning Companies 2020
FEATURE | By Cynthia Harvey,
September 22, 2020
NVIDIA and ARM: Massively Changing The AI Landscape
ARTIFICIAL INTELLIGENCE | By Rob Enderle,
September 18, 2020
Continuous Intelligence: Expert Discussion [Video and Podcast]
ARTIFICIAL INTELLIGENCE | By James Maguire,
September 14, 2020
Artificial Intelligence: Governance and Ethics [Video]
ARTIFICIAL INTELLIGENCE | By James Maguire,
September 13, 2020
IBM Watson At The US Open: Showcasing The Power Of A Mature Enterprise-Class AI
FEATURE | By Rob Enderle,
September 11, 2020
Artificial Intelligence: Perception vs. Reality
FEATURE | By James Maguire,
September 09, 2020
Anticipating The Coming Wave Of AI Enhanced PCs
FEATURE | By Rob Enderle,
September 05, 2020
The Critical Nature Of IBM’s NLP (Natural Language Processing) Effort
ARTIFICIAL INTELLIGENCE | By Rob Enderle,
August 14, 2020
Datamation is the leading industry resource for B2B data professionals and technology buyers. Datamation's focus is on providing insight into the latest trends and innovation in AI, data security, big data, and more, along with in-depth product recommendations and comparisons. More than 1.7M users gain insight and guidance from Datamation every year.
Advertise with TechnologyAdvice on Datamation and our other data and technology-focused platforms.
Advertise with Us
Property of TechnologyAdvice.
© 2025 TechnologyAdvice. All Rights Reserved
Advertiser Disclosure: Some of the products that appear on this
site are from companies from which TechnologyAdvice receives
compensation. This compensation may impact how and where products
appear on this site including, for example, the order in which
they appear. TechnologyAdvice does not include all companies
or all types of products available in the marketplace.