Redirected from Millennium Bug
Y2K (or Y2k) was the common slang for the year 2000 problem. It also went by Millennium Bug (though, strictly speaking, 2000 is the last year of the 20th century and does not start a new millennium).
It was thought computer programs could stop working or produce erroneous results because they stored years with only two digits and that the year 2000 would be represented by '00' and would not follow 1999 (i.e., '99') in numerical sequence. This would cause date comparisions to produce incorrect results. It was also thought that embedded systems, making use of similar date logic, might fail and cause utilities and other crucial infrastructure to fail.
In the years prior to 2000, some corporations and governments, when they did testing to determine the extent of the potential impact, reported that some of their critical systems really would need significant repairs or risk serious breakdowns. Throughout 1997 and 1998, there were news reports about major corporations and industries that had made uncertain estimates as to their preparedness. The vagueness of these reports, and the apparent uncertainty regarding what sort of breakdowns were possible--and the fact that literally hundreds of billions of dollars were reportedly spent in remediation efforts--were a major part of the reason for the public fear. Special committees were set up by governments to monitor remedial work and contingency planning, particularly by crucial infrastructure such a telecommunications, utilities and the like, to ensure that the most critical services had fixed their own problems and were prepared for problems with others. By early- to mid-1999, when the same corporations, industry organizations, and governments were claiming to be largely prepared, the public relations damage had been done. It was only the safe passing of the main "event horizon" itself, January 1, 2000, that fully quelled public fears.
The underlying programming problem was quite real. In the 1960s, computer memory and storage were scarce and expensive, and most data processing was done on punch cards which represented text data in 80-column records. Programming languages of the time, such as COBOL and RPG, processed numbers in their ASCII or EBCDIC representations. They occasionally used an extra bit called a "zone punch" to save one character for a minus sign on a negative number, or compressed two digits into one byte in a form called binary-coded decimal, but otherwise processed numbers as straight text. Over time the punch cards were converted to magnetic tape and then disk files and later to simple databases like ISAM, but the structure of the programs usually changed very little. Popular software like dBase[?] continued the practice of storing dates as text well into the 1980s and 1990s.
Saving two characters for every date field was a significant savings at that time. Most programmers of that time did not expect their programs to remain in use for many decades, so they did not consider it a significant problem. However this was not universally true. The problem was first recognised in 1958 by Bob Bemer[?] as a result of work on genealogical software. He spent the next twenty years trying to make programmers, IBM, the US government and the ISO care about the problem, with little result. This included the recommendation that the COBOL PICTURE clause should be used to specify four digit years for dates. This could have been done by programmers at any time from the initial release of the first COBOL compiler in 1961 onwards. However complacency and lack of foresight prevented this advice from being followed. Despite magazine articles on the subject from 1970 onwards, the majority of programmers only started recognizing Y2K as a looming problem in the 1990s, but even then, inertia and complacency caused it to be mostly ignored until the last few years of the decade.
Storage of a combined date and time within a fixed binary field is often considered a solution, but the possibility for software to mis-interpret dates remains, because such date and time representations must be relative to a defined origin. Roll-over of such systems is still a problem but can happen at varying dates and can fail in various ways. For example:
Even before January 1, 2000 arrived, there were also some worries albeit lesser compared to those generated by Y2K about September 9, 1999. This date could also be written in the numeric format, 9/9/99, which is somewhat similar to the end-of-file code, 9999, in old programming languages. It was feared that some programs might unexpectedly terminate on that date. Another related problem for the year 2000 was that it was a leap year even though years ending in '00' are normally not leap years. (A year is a leap year if it is divisible by 4 unless it is both divisible by 100 and not divisible by 400.) Fortunately, like Y2K, both fears were proven inaccurate.
Public reaction to the problem
Some industries started experiencing related problems early in the 1990s as software began to process future dates past 1999. For example, in 1993, some people with financial loans that were due in 2000 received (incorrect) notices that they were 93 years past due. As the decade progressed, more and more companies experienced problems and lost money due to erroneous date data. As another example, meat-processing companies incorrectly destroyed large amounts of good meat because the computerized inventory system identified the meat as expired. There were, in fact, many such minor "horror stories" like these, which received much play in the press as 2000 approached.
As the decade progressed, identifying and correcting or replacing affected computer systems or computerized devices became the major focus of information technology departments in most large companies and organizations. Millions of lines of programming code were reviewed and fixed during this period. Many corporations replaced major software systems with completely new ones that did not have the date processing problems. It was frequently reported that corporations had already experienced at least minor Y2K problems, and some major problems as well, due to date look-ahead functions in code and embedded systems, but it was and still is not clear what the full cost and seriousness of these problems was.
Y2K was the big media hype story of 1999. Public apprehension was tremendous, reaching, in some quarters, enormous proportions. Some individuals stockpiled canned or dried food in anticipation of food shortages. A few commentators predicted a full-scale apocalypse--three of the best known were computer consultant Edward Yourdon[?], religious commentator Gary North, and economist Edward Yardeni[?]. But when January 1, 2000 finally came, there were hardly any major problems reported, though a large number of them had been expected. Ironically, many people were upset that there appeared to be so much hype over nothing, because the vast majority of problems had been fixed correctly. Some more sophisticated critics have suggested that much preventive effort was unnecessary -- it would have been cheaper not to spend as much examining non-critical systems for flaws and simply fix the few that would have failed after the event. Such conclusions are easy to draw with the benefit of hindsight, but in any case the overhaul of many systems involved replacement with new, improved functionality anyway and thus in many cases the expenditure proved useful regardless.
Some items of interest:
Search Encyclopedia
|
Featured Article
|