The History and the Hype

  • Share
  • Read Later
CHRIS TAYLORTwo digits. That's all. Just two lousy digits. 1957, they should have written, not 57. 1970 rather than 70. Most important, 01-01-2000 would have been infinitely preferable to 01-01-00. Though most of the dire predictions connected with that date--the Year 2000 computer bug's moment of truth--are unlikely to come true, a little computer-generated chaos would provide a fitting conclusion to a 40-year story of human frailties: greed, shortsightedness and a tendency to rush into new technologies before thinking them through.How did this happen? Who is responsible for the bug we call Y2K? Conventional wisdom goes something like this: back in the 1950s, when computers were the size of office cubicles and the most advanced data-storage system came on strips of punched cardboard, several scientists, including a U.S. Navy officer named Grace Murray Hopper, begat a standard programming language called COBOL (common business-oriented language). To save precious space on the 80-column punch cards, COBOL programmers used just six digits to render the day's date: two for the day, two for the month, two for the year. It was the middle of the century, and nobody cared much about what would happen at the next click of the cosmic odometer. But today the world runs on computers, and older machines run on jury-rigged versions of COBOL that may well crash or go senile when they hit a double-zero date. So the finger of blame for the approaching crisis should point at Hopper and her COBOL cohorts, right?Wrong. Nothing, especially in the world of computing, is ever that simple. It was the fault of everybody, just everybody, says Robert Bemer, the onetime IBM whiz kid who wrote much of COBOL. If Grace Hopper and I were at fault, it was for making the language so easy that anybody could get in on the act. And anybody did, including a group of Mormons in the late '50s who wanted to enlist the newfangled machines in their massive genealogy project--clearly the kind of work that calls for thinking outside the 20th century box. Bemer obliged by inventing the picture clause, which allowed for a four-digit year. From this point on, more than 40 years ahead of schedule, the technology was available for every computer in the world to become Y2K compliant.Programmers ignored Bemer's fix. And so did his bosses at IBM, who unwittingly shipped the Y2K bug in their System/360 computers, an industry standard every bit as powerful in the '60s as Windows is today. By the end of the decade, Big Blue had effectively set the two-digit date in stone. Every machine, every manual, every maintenance guy would tell you the year was 69, not 1969. The general consensus was that this was the way you programmed, says an IBM spokesman. We recognize the potential for lawsuits on this issue.No one in the computer industry wanted to rock the boat. And no one could alter the course IBM had set, not even the International Standards Organization, which adopted the four-digit date standard in the 1970s. The Pentagon promised to adopt century-friendly dates around 1974, then sat on its hands. Bemer himself wrote the earliest published Y2K warnings--first in 1971, then again in 1979. Greeted by nothing but derision, he retired in 1982. How do you think I feel about this thing? says Bemer, now an officer at his own Y2K software firm. I made it possible to do four digits, and they screwed it up.PAGE 1|
Meanwhile, the torch of Y2K awareness passed to a new generation. In late 1977, a young Canadian named Peter de Jager signed on as a computer operator at IBM. His first task was to boot up a U.S.-wide banking system run on an IBM 370. When the machine whirred into life, it asked for the date. As De Jager, a mathematics major straight out of college, entered the number 77, a thought occurred to him. Did this machine care what century it was? With the impetuousness of youth, he marched off to his manager and informed him the computer would not work in the year 2000. The manager laughed and asked De Jager how old he was. This isn't going to be a problem until you're 45, he said. Don't worry, we'll sort it out.And that, at least for the next 13 years, was the attitude De Jager adopted. We used to joke about this at conferences, he says. Irresponsible talk, like 'We won't be around then.' But by 1991, De Jager, a self-described nobody in the industry, had decided he would be around. Four years later, he was giving more than 85 lectures a year on the topic and posting regular updates to his site, the Web's first for Y2K warnings, .And here's the curious thing. From 1995 on, Y2K awareness had a kind of critical mass. Congress, the White House and the media all got wind of the bug at about the same time. After making too little of the problem for so long, everybody began to make, if anything, too much of it.Why then, and not two decades earlier? Why De Jager, and not Bemer? Proximity to the millennium may have had something to do with it as well as the increasingly ominous tone of the warnings. This was Bemer's dry 1979 prophecy of doom: Don't drop the first two digits. The program may well fail from ambiguity. Twenty years later, here's De Jager's jeremiad: The economy worldwide would stop...you would not have water. You would not have power...This alarmist language may yet be justified. By 1999 folly has compounded folly. In many cases, the original COBOL code has been rejiggered so many times that the date locations have been lost. And even when programmers find their quarry, they aren't sure which fixes will work. The amount of code that needs to be checked has grown to a staggering 1.2 trillion lines. Estimates for the cost of the fix in the U.S. alone range from $50 billion to $600 billion. As for Y2K compliance in Asian economies still struggling with recession? Forget about it.The fact is that no one on the planet really knows what will happen when 01-01-00 rolls around. Whether we'll be glad we were panicked into action or we'll disown the doomsayers depends on how diligently the programmers do their job in the next 50 weeks. One thing is already clear. In a century in which man split the atom, spliced genes and turned silicon into data, the tale of Y2K--how we ignored it for 40 years, then flew into a tizzy--will not be remembered as our finest hour.|2