CHAPTER 12

The Integrated Circuit Chip: From the Roswell Crash Site to Silicon Valley

WITH THE NIGHT-VISION IMAGE INTENSIFIER PROJECT UNDER way at Fort Belvoir and the Project Horizon team
trying to swim upstream against the tide of civilian management of the U.S. space program, I turned my attention to the next of the Roswell crash fragments that looked especially intriguing: the charred semiconductor wafers that had broken off some larger device. I hadn’t made these my priorities at first, not knowing what they really were, until General Trudeau asked me to take a closer look.

“Talk to some of the rocket scientists down at Alamogordo about these things, Phil, “ he said. “I think they’ll know what we should do with them. “

I knew that in the days immediately following the crash, General Twining had met with the Alamogordo group of the Air Materiel Command and had described some of the debris to them. But I didn’t know how detailed his descriptions were or whether they even knew about the wafers we had in our file.

“I want to talk to some of the scientists up here, too, “ I said. “Especially, I want to see some of the engineers from the defense contractors. Maybe they can figure out what the engineering process is for these things. “
“Go over to Bell Labs, Phil, “ General Trudeau also suggested. “The transistor came out of their shop and these things look a lot like transistorized circuits.“

I’d heard that General Twining had worked very closely with both Bell Labs and Motorola on communications research during the war, afterwards at the Alamogordo test site for V2 missile launches, and after the Roswell crash. Whether he had brought them any material from the crash or showed them the tiny silicon chips was a matter of pure speculation. I only know that the entire field of circuit miniaturization took a giant leap in 1947 with the invention of the transistor and the first solid state components.

 

By the late 1950s,transistors had replaced the vacuum tube in radios and had turned the wall-sized wooden box of the 1940s into the portable plastic radio you could hear blaring away at the shore on a hot July Sunday. The electronics industry had taken a major technological jump in less than ten years, and I had to wonder privately whether any Roswell material had gotten out that I didn’t know about prior to my taking over Foreign Technology in 1961.

I didn’t realize it at first when I showed those silicon wafers to General Trudeau, but I was to become very quickly and intimately involved with the burgeoning computer industry and a very small, completely invisible, cog in an assembly line process that fifteen years later would result in the first microcomputer systems and the personal computer revolution.
Over the course of the years since I joined the army in 1942, my career took me through the stages of vacuum tube based devices, like our radios and radars in World War II, to component chassis.

 

These were large circuitry units that, if they went down, could be changed in sections, smaller sections, and finally to tiny transistors and transistorized electronic components. The first army computers I saw were room sized, clanking vacuum tube monsters that were always breaking down and, by today’s standards, took an eternity to calculate even the simplest of answers. They were simply oil filled data pots. But they amazed those of us who had never seen computers work before.


At Red Canyon and in Germany, the tracking and targeting radars we used were controlled by new transistorized chassis computers that were compact enough to fit onto a truck and travel with the battalion. So when I opened up my nut file and saw the charred matte gray quarter sized, cracker shaped silicon wafers with the gridlines etched onto them like tiny printed lines on the cover of a match book, I could make an educated guess about their function even though I’d never seen anything of the like before. I knew, however, that our rocket scientists and the university researchers who worked with the development laboratories at Bell, Motorola, and IBM would more than understand the primary function of these chips and figure out what we needed to do to make some of our own.


But first I called Professor Hermann Oberth for basic background on what, if any, development might have taken place after the Roswell crash. Dr. Oberth knew the Alamogordo scientists and probably received second hand the substance of the conversations General Twining had with his Alamogordo group in the hours after the retrieval of the vehicle. And if General Twining described some of the debris, did he describe these little silicon chips? And if he did, in those months when the ENIAC - the first working computer - was just cranking up at the Aberdeen Ordnance Testing Grounds in Maryland, what did the scientists make of those chips?

“They saw these at the Walker Field hangar, “ Dr. Oberth told me. “All of them at Alamogordo flew over to Roswell with General Twining to oversee the shipment to Wright Field. “

Oberth described what happened that day after the crash when a team of AMC rocket scientists pored over the bits and pieces of debris from the site. Some of the debris was packed for flight on B29s. Other material, especially the crates that wound up at Fort Riley, were loaded onto deuce and a halfs for the drive. Dr. Oberth said that years later, von Braun had told him how those scientists who literally had to stand in line to have their equations processed by the experimental computer in Aberdeen Maryland were in awe of the microscopic circuitry etched into the charred wafer chips that had spilled out of the craft.


Von Braun had asked General Twining whether anyone at Bell Labs was going to be contacted about this find. Twining seemed surprised at first, but when von Braun told him about the experiments on solid state components - material whose electrons don’t need to be excited by heat in order to conduct current - Twining became intrigued. What if these chips were components of a very advanced solid state circuitry? von Braun asked him. What if one of the reasons the army could find no electronic wiring on the craft were the layers of these wafers that ran throughout the ship? These circuit chips could be the nervous system of the craft, carrying signals and transmitting commands just like the nervous system in a human body.
General Twining’s only experience had been with the heavily insulated vacuum tube devices from World War II, where the multistrand wires were covered with cloth. He’d never seen metallic printed chips like these before. How did they work? he’d asked von Braun.


The German scientist wasn’t sure, although he guessed they worked on the same principle as the transistors that laboratories were trying to develop to the point where they could be manufactured commercially. It would completely transform the electronics industry, von Braun explained to General Twining, nothing short of a revolution. The Germans had been desperately trying to develop circuitry of this sort during the war, but Hitler, convinced the war would be over by 1941, told the German computer researchers that the Wehrmacht had no need for computers that had a development timetable greater than one year. They’d all be celebrating victory in Berlin before the end of the year.


But the research into solid state components that the Germans had been doing and the early work at Bell Labs was nothing compared to the marvel that Twining had shown von Braun and the other rocket scientists in New Mexico. Under the magnifying glass, the group thought they saw not just a single solid state switch but a whole system of switches integrated into each other and comprising what looked like an entire circuit or system of circuits. They couldn’t be sure because no one had ever seen anything even remotely like this before.

 

But it showed them an image of what the future of electronics could be if a way could be found to manufacture this kind of circuit on Earth. Suddenly, the huge guidance control systems necessary to control the flight of a rocket, which, in 1947, were too big to be squeezed into the fuselage of the rocket, could be miniaturized so that the rocket could have its own automatic guidance system. If we could duplicate what the EBEs had, we, too, would have the ability to explore space. In effect, the reverse engineering of solid state integrated circuitry began in the weeks and months after the crash even though William Shockley at Bell Labs was already working on a version of his transistor as early as 1946.


In the summer of 1947, the scientists at Alamogordo were only aware of the solid state circuit research under way at Bell Labs and Motorola. So they pointed Nathan Twining to research scientists at both companies and agreed to help him conduct the very early briefings into the nature of the Roswell find. The army, very covertly, turned some of the components over to research engineers for an inspection, and by the early 1950s the transistor had been invented and transistorized circuits were now turning up in consumer products as well as in military electronics systems. The era of the vacuum tube, the single piece of eighty year old technology upon which an entire generation of communications devices including television and digital computers was built, was now coming to a close with the discovery in the desert of an entirely new technology.


The radio vacuum tube was a legacy of nineteenth century experimentation with electric current. Like many historic scientific discoveries, the theory behind the vacuum tube was uncovered almost by chance, and nobody really knew what it was or cared much about it until years later. The radio vacuum tube probably reached its greatest utility from the 1930s through the 1950s, until the technology we discovered at Roswell made it all but obsolete.

 

The principle behind the radio vacuum tube, first discovered by Thomas Edison in the 1880s while he was experimenting with different components for his incandescent lightbulb, was that current, which typically flowed in either direction across a conductive material such as a wire, could be made to flow in only one direction when passed through a vacuum. This directed flow of current, called the “Edison effect, “ is the scientific principle behind the illumination of the filament material inside the vacuum of the incandescent lightbulb, a technology that has remained remarkably the same for over a hundred years.


But the lightbulb technology that Edison discovered back in the1880s, then put aside only to experiment with it again in the early twentieth century, also had another equally important function. Because the flow of electrons across the single filament wire went in only one direction, the vacuum tube was also a type of automatic switch. Excite the flow of electrons across the wire and the current flowed only in the direction you wanted it to. You didn’t need to throw a switch to turn on a circuit manually because the vacuum tube could do that for you.

 

Edison had actually discovered the first automatic switching device, which could be applied to hundreds of electronic products from the radio sets that I grew up with in the1920s to the communications networks and radar banks of World War II and to the television sets of the 1950s. In fact, the radio tube was the single component that enabled us to begin the worldwide communications network that was already in place by the early twentieth century.


Radio vacuum tubes also had another important application that wasn’t discovered until experimenters in the infant science of computers first recognized the need for them in the 1930s and then again in the 1940s. Because they were switches, opening and closing circuits, they could be programmed to reconfigure a computer to accomplish different tasks. The computer itself had, in principle, remained essentially the same type of calculating device that Charles Babbage first invented in the 1830s. It was a set of internal gears or wheels that acted as counters and a section of “memory” that stored numbers until it was their turn to be processed. Babbage’s computer was operated manually by a technician who threw mechanical switches in order to input raw numbers and execute the program that processed the numbers.


The simple principle behind the first computer, called by its inventor the “Analytical Engine, “ was that the same machine could process an infinite variety and types of calculations by reconfiguring its parts through a switching mechanism. The machine had a component for inputting numbers or instructions to the processor; the processor itself, which completed the calculations; a central control unit, or CPU, that organized and sequenced the tasks to make sure the machine was doing the right job at the right time; a memory area for storing numbers; and finally a component that output the results of the calculations to a type of printer: the same basic components you find in all computers even today.


The same machine could add, subtract, multiply, or divide and even store numbers from one arithmetical process to the next. It could even store the arithmetical computation instructions themselves from job to job. And Babbage borrowed a punch card process invented by Joseph Jacquard for programming weaving looms. Babbage’s programs could be stored on series of punch cards and fed into the computer to control the sequence of processing numbers. Though this may sound like a startling invention, it was Industrial Revolution technology that began in the late eighteenth century for the purely utilitarian challenge of processing large numbers for the British military. Yet, in concept, it was an entirely new principle in machine design that very quietly started the digital revolution.


Because Babbage’s machine was hand powered and cumbersome, little was done with it through the nineteenth century, and by the1880s, Babbage himself would be forgotten. However, the practical application of electricity to mechanical appliances and the delivery of electrical power along supply grids, invented by Thomas Edison and refined by Nikola Tesla, gave new life to the calculation machine. The concept of an automatic calculation machine would, inspire American inventors to devise their own electrically powered calculators to process large numbers in a competition to calculate the 1890 U.S. Census.

 

The winner of the competition was Herman Hollerith, whose electrically powered calculator was a monster device that not only processed numbers but displayed the progress of the process on large clocks for all to see. He was so successful that the large railroad companies hired him to process their numbers. By the turn of the century his company, the Computing Tabulating and Recording Company, had become the single largest developer of automatic calculating machines. By 1929, when Hollerith died, his company had become the automation conglomerate, IBM.


Right about the time of Hollerith’s death, a German engineer named Konrad Zuse approached some of the same challenges that had confronted Charles Babbage a hundred years earlier: how to build his own version of a universal computing machine that could reconfigure itself depending upon the type of calculation the operator wanted to perform. Zuse decided that instead of working with a machine that operated on the decimal system, which limited the types of arithmetic calculations it could perform, his machine would use only two numbers, 0 and 1, the binary system.

 

This meant that he could process any type of mathematical equation through the opening or closing of a series of electromagnetic relays, switches that would act as valves or gates either letting current through or shutting it off. These relays were the same types of devices that the large telephone companies, like the Bell system in the United States, were using as the basis of their networks. By marrying an electrical power supply and electric switches to the architecture of Babbage’s Analytical Engine and basing his computations in a binary instead of a decimal system, Zuse had come up with the European version of the first electrical digital computer, an entirely new device. It was just three years before the German invasion of Poland and the outbreak of World War II.


In the United States at about the same time as Zuse was assembling his first computer in his parents’ living room, Harvard mathematics professor Howard Aiken was trying to reconstruct a theoretical version of Babbage’s computer, also using electromagnetic relays as switching devices and relying on a binary number system. The difference between Aiken and Zuse was that Aiken had academic credentials and his background as an innovative mathematician got him into the office of Thomas Watson, president of IBM, to whom he presented his proposal for the first American digital computer. Watson was impressed, authorized a budget for $1 million, and, right before the attack on Pearl Harbor, the project design was started up at Cambridge, Massachusetts. It was then moved to IBM headquarters in New York during the war.


Because of their theoretical ability to calculate large sets of numbers in a relatively short period of time, digital computers were drafted into the war effort in the United Kingdom as a code breaking device. By 1943, at the same time that IBM’s first shiny stainless steel version of Aiken’s computer was up and running in Endicott, New York, the British were using their dedicated crypto analytical Colossus computer to break the German codes and decipher the code creating ability of the German Enigma - the code machine that the Nazis believed made their transmissions indecipherable to the Allies.

 

Unlike the IBM-Aiken computer at Harvard and Konrad Zuse’s experimental computer in Berlin, the Colossus used radio vacuum tubes as relay switches and was, therefore, hundreds of times faster than any experimental computer using electromagnetic relays. The Colossus, therefore, was a true breakthrough because it married the speed of vacuum tube technology with the component design of the Analytical Engine to create the first modern era digital computer.


The British used the Colossus so effectively that they quickly felt the need to build more of them to process the increasingly large volume of encrypted transmissions the Germans were sending, ignorant of the fact that the Allies were decoding every word and outsmarting them at every turn. I would argue even to this day that the technological advantage the Allies enjoyed in intelligence gathering apparatus, specifically code breaking computers and radar, enabled us to win the war despite Hitler’s initial successes and his early weapon advantages. The Allies’ use of the digital computer in World War II was an example of how a superior technological advantage can make the difference between victory and defeat no matter what kinds of weapons or numbers of troops the enemy is able to deploy.


The American and British experience with computers during the war and our government’s commitment to developing a viable digital computer led to the creation, in the years immediately following the war, of a computer called the Electronic Numerical Integrator and Calculator, or ENIAC. ENIAC was the brain child of Howard Aiken and one of our Army R&D brain trust advisers, the mathematician John von Neumann. Although it operated on a decimal instead of a binary system and had a very small memory, it relied on radio vacuum tube switching technology. For its time it was the first of what today are called “number crunchers. “


When measured against the way computers developed over the years since its first installation, especially the personal computers of today, ENIAC was something of a real dinosaur. It was loud, hot, cumbersome, fitful, and required the power supply of an entire town to keep it going. It couldn’t stay up for very long because the radio tubes, always unreliable even under the best working conditions, would blow out after only a few hours’ work and had to be replaced. But the machine worked, it crunched the numbers it was fed, and it showed the way for the next model, which reflected the sophisticated symbolic architectural design of John von Neumann.


Von Neumann suggested that instead of feeding the computer the programs you wanted it to run every time you turned it on, the programs themselves could be stored in the computer permanently. By treating the programs themselves as components of the machine, stored right in the hardware, the computer could change between programs, or the routines of subprograms, as necessary in order to solve problems. This meant that larger routines could be processed into subroutines, which themselves could be organized into templates to solve similar problems. In complex applications, programs could call up other programs again and again without the need of human intervention and could even change the subprograms to fit the application. von Neumann had invented block programming, the basis for the sophisticated engineering and business programming of the late 1950s and 1960s and the great, great grandmother of today’s object oriented programming.


By 1947, it had all come together: the design of the machine, the electrical power supply, the radio vacuum tube technology, the logic of machine processing, von Neumann’s mathematical architecture, and practical applications for the computer’s use. But just a few years shy of the midpoint of the century, the computer itself was the product of eighteenth and nineteenth century thinking and technology. In fact, given the short comings of the radio tube and the enormous power demands and cooling requirements to keep the computer working, the development of the computer seemed to have come to a dead end.

 

Although IBM and Bell Labs were investing huge sums of development money into designing a computer that had a lower operational and maintenance overhead, it seemed, given the technology of the digital computer circa 1947, that there was no place it could go. It was simply an expensive to build, expensive to run, lumbering elephant at the end of the line. And then an alien spacecraft fell out of the skies over Roswell, scattered across the desert floor, and in one evening everything changed.


In 1948 the first junction transistor - a microscopically thin silicon sandwich of w-type silicon, in which some of the atoms have an extra electron, and p-type silicon, in which some of the atoms have one less electron - was devised by physicist William Shockley. The invention was credited to Bell Telephone Laboratories, and, as if by magic, the dead end that had stopped the development of the dinosaur like ENIAC generation of computers melted away and an entirely new generation of miniaturized circuitry began.

 

Where the radio tube circuit required an enormous power supply to heat it up because heat generated the electricity, the transistor required very low levels of powers and no heating up time because the transistor amplified the stream of electrons that flowed into its base. Because it required only a low level of current, it could be powered by batteries. Because it didn’t rely on a heat source to generate current and it was so small, many transistors could be packed into a very small space, allowing for the miniaturization of circuitry components. Finally, because it didn’t burn out like the radio tube, it was much more reliable.

 

Thus, within months after the Roswell crash and the first exposure of the silicon wafer technology to companies already involved in the research and development of computers, the limitations on the size and power of the computer suddenly dropped like the removal of a roadblock on a highway and the next generation of computers went into development. This set up for Army R&D, especially during the years I was there, the opportunity for us to encourage that development with defense contracts calling for the implementation of integrated circuit devices into subsequent generations of weapons systems.


More than one historian of the microcomputer age has written that no one before 1947 foresaw the invention of the transistor or had even dreamed about an entirely new technology that relied upon semiconductors, which were silicon based and not carbon based like the Edison incandescent tube. Bigger than the idea of a calculating machine or an Analytical Engine or any combination of the components that made up the first computers of the 1930s and 1940s, the invention of the transistor and its natural evolution to the silicon chip of integrated circuitry was beyond what anyone could call a quantum leap of technology.

 

The entire development arc of the radio tube, from Edison’s first experiments with filament for his incandescent lightbulb to the vacuum tubes that formed the switching mechanisms of ENIAC, lasted about fifty years. The development of the silicon transistor seemed to come upon us in a matter of months. And, had I not seen the silicon wafers from the Roswell crash with my own eyes, held them in my own hands, talked about them with Hermann Oberth, Wernher von Braun, or Hans Kohler, and heard the reports from these now dead scientists of the meetings between Nathan Twining, Vannevar Bush, and researchers at Bell Labs, I would have thought the invention of the transistor was a miracle. I know now how it came about.


As history revealed, the invention of the transistor was only the beginning of an integrated circuit technology that developed through the 1950s and continues right through to the present. By the time I became personally involved in 1961, the American marketplace had already witnessed the retooling of Japan and Germany in the 1950s and Korea and Taiwan in the late 1950s through the early 1960s. General Trudeau was concerned about this, not because he considered these countries our economic enemies but because he believed that American industry would suffer as a result of its complacency about basic research and development.

 

He expressed this to me on many occasions during our meetings, and history has proved him to be correct. General Trudeau believed that the American industrial economy enjoyed a harvest of technology in the years immediately following World War II, the effects of which were still under way in the 1960s, but that it would soon slow down because R&D was an inherently costly undertaking that didn’t immediately contribute to a company’s bottom line. And you had to have a good bottom line, General Trudeau always said, to keep your stockholders happy or else they would revolt and throw the existing management team right out of the company. By throwing their efforts into the bottom line, Trudeau said, the big American industries were actually destroying themselves just like a family that spends all its savings.

“You have to keep on investing in yourself, Phil, “ the General would like to say when he’d look up from his Wall Street Journal before our morning meetings and remark about how stock analysts always liked to place their value on the wrong thing.

“Sure, these companies have to make a profit, but you look at the Japanese and the Germans and they know the value of basic research, “ he once said to me.

“American companies expect the government to pay for all their research, and that’s what you and I have to do if we want to keep them working. But there’s going to come a time when we can’t afford to pay for it any longer. Then who’s going to foot the bill?”

General Trudeau was worrying about how the drive for new electronics products based upon miniaturized circuitry was creating entirely new markets that were shutting out American companies. He said that it was becoming cheaper for American companies to have their products manufactured for them in Asia, where companies had already retooled after the war to produce transistorized components, than for American companies, which had heavily invested in the manufacturing technology of the nineteenth century, to do it themselves.

 

He knew that the requirement for space exploration, for challenging the hostile EBEs in their own territory, relied on the development of an integrated circuit technology so that the electronic components of spacecraft could be miniaturized to fit the size requirements of rocket propelled vehicles. The race to develop more intelligent missiles and ordnance also required the development of new types of circuitry that could be packed into smaller and smaller spaces. But retooled Japanese and German industries were the only ones able to take immediate advantage of what General Trudeau called the “new electronics. “


For American industry to get onto the playing field the basic research would have to be paid for by the military. It was something General Trudeau was willing to fight for at the Pentagon because he knew that was the only way we could get the weapons only a handful of us knew we needed to fight a skirmish war against aliens only a handful of us knew we were fighting.

 

Arthur Trudeau was a battlefield general engaged in a lonely military campaign that national policy and secrecy laws forbade him even to talk about. And as the gulf of time widened between the Roswell crash and the concerns over postwar economic expansion, even the people who were fighting the war alongside General Trudeau were, one by one, beginning to die away. Industry could fight the war for us, General Trudeau believed, if it was properly seeded with ideas and the money to develop them. By 1961, we had turned our attention to the integrated circuit.


Government military weapons spending and the requirements for space exploration had already heavily funded the transistorized component circuit. The radars and missiles I was commanding at Red Canyon, New Mexico, in 1958 relied on miniaturized components for their reliability and portability. New generations of tracking radars on the drawing boards in 1960 were even more sophisticated and electronically intelligent than the weapons I was aiming at Soviet targets in Germany. In the United States, Japanese and Taiwanese radios that fit into the palm of your hand were on the market.

 

Computers like ENIAC, once the size of a small warehouse, now occupied rooms no larger than closets and, while still generating heat, no longer blew out because of overheated radio vacuum tubes. Minicomputers, helped by government R&D funding, were still a few years away from market, but were already in a design phase. Television sets and stereophonic phonographs that offered solid state electronics were coming on the market, and companies like IBM, Sperry-Rand, and NCR were beginning to bring electronic office machines onto the market. It was the beginning of a new age of electronics, helped, in part, by government funding of basic research into the development and manufacture of integrated circuit products.

 

But the real prize, the development of what actually had been recovered at Roswell, was still a few years away. When it arrived, again spurred by the requirements of military weapons development and space travel, it caused another revolution.


The history of the printed circuit and the microprocessor is also the history of a technology that allowed engineers to squeeze more and more circuitry into a smaller and smaller space. It’s the history of the integrated circuit, which developed throughout the 1960s, evolved into large scale integration by the early 1970s, very large scale integration by the middle 1970s, just before the emergence of the first real personal computers, and ultra large scale integration by the early 1980s. Today’s 200 plus megahertz, multigigabyte hard drive desktop computers are the results of the integrated circuit technology that began in the 1960s and has continued to the present. The jump from the basic transistorized integrated printed circuit of the 1960s to large scale integration was made possible by the development of the microprocessor in 1972.


Once the development process of engineering a more tightly compacted circuit had been inspired by the invention of the transistor in 1948, and fueled by the need to develop better, faster, and smaller computers, it continued on a natural progression until the engineers at Intel developed the first microprocessor, a four bit central processing unit called the 4004, in 1972.

 

This year marked the beginning of the microcomputer industry, although the first personal microcomputers didn’t appear on the market until the development of Intel’s 8080ª. That computer chip was the heart of the Altair computer, the first product to package a version of a high level programming language called BASIC, which allowed non-engineering types to program the machine and create applications for it. Soon companies like Motorola and Zilog had their own microprocessors on the market, and by 1977 the Motorola 6502-powered Apple II was on the market, joined by the 8080ª Radio Shack, the Commodore PET, the Atari, and the Heathkit.

 

Operationally, at its very heart, the microprocessor shares the same binary processing functions and large arrays of digital switches as its ancestors, the big mainframes of the 1950s and 1960s and the transistorized minis of the late 1960s and early 1970s. Functionally, the microprocessor also shares the same kinds of tasks as Charles Babbage’s Analytical Engine of the 1830s: reading numbers, storing numbers, logically processing numbers, and outputting the results. The microprocessor just puts everything into a much smaller space and moves it along at a much faster speed.

In 1979, Apple Computer had begun selling the first home computer floppy disk operating system for data and program storage that kicked the microcomputer revolution into a higher gear. Not only could users input data via a keyboard or tape cassette player, they could store relatively large amounts of data, such as documents or mathematical projections, on transportable, erasable, and interchangeable Mylar disks that the computer was able to read. Now the computer reached beyond the electronics hobbyist into the work place.

 

By the end of the year, MicroPro’s introduction of the first fully functional word processor called WordStar, and Personal Software’s release of the very first electronic spreadsheet called VisiCalc, so transformed the workplace that the desktop computer became a necessity for any young executive on his or her way up the corporate ladder. And by the early 1980s, with the introduction of the Apple Macintosh and the object oriented computer environment, not only the workplace but the whole world looked like a very different place than it did in the early 1960s.

 

Even Dr. Vannevar Bush’s concept of a type of research query language based not on a linear outline but on an intellectual relationship to something embedded in a body of text became a reality with the release of a computer program by Apple called HyperCard.

 

It was as if from the year 1947 to 1980 a fundamental paradigm shift in the ability of human kind to process information took place. Computers themselves almost became something like a silicon based life form, inspiring the carbon based life forms on planet Earth to develop them, grow them, and even help them reproduce. With computer directed process control programs now in place in virtually all major industries, software that writes software, neural network based expert systems that learn from their own experience in the real world, and current experiments under way to grow almost microscopically thin silicon based chips in the weightless environment of earth orbit may be the forerunner of a time when automated orbital factories routinely grow and harvest new silicon material for microprocessors more sophisticated than we can even imagine at the present.

 

Were all of this to be true, could it not be argued that the silicon wafers we recovered from Roswell were the real masters and space travelers and the EBE creatures their hosts or servants? Once implanted successfully on Earth, our culture having reached a point of readiness through its development of the first digital computers, would not the natural development stream, starting from the invention of the transistor, have carried us to the point where we achieve a symbiotic relationship with the silicon material that carries our data and enables us to become more creative and successful?


Maybe the Roswell crash, which helped us develop the technological basis for the weapons systems to protect our planet from the EBEs, was also the mechanism for successfully implanting a completely alien non-humanoid life form that survives from host to host like a virus, a digital Ebola that we humans will carry to another planet someday. Or what if an enemy wanted to implant the perfect spying or sabotage mechanism into a culture?

 

Then the implantation of the microchip based circuit into our technology by the EBEs would be the perfect method. Was it implanted as sabotage or as something akin to the gift of fire? Maybe the Roswell crash in 1947 was an event waiting to happen, like poisoned fruit dropping from the tree into a playground. Once bitten, the poison takes effect.

“Hold your horses, Phil, “ General Trudeau would say when I would speculate too much. “Remember, you’ve got a bunch of scientists you need to talk to and the people at Bell Labs are waiting to see your report when you’ve finished talking to the Alamogordo group. “

It was 1961 and the miniaturization of computer and electronic circuitry had already begun, but my report to the general and appointments he was arranging for me at Sperry-Rand, Hughes, and Bell Labs were for meetings with scientists to determine how their respective companies were proceeding with applying miniaturized circuitry into designs for weapons systems. The inspiration for microcircuitry had fallen out of the sky at Roswell and set the development of digital computers off in an entirely new direction. It was my job now to use the process of weapons development, especially the development of guidance systems for ballistic missiles, to implement the application of microcircuitry systems to these new generations of weapons.

 

General Trudeau and I were among the first scouts in what would be the electronic battlefield of the 1980s.

“Don’t worry, General, I’ve got my appointments all set up, “ I told him. I knew how carried away I could get, but I was an intelligence officer first, and that meant you start with a blank page and fill it in. “But I think the people at Bell Labs have already seen these things before.“

And they actually did - in 1947.

Back to Contents