Computer Evolution Computer History and Development Nothing epitomizes modern life better than the computer. For better or worse, computers have Infiltrated every aspect of our society. Today computers do much more than simply compute: supermarket scanners calculate our grocery bill while keeping store Inventory; computerized telephone switching centers play traffic cop to millions of calls and keep lines of communication untangled: and automatic teller machines (ATM) let us conduct banking transactions from virtually anywhere in the world.
But where did all this technology come from and where is it heading? To fully understand and appreciate the impact computers have on our lives and promises they hold for the future, it is important to understand their evolution. Early Computing Machines and Inventors The abacus, which emerged about 5,000 years ago in Asia Minor and is still in use today, may be considered the first computer. This device allows users to make computations using a system of sliding beads arranged on a rack. Early merchants used the abacus to keep trading transactions.
But as the use of paper and pencil spread, particularly In Europe, the abacus lost Its Importance. It took nearly 12 enturies, however, for the next significant advance In computing devices to emerge. In 1642, Blaise Pascal (1623 1 662), the 18-year-old son of a French tax collector, invented what he called a numerical wheel calculator to help his father with his duties. This brass rectangular box, also called a Pascaline, used eight movable dials to add sums up to eight figures long. Pascal’s device used a base of ten to accomplish this.
For example, as one dial moved ten notches, or one complete revolution, it moved the next dial – which represented the ten’s column – one place. When the ten’s ial moved one revolution, the dial representing the hundred’s place moved one notch and so on. The drawback to the Pascallne, of course, was its limitation to addition. In 1694, a German mathematician and philosopher, Gottfried Wilhem von Leibniz (1646-1716b Improved the Pascallne by creating a machine that could also multiply. Like Its predecessor, Leibniz’s mechanical multlpller worked by a system of gears and dials.
Partly by studying Pascal’s original notes and drawings, Leibniz was able to refine his machine. The centerpiece of the machine was its stepped-drum gear design, which offered an elongated version of the simple flat gear. It wasn’t until 1820, however, that mechanical calculators gained widespread use. Charles Xavier Thomas de Colmar, a Frenchman, invented a machine that could perform the four basic arithmetic functions. Colmar’s mechanical calculator, the arithometer, presented a more practical approach to computing because it could add, subtract, multiply and divide.
With its enhanced versatility, the arithometer was widely used up until the First World War. Although later Inventors refined Colmar’s calculator, together with fellow Inventors Pascal and Leibniz, he helped define the age of mechanical computation. The real beginnings of computers as we know them today, however, lay with an English mathematics professor, Charles Babbage (1791-1871). Frustrated at the many errors he found while examining calculations for the Royal performed by steam! ” With those words, the automation of computers had begun.
By 1812, Babbage noticed a natural harmony between machines and mathematics: machines were best at performing tasks repeatedly without mistake; while mathematics, particularly the production of mathematics tables, often required the simple repetition of steps. The problem centered on applying the ability of machines o the needs of mathematics. Babbage’s first attempt at solving this problem was in 1822 when he proposed a machine to perform differential equations, called a Difference Engine. Powered by steam and large as a locomotive, the machine would have a stored program and could perform calculations and print the results automatically.
After working on the Difference Engine for 10 years, Babbage was suddenly inspired to begin work on the first general-purpose computer, which he called the Analytical Engine. Babbage’s assistant, Augusta Ada King, Countess of Lovelace (1815-1842) and daughter of English poet Lord Byron, was instrumental in he machine’s design. One of the few people who understood the Engine’s design as well as Babbage, she helped revise plans, secure funding from the British government, and communicate the specifics of the Analytical Engine to the public.
Also, Lady Lovelace’s fine understanding of the machine allowed her to create the instruction routines to be fed into the computer, making her the first female computer programmer. In the 1980’s, the U. S. Defense Department named a programming language ADA in her honor. Babbage’s steam-powered Engine, although ultimately never constructed, may seem primitive by today’s standards. However, it outlined the basic elements of a modern general purpose computer and was a breakthrough concept.
Consisting of over 50,000 components, the basic design of the Analytical Engine included input devices in the form of perforated cards containing operating instructions and a “store” for memory of 1,000 numbers of up to 50 decimal digits long. It also contained a “mill” with a control unit that allowed processing instructions in any sequence, and output devices to produce printed results. Babbage borrowed the idea of punch cards to encode the machine’s instructions from the Jacquard loom. The loom, produced in 1820 and named after its inventor, Joseph-Marie Jacquard, used punched boards that controlled the patterns to be woven.
In 1889, an American inventor, Herman Hollerith (186(:)-1929), also applied the Jacquard loom concept to computing. His first task was to find a faster way to compute the U. S. census. The previous census in 1880 had taken nearly seven years to count and with an expanding population, the bureau feared it would take 10 years to count the latest census. Unlike Babbage’s idea of using perforated cards to instruct the machine, Hollerith’s method used cards to store data information which e fed into a machine that compiled the results mechanically.
Each punch on a card represented one number, and combinations of two punches represented one letter. As many as 80 variables could be stored on a single card. Instead of ten years, census takers compiled their results in Just six weeks with Hollerith’s machine. In addition to their speed, the punch cards served as a storage method for data and they helped reduce computational errors. Hollerith brought his punch card reader into the business world, founding Tabulating Machine Company in 1896, later to become ompanies such as Remington Rand and Burroghs also manufactured punch readers for business use.
Both business and government used punch cards for data processing until the 1960’s. In the ensuing years, several engineers made other significant advances. Vannevar Bush (1890-1974) developed a calculator for solving differential equations in 1931. The machine could solve complex differential equations that had long left scientists and mathematicians baffled. The machine was cumbersome because hundreds of gears and shafts were required to represent numbers and their various relationships to each other. To eliminate this bulkiness, John V. Atanasoff (b. 903), a professor at Iowa State College (now called Iowa State University) and his graduate student, Clifford Berry, envisioned an all-electronic computer that applied Boolean algebra to computer circuitry. This approach was based on the mid-19th century work of George Boole (1815-1864) who clarified the binary system of algebra, which stated that any mathematical equations could be stated simply as either true or false. By extending this concept to electronic circuits in the form of on or off, Atanasoff and Berry had developed the first all-electronic computer by 1940.
Their project, however, lost its funding and their work was overshadowed by similar developments by other scientists. Five Generations of Modern Computers First Generation (1945-1956) With the onset of the Second World War, governments sought to develop computers to exploit their potential strategic importance. This increased funding for computer development projects hastened technical progress. By 1941 German engineer Konrad Zuse had developed a computer, the Z3, to design airplanes and missiles. The Allied forces, however, made greater strides in developing powerful computers.
In 1943, the British completed a secret code-breaking computer called Colossus to decode German messages. The Colossus’s impact on the development of the computer industry was rather limited for two important reasons. First, Colossus was not a general-purpose computer; it was only designed to decode secret messages. Second, the existence of the machine was kept secret until decades after the war. American efforts produced a broader achievement. Howard H. Aiken (1900-1973), a Harvard engineer working with 18M, succeeded in producing an all-electronic calculator by 1944.
The purpose of the computer was to create ballistic charts for the U. S. Navvy. It was about half as long as a football field and contained about 500 miles of wiring. The Harvard-IBM Automatic Sequence Controlled Calculator, or Mark I for short, was an electronic relay computer. It used electromagnetic signals to move mechanical parts. The machine was slow (taking 3-5 seconds per calculation) and inflexible (in that sequences of calculations could not change); but it could perform basic arithmetic as well as more complex equations.
Another computer development spurred by the war was the Electronic Numerical Integrator and Computer (ENIAC), produced by a partnership between the U. S. overnment and the University of Pennsylvania. Consisting of 18,000 vacuum tubes, 70,000 resistors and 5 million soldered Joints, the computer was such a massive piece of machinery that it entire section of Philadelphia. Developed by John Presper Eckert (1919-1995) and John W. Mauchly (1907-1980), ENIAC, unlike the Colossus and Mark l, was a general- purpose computer that computed at speeds 1,000 times faster than Mark l.
In the mid-1940’s John von Neumann (1903-1957) Joined the University of Pennsylvania team, initiating concepts in computer design that remained central to computer engineering for the next 40 years. Von Neumann designed the Electronic Discrete Variable Automatic Computer (EDVAC) in 1945 with a memory to hold both a stored program as well as data. This “stored memory” technique as well as the “conditional control transfer,” that allowed the computer to be stopped at any point and then resumed, allowed for greater versatility in computer programming.
The key element to the von Neumann architecture was the central processing unit, which allowed all computer functions to be coordinated through a single source. In 1951, the UNIVAC I (Universal Automatic Computer), built by Remington Rand, became one f the first commercially available computers to take advantage of these advances. Both the U. S. Census Bureau and General Electric owned UNIVACs. One of UNIVAC’s impressive early achievements was predicting the winner of the 1952 presidential election, Dwight D. Eisenhower.
First generation computers were characterized by the fact that operating instructions were made-to-order for the specific task for which the computer was to be used. Each computer had a different binary-coded program called a machine language that told it how to operate. This made the computer difficult to program and limited its versatility and speed. Other distinctive features of first generation computers were the use of vacuum tubes (responsible for their breathtaking size) and magnetic drums for data storage.
Second Generation computers (1956-1963) By 1948, the invention of the transistor greatly changed the computer’s development. The transistor replaced the large, cumbersome vacuum tube in televisions, radios and computers. As a result, the size of electronic machinery has been shrinking ever since. The transistor was at work in the computer by 1956. Coupled with early advances in magnetic-core memory, transistors led to second generation computers hat were smaller, faster, more reliable and more energy-efficient than their predecessors.
The first large-scale machines to take advantage of this transistor technology were early supercomputers, Stretch by IBM and LARC by Sperry-Rand. These computers, both developed for atomic energy laboratories, could handle an enormous amount of data, a capability much in demand by atomic scientists. The machines were costly, however, and tended to be too powerful for the business sector’s computing needs, thereby limiting their attractiveness. Only two LARCs were ever installed: one in the Lawrence Radiation Labs in Livermore, California, for which he computer was named (Livermore Atomic Research Computer) and the other at the U.
S. Navvy Research and Development Center in Washington, D. C. Second generation computers replaced machine language with assembly language, allowing abbreviated programming codes to replace long, difficult binary codes. Throughout the early 1960’s, there were a number of commercially successful second generation computers used in business, universities, and government from companies such as generation computers were also of solid state design, and contained transistors in place of vacuum tubes.
They also contained all the components we associate with the odern day computer: printers, tape storage, disk storage, memory, operating systems, and stored programs. One important example was the IBM 1401, which was universally accepted throughout industry, and is considered by many to be the Model T of the computer industry. By 1965, most large business routinely processed financial information using second generation computers. It was the stored program and programming language that gave computers the flexibility to finally be cost effective and productive for business use.
The stored program concept meant that instructions to run a computer for a specific function (known as a program) were held nside the computer’s memory, and could quickly be replaced by a different set of instructions for a different function. A computer could print customer invoices and minutes later design products or calculate paychecks. More sophisticated high-level languages such as COBOL (Common Business-oriented Language) and FORTRAN (Formula Translator) came into common use during this time, and have expanded to the current day.
These languages replaced cryptic binary machine code with words, sentences, and mathematical formulas, making it much easier to program a computer. New types of careers (programmer, analyst, and computer systems expert) nd the entire software industry began with second generation computers. Third Generation Computers (1964-1971) Though transistors were clearly an improvement over the vacuum tube, they still generated a great deal of heat, which damaged the computer’s sensitive internal parts. The quartz rock eliminated this problem.
Jack Kilby, an engineer with Texas Instruments, developed the integrated circuit (‘C) in 1958. The IC combined three electronic components onto a small silicon disc, which was made from quartz. Scientists later managed to fit even more components on a single chip, called a semiconductor. As a result, computers became ever smaller as more components were squeezed onto the chip. Another third-generation development included the use of an operating system that allowed machines to run many different programs at once with a central program that monitored and coordinated the computer’s memory.
Fourth Generation (1971 -Present) After the integrated circuits, the only place to go was down – in size, that is. Large scale integration (LSI) could fit hundreds of components onto one chip. By the 1980’s, very large scale integration (VLSI) squeezed hundreds of thousands of components onto a chip. Ultra-large scale integration (ULSI) increased that number into the millions. The ability to fit so much onto an area about half the size of a U. S. dime helped diminish the size and price of computers. It also increased their power, efficiency and reliability.
The Intel 4004 chip, developed in 1971, took the integrated circuit one step further by locating all the components of a computer (central processing unit, memory, and input and output controls) on a minuscule chip. Whereas previously the integrated circuit had had to be manufactured to fit a special purpose, now one microprocessor could be manufactured and then programmed to eet any number of demands. Soon everyday household items such as microwave microprocessors. Such condensed power allowed everyday people to harness a computer’s power.
They were no longer developed exclusively for large business or government contracts. By the mid-1970’s, computer manufacturers sought to bring computers to general consumers. These minicomputers came complete with user- friendly software packages that offered even non-technical users an array of applications, most popularly word processing and spreadsheet programs. Pioneers in this field were Commodore, Radio Shack and Apple Computers. In the early 1980’s, rcade video games such as Pac Man and home video game systems such as the Atari 2600 ignited consumer interest for more sophisticated, programmable home computers.
In 1981, IBM introduced its personal computer (PC) for use in the home, office and schools. The 1980’s saw an expansion in computer use in all three arenas as clones of the IBM PC made the personal computer even more affordable. The number of personal computers in use more than doubled from 2 million in 1981 to 5. 5 million in 1982. Ten years later, 65 million PCs were being used. Computers continued their trend toward a smaller size, working their way down from desktop to aptop computers (which could fit inside a briefcase) to palmtop (able to fit inside a breast pocket).
In direct competition with IBM’s PC was Apple’s Macintosh line, introduced in 1984. Notable for its user-friendly design, the Macintosh offered an operating system that allowed users to move screen icons instead of typing instructions. Users controlled the screen cursor using a mouse, a device that mimicked the movement of one’s hand on the computer screen. As computers became more widespread in the workplace, new ways to harness their potential developed. As smaller computers became more powerful, they could be linked ogether, or networked, to share memory space, software, information and communicate with each other.
As opposed to a mainframe computer, which was one powerful computer that shared time with many terminals for many applications, networked computers allowed individual computers to form electronic co-ops. Using either direct wiring, called a Local Area Network (LAN), or telephone lines, these networks could reach enormous proportions. A global web of computer circuitry, the Internet, for example, links computers worldwide into a single network of information. During the 1992 U. S. residential election, vice-presidential candidate A1 Gore promised to make the development of this so-called “information superhighway” an administrative priority. Though the possibilities envisioned by Gore and others for such a large network are often years (if not decades) away from realization, the most popular use today for computer networks such as the Internet is electronic mail, or E-mail, which allows users to type in a computer address and send messages through networked terminals across the office or across the world.
Fifth Generation (Present and Beyond) Defining the fifth generation of computers is somewhat difficult because the field is n its infancy. The most famous example of a fifth generation computer is the fictional HAL9000 from Arthur C. Clarke’s novel, 2001: A Space Odyssey. HAL performed all of the functions currently envisioned for real-life fifth generation computers. With artificial intelligence, HAL could reason well enough to hold conversations with its human operators, use visual input, and learn from its own experiences. commandeering a spaceship and killing most humans on board. Though the wayward HAL9000 may be far from the reach of real-life computer designers, many of its functions are not. Using recent engineering advances, computers are able to accept spoken word instructions (voice recognition) and imitate human reasoning. The ability to translate a foreign language is also moderately possible with fifth generation computers. This feat seemed a simple objective at first, but appeared much more difficult when programmers realized that human understanding relies as much on context and meaning as it does on the simple translation of words.
Many advances in the science of computer design and technology are coming together to enable the creation of fifth-generation computers. Two such engineering advances re parallel processing, which replaces von Neumann’s single central processing unit design with a system harnessing the power of many CPUs to work as one. Another advance is superconductor technology, which allows the flow of electricity with little or no resistance, greatly improving the speed of information flow. Computers today have some attributes of fifth generation computers.
For example, expert systems assist doctors in making diagnoses by applying the problem-solving steps a doctor might use in assessing a patient’s needs. It will take several more years of development before expert systems are in widespread use. Personal Computers History and Development The personal computer (PC) has revolutionized business and personal activities and even the way people talk and think; however, its development has been less of a revolution than an evolution and convergence of three critical elements – thought, hardware, and software.
Although the PC traces its lineage to the mainframe and minicomputers of the 1950s and 1960s, the conventional thought that was prevalent during the first thirty years of the computer age saw no value in a small computer that could be used by individuals. A PC is a microcomputer, so named because it is maller than a minicomputer, which in turn is smaller than a mainframe computer. While early mainframes and their peripheral devices often took up the floor space of a house, minicomputers are about the size of a refrigerator and stove.
The microcomputer, whose modern development traces back to the early 1970s, and fits on a desk. From the start, the creation of the computer was centered around the concept that a single unit would be used to perform complex calculations with greater speed and accuracy than humans could achieve. The Transistor On December 23, 1947, one of the most far-reaching technologies of the 20th Century as developed at Bell Laboratories by John Bardeen, Walter Brattain, and William Shockley – the transistor. But the transistor wasn’t available to U. S. anufacturers until 1956, when a seven year-old antitrust law suit against AT&T, the owners of Bell Labs, was settled. The Judgment required that AT&T give away licenses to manufacture the transistor to American companies. Following this decision, the transistor was used to replace thousands of vacuum tubes in computers and began the miniaturization of electronics. Because it drastically reduced the size and heat onsiderations of the large vacuum tubes, the transistor enabled the computer to become a viable tool for business and government.
The Computer Mystique corporate and government offices and on university campuses, information processing departments sprouted up to serve the computer. The IBM 701, which was introduced in 1952 as a business computer, was comprised of several units that could be shipped and connected at a customer’s location, rather than the earlier massive units that had to be assembled on site. In 1953, IBM began shipping the first mass-produced computer, the IBM 650. IBM introduced the first solid-state transistorized) computer in 1959, the IBM 7090.
Then in 1964, IBM culminated over $1 billion in research when it brought out the System/360 series of computers. Unlike other mainframes, the System/360 computers were compatible with each other. By 1960, the computer was king. Companies hired armies of technicians and programmers to write its operating programs and software, fix it, and allocate the precious computer time. The capability of the machines was more than a mere mortal could fathom, but gathering raw data and “keying” it in so the computer could “crunch the numbers” was a complicated and time-consuming task.
Frustrations abounded, computer errors were called “glitches,” and the phrases “garbage in/ garbage out,” “It’s a computer mistake,” and “Sorry, the computer’s down and we cant do anything,” were introduced into the lexicon. On college campuses in the 1960s, students carried bundles of computer cards to and from class, hoping that their share of the valuable computer time would not be bumped or allocated to someone else. The term, “Do not fold, spindle or mutilate,” was coined so people wouldn’t disable the process of feeding the punched computer cards into punch card readers, where the intricate patterns of holes were decoded.
The computer mystique was reinforced in people every time they heard of some new accomplishment. In 1961, a computer calculated the value of pi to 100,000 decimal places. A computer could play checkers, and in 1967 a chess playing computer program was made an honorary member of the United States Chess Federation. Banks began printing checks with magnetic ink so they could be processed by the computers. A Small Change in Thought Until 1971, nobody even thought of a computer as anything but a big, fast, electronic brain that resided in a climate-controlled room and consumed data and electricity in assive quantities.
In 1971, an Intel 4004 chip containing 4004 transistors was programmed to perform complex mathematical calculations; the hand-held calculator was born. Suddenly, scientists and engineers could carry the computational power of a computer with them to Job sites, classrooms, and laboratories; but the hand-held calculator, like the ENIAC before it, was not yet a computer. The microprocessor was developed by Robert Noyce, the founder of Intel and one of the inventors of the integrated circuit, and brought with it a change in the way people worked.
New Technologies and New Ideas Small, hand-held calculators had provided an idea, or at least a “what if,” to some people. Still, in the early 1970s, computers were used for number crunching and printing out long streams of green and white paper. IBM Selectric typewriters were the top of the line “word processors” and Xerox copiers churned out photocopies. Most people never imagined that a computer could process data in real time, be of processing 8-bits of data, enough to convey numbers and letters of the alphabet.
In that same year, Xerox began working on a personal computer at their Palo Alto Research Center. For the next several years, a team of Xerox scientists worked on the “Alto,” a small computer that would have become the first PC if only the development team had been able to convince someone of its usefulness. Likewise, in 1972 Digital Equipment Corporation (DEC), a minicomputer manufacturing company headed by Kenneth Olsen, had a group of product engineers developing the DEC Datacenter. This PC incorporated not only the computer hardware but the desk as well.
The DEC Datacenter could have put tremendous computing capability in the home or at work, but management saw no value to the product and halted its development. In the end, none of the giant companies whose names had been synonymous with computers would introduce the PC to the world. There seemed to be no future in an inexpensive product that would replace the million dollar “Big Iron” that they were selling as fast as they could make them. The people who eventually introduced the PC were rebels. Many had spent time in the bowels of the big companies and were frustrated by the lack of vision they encountered.
They retreated into their own garages and attended meetings with other “computer nuts” who saw a much different future than the one aid out over the previous 30 years by the giants of the computer industry. The pc Born In 1975, Rubik’s Cube was put on store shelves and proved to many that the human brain was incapable of complex problem solving. But a ray of hope also appeared; the first PC was introduced. Micro Instrumentation and Telemetry Systems, Inc. (MITS) sold a kit for the MITS Altair 8800 that enabled computer hobbyists to assemble their own computers.
It had no monitor, no keyboard, no printer, and couldn’t store data, but the demand for it, like Rubik’s Cube, was overwhelming. The Altair proved that a PC was both possible and popular, but only with those people who would spend hours in their basements with soldering irons and wire strippers. The Altair, which looked like a control panel for a sprinkler system, didn’t last, but it helped launch one of the largest companies in the computer world and gave a couple of young software programmers a start.
In 1974, Bill Gates and Paul Allen wrote a version of BASIC for the Altair and started a company called Microsoft Corporation. In 1976, another computer kit was sold to hobbyists – the Apple l. Stephen Wozniak sold his Volkswagen and Steve Jobs sold his programmable calculator to get enough money to start Apple. In 1977, they introduced the Apple II, a pre-assembled PC with a color monitor, sound, and graphics. It was popular, but everyone knew that a serious computer didn’t need any of this. The kits were Just a hobby and the Apple II was seen as a toy.
Even the Apple name wasnt a serious, corporate sounding name like 18M, Digital Equipment Corporation, or Control Data. But 1977 also brought competition. The Zilog Z-80 microprocessor, which had been introduced in 1975, was used in the Tandy Radio Shack TRS-80, affectionately called the “Trash 80. Apple, Commodore, and Tandy dominated the PC marketplace. The Apple II had 16K bytes of RAM and 16K bytes of ROM; Commodore Business Machines’ Personal Electronic Transactor (PET) included 4K RAM and 14K ROM; and the TRS-80 had 4K RAM and 4K ROM.
Also in 1977, the Central Program for Microprocessors (CP/M) operating system CP/M was used in most PCs, but even that did not guarantee that a program or document could be written on one machine and read on another because each manufacturer used different floppy disk drives. Apple introduced the floppy disk drive in 1978, allowing Apple II users to store data on something other than the umbersome and unreliable tape cassettes that had been used up to that point. But despite the popularity of the three PCs, non-computer people still saw little reason to buy an expensive calculator when there were other ways to do the same things.
In 1979, that all changed. When VisiCalc was introduced for the Apple II, non-computer people suddenly saw a reason to buy a computer. VisiCalc, a spreadsheet program created by Dan Bricklin and Bob Frankston, allowed people to change one number in a budget and watch the effect it had on the entire budget. It was something new and valuable that could only be done with a computer. For thousands of people, the toy, the computer few could find a use for, had been transformed into a device that could actually do something worthwhile.
Microprocessors and high-tech gadgets were gradually worming their way into people’s lives. In 1978, Sony introduced the Beta format video cassette recorder, and a year later the VHS video recorder and the Sony Walkman. And to remind everyone of how far we had to go, Star Trek: The Motion Picture came to theaters in 1979. The Sinclair ZX-80 PC, which hit the market in 1980, used the same Z-80 chip as Commodore’s PET and the Tandy TRS-80. The ZX-80 had 1K RAM and 4K ROM. Developed by British entrepreneur Clive Sinclair, the ZX-80 meant that people could enter the computer revolution for under $200.
Its small size and price attracted people who had never thought about owning a PC. The Commodore VIC-20, also introduced in 1980, had a color monitor and would eventually become the first PC to sell more than one million units. Even with all of the success the early PC manufacturers had in the late 1970s and early 1980s, the advances in microprocessor speeds, and the creation of software, the PC was still not seen as a serious business tool. Unknown to everyone in the computer industry; owever, a huge oak tree was about to drop an acorn that would fall close to the tree and change everything.
Validation of the PC Two events occurred in 1981 that would have a tremendous impact on the future of the PC. In 1980, IBM had started a secret project in Boca Raton, Florida called “Acorn. ” Thirteen months later, in 1981, IBM introduced the IBM PC, a product that validated the PC as a legitimate business tool. For many people, even those who prided themselves on being able to operate the “Big Iron,” if IBM was making PCs then the small desk-top units were worthy of respect. When the IBM PC hit the market, it was a ompete system.