
Gheorge Hagi
Anasayfa
Ben Kimim
Bilgisayarin Tarihcesi
Oyunlar
Super Lig
Bana Mail Atin
Sunum
|
|
Five Generations of Modern Computers
First Generation (1945-1956)
With the onset of the Second World War, governments sought to develop computers
to exploit their potential strategic importance. This increased funding for
computer development projects hastened technical progress. By 1941 German
engineer Konrad Zuse had developed a computer, the Z3, to design airplanes and
missiles. The Allied forces, however, made greater strides in developing
powerful computers. In 1943, the British completed a secret code-breaking
computer called Colossus to decode German messages. The Colossus's impact on the
development of the computer industry was rather limited for two important
reasons. First, Colossus was not a general-purpose computer; it was only
designed to decode secret messages. Second, the existence of the machine was
kept secret until decades after the war.
American efforts produced a broader achievement. Howard H. Aiken (1900-1973), a
Harvard engineer working with IBM, succeeded in producing an all-electronic
calculator by 1944. The purpose of the computer was to create ballistic charts
for the U.S. Navy. It was about half as long as a football field and contained
about 500 miles of wiring. The Harvard-IBM Automatic Sequence Controlled
Calculator, or Mark I for short, was a electronic relay computer. It used
electromagnetic signals to move mechanical parts. The machine was slow (taking
3-5 seconds per calculation) and inflexible (in that sequences of calculations
could not change); but it could perform basic arithmetic as well as more complex
equations.
Another computer development spurred by the war was the Electronic Numerical
Integrator and Computer (ENIAC), produced by a partnership between the U.S.
government and the University of Pennsylvania. Consisting of 18,000 vacuum
tubes, 70,000 resistors and 5 million soldered joints, the computer was such a
massive piece of machinery that it consumed 160 kilowatts of electrical power,
enough energy to dim the lights in an entire section of Philadelphia. Developed
by John Presper Eckert (1919-1995) and John W. Mauchly (1907-1980), ENIAC,
unlike the Colossus and Mark I, was a general-purpose computer that computed at
speeds 1,000 times faster than Mark I.
In the mid-1940's John von Neumann (1903-1957) joined the University of
Pennsylvania team, initiating concepts in computer design that remained central
to computer engineering for the next 40 years. Von Neumann designed the
Electronic Discrete Variable Automatic Computer (EDVAC) in 1945 with a memory to
hold both a stored program as well as data. This "stored memory" technique as
well as the "conditional control transfer," that allowed the computer to be
stopped at any point and then resumed, allowed for greater versatility in
computer programming. The key element to the von Neumann architecture was the
central processing unit, which allowed all computer functions to be coordinated
through a single source. In 1951, the UNIVAC I (Universal Automatic Computer),
built by Remington Rand, became one of the first commercially available
computers to take advantage of these advances. Both the U.S. Census Bureau and
General Electric owned UNIVACs. One of UNIVAC's impressive early achievements
was predicting the winner of the 1952 presidential election, Dwight D.
Eisenhower.
First generation computers were characterized by the fact that operating
instructions were made-to-order for the specific task for which the computer was
to be used. Each computer had a different binary-coded program called a machine
language that told it how to operate. This made the computer difficult to
program and limited its versatility and speed. Other distinctive features of
first generation computers were the use of vacuum tubes (responsible for their
breathtaking size) and magnetic drums for data storage.
Second Generation Computers (1956-1963)
By 1948, the invention of the transistor greatly changed the computer's
development. The transistor replaced the large, cumbersome vacuum tube in
televisions, radios and computers. As a result, the size of electronic machinery
has been shrinking ever since. The transistor was at work in the computer by
1956. Coupled with early advances in magnetic-core memory, transistors led to
second generation computers that were smaller, faster, more reliable and more
energy-efficient than their predecessors. The first large-scale machines to take
advantage of this transistor technology were early supercomputers, Stretch by
IBM and LARC by Sperry-Rand. These computers, both developed for atomic energy
laboratories, could handle an enormous amount of data, a capability much in
demand by atomic scientists. The machines were costly, however, and tended to be
too powerful for the business sector's computing needs, thereby limiting their
attractiveness. Only two LARCs were ever installed: one in the Lawrence
Radiation Labs in Livermore, California, for which the computer was named
(Livermore Atomic Research Computer) and the other at the U.S. Navy Research and
Development Center in Washington, D.C. Second generation computers replaced
machine language with assembly language, allowing abbreviated programming codes
to replace long, difficult binary codes.
Throughout the early 1960's, there were a number of commercially successful
second generation computers used in business, universities, and government from
companies such as Burroughs, Control Data, Honeywell, IBM, Sperry-Rand, and
others. These second generation computers were also of solid state design, and
contained transistors in place of vacuum tubes. They also contained all the
components we associate with the modern day computer: printers, tape storage,
disk storage, memory, operating systems, and stored programs. One important
example was the IBM 1401, which was universally accepted throughout industry,
and is considered by many to be the Model T of the computer industry. By 1965,
most large business routinely processed financial information using second
generation computers.
It was the stored program and programming language that gave computers the
flexibility to finally be cost effective and productive for business use. The
stored program concept meant that instructions to run a computer for a specific
function (known as a program) were held inside the computer's memory, and could
quickly be replaced by a different set of instructions for a different function.
A computer could print customer invoices and minutes later design products or
calculate paychecks. More sophisticated high-level languages such as COBOL
(Common Business-Oriented Language) and FORTRAN (Formula Translator) came into
common use during this time, and have expanded to the current day. These
languages replaced cryptic binary machine code with words, sentences, and
mathematical formulas, making it much easier to program a computer. New types of
careers (programmer, analyst, and computer systems expert) and the entire
software industry began with second generation computers.
|
|
Third Generation Computers (1964-1971)
Though transistors were clearly an improvement over the vacuum tube, they still
generated a great deal of heat, which damaged the computer's sensitive internal
parts. The quartz rock eliminated this problem. Jack Kilby, an engineer with
Texas Instruments, developed the integrated circuit (IC) in 1958. The IC
combined three electronic components onto a small silicon disc, which was made
from quartz. Scientists later managed to fit even more components on a single
chip, called a semiconductor. As a result, computers became ever smaller as more
components were squeezed onto the chip. Another third-generation development
included the use of an operating system that allowed machines to run many
different programs at once with a central program that monitored and coordinated
the computer's memory.
Fourth Generation (1971-Present)
After the integrated circuits, the only place to go was down - in size, that is.
Large scale integration (LSI) could fit hundreds of components onto one chip. By
the 1980's, very large scale integration (VLSI) squeezed hundreds of thousands
of components onto a chip. Ultra-large scale integration (ULSI) increased that
number into the millions. The ability to fit so much onto an area about half the
size of a U.S. dime helped diminish the size and price of computers. It also
increased their power, efficiency and reliability. The Intel 4004 chip,
developed in 1971, took the integrated circuit one step further by locating all
the components of a computer (central processing unit, memory, and input and
output controls) on a minuscule chip. Whereas previously the integrated circuit
had had to be manufactured to fit a special purpose, now one microprocessor
could be manufactured and then programmed to meet any number of demands. Soon
everyday household items such as microwave ovens, television sets and
automobiles with electronic fuel injection incorporated microprocessors.
Such condensed power allowed everyday people to harness a computer's power. They
were no longer developed exclusively for large business or government contracts.
By the mid-1970's, computer manufacturers sought to bring computers to general
consumers. These minicomputers came complete with user-friendly software
packages that offered even non-technical users an array of applications, most
popularly word processing and spreadsheet programs. Pioneers in this field were
Commodore, Radio Shack and Apple Computers. In the early 1980's, arcade video
games such as Pac Man and home video game systems such as the Atari 2600 ignited
consumer interest for more sophisticated, programmable home computers.
In 1981, IBM introduced its personal computer (PC) for use in the home, office
and schools. The 1980's saw an expansion in computer use in all three arenas as
clones of the IBM PC made the personal computer even more affordable. The number
of personal computers in use more than doubled from 2 million in 1981 to 5.5
million in 1982. Ten years later, 65 million PCs were being used. Computers
continued their trend toward a smaller size, working their way down from desktop
to laptop computers (which could fit inside a briefcase) to palmtop (able to fit
inside a breast pocket). In direct competition with IBM's PC was Apple's
Macintosh line, introduced in 1984. Notable for its user-friendly design, the
Macintosh offered an operating system that allowed users to move screen icons
instead of typing instructions. Users controlled the screen cursor using a
mouse, a device that mimicked the movement of one's hand on the computer screen.
As computers became more widespread in the workplace, new ways to harness their
potential developed. As smaller computers became more powerful, they could be
linked together, or networked, to share memory space, software, information and
communicate with each other. As opposed to a mainframe computer, which was one
powerful computer that shared time with many terminals for many applications,
networked computers allowed individual computers to form electronic co-ops.
Using either direct wiring, called a Local Area Network (LAN), or telephone
lines, these networks could reach enormous proportions. A global web of computer
circuitry, the Internet, for example, links computers worldwide into a single
network of information. During the 1992 U.S. presidential election,
vice-presidential candidate Al Gore promised to make the development of this
so-called "information superhighway" an administrative priority. Though the
possibilities envisioned by Gore and others for such a large network are often
years (if not decades) away from realization, the most popular use today for
computer networks such as the Internet is electronic mail, or E-mail, which
allows users to type in a computer address and send messages through networked
terminals across the office or across the world.
Fifth Generation (Present and Beyond)
Defining the fifth generation of computers is somewhat difficult because the
field is in its infancy. The most famous example of a fifth generation computer
is the fictional HAL9000 from Arthur C. Clarke's novel, 2001: A Space Odyssey.
HAL performed all of the functions currently envisioned for real-life fifth
generation computers. With artificial intelligence, HAL could reason well enough
to hold conversations with its human operators, use visual input, and learn from
its own experiences. (Unfortunately, HAL was a little too human and had a
psychotic breakdown, commandeering a spaceship and killing most humans on
board.)
Though the wayward HAL9000 may be far from the reach of real-life computer
designers, many of its functions are not. Using recent engineering advances,
computers are able to accept spoken word instructions (voice recognition) and
imitate human reasoning. The ability to translate a foreign language is also
moderately possible with fifth generation computers. This feat seemed a simple
objective at first, but appeared much more difficult when programmers realized
that human understanding relies as much on context and meaning as it does on the
simple translation of words.
Many advances in the science of computer design and technology are coming
together to enable the creation of fifth-generation computers. Two such
engineering advances are parallel processing, which replaces von Neumann's
single central processing unit design with a system harnessing the power of many
CPUs to work as one. Another advance is superconductor technology, which allows
the flow of electricity with little or no resistance, greatly improving the
speed of information flow. Computers today have some attributes of fifth
generation computers. For example, expert systems assist doctors in making
diagnoses by applying the problem-solving steps a doctor might use in assessing
a patient's needs. It will take several more years of development before expert
systems are in widespread use.
|