28 July 2009
In vitro plant culture
The invention: Method for propagating plants in artificial media
that has revolutionized agriculture.
The people behind the invention:
Georges Michel Morel (1916-1973), a French physiologist
Philip Cleaver White (1913- ), an American chemist
Plant Tissue Grows “In Glass”
In the mid-1800’s, biologists began pondering whether a cell isolated
from a multicellular organism could live separately if it were
provided with the proper environment. In 1902, with this question in
mind, the German plant physiologist Gottlieb Haberlandt attempted
to culture (grow) isolated plant cells under sterile conditions on an artificial
growth medium. Although his cultured cells never underwent
cell division under these “in vitro” (in glass) conditions, Haberlandt
is credited with originating the concept of cell culture.
Subsequently, scientists attempted to culture plant tissues and
organs rather than individual cells and tried to determine the medium
components necessary for the growth of plant tissue in vitro.
In 1934, Philip White grew the first organ culture, using tomato
roots. The discovery of plant hormones, which are compounds that
regulate growth and development, was crucial to the successful culture
of plant tissues; in 1939, Roger Gautheret, P. Nobécourt, and
White independently reported the successful culture of plant callus
tissue. “Callus” is an irregular mass of dividing cells that often results
from the wounding of plant tissue. Plant scientists were fascinated
by the perpetual growth of such tissue in culture and spent
years establishing optimal growth conditions and exploring the nutritional
and hormonal requirements of plant tissue.
Plants by the Millions
A lull in botanical research occurred during World War II, but
immediately afterward there was a resurgence of interest in applying
tissue culture techniques to plant research. Georges Morel, a plant physiologist at the National Institute for Agronomic Research
in France, was one of many scientists during this time who
had become interested in the formation of tumors in plants as well
as in studying various pathogens such as fungi and viruses that
cause plant disease.
To further these studies, Morel adapted existing techniques in order
to grow tissue from a wider variety of plant types in culture, and
he continued to try to identify factors that affected the normal
growth and development of plants. Morel was successful in culturing
tissue from ferns and was the first to culture monocot plants.
Monocots have certain features that distinguish them fromthe other
classes of seed-bearing plants, especially with respect to seed structure.
More important, the monocots include the economically important
species of grasses (the major plants of range and pasture)
and cereals.
For these cultures, Morel utilized a small piece of the growing tip
of a plant shoot (the shoot apex) as the starting tissue material. This
tissue was placed in a glass tube, supplied with a medium containing
specific nutrients, vitamins, and plant hormones, and allowed
to grow in the light. Under these conditions, the apex tissue grew
roots and buds and eventually developed into a complete plant.
Morel was able to generate whole plants from pieces of the shoot
apex that were only 100 to 250 micrometers in length.
Morel also investigated the growth of parasites such as fungi and
viruses in dual culture with host-plant tissue. Using results from
these studies and culture techniques that he had mastered, Morel
and his colleague Claude Martin regenerated virus-free plants from
tissue that had been taken from virally infected plants. Tissues from
certain tropical species, dahlias, and potato plants were used for the
original experiments, but after Morel adapted the methods for the
generation of virus-free orchids, plants that had previously been
difficult to propagate by any means, the true significance of his
work was recognized.
Morel was the first to recognize the potential of the in vitro culture
methods for the mass propagation of plants. He estimated that several
million plants could be obtained in one year from a single small
piece of shoot-apex tissue. Plants generated in this manner were
clonal (genetically identical organisms prepared from a single plant).With other methods of plant propagation, there is often a great variation
in the traits of the plants produced, but as a result of Morel’s
ideas, breeders could select for some desirable trait in a particular
plant and then produce multiple clonal plants, all of which expressed
the desired trait. The methodology also allowed for the production of
virus-free plant material, which minimized both the spread of potential
pathogens during shipping and losses caused by disease.
Consequences
Variations on Morel’s methods are used to propagate plants used
for human food consumption; plants that are sources of fiber, oil,
and livestock feed; forest trees; and plants used in landscaping and
in the floral industry. In vitro stocks are preserved under deepfreeze
conditions, and disease-free plants can be proliferated quickly
at any time of the year after shipping or storage.
The in vitro multiplication of plants has been especially useful
for species such as coconut and certain palms that cannot be propagated
by other methods, such as by sowing seeds or grafting, and
has also become important in the preservation and propagation of rare plant species that might otherwise have become extinct. Many
of these plants are sources of pharmaceuticals, oils, fragrances, and
other valuable products.
The capability of regenerating plants from tissue culture has also
been crucial in basic scientific research. Plant cells grown in culture
can be studied more easily than can intact plants, and scientists have
gained an in-depth understanding of plant physiology and biochemistry
by using this method. This information and the methods
of Morel and others have made possible the genetic engineering and
propagation of crop plants that are resistant to disease or disastrous
environmental conditions such as drought and freezing. In vitro
techniques have truly revolutionized agriculture.
IBM Model 1401 Computer
The invention: A relatively small, simple, and inexpensive computer
that is often credited with having launched the personal
computer age.
The people behind the invention:
Howard H. Aiken (1900-1973), an American mathematician
Charles Babbage (1792-1871), an English mathematician and
inventor
Herman Hollerith (1860-1929), an American inventor
Computers: From the Beginning
Computers evolved into their modern form over a period of
thousands of years as a result of humanity’s efforts to simplify the
process of counting. Two counting devices that are considered to be
very simple, early computers are the abacus and the slide rule.
These calculating devices are representative of digital and analog
computers, respectively, because an abacus counts numbers of things,
while the slide rule calculates length measurements.
The first modern computer, which was planned by Charles Babbage
in 1833, was never built. It was intended to perform complex
calculations with a data processing/memory unit that was controlled
by punched cards. In 1944, Harvard University’s Howard H.
Aiken and the International Business Machines (IBM) Corporation
built such a computer—the huge, punched-tape-controlled Automatic
Sequence Controlled Calculator, or Mark I ASCC, which
could perform complex mathematical operations in seconds. During
the next fifteen years, computer advances produced digital computers
that used binary arithmetic for calculation, incorporated
simplified components that decreased the sizes of computers, had
much faster calculating speeds, and were transistorized.
Although practical computers had become much faster than
they had been only a few years earlier, they were still huge and extremely
expensive. In 1959, however, IBM introduced the Model
1401 computer. Smaller, simpler, and much cheaper than the multimillion-dollar computers that were available, the IBM Model 1401
computer was also relatively easy to program and use. Its low cost,
simplicity of operation, and very wide use have led many experts
to view the IBM Model 1401 computer as beginning the age of the
personal computer.
Computer Operation and IBM’s Model 1401
Modern computers are essentially very fast calculating machines
that are capable of sorting, comparing, analyzing, and outputting information,
as well as storing it for future use. Many sources credit
Aiken’s Mark I ASCC as being the first modern computer to be built.
This huge, five-ton machine used thousands of relays to perform complex
mathematical calculations in seconds. Soon after its introduction,
other companies produced computers that were faster and more versatile
than the Mark I. The computer development race was on.
All these early computers utilized the decimal system for calculations
until it was found that binary arithmetic, whose numbers are
combinations of the binary digits 1 and 0, was much more suitable
for the purpose. The advantage of the binary system is that the electronic
switches that make up a computer (tubes, transistors, or
chips) can be either on or off; in the binary system, the on state can
be represented by the digit 1, the off state by the digit 0. Strung together
correctly, binary numbers, or digits, can be inputted rapidly
and used for high-speed computations. In fact, the computer term
bit is a contraction of the phrase “binary digit.”
A computer consists of input and output devices, a storage device
(memory), arithmetic and logic units, and a control unit. In
most cases, a central processing unit (CPU) combines the logic,
arithmetic, memory, and control aspects. Instructions are loaded
into the memory via an input device, processed, and stored. Then,
the CPU issues commands to the other parts of the system to carry
out computations or other functions and output the data as needed.
Most output is printed as hard copy or displayed on cathode-ray
tube monitors, or screens.
The early modern computers—such as the Mark I ASCC—were
huge because their information circuits were large relays or tubes.
Computers became smaller and smaller as the tubes were replaced first with transistors, then with simple integrated circuits, and then
with silicon chips. Each technological changeover also produced
more powerful, more cost-effective computers.
In the 1950’s, with reliable transistors available, IBM began the
development of two types of computers that were completed by
about 1959. The larger version was the Stretch computer, which was
advertised as the most powerful computer of its day. Customized
for each individual purchaser (for example, the Atomic Energy
Commission), a Stretch computer cost $10 million or more. Some innovations
in Stretch computers included semiconductor circuits,
new switching systems that quickly converted various kinds of data
into one language that was understood by the CPU, rapid data readers,
and devices that seemed to anticipate future operations.
Consequences
The IBM Model 1401 was the first computer sold in very large
numbers. It led IBM and other companies to seek to develop less expensive,
more versatile, smaller computers that would be sold to
small businesses and to individuals. Six years after the development
of the Model 1401, other IBM models—and those made by
other companies—became available that were more compact and
had larger memories. The search for compactness and versatility
continued. A major development was the invention of integrated
circuits by Jack S. Kilby of Texas Instruments; these integrated circuits
became available by the mid-1960’s. They were followed by
even smaller “microprocessors” (computer chips) that became available
in the 1970’s. Computers continued to become smaller and more
powerful.
Input and storage devices also decreased rapidly in size. At first,
the punched cards invented by Herman Hollerith, founder of the
Tabulation Machine Company (which later became IBM), were read
by bulky readers. In time, less bulky magnetic tapes and more compact
readers were developed, after which magnetic disks and compact
disc drives were introduced.
Many other advances have been made. Modern computers can
talk, create art and graphics, compose music, play games, and operate
robots. Further advancement is expected as societal needs change. Many experts believe that it was the sale of large numbers
of IBM Model 1401 computers that began the trend.
20 July 2009
Hydrogen bomb
The invention: Popularly known as the “H-Bomb,” the hydrogen
bomb differs from the original atomic bomb in using fusion,
rather than fission, to create a thermonuclear explosion almost a
thousand times more powerful.
The people behind the invention:
Edward Teller (1908- ), a Hungarian-born theoretical
physicist
Stanislaw Ulam (1909-1984), a Polish-born mathematician
Crash Development
Afew months before the 1942 creation of the Manhattan Project,
the United States-led effort to build the atomic (fission) bomb, physicist
Enrico Fermi suggested to Edward Teller that such a bomb
could release more energy by the process of heating a mass of the
hydrogen isotope deuterium and igniting the fusion of hydrogen
into helium. Fusion is the process whereby two atoms come together
to form a larger atom, and this process usually occurs only in stars,
such as the Sun. Physicists Hans Bethe, George Gamow, and Teller
had been studying fusion since 1934 and knew of the tremendous
energy than could be released by this process—even more energy
than the fission (atom-splitting) process that would create the atomic
bomb. Initially, Teller dismissed Fermi’s idea, but later in 1942, in
collaboration with Emil Konopinski, he concluded that a hydrogen
bomb, or superbomb, could be made.
For practical considerations, it was decided that the design of the
superbomb would have to wait until after the war. In 1946, a secret
conference on the superbomb was held in Los Alamos, New Mexico,
that was attended by, among other Manhattan Project veterans,
Stanislaw Ulam and Klaus Emil Julius Fuchs. Supporting the investigation
of Teller’s concept, the conferees requested a more complete
mathematical analysis of his own admittedly crude calculations
on the dynamics of the fusion reaction. In 1947, Teller believed
that these calculations might take years. Two years later, however,the Soviet explosion of an atomic bomb convinced Teller that America’s
ColdWar adversary was hard at work on its own superbomb.
Even when new calculations cast further doubt on his designs,
Teller began a vigorous campaign for crash development of the hydrogen
bomb, or H-bomb.
The Superbomb
Scientists knew that fusion reactions could be induced by the explosion
of an atomic bomb. The basic problem was simple and formidable:
How could fusion fuel be heated and compressed long
enough to achieve significant thermonuclear burning before the
atomic fission explosion blew the assembly apart? A major part of
the solution came from Ulam in 1951. He proposed using the energy
from an exploding atomic bomb to induce significant thermonuclear
reactions in adjacent fusion fuel components.
This arrangement, in which the A-bomb (the primary) is physically
separated from the H-bomb’s (the secondary’s) fusion fuel, became
known as the “Teller-Ulam configuration.” All H-bombs are
cylindrical, with an atomic device at one end and the other components
filling the remaining space. Energy from the exploding primary
could be transported by X rays and would therefore affect the
fusion fuel at near light speed—before the arrival of the explosion.
Frederick de Hoffman’s work verified and enriched the new concept.
In the revised method, moderated X rays from the primary irradiate
a reactive plastic medium surrounding concentric and generally
cylindrical layers of fusion and fission fuel in the secondary.
Instantly, the plastic becomes a hot plasma that compresses and
heats the inner layer of fusion fuel, which in turn compresses a central
core of fissile plutonium to supercriticality. Thus compressed,
and bombarded by fusion-produced, high-energy neutrons, the fission
element expands rapidly in a chain reaction from the inside
out, further compressing and heating the surrounding fusion fuel,
releasing more energy and more neutrons that induce fission in a
fuel casing-tamper made of normally stable uranium 238.
With its equipment to refrigerate the hydrogen isotopes, the device
created to test Teller’s new concept weighed more than sixty
tons. During Operation Ivy, it was tested at Elugelab in the Marshall Islands on November 1, 1952. Exceeding the expectations of all concerned
and vaporizing the island, the explosion equaled 10.4 million
tons of trinitrotoluene (TNT), which meant that it was about
seven hundred times more powerful than the atomic bomb dropped
on Hiroshima, Japan, in 1945. A version of this device weighing
about 20 tons was prepared for delivery by specially modified Air
Force B-36 bombers in the event of an emergency during wartime.
In development at Los Alamos before the 1952 test was a device
weighing only about 4 tons, a “dry bomb” that did not require refrigeration
equipment or liquid fusion fuel; when sufficiently compressed
and heated in its molded-powder form, the new fusion fuel
component, lithium-6 deutride, instantly produced tritium, an isotope
of hydrogen. This concept was tested during Operation Castle
at Bikini atoll in 1954 and produced a yield of 15 million tons of TNT,
the largest-ever nuclear explosion created by the United States.
Consequences
Teller was not alone in believing that the world could produce
thermonuclear devices capable of causing great destruction. Months
before Fermi suggested to Teller the possibility of explosive thermonuclear
reactions on Earth, Japanese physicist Tokutaro Hagiwara
had proposed that a uranium 235 bomb could ignite significant fusion
reactions in hydrogen. The Soviet Union successfully tested an
H-bomb dropped from an airplane in 1955, one year before the
United States did so.
Teller became the scientific adviser on nuclear affairs of many
presidents, from Dwight D. Eisenhower to Ronald Reagan. The
widespread blast and fallout effects of H-bombs assured the mutual
destruction of the users of such weapons. During the Cold War
(from about 1947 to 1981), both the United States and the Soviet
Union possessed H-bombs. “Testing” these bombs made each side
aware of how powerful the other side was. Everyone wanted to
avoid nuclear war. It was thought that no one would try to start a
war that would end in the world’s destruction. This theory was
called deterrence: The United States wanted to let the Soviet Union
know that it had just as many bombs, or more, than it did, so that the
leaders of the Sovet Union would be deterred from starting a war.Teller knew that the availability of H-bombs on both sides was
not enough to guarantee that such weapons would never be used. It
was also necessary to make the Soviet Union aware of the existence
of the bombs through testing. He consistently advised against U.S.
participation with the Soviet Union in a moratorium (period of
waiting) on nuclear weapons testing. Largely based on Teller’s urging
that underground testing be continued, the United States rejected
a total moratorium in favor of the 1963 Atmospheric Test Ban
Treaty.
During the 1980’s, Teller, among others, convinced President
Reagan to embrace the Strategic Defense Initiative (SDI). Teller argued
that SDI components, such as the space-based “Excalibur,” a
nuclear bomb-powered X-ray laser weapon proposed by the Lawrence-
Livermore National Laboratory, would make thermonuclear
war not unimaginable, but theoretically impossible.
19 July 2009
Hovercraft
The invention: A vehicle requiring no surface contact for traction
that moves freely over a variety of surfaces—particularly
water—while supported on a self-generated cushion of air.
The people behind the invention:
Christopher Sydney Cockerell (1910- ), a British engineer
who built the first hovercraft
Ronald A. Shaw (1910- ), an early pioneer in aerodynamics
who experimented with hovercraft
Sir John Isaac Thornycroft (1843-1928), a Royal Navy architect
who was the first to experiment with air-cushion theory
Air-Cushion Travel
The air-cushion vehicle was first conceived by Sir John Isaac
Thornycroft of Great Britain in the 1870’s. He theorized that if a
ship had a plenum chamber (a box open at the bottom) for a hull
and it were pumped full of air, the ship would rise out of the water
and move faster, because there would be less drag. The main problem
was keeping the air from escaping from under the craft.
In the early 1950’s, Christopher Sydney Cockerell was experimenting
with ways to reduce both the wave-making and frictional
resistance that craft had to water. In 1953, he constructed a punt
with a fan that supplied air to the bottom of the craft, which could
thus glide over the surface with very little friction. The air was contained
under the craft by specially constructed side walls. In 1955,
the first true “hovercraft,” as Cockerell called it, was constructed of
balsa wood. It weighed only 127 grams and traveled over water at a
speed of 13 kilometers per hour.
On November 16, 1956, Cockerell successfully demonstrated
his model hovercraft at the patent agent’s office in London. It was
immediately placed on the “secret” list, and Saunders-Roe Ltd.
was given the first contract to build hovercraft in 1957. The first experimental
piloted hovercraft, the SR.N1, which had a weight of
3,400 kilograms and could carry three people at the speed of 25 knots, was completed on May 28, 1959, and publicly demonstrated
on June 11, 1959.
Ground Effect Phenomenon
In a hovercraft, a jet airstream is directed downward through a
hole in a metal disk, which forces the disk to rise. The jet of air has a
reverse effect of its own that forces the disk away from the surface.
Some of the air hitting the ground bounces back against the disk to
add further lift. This is called the “ground effect.” The ground effect
is such that the greater the under-surface area of the hovercraft, the
greater the reverse thrust of the air that bounces back. This makes
the hovercraft a mechanically efficient machine because it provides
three functions.
First, the ground effect reduces friction between the craft and the
earth’s surface. Second, it acts as a spring suspension to reduce
some of the vertical acceleration effects that arise from travel over
an uneven surface. Third, it provides a safe and comfortable ride at
high speed, whatever the operating environment. The air cushion
can distribute the weight of the hovercraft over almost its entire area
so that the cushion pressure is low.
The basic elements of the air-cushion vehicle are a hull, a propulsion
system, and a lift system. The hull, which accommodates the
crew, passengers, and freight, contains both the propulsion and lift
systems. The propulsion and lift systems can be driven by the same
power plant or by separate power plants. Early designs used only
one unit, but this proved to be a problem when adequate power was
not achieved for movement and lift. Better results are achieved
when two units are used, since far more power is used to lift the vehicle
than to propel it.
For lift, high-speed centrifugal fans are used to drive the air
through jets that are located under the craft. A redesigned aircraft
propeller is used for propulsion. Rudderlike fins and an air fan that
can be swiveled to provide direction are placed at the rear of the
craft.
Several different air systems can be used, depending on whether
a skirt system is used in the lift process. The plenum chamber system,
the peripheral jet system, and several types of recirculating air systems have all been successfully tried without skirting. Avariety
of rigid and flexible skirts have also proved to be satisfactory, depending
on the use of the vehicle.
Skirts are used to hold the air for lift. Skirts were once hung like curtains around hovercraft. Instead of simple curtains to contain the air,
there are now complicated designs that contain the cushion, duct the
air, and even provide a secondary suspension. The materials used in
the skirting have also changed from a rubberized fabric to pure rubber
and nylon and, finally, to neoprene, a lamination of nylon and plastic.
The three basic types of hovercraft are the amphibious, nonamphibious,
and semiamphibious models. The amphibious type can
travel over water and land, whereas the nonamphibious type is restricted
to water travel. The semiamphibious model is also restricted
to water travel but may terminate travel by nosing up on a prepared
ramp or beach. All hovercraft contain built-in buoyancy tanks in the
side skirting as a safety measure in the event that a hovercraft must
settle on the water. Most hovercraft are equipped with gas turbines
and use either propellers or water-jet propulsion.
Impact
Hovercraft are used primarily for short passenger ferry services.
Great Britain was the only nation to produce a large number of hovercraft.
The British built larger and faster craft and pioneered their
successful use as ferries across the English Channel, where they
could reach speeds of 111 kilometers per hour (160 knots) and carry
more than four hundred passengers and almost one hundred vehicles.
France and the former Soviet Union have also effectively demonstrated
hovercraft river travel, and the Soviets have experimented
with military applications as well.
The military adaptations of hovercraft have been more diversified.
Beach landings have been performed effectively, and the United
States used hovercraft for river patrols during the Vietnam War.
Other uses also exist for hovercraft. They can be used as harbor pilot
vessels and for patrolling shores in a variety of police-and customs-
related duties. Hovercraft can also serve as flood-rescue craft
and fire-fighting vehicles. Even a hoverfreighter is being considered.
The air-cushion theory in transport systems is rapidly developing.
It has spread to trains and smaller people movers in many
countries. Their smooth, rapid, clean, and efficient operation makes
hovercraft attractive to transportation designers around the world.
16 July 2009
Holography
The invention: A lensless system of three-dimensional photography
that was one of the most important developments in twentieth
century optical science.
The people behind the invention:
Dennis Gabor (1900-1979), a Hungarian-born inventor and
physicist who was awarded the 1971 Nobel Prize in Physics
Emmett Leith (1927- ), a radar researcher who, with Juris
Upatnieks, produced the first laser holograms
Juris Upatnieks (1936- ), a radar researcher who, with
Emmett Leith, produced the first laser holograms
Easter Inspiration
The development of photography in the early 1900’s made possible
the recording of events and information in ways unknown before
the twentieth century: the photographing of star clusters, the
recording of the emission spectra of heated elements, the storing of
data in the form of small recorded images (for example, microfilm),
and the photographing of microscopic specimens, among other
things. Because of its vast importance to the scientist, the science of
photography has developed steadily.
An understanding of the photographic and holographic processes
requires some knowledge of the wave behavior of light. Light is an
electromagnetic wave that, like a water wave, has an amplitude and a
phase. The amplitude corresponds to the wave height, while the
phase indicates which part of the wave is passing a given point at a
given time. A cork floating in a pond bobs up and down as waves
pass under it. The position of the cork at any time depends on both
amplitude and phase: The phase determines on which part of the
wave the cork is floating at any given time, and the amplitude determines
how high or low the cork can be moved. Waves from more
than one source arriving at the cork combine in ways that depend on
their relative phases. If the waves meet in the same phase, they add
and produce a large amplitude; if they arrive out of phase, they subtract and produce a small amplitude. The total amplitude, or intensity,
depends on the phases of the combining waves.
Dennis Gabor, the inventor of holography, was intrigued by the
way in which the photographic image of an object was stored by a
photographic plate but was unable to devote any consistent research
effort to the question until the 1940’s. At that time, Gabor was involved
in the development of the electron microscope. On Easter
morning in 1947, as Gabor was pondering the problem of how to
improve the electron microscope, the solution came to him. He
would attempt to take a poor electron picture and then correct it optically.
The process would require coherent electron beams—that is,
electron waves with a definite phase.
This two-stage method was inspired by the work of Lawrence
Bragg. Bragg had formed the image of a crystal lattice by diffracting
the photographic X-ray diffraction pattern of the original lattice.
This double diffraction process is the basis of the holographic process.
Bragg’s method was limited because of his inability to record
the phase information of the X-ray photograph. Therefore, he could
study only those crystals for which the phase relationship of the reflected
waves could be predicted.
Waiting for the Laser
Gabor devised a way of capturing the phase information after he
realized that adding coherent background to the wave reflected from
an object would make it possible to produce an interference pattern
on the photographic plate. When the phases of the two waves are
identical, a maximum intensity will be recorded; when they are out of
phase, a minimum intensity is recorded. Therefore, what is recorded
in a hologram is not an image of the object but rather the interference
pattern of the two coherent waves. This pattern looks like a collection
of swirls and blank spots. The hologram (or photograph) is then illuminated
by the reference beam, and part of the transmitted light is a
replica of the original object wave. When viewing this object wave,
one sees an exact replica of the original object.
The major impediment at the time in making holograms using
any form of radiation was a lack of coherent sources. For example,
the coherence of the mercury lamp used by Gabor and his assistant IvorWilliams was so short that they were able to make holograms of
only about a centimeter in diameter. The early results were rather
poor in terms of image quality and also had a double image. For this
reason, there was little interest in holography, and the subject lay almost
untouched for more than ten years.
Interest in the field was rekindled after the laser (light amplification
by stimulated emission of radiation) was developed in 1962.
Emmett Leith and Juris Upatnieks, who were conducting radar research
at the University of Michigan, published the first laser holographs
in 1963. The laser was an intense light source with a very
long coherence length. Its monochromatic nature improved the resolution
of the images greatly. Also, there was no longer any restriction
on the size of the object to be photographed.
The availability of the laser allowed Leith and Upatnieks to propose
another improvement in holographic technique. Before 1964,
holograms were made of only thin transparent objects. A small region
of the hologram bore a one-to-one correspondence to a region
of the object. Only a small portion of the image could be viewed at
one time without the aid of additional optical components. Illuminating
the transparency diffusely allowed the whole image to be
seen at one time. This development also made it possible to record
holograms of diffusely reflected three-dimensional objects. Gabor
had seen from the beginning that this should make it possible to create
three-dimensional images.
After the early 1960’s, the field of holography developed very
quickly. Because holography is different from conventional photography,
the two techniques often complement each other. Gabor saw
his idea blossom into a very important technique in optical science.
Impact
The development of the laser and the publication of the first laser
holograms in 1963 caused a blossoming of the new technique in
many fields. Soon, techniques were developed that allowed holograms
to be viewed with white light. It also became possible for holograms
to reconstruct multicolored images. Holographic methods
have been used to map terrain with radar waves and to conduct surveillance
in the fields of forestry, agriculture, and meteorology.By the 1990’s, holography had become a multimillion-dollar industry,
finding applications in advertising, as an art form, and in security
devices on credit cards, as well as in scientific fields. An alternate
form of holography, also suggested by Gabor, uses sound
waves. Acoustical imaging is useful whenever the medium around
the object to be viewed is opaque to light rays—for example, in
medical diagnosis. Holography has affected many areas of science,
technology, and culture.
13 July 2009
Heat pump
The invention:
A device that warms and cools buildings efficiently
and cheaply by moving heat from one area to another.
The people behind the invention:
T. G. N. Haldane, a British engineer
Lord Kelvin (William Thomson, 1824-1907), a British
mathematician, scientist, and engineer
Sadi Carnot (1796-1832), a French physicist and
thermodynamicist
Heart-lung machine
The invention: The first artificial device to oxygenate and circulate
blood during surgery, the heart-lung machine began the era of
open-heart surgery.
The people behind the invention:
John H. Gibbon, Jr. (1903-1974), a cardiovascular surgeon
Mary Hopkinson Gibbon (1905- ), a research technician
Thomas J. Watson (1874-1956), chairman of the board of IBM
T. L. Stokes and J. B. Flick, researchers in Gibbon’s laboratory
Bernard J. Miller (1918- ), a cardiovascular surgeon and
research associate
Cecelia Bavolek, the first human to undergo open-heart surgery
successfully using the heart-lung machine
A Young Woman’s Death
In the first half of the twentieth century, cardiovascular medicine
had many triumphs. Effective anesthesia, antiseptic conditions, and
antibiotics made surgery safer. Blood-typing, anti-clotting agents,
and blood preservatives made blood transfusion practical. Cardiac
catheterization (feeding a tube into the heart), electrocardiography,
and fluoroscopy (visualizing living tissues with an X-ray machine)
made the nonsurgical diagnosis of cardiovascular problems possible.
As of 1950, however, there was no safe way to treat damage or defects
within the heart. To make such a correction, this vital organ’s
function had to be interrupted. The problem was to keep the body’s
tissues alive while working on the heart. While some surgeons practiced
so-called blind surgery, in which they inserted a finger into the
heart through a small incision without observing what they were attempting
to correct, others tried to reduce the body’s need for circulation
by slowly chilling the patient until the heart stopped. Still other
surgeons used “cross-circulation,” in which the patient’s circulation
was connected to a donor’s circulation. All these approaches carried
profound risks of hemorrhage, tissue damage, and death.
In February of 1931, Gibbon witnessed the death of a young woman whose lung circulation was blocked by a blood clot. Because
her blood could not pass through her lungs, she slowly lost
consciousness from lack of oxygen. As he monitored her pulse and
breathing, Gibbon thought about ways to circumvent the obstructed
lungs and straining heart and provide the oxygen required. Because
surgery to remove such a blood clot was often fatal, the woman’s
surgeons operated only as a last resort. Though the surgery took
only six and one-half minutes, she never regained consciousness.
This experience prompted Gibbon to pursue what few people then
considered a practical line of research: a way to circulate and oxygenate
blood outside the body.
A Woman’s Life Restored
Gibbon began the project in earnest in 1934, when he returned to
the laboratory of Edward D. Churchill at Massachusetts General
Hospital for his second surgical research fellowship. He was assisted
by Mary Hopkinson Gibbon. Together, they developed, using
cats, a surgical technique for removing blood froma vein, supplying
the blood with oxygen, and returning it to an artery using tubes inserted
into the blood vessels. Their objective was to create a device
that would keep the blood moving, spread it over a very thin layer
to pick up oxygen efficiently and remove carbon dioxide, and avoid
both clotting and damaging blood cells. In 1939, they reported that
prolonged survival after heart-lung bypass was possible in experimental
animals.
WorldWar II (1939-1945) interrupted the progress of this work; it
was resumed by Gibbon at Jefferson Medical College in 1944. Shortly
thereafter, he attracted the interest of Thomas J.Watson, chairman of
the board of the International Business Machines (IBM) Corporation,
who provided the services of IBM’s experimental physics laboratory
and model machine shop as well as the assistance of staff engineers.
IBM constructed and modified two experimental machines
over the next seven years, and IBM engineers contributed significantly
to the evolution of a machine that would be practical in humans.
Gibbon’s first attempt to use the pump-oxygenator in a human
being was in a fifteen-month-old baby. This attempt failed, not because of a malfunction or a surgical mistake but because of a misdiagnosis.
The child died following surgery because the real problem
had not been corrected by the surgery.
On May 6, 1953, the heart-lung machine was first used successfully
on Cecelia Bavolek. In the six months before surgery, Bavolek
had been hospitalized three times for symptoms of heart failure
when she tried to engage in normal activity. While her circulation
was connected to the heart-lung machine for forty-five minutes, the
surgical team headed by Gibbon was able to close an opening between
her atria and establish normal heart function. Two months
later, an examination of the defect revealed that it was fully closed;
Bavolek resumed a normal life. The age of open-heart surgery had
begun.
Consequences
The heart-lung bypass technique alone could not make openheart
surgery truly practical. When it was possible to keep tissues
alive by diverting blood around the heart and oxygenating it, other
questions already under investigation became even more critical:
how to prolong the survival of bloodless organs, how to measure
oxygen and carbon dioxide levels in the blood, and how to prolong
anesthesia during complicated surgery. Thus, following the first
successful use of the heart-lung machine, surgeons continued to refine
the methods of open-heart surgery.
The heart-lung apparatus set the stage for the advent of “replacement
parts” for many types of cardiovascular problems. Cardiac
valve replacement was first successfully accomplished in 1960 by
placing an artificial ball valve between the left atrium and ventricle.
In 1957, doctors performed the first coronary bypass surgery, grafting
sections of a leg vein into the heart’s circulation system to divert
blood around clogged coronary arteries. Likewise, the first successful
heart transplant (1967) and the controversial Jarvik-7 artificial
heart implantation (1982) required the ability to stop the heart and
keep the body’s tissues alive during time-consuming and delicate
surgical procedures. Gibbon’s heart-lung machine paved the way
for all these developments.
09 July 2009
Hearing aid
The invention: Miniaturized electronic amplifier worn inside the
ears of hearing-impaired persons.
The organization behind the invention:
Bell Labs, the research and development arm of the American
Telephone and Telegraph Company
Trapped in Silence
Until the middle of the twentieth century, people who experienced
hearing loss had little hope of being able to hear sounds without the
use of large, awkward, heavy appliances. For many years, the only
hearing aids available were devices known as ear trumpets. The ear
trumpet tried to compensate for hearing loss by increasing the number
of sound waves funneled into the ear canal. A wide, bell-like
mouth similar to the bell of a musical trumpet narrowed to a tube that
the user placed in his or her ear. Ear trumpets helped a little, but they
could not truly increase the volume of the sounds heard.
Beginning in the nineteenth century, inventors tried to develop
electrical devices that would serve as hearing aids. The telephone
was actually a by-product of Alexander Graham Bell’s efforts to
make a hearing aid. Following the invention of the telephone, electrical
engineers designed hearing aids that employed telephone
technology, but those hearing aids were only a slight improvement
over the old ear trumpets. They required large, heavy battery packs
and used a carbon microphone similar to the receiver in a telephone.
More sensitive than purely physical devices such as the ear trumpet,
they could transmit a wider range of sounds but could not amplify
them as effectively as electronic hearing aids now do.
Transistors Make Miniaturization Possible
Two types of hearing aids exist: body-worn and head-worn.
Body-worn hearing aids permit the widest range of sounds to be
heard, but because of the devices’ larger size, many hearing impaired persons do not like to wear them. Head-worn hearing
aids, especially those worn completely in the ear, are much less conspicuous.
In addition to in-ear aids, the category of head-worn hearing
aids includes both hearing aids mounted in eyeglass frames and
those worn behind the ear.
All hearing aids, whether head-worn or body-worn, consist of
four parts: a microphone to pick up sounds, an amplifier, a receiver,
and a power source. The microphone gathers sound waves and converts
them to electrical signals; the amplifier boosts, or increases,
those signals; and the receiver then converts the signals back into
sound waves. In effect, the hearing aid is a miniature radio. After
the receiver converts the signals back to sound waves, those waves
are directed into the ear canal through an earpiece or ear mold. The
ear mold generally is made of plastic and is custom fitted from an
impression taken from the prospective user’s ear.
Effective head-worn hearing aids could not be built until the
electronic circuit was developed in the early 1950’s. The same invention—
the transistor—that led to small portable radios and tape
players allowed engineers to create miniaturized, inconspicuous
hearing aids. Depending on the degree of amplification required,
the amplifier in a hearing aid contains three or more transistors.
Transistors first replaced vacuum tubes in devices such as radios
and phonographs, and then engineers realized that they could be
used in devices for the hearing-impaired.
The research at Bell Labs that led to the invention of the transistor
rose out of military research duringWorldWar II. The vacuum tubes
used in, for example, radar installations to amplify the strength of electronic
signals were big, were fragile because they were made of
blown glass, and gave off high levels of heat when they were used.
Transistors, however, made it possible to build solid-state, integrated
circuits. These are made from crystals of metals such as germanium
or arsenic alloys and therefore are much less fragile than glass. They
are also extremely small (in fact, some integrated circuits are barely
visible to the naked eye) and give off no heat during use.
The number of transistors in a hearing aid varies depending upon
the amount of amplification required. The first transistor is the most
important for the listener in terms of the quality of sound heard. If the
frequency response is set too high—that is, if the device is too sensitive—the listener will be bothered by distracting background noise.
Theoretically, there is no limit on the amount of amplification that a
hearing aid can be designed to provide, but there are practical limits.
The higher the amplification, the more power is required to operate
the hearing aid. This is why body-worn hearing aids can convey a
wider range of sounds than head-worn devices can. It is the power
source—not the electronic components—that is the limiting factor. A
body-worn hearing aid includes a larger battery pack than can be
used with a head-worn device. Indeed, despite advances in battery
technology, the power requirements of a head-worn hearing aid are
such that a 1.4-volt battery that could power a wristwatch for several
years will last only a few days in a hearing aid.
Consequences
The invention of the electronic hearing aid made it possible for
many hearing-impaired persons to participate in a hearing world.
Prior to the invention of the hearing aid, hearing-impaired children
often were unable to participate in routine school activities or function
effectively in mainstream society. Instead of being able to live at
home with their families and enjoy the same experiences that were
available to other children their age, often they were forced to attend
special schools operated by the state or by charities.
Hearing-impaired people were singled out as being different and
were limited in their choice of occupations. Although not every
hearing-impaired person can be helped to hear with a hearing aid—
particularly in cases of total hearing loss—the electronic hearing aid
has ended restrictions for many hearing-impaired people. Hearingimpaired
children are now included in public school classes, and
hearing-impaired adults can now pursue occupations from which
they were once excluded.
Today, many deaf and hearing-impaired persons have chosen to
live without the help of a hearing aid. They believe that they are not
disabled but simply different, and they point out that their “disability”
often allows them to appreciate and participate in life in unique
and positive ways. For them, the use of hearing aids is a choice, not a
necessity. For those who choose, hearing aids make it possible to
participate in the hearing world.
Hard disk
The invention: A large-capacity, permanent magnetic storage device
built into most personal computers.
The people behind the invention:
Alan Shugart (1930- ), an engineer who first developed the
floppy disk
Philip D. Estridge (1938?-1985), the director of IBM’s product
development facility
Thomas J. Watson, Jr. (1914-1993), the chief executive officer of
IBM
The Personal Oddity
When the International Business Machines (IBM) Corporation
introduced its first microcomputer, called simply the IBM PC (for
“personal computer”), the occasion was less a dramatic invention
than the confirmation of a trend begun some years before. A number
of companies had introduced microcomputers before IBM; one
of the best known at that time was Apple Corporation’s Apple II, for
which software for business and scientific use was quickly developed.
Nevertheless, the microcomputer was quite expensive and
was often looked upon as an oddity, not as a useful tool.
Under the leadership of Thomas J. Watson, Jr., IBM, which had
previously focused on giant mainframe computers, decided to develop
the PC. A design team headed by Philip D. Estridge was assembled
in Boca Raton, Florida, and it quickly developed its first,
pacesetting product. It is an irony of history that IBM anticipated
selling only one hundred thousand or so of these machines, mostly
to scientists and technically inclined hobbyists. Instead, IBM’s product
sold exceedingly well, and its design parameters, as well as its
operating system, became standards.
The earliest microcomputers used a cassette recorder as a means
of mass storage; a floppy disk drive capable of storing approximately
160 kilobytes of data was initially offered only as an option.
While home hobbyists were accustomed to using a cassette recorder for storage purposes, such a system was far too slow and awkward
for use in business and science. As a result, virtually every IBM PC
sold was equipped with at least one 5.25-inch floppy disk drive.
Memory Requirements
All computers require memory of two sorts in order to carry out
their tasks. One type of memory is main memory, or random access
memory (RAM), which is used by the computer’s central processor
to store data it is using while operating. The type of memory used
for this function is built typically of silicon-based integrated circuits
that have the advantage of speed (to allow the processor to fetch or
store the data quickly), but the disadvantage of possibly losing or
“forgetting” data when the electric current is turned off. Further,
such memory generally is relatively expensive.
To reduce costs, another type of memory—long-term storage
memory, known also as “mass storage”—was developed. Mass
storage devices include magnetic media (tape or disk drives) and
optical media (such as the compact disc, read-only memory, or CDROM).
While the speed with which data may be retrieved from or
stored in such devices is rather slow compared to the central processor’s
speed, a disk drive—the most common form of mass storage
used in PCs—can store relatively large amounts of data quite inexpensively.
Early floppy disk drives (so called because the magnetically
treated material on which data are recorded is made of a very flexible
plastic) held 160 kilobytes of data using only one side of the
magnetically coated disk (about eighty pages of normal, doublespaced,
typewritten information). Later developments increased
storage capacities to 360 kilobytes by using both sides of the disk
and later, with increasing technological ability, 1.44 megabytes (millions
of bytes). In contrast, mainframe computers, which are typically
connected to large and expensive tape drive storage systems,
could store gigabytes (millions of megabytes) of information.
While such capacities seem large, the needs of business and scientific
users soon outstripped available space. Since even the mailing
list of a small business or a scientist’s mathematical model of a
chemical reaction easily could require greater storage potential than early PCs allowed, the need arose for a mass storage device that
could accommodate very large files of data.
The answer was the hard disk drive, also known as a “fixed disk
drive,” reflecting the fact that the disk itself is not only rigid but also
permanently installed inside the machine. In 1955, IBM had envisioned
the notion of a fixed, hard magnetic disk as a means of storing
computer data, and, under the direction of Alan Shugart in the
1960’s, the floppy disk was developed as well.
As the engineers of IBM’s facility in Boca Raton refined the idea
of the original PC to design the new IBM PC XT, it became clear that
chief among the needs of users was the availability of large-capability
storage devices. The decision was made to add a 10-megabyte
hard disk drive to the PC. On March 8, 1983, less than two years after
the introduction of its first PC, IBM introduced the PC XT. Like
the original, it was an evolutionary design, not a revolutionary one.
The inclusion of a hard disk drive, however, signaled that mass storage
devices in personal computers had arrived.
Consequences
Above all else, any computer provides a means for storing, ordering,
analyzing, and presenting information. If the personal computer
is to become the information appliance some have suggested
it will be, the ability to manipulate very large amounts of data will
be of paramount concern. Hard disk technology was greeted enthusiastically
in the marketplace, and the demand for hard drives has
seen their numbers increase as their quality increases and their
prices drop.
It is easy to understand one reason for such eager acceptance:
convenience. Floppy-bound computer users find themselves frequently
changing (or “swapping”) their disks in order to allow programs
to find the data they need. Moreover, there is a limit to how
much data a single floppy disk can hold. The advantage of a hard
drive is that it allows users to keep seemingly unlimited amounts of
data and programs stored in their machines and readily available.
Also, hard disk drives are capable of finding files and transferring
their contents to the processor much more quickly than a
floppy drive. A user may thus create exceedingly large files, keep them on hand at all times, and manipulate data more quickly than
with a floppy. Finally, while a hard drive is a slow substitute for
main memory, it allows users to enjoy the benefits of larger memories
at significantly lower cost.
The introduction of the PC XT with its 10-megabyte hard drive
was a milestone in the development of the PC. Over the next two decades,
the size of computer hard drives increased dramatically. By
2001, few personal computers were sold with hard drives with less
than three gigabytes of storage capacity, and hard drives with more
than thirty gigabytes were becoming the standard. Indeed, for less
money than a PC XT cost in the mid-1980’s, one could buy a fully
equipped computer with a hard drive holding sixty gigabytes—a
storage capacity equivalent to six thousand 10-megabyte hard drives.
Gyrocompass
The invention: The first practical navigational device that enabled
ships and submarines to stay on course without relying on the
earth’s unreliable magnetic poles.
The people behind the invention:
Hermann Anschütz-Kaempfe (1872-1931), a German inventor
and manufacturer
Jean-Bernard-Léon Foucault (1819-1868), a French experimental
physicist and inventor
Elmer Ambrose Sperry (1860-1930), an American engineer and
inventor
From Toys to Tools
A gyroscope consists of a rapidly spinning wheel mounted in a
frame that enables the wheel to tilt freely in any direction. The
amount of momentum allows the wheel to maintain its “attitude”
even when the whole device is turned or rotated.
These devices have been used to solve problems arising in such
areas as sailing and navigation. For example, a gyroscope aboard a
ship maintains its orientation even while the ship is rolling. Among
other things, this allows the extent of the roll to be measured accurately.
Moreover, the spin axis of a free gyroscope can be adjusted to
point toward true north. It will (with some exceptions) stay that
way despite changes in the direction of a vehicle in which it is
mounted. Gyroscopic effects were employed in the design of various
objects long before the theory behind them was formally
known. A classic example is a child’s top, which balances, seemingly
in defiance of gravity, as long as it continues to spin. Boomerangs
and flying disks derive stability and accuracy from the spin
imparted by the thrower. Likewise, the accuracy of rifles improved
when barrels were manufactured with internal spiral grooves that
caused the emerging bullet to spin.
In 1852, the French inventor Jean-Bernard-Léon Foucault built
the first gyroscope, a measuring device consisting of a rapidly spinning
wheel mounted within concentric rings that allowed the wheel to move freely about two axes. This device, like the Foucault pendulum,
was used to demonstrate the rotation of the earth around its
axis, since the spinning wheel, which is not fixed, retains its orientation
in space while the earth turns under it. The gyroscope had a related
interesting property: As it continued to spin, the force of the
earth’s rotation caused its axis to rotate gradually until it was oriented
parallel to the earth’s axis, that is, in a north-south direction. It
is this property that enables the gyroscope to be used as a compass.
When Magnets Fail
In 1904, Hermann Anschütz-Kaempfe, a German manufacturer
working in the Kiel shipyards, became interested in the navigation
problems of submarines used in exploration under the polar ice cap.
By 1905, efficient working submarines were a reality, and it was evident
to all major naval powers that submarines would play an increasingly
important role in naval strategy.
Submarine navigation posed problems, however, that could not
be solved by instruments designed for surface vessels. Asubmarine
needs to orient itself under water in three dimensions; it has no automatic
horizon with respect to which it can level itself. Navigation
by means of stars or landmarks is impossible when the submarine is
submerged. Furthermore, in an enclosed metal hull containing machinery
run by electricity, a magnetic compass is worthless. To a
lesser extent, increasing use of metal, massive moving parts, and
electrical equipment had also rendered the magnetic compass unreliable
in conventional surface battleships.
It made sense for Anschütz-Kaempfe to use the gyroscopic effect
to design an instrument that would enable a ship to maintain its
course while under water. Yet producing such a device would not be
easy. First, it needed to be suspended in such a way that it was free to
turn in any direction with as little mechanical resistance as possible.
At the same time, it had to be able to resist the inevitable pitching and
rolling of a vessel at sea. Finally, a continuous power supply was required
to keep the gyroscopic wheels spinning at high speed.
The original Anschütz-Kaempfe gyrocompass consisted of a pair
of spinning wheels driven by an electric motor. The device was connected
to a compass card visible to the ship’s navigator. Motor, gyroscope, and suspension system were mounted in a frame that allowed
the apparatus to remain stable despite the pitch and roll of the ship.
In 1906, the German navy installed a prototype of the Anschütz-
Kaempfe gyrocompass on the battleship Undine and subjected it to
exhaustive tests under simulated battle conditions, sailing the ship
under forced draft and suddenly reversing the engines, changing the
position of heavy turrets and other mechanisms, and firing heavy
guns. In conditions under which a magnetic compass would have
been worthless, the gyrocompass proved a satisfactory navigational
tool, and the results were impressive enough to convince the German
navy to undertake installation of gyrocompasses in submarines and
heavy battleships, including the battleship Deutschland.
Elmer Ambrose Sperry, a New York inventor intimately associated
with pioneer electrical development, was independently working on a design for a gyroscopic compass at about the same time.
In 1907, he patented a gyrocompass consisting of a single rotor
mounted within two concentric shells, suspended by fine piano
wire from a frame mounted on gimbals. The rotor of the Sperry
compass operated in a vacuum, which enabled it to rotate more
rapidly. The Sperry gyrocompass was in use on larger American
battleships and submarines on the eve ofWorldWar I (1914-1918).
Impact
The ability to navigate submerged submarines was of critical
strategic importance in World War I. Initially, the German navy
had an advantage both in the number of submarines at its disposal
and in their design and maneuverability. The German U-boat fleet
declared all-out war on Allied shipping, and, although their efforts
to blockade England and France were ultimately unsuccessful, the
tremendous toll they inflicted helped maintain the German position
and prolong the war. To a submarine fleet operating throughout
the Atlantic and in the Caribbean, as well as in near-shore European
waters, effective long-distance navigation was critical.
Gyrocompasses were standard equipment on submarines and
battleships and, increasingly, on larger commercial vessels during
World War I, World War II (1939-1945), and the period between the
wars. The devices also found their way into aircraft, rockets, and
guided missiles. Although the compasses were made more accurate
and easier to use, the fundamental design differed little from that invented
by Anschütz-Kaempfe.
05 July 2009
Geothermal power
The invention: Energy generated from the earth’s natural hot
springs.
The people behind the invention:
Prince Piero Ginori Conti (1865-1939), an Italian nobleman and
industrialist
Sir Charles Parsons (1854-1931), an English engineer
B. C. McCabe, an American businessman
Developing a Practical System
The first successful use of geothermal energy was at Larderello in
northern Italy. The Larderello geothermal field, located near the city
of Pisa about 240 kilometers northwest of Rome, contains many hot
springs and fumaroles (steam vents). In 1777, these springs were
found to be rich in boron, and in 1818, Francesco de Larderel began
extracting the useful mineral borax from them. Shortly after 1900,
Prince Piero Ginori Conti, director of the Larderello borax works,
conceived the idea of using the steam for power production. An experimental
electrical power plant was constructed at Larderello in
1904 to provide electric power to the borax plant. After this initial
experiment proved successful, a 250-kilowatt generating station
was installed in 1913 and commercial power production began.
As the Larderello field grew, additional geothermal sites throughout
the region were prospected and tapped for power. Power production
grew steadily until the 1940’s, when production reached
130 megawatts; however, the Larderello power plants were destroyed
late inWorldWar II (1939-1945). After the war, the generating
plants were rebuilt, and they were producing more than 400
megawatts by 1980.
The Larderello power plants encountered many of the technical
problems that were later to concern other geothermal facilities. For
example, hydrogen sulfide in the steam was highly corrosive to copper,
so the Larderello power plant used aluminum for electrical connections
much more than did conventional power plants of the time. Also, the low pressure of the steam in early wells at Larderello
presented problems. The first generators simply used steam to drive
a generator and vented the spent steam into the atmosphere. Asystem
of this sort, called a “noncondensing system,” is useful for small
generators but not efficient to produce large amounts of power.
Most steam engines derive power not only from the pressure of
the steam but also from the vacuum created when the steam is condensed
back to water. Geothermal systems that generate power
from condensation, as well as direct steam pressure, are called “condensing
systems.” Most large geothermal generators are of this
type. Condensation of geothermal steam presents special problems
not present in ordinary steam engines: There are other gases present
that do not condense. Instead of a vacuum, condensation of steam
contaminated with other gases would result in only a limited drop
in pressure and, consequently, very low efficiency.
Initially, the operators of Larderello tried to use the steam to heat
boilers that would, in turn, generate pure steam. Eventually, a device
was developed that removed most of the contaminating gases from
the steam. Although later wells at Larderello and other geothermal
fields produced steam at greater pressure, these engineering innovations
improved the efficiency of any geothermal power plant.
Expanding the Idea
In 1913, the English engineer Sir Charles Parsons proposed drilling
an extremely deep (12-kilometer) hole to tap the earth’s deep
heat. Power from such a deep hole would not come from natural
steam as at Larderello but would be generated by pumping fluid
into the hole and generating steam (as hot as 500 degrees Celsius) at
the bottom. In modern terms, Parsons proposed tapping “hot dryrock”
geothermal energy. (No such plant has been commercially operated
yet, but research is being actively pursued in several countries.)
The first use of geothermal energy in the United States was for direct
heating. In 1890, the municipal water company of Boise, Idaho,
began supplying hot water from a geothermal well. Water was
piped from the well to homes and businesses along appropriately
namedWarm Springs Avenue. At its peak, the system served more than four hundred customers, but as cheap natural gas became
available, the number declined.
Although Larderello was the first successful geothermal electric
power plant, the modern era of geothermal electric power began
with the opening of the Geysers Geothermal Field in California.
Early attempts began in the 1920’s, but it was not until 1955 that B.
C. McCabe, a Los Angeles businessman, leased 14.6 square kilometers
in the Geysers area and founded the Magma Power Company.
The first 12.5-megawatt generator was installed at the Geysers in
1960, and production increased steadily from then on. The Geysers
surpassed Larderello as the largest producing geothermal field in
the 1970’s, and more than 1,000 megawatts were being generated by
1980. By the end of 1980, geothermal plants had been installed in
thirteen countries, with a total capacity of almost 2,600 megawatts,
and projects with a total capacity of more than 15,000 megawatts
were being planned in more than twenty countries.
Impact
Geothermal power has many attractive features. Because the
steam is naturally heated and under pressure, generating equipment
can be simple, inexpensive, and quickly installed. Equipment
and installation costs are offset by savings in fuel. It is economically
practical to install small generators, a fact that makes geothermal
plants attractive in remote or underdeveloped areas. Most important
to a world faced with a variety of technical and environmental
problems connected with fossil fuels, geothermal power does not
deplete fossil fuel reserves, produces little pollution, and contributes
little to the greenhouse effect.
Despite its attractive features, geothermal power has some limitations.
Geologic settings suitable for easy geothermal power production
are rare; there must be a hot rock or magma body close to
the surface. Although it is technically possible to pump water from
an external source into a geothermal well to generate steam, most
geothermal sites require a plentiful supply of natural underground
water that can be tapped as a source of steam. In contrast, fossil-fuel
generating plants can be at any convenient location.
Genetically engineered insulin
The invention: Artificially manufactured human insulin (Humulin)
as a medication for people suffering from diabetes.
The people behind the invention:
Irving S. Johnson (1925- ), an American zoologist who was
vice president of research at Eli Lilly Research Laboratories
Ronald E. Chance (1934- ), an American biochemist at Eli
Lilly Research Laboratories
What Is Diabetes?
Carbohydrates (sugars and related chemicals) are the main food
and energy source for humans. In wealthy countries such as the
United States, more than 50 percent of the food people eat is made
up of carbohydrates, while in poorer countries the carbohydrate
content of diets is higher, from 70 to 90 percent.
Normally, most carbohydrates that a person eats are used (or metabolized)
quickly to produce energy. Carbohydrates not needed for
energy are either converted to fat or stored as a glucose polymer
called “glycogen.” Most adult humans carry about a pound of body
glycogen; this substance is broken down to produce energy when it
is needed.
Certain diseases prevent the proper metabolism and storage of
carbohydrates. The most common of these diseases is diabetes mellitus,
usually called simply “diabetes.” It is found in more than seventy
million people worldwide. Diabetic people cannot produce or
use enough insulin, a hormone secreted by the pancreas. When their
condition is not treated, the eyes may deteriorate to the point of
blindness. The kidneys may stop working properly, blood vessels
may be damaged, and the person may fall into a coma and die. In
fact, diabetes is the third most common killer in the United States.
Most of the problems surrounding diabetes are caused by high levels
of glucose in the blood. Cataracts often form in diabetics, as excess
glucose is deposited in the lens of the eye.
Important symptoms of diabetes include constant thirst, excessive urination, and large amounts of sugar in the blood and in the
urine. The glucose tolerance test (GTT) is the best way to find out
whether a person is suffering from diabetes. People given a GTT are
first told to fast overnight. In the morning their blood glucose level
is measured; then they are asked to drink about a fourth of a pound
of glucose dissolved in water. During the next four to six hours, the
blood glucose level is measured repeatedly. In nondiabetics, glucose
levels do not rise above a certain amount during a GTT, and the
level drops quickly as the glucose is assimilated by the body. In diabetics,
the blood glucose levels rise much higher and do not drop as
quickly. The extra glucose then shows up in the urine.
Treating Diabetes
Until the 1920’s, diabetes could be controlled only through a diet
very low in carbohydrates, and this treatment was not always successful.
Then Sir Frederick G. Banting and Charles H. Best found a
way to prepare purified insulin from animal pancreases and gave it
to patients. This gave diabetics their first chance to live a fairly normal
life. Banting and his coworkers won the 1923 Nobel Prize in
Physiology or Medicine for their work.
The usual treatment for diabetics became regular shots of insulin.
Drug companies took the insulin from the pancreases of cattle and
pigs slaughtered by the meat-packing industry. Unfortunately, animal
insulin has two disadvantages. First, about 5 percent of diabetics
are allergic to it and can have severe reactions. Second, the world
supply of animal pancreases goes up and down depending on how
much meat is being bought. Between 1970 and 1975, the supply of
insulin fell sharply as people began to eat less red meat, yet the
numbers of diabetics continued to increase. So researchers began to
look for a better way to supply insulin.
Studying pancreases of people who had donated their bodies to
science, researchers found that human insulin did not cause allergic
reactions. Scientists realized that it would be best to find a chemical
or biological way to prepare human insulin, and pharmaceutical
companies worked hard toward this goal. Eli Lilly and Company
was the first to succeed, and on May 14, 1982, it filed a new drug application
with the Food and Drug Administration (FDA) for the human insulin preparation it named “Humulin.”
Humulin is made by genetic engineering. Irving S. Johnson, who
worked on the development of Humulin, described Eli Lilly’s method
for producing Humulin. The common bacterium Escherichia coli
is used. Two strains of the bacterium are produced by genetic engineering:
The first strain is used to make a protein called an “A
chain,” and the second strain is used to make a “B chain.” After the
bacteria are harvested, the Aand B chains are removed and purified
separately. Then the two chains are combined chemically. When
they are purified once more, the result is Humulin, which has been
proved by Ronald E. Chance and his Eli Lilly coworkers to be chemically,
biologically, and physically identical to human insulin.
Consequences
The FDA and other regulatory agencies around the world approved
genetically engineered human insulin in 1982. Humulin
does not trigger allergic reactions, and its supply does not fluctuate.
It has brought an end to the fear that there would be a worldwide
shortage of insulin.
Humulin is important as well in being the first genetically engineered
industrial chemical. It began an era in which such advanced
technology could be a source for medical drugs, chemicals used in
farming, and other important industrial products. Researchers hope
that genetic engineering will help in the understanding of cancer
and other diseases, and that it will lead to ways to grow enough
food for a world whose population continues to rise.
29 June 2009
Genetic “fingerprinting”
The invention: Atechnique for using the unique characteristics of
each human being’s DNA to identify individuals, establish connections
among relatives, and identify criminals.
The people behind the invention:
Alec Jeffreys (1950- ), an English geneticist
Victoria Wilson (1950- ), an English geneticist
Swee Lay Thein (1951- ), a biochemical geneticist
Microscopic Fingerprints
In 1985, Alec Jeffreys, a geneticist at the University of Leicester in
England, developed a method of deoxyribonucleic acid (DNA)
analysis that provides a visual representation of the human genetic
structure. Jeffreys’s discovery had an immediate, revolutionary impact
on problems of human identification, especially the identification
of criminals. Whereas earlier techniques, such as conventional
blood typing, provide evidence that is merely exclusionary (indicating
only whether a suspect could or could not be the perpetrator of a
crime), DNA fingerprinting provides positive identification.
For example, under favorable conditions, the technique can establish
with virtual certainty whether a given individual is a murderer
or rapist. The applications are not limited to forensic science;
DNA fingerprinting can also establish definitive proof of parenthood
(paternity or maternity), and it is invaluable in providing
markers for mapping disease-causing genes on chromosomes. In
addition, the technique is utilized by animal geneticists to establish
paternity and to detect genetic relatedness between social groups.
DNAfingerprinting (also referred to as “genetic fingerprinting”)
is a sophisticated technique that must be executed carefully to produce
valid results. The technical difficulties arise partly from the
complex nature of DNA. DNA, the genetic material responsible for
heredity in all higher forms of life, is an enormously long, doublestranded
molecule composed of four different units called “bases.”
The bases on one strand of DNApair with complementary bases on the other strand. A human being contains twenty-three pairs of
chromosomes; one member of each chromosome pair is inherited
fromthe mother, the other fromthe father. The order, or sequence, of
bases forms the genetic message, which is called the “genome.” Scientists
did not know the sequence of bases in any sizable stretch of
DNA prior to the 1970’s because they lacked the molecular tools to
split DNA into fragments that could be analyzed. This situation
changed with the advent of biotechnology in the mid-1970’s.
The door toDNAanalysis was opened with the discovery of bacterial
enzymes called “DNA restriction enzymes.” A restriction enzyme
binds to DNA whenever it finds a specific short sequence of
base pairs (analogous to a code word), and it splits the DNAat a defined
site within that sequence. A single enzyme finds millions of
cutting sites in human DNA, and the resulting fragments range in
size from tens of base pairs to hundreds or thousands. The fragments
are exposed to a radioactive DNA probe, which can bind to
specific complementary DNA sequences in the fragments. X-ray
film detects the radioactive pattern. The developed film, called an
“autoradiograph,” shows a pattern of DNA fragments, which is
similar to a bar code and can be compared with patterns from
known subjects.
The Presence of Minisatellites
The uniqueness of a DNA fingerprint depends on the fact that,
with the exception of identical twins, no two human beings have
identical DNA sequences. Of the three billion base pairs in human
DNA, many will differ from one person to another.
In 1985, Jeffreys and his coworkers, Victoria Wilson at the University
of Leicester and Swee Lay Thein at the John Radcliffe Hospital
in Oxford, discovered a way to produce a DNA fingerprint.
Jeffreys had found previously that human DNA contains many repeated
minisequences called “minisatellites.” Minisatellites consist
of sequences of base pairs repeated in tandem, and the number of
repeated units varies widely from one individual to another. Every
person, with the exception of identical twins, has a different number
of tandem repeats and, hence, different lengths of minisatellite
DNA. By using two labeled DNA probes to detect two different minisatellite sequences, Jeffreys obtained a unique fragment band
pattern that was completely specific for an individual.
The power of the technique derives from the law of chance,
which indicates that the probability (chance) that two or more unrelated
events will occur simultaneously is calculated as the multiplication
product of the two separate probabilities. As Jeffreys discovered,
the likelihood of two unrelated people having completely
identical DNAfingerprints is extremely small—less than one in ten
trillion. Given the population of the world, it is clear that the technique
can distinguish any one person from everyone else. Jeffreys
called his band patterns “DNAfingerprints” because of their ability
to individualize. As he stated in his landmark research paper, published
in the English scientific journal Nature in 1985, probes to
minisatellite regions of human DNA produce “DNA ‘fingerprints’
which are completely specific to an individual (or to his or her identical
twin) and can be applied directly to problems of human identification,
including parenthood testing.”
Consequences
In addition to being used in human identification, DNA fingerprinting
has found applications in medical genetics. In the search
for a cause, a diagnostic test for, and ultimately the treatment of an
inherited disease, it is necessary to locate the defective gene on a human
chromosome. Gene location is accomplished by a technique
called “linkage analysis,” in which geneticists use marker sections
of DNA as reference points to pinpoint the position of a defective
gene on a chromosome. The minisatellite DNA probes developed
by Jeffreys provide a potent and valuable set of markers that are of
great value in locating disease-causing genes. Soon after its discovery,
DNA fingerprinting was used to locate the defective genes responsible
for several diseases, including fetal hemoglobin abnormality
and Huntington’s disease.
Genetic fingerprinting also has had a major impact on genetic
studies of higher animals. BecauseDNAsequences are conserved in
evolution, humans and other vertebrates have many sequences in
common. This commonality enabled Jeffreys to use his probes to
human minisatellites to bind to the DNA of many different vertebrates, ranging from mammals to birds, reptiles, amphibians, and
fish; this made it possible for him to produce DNA fingerprints of
these vertebrates. In addition, the technique has been used to discern
the mating behavior of birds, to determine paternity in zoo primates,
and to detect inbreeding in imperiled wildlife. DNA fingerprinting
can also be applied to animal breeding problems, such as
the identification of stolen animals, the verification of semen samples
for artificial insemination, and the determination of pedigree.
The technique is not foolproof, however, and results may be far
from ideal. Especially in the area of forensic science, there was a
rush to use the tremendous power of DNA fingerprinting to identify
a purported murderer or rapist, and the need for scientific standards
was often neglected. Some problems arose because forensic
DNA fingerprinting in the United States is generally conducted in
private, unregulated laboratories. In the absence of rigorous scientific
controls, the DNA fingerprint bands of two completely unknown
samples cannot be matched precisely, and the results may be
unreliable.
Geiger counter
The invention: the first electronic device able to detect and measure
radioactivity in atomic particles.
The people behind the invention:
Hans Geiger (1882-1945), a German physicist
Ernest Rutherford (1871-1937), a British physicist
Sir John Sealy Edward Townsend (1868-1957), an Irish physicist
Sir William Crookes (1832-1919), an English physicist
Wilhelm Conrad Röntgen (1845-1923), a German physicist
Antoine-Henri Becquerel (1852-1908), a French physicist
Discovering Natural Radiation
When radioactivity was discovered and first studied, the work
was done with rather simple devices. In the 1870’s, Sir William
Crookes learned how to create a very good vacuum in a glass tube.
He placed electrodes in each end of the tube and studied the passage
of electricity through the tube. This simple device became known as
the “Crookes tube.” In 1895, Wilhelm Conrad Röntgen was experimenting
with a Crookes tube. It was known that when electricity
went through a Crookes tube, one end of the glass tube might glow.
Certain mineral salts placed near the tube would also glow. In order
to observe carefully the glowing salts, Röntgen had darkened the
room and covered most of the Crookes tube with dark paper. Suddenly,
a flash of light caught his eye. It came from a mineral sample
placed some distance from the tube and shielded by the dark paper;
yet when the tube was switched off, the mineral sample went dark.
Experimenting further, Röntgen became convinced that some ray
from the Crookes tube had penetrated the mineral and caused it to
glow. Since light rays were blocked by the black paper, he called the
mystery ray an “X ray,” with “X” standing for unknown.
Antoine-Henri Becquerel heard of the discovery of X rays and, in
February, 1886, set out to discover if glowing minerals themselves
emitted X rays. Some minerals, called “phosphorescent,” begin to
glow when activated by sunlight. Becquerel’s experiment involved wrapping photographic film in black paper and setting various
phosphorescent minerals on top and leaving them in the sun. He
soon learned that phosphorescent minerals containing uranium
would expose the film.
Aseries of cloudy days, however, brought a great surprise. Anxious
to continue his experiments, Becquerel decided to develop film
that had not been exposed to sunlight. He was astonished to discover
that the film was deeply exposed. Some emanations must be
coming from the uranium, he realized, and they had nothing to do
with sunlight. Thus, natural radioactivity was discovered by accident
with a simple piece of photographic film.
Rutherford and Geiger
Ernest Rutherford joined the world of international physics at
about the same time that radioactivity was discovered. Studying the
“Becquerel rays” emitted by uranium, Rutherford eventually distinguished
three different types of radiation, which he named “alpha,”
“beta,” and “gamma” after the first three letters of the Greek alphabet.
He showed that alpha particles, the least penetrating of the three, are
the nuclei of helium atoms (a group of two neutrons and a proton
tightly bound together). It was later shown that beta particles are electrons.
Gamma rays, which are far more penetrating than either alpha
or beta particles, were shown to be similar to X rays, but with higher
energies.
Rutherford became director of the associated research laboratory
at Manchester University in 1907. Hans Geiger became an assistant.
At this time, Rutherford was trying to prove that alpha particles
carry a double positive charge. The best way to do this was to measure
the electric charge that a stream of alpha particles would bring
to a target. By dividing that charge by the total number of alpha particles
that fell on the target, one could calculate the charge of a single
alpha particle. The problem lay in counting the particles and in
proving that every particle had been counted.
Basing their design upon work done by Sir John Sealy Edward
Townsend, a former colleague of Rutherford, Geiger and Rutherford
constructed an electronic counter. It consisted of a long brass
tube sealed at both ends from which most of the air had been pumped. A thin wire, insulated from the brass, was suspended
down the middle of the tube. This wire was connected to batteries
producing about thirteen hundred volts and to an electrometer, a
device that could measure the voltage of the wire. This voltage
could be increased until a spark jumped between the wire and the
tube. If the voltage was turned down a little, the tube was ready to
operate. An alpha particle entering the tube would ionize (knock
some electrons away from) at least a few atoms. These electrons
would be accelerated by the high voltage and, in turn, would ionize
more atoms, freeing more electrons. This process would continue
until an avalanche of electrons struck the central wire and the electrometer
registered the voltage change. Since the tube was nearly
ready to arc because of the high voltage, every alpha particle, even if
it had very little energy, would initiate a discharge. The most complex
of the early radiation detection devices—the forerunner of the
Geiger counter—had just been developed. The two physicists reported
their findings in February, 1908.
Impact
Their first measurements showed that one gram of radium
emitted 34 thousand million alpha particles per second. Soon, the
number was refined to 32.8 thousand million per second. Next,
Geiger and Rutherford measured the amount of charge emitted
by radium each second. Dividing this number by the previous
number gave them the charge on a single alpha particle. Just as
Rutherford had anticipated, the charge was double that of a hydrogen
ion (a proton). This proved to be the most accurate determination
of the fundamental charge until the American physicist
Robert Andrews Millikan conducted his classic oil-drop experiment
in 1911.
Another fundamental result came froma careful measurement of
the volume of helium emitted by radium each second. Using that
value, other properties of gases, and the number of helium nuclei
emitted each second, they were able to calculate Avogadro’s number
more directly and accurately than had previously been possible.
(Avogadro’s number enables one to calculate the number of atoms
in a given amount of material.)The true Geiger counter evolved when Geiger replaced the central
wire of the tube with a needle whose point lay just inside a thin
entrance window. This counter was much more sensitive to alpha
and beta particles and also to gamma rays. By 1928, with the assistance
of Walther Müller, Geiger made his counter much more efficient,
responsive, durable, and portable. There are probably few radiation
facilities in the world that do not have at least one Geiger
counter or one of its compact modern relatives.
26 June 2009
Gas-electric car
The invention:
A hybrid automobile with both an internal combustion engine and an electric motor.
The people behind the invention:
Victor Wouk - an American engineer Tom Elliott, executive vice president of
American Honda Motor Company
Hiroyuki Yoshino - president and chief executive officer of Honda Motor Company
Fujio Cho - president of Toyota Motor Corporation
23 June 2009
Fuel cell
The invention: An electrochemical cell that directly converts energy
from reactions between oxidants and fuels, such as liquid
hydrogen, into electrical energy.
The people behind the invention:
Francis Thomas Bacon (1904-1992), an English engineer
Sir William Robert Grove (1811-1896), an English inventor
Georges Leclanché (1839-1882), a French engineer
Alessandro Volta (1745-1827), an Italian physicist
The Earth’s Resources
Because of the earth’s rapidly increasing population and the
dwindling of fossil fuels (natural gas, coal, and petroleum), there is
a need to design and develop new ways to obtain energy and to encourage
its intelligent use. The burning of fossil fuels to create energy
causes a slow buildup of carbon dioxide in the atmosphere,
creating pollution that poses many problems for all forms of life on
this planet. Chemical and electrical studies can be combined to create
electrochemical processes that yield clean energy.
Because of their very high rate of efficiency and their nonpolluting
nature, fuel cells may provide the solution to the problem of
finding sufficient energy sources for humans. The simple reaction of
hydrogen and oxygen to form water in such a cell can provide an
enormous amount of clean (nonpolluting) energy. Moreover, hydrogen
and oxygen are readily available.
Studies by Alessandro Volta, Georges Leclanché, and William
Grove preceded the work of Bacon in the development of the fuel
cell. Bacon became interested in the idea of a hydrogen-oxygen fuel
cell in about 1932. His original intent was to develop a fuel cell that
could be used in commercial applications.
The Fuel Cell Emerges
In 1800, the Italian physicist Alessandro Volta experimented
with solutions of chemicals and metals that were able to conduct electricity. He found that two pieces of metal and such a solution
could be arranged in such a way as to produce an electric current.
His creation was the first electrochemical battery, a device that produced
energy from a chemical reaction. Studies in this area were
continued by various people, and in the late nineteenth century,
Georges Leclanché invented the dry cell battery, which is now commonly
used.
The work of William Grove followed that of Leclanché. His first
significant contribution was the Grove cell, an improved form of the
cells described above, which became very popular. Grove experimented
with various forms of batteries and eventually invented the
“gas battery,” which was actually the earliest fuel cell. It is worth
noting that his design incorporated separate test tubes of hydrogen
and oxygen, which he placed over strips of platinum.
After studying the design of Grove’s fuel cell, Bacon decided
that, for practical purposes, the use of platinum and other precious
metals should be avoided. By 1939, he had constructed a cell in
which nickel replaced the platinum used.
The theory behind the fuel cell can be described in the following
way. If a mixture of hydrogen and oxygen is ignited, energy is released
in the form of a violent explosion. In a fuel cell, however, the
reaction takes place in a controlled manner. Electrons lost by the hydrogen
gas flow out of the fuel cell and return to be taken up by the
oxygen in the cell. The electron flow provides electricity to any device
that is connected to the fuel cell, and the water that the fuel cell
produces can be purified and used for drinking.
Bacon’s studies were interrupted byWorldWar II. After the war
was over, however, Bacon continued his work. Sir Eric Keightley
Rideal of Cambridge University in England supported Bacon’s
studies; later, others followed suit. In January, 1954, Bacon wrote an
article entitled “Research into the Properties of the Hydrogen/ Oxygen
Fuel Cell” for a British journal. He was surprised at the speed
with which news of the article spread throughout the scientific
world, particularly in the United States.
After a series of setbacks, Bacon demonstrated a forty-cell unit
that had increased power. This advance showed that the fuel cell
was not merely an interesting toy; it had the capacity to do useful
work. At this point, the General Electric Company (GE), an American corporation, sent a representative to England to offer employment
in the United States to senior members of Bacon’s staff. Three scientists
accepted the offer.
A high point in Bacon’s career was the announcement that the
American Pratt and Whitney Aircraft company had obtained an order
to build fuel cells for the Apollo project, which ultimately put
two men on the Moon in 1969. Toward the end of his career in 1978,
Bacon hoped that commercial applications for his fuel cells would
be found.Impact
Because they are lighter and more efficient than batteries, fuel
cells have proved to be useful in the space program. Beginning with
the Gemini 5 spacecraft, alkaline fuel cells (in which a water solution
of potassium hydroxide, a basic, or alkaline, chemical, is placed)
have been used for more than ten thousand hours in space. The fuel
cells used aboard the space shuttle deliver the same amount of power
as batteries weighing ten times as much. On a typical seven-day
mission, the shuttle’s fuel cells consume 680 kilograms (1,500 pounds)
of hydrogen and generate 719 liters (190 gallons) of water that can
be used for drinking.
Major technical and economic problems must be overcome in order
to design fuel cells for practical applications, but some important
advancements have been made.Afew test vehicles that use fuel cells as a source of power have been constructed. Fuel cells using
hydrogen as a fuel and oxygen to burn the fuel have been used in a
van built by General Motors Corporation. Thirty-two fuel cells are
installed below the floorboards, and tanks of liquid oxygen are carried
in the back of the van. A power plant built in New York City
contains stacks of hydrogen-oxygen fuel cells, which can be put on
line quickly in response to power needs. The Sanyo Electric Company
has developed an electric car that is partially powered by a
fuel cell.
These tremendous technical advances are the result of the singleminded
dedication of Francis Thomas Bacon, who struggled all of
his life with an experiment he was convinced would be successful.
Freeze-drying
The invention:
Method for preserving foods and other organic matter by freezing them and using a vacuum to remove their water content without damaging their solid matter.
The people behind the invention:
Earl W. Flosdorf (1904- ), an American physician
Ronald I. N. Greaves (1908- ), an English pathologist
Jacques Arsène d’Arsonval (1851-1940), a French physicist
Method for preserving foods and other organic matter by freezing them and using a vacuum to remove their water content without damaging their solid matter.
The people behind the invention:
Earl W. Flosdorf (1904- ), an American physician
Ronald I. N. Greaves (1908- ), an English pathologist
Jacques Arsène d’Arsonval (1851-1940), a French physicist
Subscribe to:
Posts (Atom)