Anthropology

advertisement
1
TOPICS
Anthropology
What is anthropology
A day in Samoa
The economic process in primitive societies
The early education of Manus children
Moral standards and social organization
Production in primitive societies
The rules of good fieldwork
The science of custom
Survival in the cage
Gestures
Regional signals
The voices of time
Biology
Evolution and natural selection.
The laws of heredity: Mendel
Banting and the discovery of insulin
The Galapagos
On the origin of species
A modern look at monsters
Business
Brands Up
How to be a great manager
Derivatives - the beauty
www.seyfihoca.com
2
Motives
Research & development
SONY
American and Japanese styles
Chemistry
Metallurgy: Making alloys
Electricity helps Chemistry: Electro-plating
Economics
The conventional wisdom
Markets
Investment
Barter
Productivity as a guide to wages
The failure of the classical theory of commercial policy
Education
The personal qualities of a teacher
Rousseau's Emile
The beginnings of scientific and technical education
Supposed mental faculties and their training
The concept of number
English in the primary school
Britain and Japan: Two roads to higher education
What types of students do you have to teach?
Spoon-fed feel lost at the cutting edge
Geology/Geography
www.seyfihoca.com
3
The age of the earth
Oils
The use of land
History
The nature, object and purpose of history
The expansion of Western civilization
The career of Jenghis Khan
The trial and execution of Charles I
Peter the Great
The United States in 1790
Civilisation and history
Coal
Language
'Primitiveness' in language
English in the fifteenth century
An international language
Language as symbolism
From word symbol to phoneme symbol
Law
Modern constitutions
The functions of government
Initiative and referendum
Law report
The legal character of international law
The law of negligence
www.seyfihoca.com
4
The death penalty
Mathematics
On different degrees of smallness
Chance or probability
Philosophy
Definition and some of its difficulties
The subject matter of philosophy
What can we communicate?
Ethics
Aristotle's Ethics
The road to happiness
Logic
Inductive and deductive logic
Physics
The origin of the sun and the planets.
Can life exist on the planets?
The theory of continuous creation.
The creation of the universe
Atomic radiation and life
Marconi and the invention of radio
Particles or waves?
Matter, mass and energy
Structure of matter
The quantum theory of radiation
Footprints of the atom
www.seyfihoca.com
5
Splitting the atom
The development of electricity
The discovery of x-rays
Politics
Crowds
Diplomacy
What future for Africa?
Nationalism
Democracy
Locke's political theory
The search for world order
The declaration of independence
The rights of man
Psychology
Society and intelligence
The pressure to conform
Learning to live with the computer
Forgetting
Adolescence
Body language
Distance regulation in animals
An observation and an explanation
Gestures
Adaptive control of reading rate
Sociology
www.seyfihoca.com
6
Rational and irrational elements in contemporary society
Social life in a provincial university
The menace of over-population
Changes in English social life after 1918
Scientific method in the social sciences
Shopping in Russia
Technology
Seduced by technology
Blowing hot and cold on British windmills
Direct uses of solar radiation
Industrial robots
Tomorrow's phone calls
Coal
The medium is the message
The development of electricity
The autonomous house
Twentieth Century Discovery
Discovery of insecticides and pesticides
The origin of life
The structure of matter
Distance in our solar system
Space travel
The Artificial World Around Us
What men are doing to things
The nose does not know
www.seyfihoca.com
7
"This movie smells"
The artificial air
The odor-makers
The truth about tastes
Inside the flavor factory
Let us have "Nothing" to eat
What is happening to the steel age?
Diamonds of the laboratory
Who made this ruby?
The synthetic future
www.seyfihoca.com
8
ANTROPOLOGY
What Is Anthropology?
Anthropology is the study of humankind, especially of Homo sapiens, the biological species
to which we human beings belong. It is the study of how our species evolved from more
primitive organisms; it is also the study of how our species developed a mode of
communication known as language and a mode of social life known as culture. It is the study
of how culture evolved and diversified. And finally, it is the study of how culture, people, and
nature Interact wherever human beings are found.
This book is an Introduction to general anthropology, which is an amalgam of four fields of
study traditionally found within departments of anthropology at major universities. The four
fields are cultural anthropology (sometimes called social anthropology), archaeology,
anthropological linguistics, and physical anthropology. The collaborative effort of these four
fields is needed in order to study our species in evolutionary perspective and in relation to
diverse habitats and cultures.
Cultural anthropology deals with the description and analysis of the forms and styles of social
life of past and present ages. Its subdiscipline, ethnography, systematically describes
contemporary societies and cultures. Comparison of these descriptions provides the basis for
hypotheses and theories about the causes of human lifestyles.
Archaeology adds a crucial dimension to this endeavor. By digging up the remains of cultures
of past ages, archaeology studies sequences of social and cultural evolution under diverse
natural and cultural conditions. In the quest for understanding the present-day characteristics
of human existence, for validating or invalidating proposed theories of historical causation,
the great temporal depth of the archaeological record is indispensable.
Anthropological linguistics provides another crucial perspective:
the study of the totality of languages spoken by human beings. Linguistics attempts to
reconstruct the historical changes that have led to the formation of individual languages and
families of languages. More fundamentally, anthropological linguistics is concerned with the
nature of language and Its functions and the way language Influences and is Influenced by
other aspects of cultural life. Anthropological linguistics is concerned with the origin of
language and the relationship between the evolution of language and the evolution of Homo
sapiens. And finally, anthropological linguistics is concerned with the relationship between
the evolution of languages and the evolution and differentiation of human cultures.
Physical anthropology grounds the work of the other anthropological fields in our animal
origins and our genetically determined nature. Physical anthropology seeks to reconstruct the
course of human evolution by studying the fossil remains of ancient human and infrahuman
species. Physical anthropology seeks to describe the distribution of hereditary variations
among contemporary populations and to sort out and measure the relative contributions made
by heredity, environment, and culture to human biology.
www.seyfihoca.com
9
Because of Its combination of biological, archaeological, and ethnographic perspectives,
general anthropology is uniquely suited to the study of many problems of vital Importance to
the survival and well-being of our species.
To be sure, disciplines other than anthropology are concerned with the study of human beings.
Our animal nature is the subject of intense research by biologists, geneticists, and
physiologists. In medicine alone, hundreds of additional specialists investigate the human
body, and psychiatrists and psychologists, rank upon rank, seek the essence of the human
mind and soul. Many other disciplines examine our cultural, intellectual, and aesthetic
behavior. These disciplines include sociology, human geography, social psychology, political
science, economics, linguistics, theology, philosophy, musicology, art, literature, and
architecture. There are also many “area specialists,” who study the languages and life-styles
of particular peoples, nations, or regions: “Latin Americanists,” “Indianists,” “Sinologists,”
and so on. In view of this profusion of disciplines that describe, explain, and Interpret aspects
of human life, what justification can there be for a single discipline that claims to be the
general science of the human species?
The Importance of General Anthropology
Research and publications are accumulating in each of the four fields of anthropology at an
exponential rate. Few anthropologists nowadays master more than one field. And
anthropologists increasingly find themselves working not with fellow anthropologists of
another field but with members of entirely different scientific or humanistic specialties. For
example, cultural anthropologists interested in the relationship between cultural practices and
the natural environment may be obliged to pay closer attention to agronomy or ecology than
to linguistics. Physical anthropologists interested in the relationship between human and
protohuman fossils may, because of the Importance of teeth in the fossil record, become more
familiar with dentistry journals than with journals devoted to ethnography or linguistics.
Cultural anthropologists interested in the relationship between culture and individual
personality are sometimes more at home professionally with psychiatrists and social
psychologists than with the archaeologists in theIr own university departments. Hence, many
more than four fields are represented in the ongoing research of modern anthropology.
The specialized nature of most anthropological research makes it Imperative that the general
significance of anthropological facts and theories be preserved. This is the task of general
anthropology.
General anthropology does not pretend to survey the entire subject matter of physical,
cultural, archaeological, and linguistic anthropology. Much less does It pretend to survey the
work of the legions of scholars in other disciplines who also study the biological, linguistic,
and cultural aspects of human existence. Rather, it strives to achieve a particular orientation
toward all the human sciences, disciplines, and fields. Perhaps the best word for this
orientation is ecumenical. General anthropology does not teach all that one must know in
order to master the four fields or all that one must know in order to become an anthropologist.
Instead, general anthropology teaches how to evaluate facts and theories about human nature
and human culture by placing them in a total, universalist perspective. In the words of
Frederica De Laguna,
Anthropology is the only discipline that offers a conceptual schema for the whole context of
human experience…. It is like the carrying frame onto which may be fitted all the several
www.seyfihoca.com
10
subjects of a liberal education, and by organizing the load, making it more wieldy and capable
of being carried. (1968, p. 475)
I believe that the importance of general anthropology is that It is panhuman, evolutionary, and
comparative. The previously mentioned disciplines are concerned with only a particular
segment of human experience or a particular time or phase of our cultural or biological
development. But general anthropology is systematically and uncompromisingly comparative.
Its findings are never based upon the study of a single population, race, “tribe,” class, or
nation. General anthropology insists first and foremost that conclusions based upon the study
of one particular human group or civilization be checked against the evidence of other groups
or civilizations under both similar and different conditions. In this way the relevance of
general anthropology transcends the interests of any particular “tribe,” race, nation, or culture.
In anthropological perspective, all peoples and civilizations are fundamentally local and
evanescent. Thus general anthropology is implacably opposed to the insularity and mental
constriction of those who would have themselves and none other represent humanity, stand at
the pinnacle of progress, or be chosen by God or history to fashion the world in their own
Image.
Therefore general anthropology is “relevant” even when It deals with fragments of fossils,
extinct civilizations, remote villages, or exotic customs. The proper study of humankind
requires a knowledge of distant as well as near lands and of remote as well as present times.
Only in this way can we humans hope to tear off the blinders of our local life-styles to look
upon the human condition without prejudice.
Because of Its multidisciplinary, comparative, and diachronic perspective, anthropology holds
the key to many fundamental questions of recurrent and contemporary relevance. It lies
peculiarly within the competence of general anthropology to explicate our species’ animal
heritage, to define what is distinctively human about human nature, and to differentiate the
natural and the cultural conditions responsible for competition, conflict, and war. General
anthropology is also strategically equipped to probe the significance of racial factors in the
evolution of culture and in the conduct of contemporary human affairs. General anthropology
holds the key to an understanding of the origins of social inequality - of racism, exploitation,
poverty, and underdevelopment. Overarching all of general anthropology’s contributions is
the search for the causes of social and cultural differences and similarities. What is the nature
of the determinism that operates in human history, and what are the consequences of this
determinism for individual freedom of thought and action? To answer these questions is to
begin to understand the extent to which we can increase humanity’s freedom and well-being
by conscious intervention in the processes of cultural evolution.
A Day in Samoa
The life of the day begins at dawn, or if the moon has shown until daylight, the shouts of the
young men may be heard before dawn from the hillside. Uneasy in the night, populous with
ghosts, they shout lustily to one another as they hasten with their work. As the dawn begins to
fall among the soft brown roofs and the slender palm trees stand out against a colourless,
gleaming sea, lovers slip home from trysts beneath the palm trees or in the shadow of beached
canoes, that the light may find each sleeper in his appointed place. Cocks crow, negligently,
and a shrill-voiced bird cries from the breadfruit trees. The insistent roar of the reef seems
muted to an undertone for the sounds of a waking village. Babies cry, a few short wails before
www.seyfihoca.com
11
sleepy mothers give them the breast. Restless little children roll out of their sheets and wander
drowsily down to the beach to freshen their faces in the sea. Boys, bent upon an early fishing,
start collecting their tackle and go to rouse their more laggard companions. Fires are lit, here
and there, the white smoke hardly visible against the paleness of the dawn. The whole village,
sheeted and frowsy, stirs, rubs its eyes, and stumbles towards the beach. 'Talofa!' 'Talofa!'
'Will the journey start today?' 'Is it bonito fishing your lordship is going?' Girls stop to giggle
over some young ne'er-do-well who escaped during the night from an angry father's pursuit
and to venture a shrewd guess that the daughter knew more about his presence than she told.
The boy who is taunted by another, who has succeeded him in his sweetheart's favour,
grapples with his rival, his foot slipping in the wet sand. From the other end of the village
comes a long-drawn-out, piercing wail. A messenger has just brought word of the death of
some relative in another village. Half-clad, unhurried women, with babies at their breasts or
astride their hips, pause in their tale of Losa's outraged departure from her father's house to
the greater kindness in the home of her uncle, to wonder who is dead. Poor relatives whisper
their requests to rich relatives, men make plans to set a fish-trap together, a woman begs a bit
of yellow dye from a kinswoman, and through the village sounds the rhythmic tattoo which
calls the young men together. They gather from all parts of the village, diggingsticks in hand,
ready to start inland to the plantation. The older men set off upon their more lonely
occupations, and each household, reassembled under its peaked roof, settles down to the
routine of the morning. Little children, too hungry to wait for the late breakfast, beg lumps of
cold taro which they munch greedily. Women carry piles of washing to the sea or to the
spring at the far end of the village, or set off inland after weaving materials. The older girls go
fishing on the reef, or perhaps set themselves to weaving a new set of Venetian blinds.
In the houses, where the pebbly floors have been swept bare with a stiff, long-handled broom,
the women great with child and the nursing mothers sit and gossip with one another. Old men
sit apart, unceasingly twisting palm husk on their bare thighs and muttering old tales under
their breath. The carpenters begin work on the new house, while the owner bustles about
trying to keep them in a good humour. Families who will cook today are hard at work; the
taro, yams, and bananas have already been brought from inland; the children are scuttling
back and forth, fetching sea water, or leaves to stuff the pig. As the sun rises higher in the sky,
the shadows deepen under the thatched roofs, the sand is burning to the touch, the hibiscus
flowers wilt on the hedges, and little children bid the smaller ones, 'Come out of the sun.'
Those whose excursions have been short return to the village, the women with strings of
crimson jellyfish, or baskets of shellfish, the men with coconuts, carried in baskets slung on a
shoulder-pole. The women and children eat their breakfast, just hot from the oven, if this is
cook day, and the young men work swiftly in the midday heat, preparing the noon feast for
their elders.
It is high noon. The sand burns the feet of the little children, who leave their palm-leaf balls
and their pinwheels of frangipani blossoms to wither in the sun, as they creep into the shade
of the houses. The women who must go abroad carry great banana leaves as sunshades or
wind wet cloths about their heads. Lowering a few blinds against the slanting sun all who are
left in the village wrap their heads in sheets and go to sleep. Only a few adventurous children
may slip away for a swim in the shadow of a high rock, some industrious woman continues
with her weaving, or a close little group of women bend anxiously over a woman in labour.
The village is dazzling and dead; any sound seems oddly loud and out of place. Words have to
cut through the solid heat slowly. And then the sun gradually sinks over the sea.
www.seyfihoca.com
12
A second time the sleeping people stir, roused perhaps by the cry of 'A boat!' resounding
through the village. The fishermen beach their canoes, weary and spent from the heat, in spite
of the slaked lime on their heads, with which they have sought to cool their brains and redden
their hair. The brightly coloured fishes are spread out on the floor, or piled in front of the
houses until the women pour water over them to free them from taboo. Regretfully, the young
fishermen separate out the 'taboo fish', which must be sent to the chief, or proudly they pack
the little palm-leaf baskets with offerings offish to take to their sweethearts. Men come home
from the bush, grimy and heavy laden, shouting as they come, greeted in a sonorous rising
cadence by those who have remained at home. They gather in the guest house for their
evening kava drinking. The soft clapping of hands, the highpitched intoning of the talking
chief who serves the kava echo through the village. Girls gather flowers to weave into
necklaces; children, lusty from their naps and bound to no particular task, play circular games
in the half shade of the late afternoon. Finally the sun sets, in a flame which stretches from the
mountain behind to the horizon on the sea; the last bather comes up from the beach, children
straggle home, dark little figures etched against the sky; lights shine in the houses, and each
household gathers for its evening meal. The suitor humbly presents his offering, the children
have been summoned from their noisy play, perhaps there is an honoured guest who must be
served first, after the soft, barbaric singing of Christian hymns and the brief and graceful
evening prayer. In front of a house at the end of the village, a father cries out the birth of a
son. In some family circles a face is missing, in others little runaways have found a haven.
Again quiet settles upon the village, as first the head of the household, then the women and
children, and last of all the patient boys, eat their supper.
After supper the old people and the little children are bundled off to bed. If the young people
have guests, the front of the house is yielded to them. For day is the time for the councils of
old men and the labours of youth, and night is the time for lighter things. Two kinsmen, or a
chief and his councillor, sit and gossip over the day's events or make plans for the morrow.
Outside a crier goes through the village announcing that the communal breadfruit pit will be
opened in the morning, or that the village will make a great fish-trap. If it is moonlight,
groups of young men, women by twos and threes, wander through the village, and crowds of
children hunt for land crabs or chase each other among the breadfruit trees. Half the village
may go fishing by torchlight, and the curving reef will gleam with wavering lights and echo
with shouts of triumph or disappointment, teasing words or smothered cries of outraged
modesty. Or a group of youths may dance for the pleasure of some visiting maiden.
Many of those who have retired to sleep, drawn by the merry music, will wrap their sheets
about them and set out to find the dancing. A white-clad, ghostly throng will gather in a circle
about the gaily lit house, a circle from which every now and then a few will detach
themselves and wander away among the trees. Sometimes sleep will not descend upon the
village until long past midnight; then at last there is only the mellow thunder of the reef and
the whisper of lovers, as the village rests until dawn.
(From Coming of age in Samoa by Margaret Mead (1928))
The Economic Process in Primitive Societies
In our own economic system money gives a universal measure of values, a convenient
medium of exchange through which we can buy or sell almost anything, and also a standard
by which payments at one time can be expressed as commitments for the future. In a wider
sense it allows for the measurement of services against things, and promotes the flow of the
www.seyfihoca.com
13
economic process. In a primitive society without money we might expect all this to be absent,
yet the economic process goes on. There is a recognition of services, and payment is made for
them; there are means of absorbing people into the productive process, and values are
expressed in quantitative terms, measured by traditional standards.
Let us examine, to begin with, a situation of simple distribution such as occurs when an
animal is killed in a hunt. Do the hunters fall on the carcass and cut it to pieces, the largest
piece to the strongest man? This is hardly ever the case. The beast is normally divided
according to recognized principles. Since the killing of an animal is usually a co-operative
activity one might expect to find it portioned out according to the amount of work done by
each hunter to obtain it. To some extent this principle is followed, but other people have their
rights as well. In many parts of Australia each person in the camp gets a share depending
upon his or her relation to the hunters. The worst parts may even be kept by the hunters
themselves. In former times, at Alice Springs, according to Palmer, when a kangaroo was
killed the hunter had to give the left hind leg to his brother, the tail to his father's brother's
son, the loins and fat to his father-in-law, the ribs to his mother-in-law, the forelegs to his
father's younger sister, the head to his wife, and he kept for himself the entrails and the blood.
In different areas the portions assigned to such kinsfolk differ. When grumbles and fights
occur, as they often do, it is not because the basic principles of distribution are questioned, but
because it is thought they are not being properly followed. Though the hunter, his wife, and
children seem to fare badly, this inequality is corrected by their getting in their turn better
portions from kills by other people. The result is a criss-cross set of exchanges always in
progress. The net result in the long run is substantially the same to each person, but through
this system the principles of kinship obligation and the morality of sharing food have been
emphasized.
We see from this that though the principle that a person should get a reward for his labour is
not ignored, this principle is caught up into a wider set of codes which recognize that kinship
ties, positions, or privilege, and ritual ideas should be supported on an economic basis. As
compared with our own society, primitive societies make direct allowance for the dependants
upon producers as well as for the immediate producers themselves.
These same principles come out in an even more striking way in the feasts which are such an
important part of much primitive life. The people who produce the food, or who own it,
deliberately often hand over the best portions to others.
A feast may be the means of repaying the labour of others; of setting the seal on an important
event, such as initiation or marriage; or of cementing an alliance between groups. Prestige is
usually gained by the giver of the feast, but where personal credit and renown are linked most
closely with the expenditure of wealth, the giving of a feast is a step upon the ladder of social
status. In the Banks Islands and other parts of Melanesia such feasts are part of the ceremonial
of attaining the various ranks of the men's society, which is an important feature of native life.
In Polynesia these graded feasts do not normally occur, but in Tikopia a chief is expected to
mark the progress of his reign by a feast every decade or so. The 'feasts of merit' of the Nagas
of Assam are not so much assertion against social competitors as means of gaining certain
recognized ranks in the society.
(From Human Types, by Raymond Firth.
The Early Education of Manus Children
www.seyfihoca.com
14
For the first few months after he has begun to accompany his mother about the village the
baby rides quietly on her neck or sits in the bow of the canoe while his mother punts in the
stern some ten feet away. The child sits quietly, schooled by the hazards to which he has been
earlier exposed. There are no straps, no baby harness to detain him in his place. At the same
time, if he should tumble overboard, there would be no tragedy. The fall into the water is
painless. The mother or father is there to pick him up. Babies under two and a half or three are
never trusted with older children or even with young people. The parents demand a speedy
physical adjustment from the child, but they expose him to no unnecessary risks. He is never
allowed to stray beyond the limits of safety and watchful adult care.
So the child confronts duckings, falls, dousings of cold water, or entanglements in slimy
seaweed, but he never meets with the type of accident which will make him distrust the
fundamental safety of his world. Although he himself may not yet have mastered the physical
technique necessary for perfect comfort in the water his parents have. A lifetime of dwelling
on the water has made them perfectly at home there. They are sure-footed, clear-eyed, quick
handed. A baby is never dropped; his mother never lets him slip from her arms or carelessly
bumps his head against door post or shelf. In the physical care of the child she makes no
clumsy blunders. Her every move is a reassurance to the child, counteracting any doubts
which he may have accumulated in the course of his own less sure-footed progress. So
thoroughly do Manus children trust their parents that a child will leap from any height into an
adult's outstretched arms, leap blindly and with complete confidence of being safely caught.
Side by side with the parents' watchfulness and care goes the demand that the child himself
should make as much effort, acquire as much physical dexterity, as possible. Every gain a
child makes is noted, and the child is inexorably held to his past record. There are no cases of
children who toddle a few steps, fall, bruise their noses, and refuse to take another step for
three months. The rigorous way of life demands that the children be self-sufficient as early as
possible. Until a child has learned to handle his own body, he is not safe in the house, in a
canoe or on the small islands. His mother or aunt is a slave, unable to leave him for a minute,
never free of watching his wandering steps. So every new proficiency is encouraged and
insisted upon. Whole groups of busy men and women cluster about the baby's first step, but
there is no such delightful audience to bemoan his first fall. He is set upon his feet gently but
firmly and told to try again. The only way in which he can keep the interest of his admiring
audience is to try again. So self-pity is stifled and another step is attempted.
As soon as the baby can toddle uncertainly he is put down into the water at low tide when
parts of the lagoon are high and others only a few inches under water. Here the baby sits and
plays in the water or takes a few hesitating steps in the yielding spongy mud. The mother does
not leave his side, nor does she leave him there long enough to weary him. As he grows older,
he is allowed to wade about at low tide. His elders keep a sharp lookout that he does not stray
into deep water until he is old enough to swim. But the supervision is unobtrusive. Mother is
always there if the child gets into difficulties, but he is not nagged and plagued with continual
'don'ts'. His whole play-world is so arranged that he is permitted to make small mistakes from
which he may learn better judgment and greater circumspection, but he is never allowed to
make mistakes which are serious enough to frighten him permanently or inhibit his activity.
He is a tight-rope walker, learning feats which we would count outrageously difficult for little
children, but his tight-rope is stretched above a net of expert parental solicitude.... Expecting
children to swim at three, to climb about like young monkeys even before that age, may look
to us like forcing them; really it is simply a quiet insistence upon their exerting every particle
of energy and strength which they possess. Swimming is not taught: the small waders imitate
www.seyfihoca.com
15
their slightly older brothers and sisters, and after floundering about in waist deep water begin
to strike out for themselves. Sure-footedness on land and swimming come almost together, so
that the charm which is recited over a newly delivered woman says, 'May you not have
another child until this one can walk and swim.'
(From Growing up in New Guinea, by Margaret Mead.)
Moral Standards and Social Organization
What the anthropologist does in the study of moral systems is to examine for particular
societies the ideas of right and wrong that are held, and their social circumstances.
Consideration of material from some of the more primitive societies, and a contrast of it with
Western patterns, will help to bring out some of the basic moral aspects of social action.
A simple way of introducing this subject is to mention a personal experience. It concerns the
morality of giving, which has important problems in all human societies.
When I went to the isolated island of Tikopia I was dependent, as every anthropologist is, on
the local people for information and for guidance. This they gave, freely in some respects, but
with reservation in others, particularly on religious matters. Almost without exception, too,
they showed themselves greedy for material goods such as knives, fish-hooks, calico, pipes
and tobacco, and adept at many stratagems for obtaining them. In particular, they used the
forms of friendship. They made me gifts in order to play upon the sense of obligation thus
aroused in me. They lured me to their houses by generous hospitality which it was difficult to
refuse, and then paraded their poverty before me. The result of a month or two of this was that
I became irritated and weary. My stocks of goods were not unlimited, and I did not wish to
exhaust them in this casual doling out to people from whom I got no special anthropological
return. I foresaw the time when I would wish to reward people for ethnographic data and help
of a scientific kind and I would either have debased my currency or exhausted it. Moreover I
came to the conclusion that there was no such thing as friendship or kindliness among these
people. Everything they did for me seemed to be in expectation of some return. What was
worse, they were apt to ask for such return at the time, or even in advance of their service.
Then I began to reflect. What was this disinterested friendship and kindness which I expected
to find? Why, indeed, should these people do many services for me, a perfect stranger,
without return? Why should they be content to leave it to me to give them what I wanted
rather than express their own ideas upon what they themselves wanted? In our European
society how far can we say disinterestedness goes? How far do we use this term for what is
really one imponderable item in a whole series of interconnected services and obligations? A
Tikopia, like anyone else, will help to pick a person up if he slips to the ground, bring him a
drink, or do many other small things without any mention of reciprocation. But many other
services which involve him in time and trouble he regards as creating an obligation. This is
just what they do themselves. He thinks it right to get a material reward, and right that he
should be able to ask for it. Is he wrong in this? Was my moral indignation at his self-seeking
justified?
So I revised my procedure. At first I had expected a man to do me a service and wait, until, in
my own good time, I made him a freewill gift. Now I abandoned the pretence of disinterested
friendliness. When a gift was made to me or a service done, I went at once to my stores,
www.seyfihoca.com
16
opened them, and made the giver a present roughly commensurate to the value of that
received.
But more important than the change in my procedure was the change in my moral attitudes. I
was no longer indignant at the behaviour of these calculating savages, to whom friendship
seemed to be expressed only in material terms. It was pleasant and simple to adopt their
method. If one was content to drop the search for 'pure' or 'genuine' sentiments and accept the
fact that to people of another culture, especially when they had not known one long, the most
obvious foundation of friendship was material reciprocity, the difficulties disappeared. When
the obligation to make a material return was dragged into the light, it did not inhibit the
development of sentiments of friendship, but fostered it.
What I have shown of the material elements of friendship in Tikopia is intelligible in a society
where no very clear-cut line is drawn between social service and economic service, where
there is no sale or even barter of goods, but only borrowing and exchange in friendly or
ceremonial form. In European culture we demarcate the sphere of business from that of
friendship. The former insists on the rightness of obtaining the best bargain possible, while
the latter refuses to treat in terms of bargains at all. Yet there is an intermediate sphere.
Business has its social morality. Things are done 'as a favour', there are concepts of 'fair'
prices, and sharp practice and profiteering are judged as wrong. On the other hand, friendship
does not necessarily ignore the material aspects. 'One good turn deserves another' epitomizes
regard for reciprocity which underlies many friendly actions.
(From Elements of Social Organization, by Raymond Firth.)
Production in Primitive Societies
The exploitation of the natural resources of the environment constitutes the productive system
of any people, and the organization of this system in primitive society differs in several
important respects from our own. The first point which must be mentioned is the character of
work. As we have said, most economic effort in primitive society is devoted to the production
of food. The activities involved in this have, quite apart from the stimulus of real or potential
hunger, a spontaneous interest lacking in the ordinary work of an office or factory in
contemporary civilization. This will become clear when we reflect that most of the foodgetting activities of primitive peoples, such as fishing, hunting and gardening, are recreations
among ourselves. It does not follow that primitive man takes an undiluted pleasure in such
activities-much of the labour connected with them is heavy, monotonous or hazardous. But
they do possess an inherent interest lacking in most of the economic labour in modern
civilization, and much the same applies to primitive technology, in which the craftsman
himself creates an artefact, rather than being merely a human cog in the machinery of
production.
The spontaneous interest of work under primitive conditions is reinforced by a number of
social values attached to it. Skill and industry are honoured and laziness condemned, a
principle exemplified in the folk songs and proverbs of the Maori. From childhood onwards
the virtues of industry are extolled, as in the term ihu puku, literally 'dirty nose', applied as a
compliment to an industrious man because it implies that he is continually occupied in
cultivation with his face to the ground; on the other hand, the twin vices of greed and laziness
are condemned in the saying: 'Deep throat, shallow muscles'. Such social evaluations as these
www.seyfihoca.com
17
give pride in successful and energetic work, and stimulate potential laggards to play their part
in productive effort.
The interest of primitive work is increased, and its drudgery mitigated, by the fact that it is
often co-operative. Major undertakings, such as house-building or the construction of large
canoes, usually require the labour of more than one person. And even when the task
concerned could be done individually, primitive peoples often prefer collective labour. Thus
in Hehe agriculture much of the cultivation is done individually or by small family groups.
But at the time of the annual hoeing of the ground, it is customary for a man to announce that
on a certain day his wife will brew beer. His relatives and neighbours attend, help with the
hoeing, and are rewarded with beer in the middle of the day and in the evening. This is not to
be regarded as payment, since casual visitors who have not helped with the hoeing may also
take part in the beer drink. Under this system, each man helps others and is helped by them in
turn. From the purely economic point of view, the system has no advantage, since each man
could quite well hoe his own ground and the preparation of beer adds substantially to the
work involved. But the system does possess psychological advantages. The task of hoeing
might well appear endless if undertaken by each individual separately. Collective labour, and
the collateral activity of beer-drinking, changes a dreary task into a social occasion. The same
principle applies to collective labour in general in primitive society, and to the social activities
of feasting, dancing and other forms of collective enjoyment which frequently accompany it
or mark its conclusion.
(From An Introduction to Social Anthropology, by Ralph Piddington.)
Production in Primitive Societies
The exploitation of the natural resources of the environment constitutes the productive system
of any people, and the organization of this system in primitive society differs in several
important respects from our own. The first point which must be mentioned is the character of
work. As we have said, most economic effort in primitive society is devoted to the production
of food. The activities involved in this have, quite apart from the stimulus of real or potential
hunger, a spontaneous interest lacking in the ordinary work of an office or factory in
contemporary civilization. This will become clear when we reflect that most of the foodgetting activities of primitive peoples, such as fishing, hunting and gardening, are recreations
among ourselves. It does not follow that primitive man takes an undiluted pleasure in such
activities-much of the labour connected with them is heavy, monotonous or hazardous. But
they do possess an inherent interest lacking in most of the economic labour in modern
civilization, and much the same applies to primitive technology, in which the craftsman
himself creates an artefact, rather than being merely a human cog in the machinery of
production.
The spontaneous interest of work under primitive conditions is reinforced by a number of
social values attached to it. Skill and industry are honoured and laziness condemned, a
principle exemplified in the folk songs and proverbs of the Maori. From childhood onwards
the virtues of industry are extolled, as in the term ihu puku, literally 'dirty nose', applied as a
compliment to an industrious man because it implies that he is continually occupied in
cultivation with his face to the ground; on the other hand, the twin vices of greed and laziness
are condemned in the saying: 'Deep throat, shallow muscles'. Such social evaluations as these
www.seyfihoca.com
18
give pride in successful and energetic work, and stimulate potential laggards to play their part
in productive effort.
The interest of primitive work is increased, and its drudgery mitigated, by the fact that it is
often co-operative. Major undertakings, such as house-building or the construction of large
canoes, usually require the labour of more than one person. And even when the task
concerned could be done individually, primitive peoples often prefer collective labour. Thus
in Hehe agriculture much of the cultivation is done individually or by small family groups.
But at the time of the annual hoeing of the ground, it is customary for a man to announce that
on a certain day his wife will brew beer. His relatives and neighbours attend, help with the
hoeing, and are rewarded with beer in the middle of the day and in the evening. This is not to
be regarded as payment, since casual visitors who have not helped with the hoeing may also
take part in the beer drink. Under this system, each man helps others and is helped by them in
turn. From the purely economic point of view, the system has no advantage, since each man
could quite well hoe his own ground and the preparation of beer adds substantially to the
work involved. But the system does possess psychological advantages. The task of hoeing
might well appear endless if undertaken by each individual separately. Collective labour, and
the collateral activity of beer-drinking, changes a dreary task into a social occasion. The same
principle applies to collective labour in general in primitive society, and to the social activities
of feasting, dancing and other forms of collective enjoyment which frequently accompany it
or mark its conclusion.
(From An Introduction to Social Anthropology, by Ralph Piddington.)
The Science of Custom
Anthropology is the study of human beings as creatures of society. It fastens its attention upon
those physical characteristics and industrial techniques, those conventions and values, which
distinguish one community from all others that belong to a different tradition.
The distinguishing mark of anthropology among the social sciences is that it includes for
serious study other societies than our own. For its purposes any social regulation of mating
and reproduction is as significant as our own, though it may be that of the Sea Dyaks, and
have no possible historical relation to that of our civilization. To the anthropologist, our
customs and those of a New Guinea tribe are two possible social schemes for dealing with a
common problem, and in so far as he remains an anthropologist he is bound to avoid any
weighting of one in favour of the other. He is interested in human behaviour, not as it is
shaped by one tradition, our own, but as it has been shaped by any tradition whatsoever. He is
interested in the great gamut of custom that is found in various cultures, and his object is to
understand the way in which these cultures change and differentiate, the different forms
through which they express themselves, and the manner in which the customs of any peoples
function in the lives of the individuals who compose them.
Now custom has not been commonly regarded as a subject of any great moment. The inner
workings of our own brains we feel to be uniquely worthy of investigation, but custom, we
have a way of thinking, is behaviour at its most commonplace. As a matter of fact, it is the
other way round. Traditional custom, taken the world over, is a mass of detailed behaviour
more astonishing than what any one person can ever evolve in individual actions no matter
how aberrant. Yet that is a rather trivial aspect of the matter. The fact of first-rate importance
www.seyfihoca.com
19
is the predominant role that custom plays in experience and belief, and the very great varieties
it may manifest.
No man ever looks at the world with pristine eyes. He sees it edited by a definite set of
customs and institutions and ways of thinking. Even in his philosophical probings he cannot
go behind these stereotypes; his very concepts of the true and the false will still have
reference to his particular traditional customs. John Dewey has said in all seriousness that the
part played by custom in shaping the behaviour of the individual as over against any way in
which he can affect traditional custom, is as the proportion of the total vocabulary of his
mother tongue over against those words of his own baby talk that are taken up into the
vernacular of his family. When one seriously studies social orders that have had the
opportunity to develop autonomously, the figure becomes no more than exact and matter-offact observation. The life-history of the individual is first and foremost an accommodation to
the patterns and standards traditionally handed down in his community. From the moment of
his birth the customs into which he is born shape his experience and behaviour. By the time
he can talk, he is the little creature of his culture, and by the time he is grown and able to take
part in its activities, its habits are his habits, its beliefs his beliefs, its impossibilities his
impossibilities. Every child that is born into his group will share them with him, and no child
born into one on the opposite side of the globe will ever achieve the thousandth part. There is
no social problem it is more incumbent upon us to understand than this of the role of custom.
Until we are intelligent as to its laws and varieties, the main complicating facts of human life
must remain unintelligible.
(From Patterns of Culture, by Ruth Benedict.)
Survival in the Cage
Most casual visitors to zoos are convinced, as they stroll from cage to cage, that the antics of
the inmates are no more than an obliging performance put on solely for their entertainment.
Unfortunately for our consciences, however, this sanguine view of the contented, playful,
caged animal could in many cases be hardly farther from the truth. Recent research at London
Zoo has amply demonstrated that many caged animals are in fact facing a survival problem as
severe as that of their cousins in the wild-a struggle to survive, simply, against the monotony
of their environment. Well fed, well housed, well cared for, and protected from its natural
enemies, the zoo animal in its super-Welfare State existence is bored, sometimes literally to
death.
The extraordinary and subtle lengths to which some animals go to overcome this problem, and
the surprising behaviour patterns which arise as a result, were vividly described by Dr
Desmond Morris (Curator of Mammals, London Zoo) at a conference on `The biology of
survival' held in the rooms of the Zoological Society. As he and other speakers pointed out,
the problem of surviving in a monotonous and restricted environment is not confined to the
animal cage. Apart from the obvious examples of human prisoners or the astronaut, the
number of situations in which human beings have to face boredom and confinement for long
stretches is growing rather than decreasing. More to the point, many of the ways in which
animals respond to these conditions have striking analogies in many forms of obsessional or
neurotic behaviour in humans: the psychiatrist could well learn from the apes.
The animals which seem to react most strongly to this monotony are the higher 'nonspecialists' - those that do not rely on one or two highly developed adaptations or `tricks' to
www.seyfihoca.com
20
survive in the wild. Normally seizing every opportunity to exploit the chances and variety of
their surroundings, they are constantly investigating and exploring; in short, they are neophilic
(loving the new). Most species seem to show a balance between this active, curious behaviour
and its opposite, or neophobic; but with the primates and members of the dog and cat families,
for example, the neophilic pattern is usually overwhelmingly predominant.
It is not surprising that when such species are placed in the highly non-variable environment
of a zoo cage, where there are few novel stimuli, they cannot accept-and indeed actually fight
against-any kind of enforced inactivity (apart from those times when they obviously just give
up and relax). As Dr Morris has remarked, how they do this is a great testimony to their
ingenuity.
Observations by Dr Morris and the staff of London Zoo have revealed that there are probably
five main ways in which animals try to overcome their monotony. The first is to invent new
motor patterns for themselves-new exercises, gymnastics, and so forth. They may also try to
increase the complexity of their environment by creating new stimulus situations: many
carnivores, such as the large cats, play with their food as though it were a living animal,
throwing up dead birds in the air, pursuing the carcass, and pouncing on it to kill.
Alternatively the animal may increase the quantity of its reaction to normal stimuli.
Hypersexuality is one common response to this type of behaviour. A fourth method, akin to
some kinds of obsessional behaviour in man, is to increase the variability of its response to
stimuli such as food. Many animals can be seen playing, pawing, advancing, and retreating
from their food before eating it: some even go further by regurgitating it once eaten and then
devouring it again, and so on. Lastly-and this kind of behaviour can most nearly be called
neurotic-is the development of full, normal responses to subnormal stimuli, such as the
camel's expression of sexual arousal when cigarette smoke is blown in its face, or the making
of mother substitutes out of straw, wood, and suchlike.
No one claims that the observations of animals under these conditions are anything but
fragmentary. But at least enough is now known about them to persuade zoologists that these
bizarre behaviour patterns are not just haphazard, neurotic responses but are genuine attempts
to put back some kind of values into the animal's surroundings, attempts which are beginning
to show consistent patterns. It is also too early to say how far studies of this sort can throw
light on human behaviour under similar conditions (though as one zoologist remarked, they
do show that the best way of surviving a prison sentence is to turn oneself utterly neophobic
and take up an advanced course on economics). Yet there is a growing realization that the
human environment in the future will become more like that of the zoo animal rather than
less, so that the kind of observations mentioned above might well have a growing relevance.
Fairly recent studies of coalminers, for example, have shown that in spite of their
phenomenally high daily energy output they spend about seventeen hours sitting or lying
down, and sedentary workers some twenty to twenty-one hours. With the spread of
automation and the growth of white collar workers these figures are likely to increase. With
astronauts, polar scientists, and long-range aircraft crews the problem already exists: a recent
study of Antarctic scientists produced the remarkable fact that on average they spent only four
per cent of their time in winter outside the confines of their living quarters. With this
continued confinement and the extreme uniformity of the outside environment many odd
behaviour patterns were developed.
www.seyfihoca.com
21
(From an article by Gerald Leach in The Guardian, Tuesday, May 14th, 1963.)
GESTURES
A gesture is any action that sends a visual signal to an onlooker. To become a gesture, an act
has to be seen by someone else and has to communicate some piece of information to them. It
can do this either because the gesturer deliberately sets out to send a signal - as when he
waves his hand - or it can do it only incidentally - as when he sneezes. The hand-wave is a
Primary Gesture, because it has no other existence or function. It is a piece of communication
from start to finish. The sneeze, by contrast, is a secondary, or Incidental Gesture. Its primary
function is mechanical and is concerned with the sneezer’s personal breathing problem. In its
secondary role, however, it cannot help but transmit a message to his companions, warning
them that he may have caught a cold.
Most people tend to limit their use of the term ‘gesture’ to the primary form - the hand-wave
type - but this misses an important point. What matters with gesturing is not what signals we
think we are sending out, but what signals are being received. The observers of our acts will
make no distinction between our intentional Primary Gestures and our unintentional,
incidental ones. In some ways, our Incidental Gestures are the more illuminating of the two, if
only for the very fact that we do not think of them as gestures, and therefore do not censor and
manipulate them so strictly. This is why it is preferable to use the term ‘gesture’ in its wider
meaning as an ‘observed action’.
A convenient way to distinguish between Incidental and Primary Gestures is to ask the
question: Would I do it if I were completely alone? If the answer is No, then it is a Primary
Gesture. We do not wave, wink, or point when we are by ourselves; not, that is, unless we
have reached the unusual condition of talking animatedly to ourselves.
INCIDENTAL GESTURES
Mechanical actions with secondary messages
Many of our actions are basically non-social, having to do with problems of personal body
care, body comfort and body transportation; we clean and groom ourselves with a variety of
scratchings, rubbings and wipings; we cough, yawn and stretch our limbs; we eat and drink;
we prop ourselves up in restful postures, folding our arms and crossing our legs; we sit, stand,
squat and recline, in a whole range of different positions; we crawl, walk and run in varying
gaits and styles. But although we do these things for our own benefit, we are not always
unaccompanied when we do them. Our companions learn a great deal about us from these
‘personal’ actions - not merely that we are scratching because we itch or that we are running
because we are late, but also, from the way we do them, what kind of personalities we possess
and what mood we are in at the time.
Sometimes the mood-signal transmitted unwittingly in this way is one that we would rather
conceal, if we stopped to think about it. Occasionally we do become self-consciously aware of
the ‘mood broadcasts’ and ‘personality displays’ we are making and we may then try to check
ourselves. But often we do not, and the message goes out loud and clear.
For instance, if a student props his head on his hands while listening to a boring lecture, his
head-on-hands action operates both mechanically and gesturally. As a mechanical act, it is
www.seyfihoca.com
22
simply a case of supporting a tired head - a physical act that concerns no one but the student
himself. At the same time, though, it cannot help operating as a gestural act, beaming out a
visual signal to his companions, and perhaps to the lecturer himself, telling them that he is
bored.
In such a case his gesture was not deliberate and he may not even have been aware that he
was transmitting it. If challenged, he would claim that he was not bored at all, but merely
tired. If he were honest - or impolite - he would have to admit that excited attention easily
banishes tiredness, and that a really fascinating speaker need never fear to see a slumped,
head-propped figure like his in the audience.
In the schoolroom, the teacher who barks at his pupils to ‘sit up straight’ is demanding, by
right, the attention-posture that he should have gained by generating interest in his lesson. It
says a great deal for the power of gesture-signals that he feels more ‘attended-to’ when he
sees his pupils sitting up straight, even though he is consciously well aware of the fact that
they have just been forcibly un-slumped, rather than genuinely excited by his teaching.
Many of our Incidental Gestures provide mood information of a kind that neither we nor our
companions become consciously alerted to. It is as if there is an underground communication
system operating just below the surface of our social encounters. We perform an act and it is
observed. Its meaning is read, but not out loud. We ‘feel’ the mood, rather than analyse it.
Occasionally an action of this type becomes so characteristic of a particular situation that we
do eventually identify it - as when we say of a difficult problem: ‘That will make him scratch
his head’, indicating that we do understand the link that exists between puzzlement and the
Incidental Gesture of head-scratching. But frequently this type of link operates below the
conscious level, or is missed altogether.
Where the links are clearer, we can, of course, manipulate the situation and use our Incidental
Gestures in a contrived way. If a student listening to a lecture is not tired, but wishes to insult
the speaker, he can deliberately adopt a bored, slumped posture, knowing that its message will
get across. This is a Stylized Incidental Gesture - a mechanical action that is being artificially
employed as a pure signal. Many of the common ‘courtesies’ also fall into this category - as
when we greedily eat up a plate of food that we do not want and which we do not like, merely
to transmit a suitably grateful signal to our hosts. Controlling our Incidental Gestures in this
way is one of the processes that every child must learn as it grows up and learns to adapt to
the rules of conduct of the society in which it lives.
EXPRESSIVE GESTURES
Biological gestures of the kind we share with other animals
Primary Gestures fall into six main categories. Five of these are unique to man, and depend on
his complex, highly evolved brain. The exception is the category I called Expressive Gestures.
These are gestures of the type which all men, everywhere, share with one another, and which
other animals also perform. They include the important signals of Facial Expression, so
crucial to daily human interaction.
All primates are facially expressive and among the higher species the facial muscles become
increasingly elaborate, making possible the performance of a whole range of subtly varying
www.seyfihoca.com
23
facial signals. In man this trend reaches its peak, and it is true to say that the bulk of nonverbal signalling is transmitted by the human face.
The human hands are also important, having been freed from their ancient locomotion duties,
and are capable, with their Manual Gesticulations, of transmitting many small mood changes
by shifts in their postures and movements, especially during conversational encounters. I am
defining the word ‘gesticulation’, as distinct from ‘gesture’, as a manual action performed
unconsciously during social interactions, when the gesticulator is emphasizing a verbal point
he is making.
These natural gestures are usually spontaneous and very much taken for granted. Yes, we say,
he made a funny face. But which way did his eyebrows move? We cannot recall. Yes, we say,
he was waving his arms about as he spoke. But what shape did his fingers make? We cannot
remember. Yet we were not inattentive. We saw it all and our brains registered what we saw.
We simply did not need to analyse the actions, any more than we had to spell out the words
we heard, in order to understand them. In this respect they are similar to the Incidental
Gestures of the previous category, but they differ, because here there is no mechanical
function - only signalling. This is the world of smiles and sneers, shrugs and pouts, laughs and
winces, blushes and blanches, waves and beckons, nods and glares, frowns and snarls. These
are the gestures that nearly everyone performs nearly everywhere in the world. They may
differ in detail and in context from place to place, but basically they are actions we all share.
We all have complex facial muscles whose sole job it is to make expressions, and we all stand
on two feet rather than four, freeing our hands and letting them dance in the air evocatively as
we explain, argue and joke our way through our social encounters. We may have lost our
twitching tails and our bristling fur, but we more than make up for it with our marvellously
mobile faces and our twisting, spreading, fluttering hands.
In origin, our Expressive Gestures are closely related to our Incidental Gestures, because their
roots also lie in primarily non-communicative actions. The clenched fist of the gesticulator
owes its origin to an intention movement of hitting an opponent, just as the frown on the face
of a worried man can be traced back to an ancient eye-protection movement of an animal
anticipating physical attack. But the difference is that in these cases the link between the
primary physical action and its ultimate descendant, the Expressive Gesture, has been broken.
Smiles, pouts, winces, gapes, smirks, and the rest, are now, for all practical purposes, pure
gestures and exclusively communicative in function.
Despite their worldwide distribution, Expressive Gestures are nevertheless subject to
considerable cultural influences. Even though we all have an evolved set of smiling muscles,
we do not all smile in precisely the same way, to the same extent, or on the same occasions.
For example, all children may start out as easy-smilers and easy-laughers, but a local tradition
may insist that, as the youngsters mature, they must hide their feelings, and their adult
laughter may become severely muted as a result. These local Display Rules, varying from
place to place, often give the false impression that Expressive Gestures are local inventions
rather than modified, but universal, behaviour patterns.
MIMIC GESTURES
Gestures which transmit signals by imitation
www.seyfihoca.com
24
Mimic Gestures are those in which the performer attempts to imitate, as accurately as
possible, a person, an object or an action. Here we leave our animal heritage behind and enter
an exclusively human sphere. The essential quality of a Mimic Gesture is that it attempts to
copy the thing it is trying to portray. No stylized conventions are applied. A successful Mimic
Gesture is therefore understandable to someone who has never seen it performed before. No
prior knowledge should be required and there need be no set tradition concerning the way in
which a particular item is represented. There are four kinds of Mimic Gesture:
First, there is Social Mimicry, or ‘putting on a good face’. We have all done this. We have all
smiled at a party when really we feel sad, and perhaps looked sadder at a funeral than we feel,
simply because it is expected of us. We lie with simulated gestures to please others. This
should not be confused with what psychologists call ‘role-playing’. When indulging in Social
Mimicry we deceive only others, but when role-playing we deceive ourselves as well.
Second, there is Theatrical Mimicry - the world of actors and actresses, who simulate
everything for our amusement. Essentially it embraces two distinct techniques. One is the
calculated attempt to imitate specifically observed actions. The actor who is to play a general,
say, will spend long hours watching films of military scenes in which he can analyse every
tiny movement and then consciously copy them and incorporate them into his final portrayal.
The other technique is to concentrate instead on the imagined mood of the character to be
portrayed, to attempt to take on that mood, and to rely upon it to produce, unconsciously, the
necessary style of body actions.
In reality, all actors use a combination of both these techniques, although in explaining their
craft they may stress one or other of the two methods. In the past, acting performances were
usually highly stylized, but today, except in pantomime, opera and farce, extraordinary
degrees of realism are reached and the formal, obtrusive audience has become instead a
shadowy group of eavesdroppers. Gone are the actor’s asides, gone are the audience
participations. We must all believe that it is really happening. In other words, Theatrical
Mimicry has at last become as realistic as day-to-day Social Mimicry. In this respect, these
first two types of mimic activity contrast sharply with the third, which can be called Partial
Mimicry.
In Partial Mimicry the performer attempts to imitate something which he is not and never can
be, such as a bird, or raindrops. Usually only the hands are involved, but these make the most
realistic approach to the subject they can manage. If a bird, they flap their ‘wings’ as best they
can; if raindrops, they describe a sprinkling descent as graphically as possible. Widely used
mimic gestures of this kind are those which convert the hand into a ‘gun’, an animal of some
sort, or the foot of an animal; or those which use the movements of the hand to indicate the
outline shape of an object of some kind.
The fourth kind of Mimic Gesture can best be called Vacuum Mimicry, because the action
takes place in the absence of the object to which it is related. If I am hungry, for example, I
can go through the motions of putting imaginary food into my mouth. If I am thirsty, I can
raise my hand as if holding an invisible glass, and gulp invisible liquid from it.
The important feature of Partial Mimicry and Vacuum Mimicry is that, like Social and
Theatrical Mimicry, they strive for reality. Even though they are doomed to failure, they make
an attempt. This means that they can be understood internationally. In this respect they
contrast strongly with the next two types of gesture, which show marked cultural restrictions.
www.seyfihoca.com
25
SCHEMATIC GESTURES
Imitations that become abbreviated or abridged
Schematic Gestures are abbreviated or abridged versions of Mimic Gestures. They attempt to
portray something by taking just one of its prominent features and then performing that alone.
There is no longer any attempt at realism.
Schematic Gestures usually arise as a sort of gestural shorthand because of the need to
perform an imitation quickly and on many occasions. Just as, in ordinary speech, we reduce
the word ‘cannot’ to ‘can’t’, so an elaborate miming of a charging bull becomes reduced
simply to a pair of horns jabbed in the air as a pair of fingers.
When one element of a mime is selected and retained in this way, and the other elements are
reduced or omitted, the gesture may still be easy to understand, when seen for the first time,
but the stylization may go so far that it becomes meaningless to those not ‘in the know’. The
Schematic Gesture then becomes a local tradition with a limited geographical range. If the
original mime was complex and involved several distinctive features, different localities may
select different key features for their abridged versions. Once these different forms of
shorthand have become fully established in each region, then the people who use them will
become less and less likely to recognize the foreign forms. The local gesture becomes ‘the’
gesture, and there quickly develops, in gesture communication, a situation similar to that
found in linguistics. Just as each region has its own verbal language, so it also has its own set
of Schematic Gestures.
To give an example: the American Indian sign for a horse consists of a gesture in which two
fingers of one hand ‘sit astride’ the fingers of the other hand. A Cistercian monk would
instead signal ‘horse’ by lowering his head slightly and pulling at an imaginary tuft of hair on
his forehead. An Englishman would probably crouch down like a jockey and pull at imaginary
reins. The Englishman’s version, being closer to a Vacuum Mimic Gesture, might be
understood by the other two, but their gestures, being highly schematic, might well prove
incomprehensible to anyone outside their groups.
Some objects, however, have one special feature that is so strongly characteristic of them that,
even with Schematic Gestures, there is little doubt about what is being portrayed. The bull,
mentioned above, is a good example of this. Cattle are nearly always indicated by their horns
alone, and the two horns are always represented by two digits. In fact, if an American Indian,
a Hindu dancer, and an Australian Aborigine met, they would all understand one another’s
cattle signs, and we would understand all three of them. This does not mean that the signs are
all identical. The American Indian’s cattle sign would represent the bison, and the horns of
bison do not curve forward like those of domestic cattle, but inward, towards each other. The
American Indian’s sign reflects this, his hands being held to his temples and his forefingers
being pointed inward. The Australian Aborigine instead points his forefingers forward. The
Hindu dancer also points forward, but rather than using two forefingers up at the temples,
employs the forefinger and little finger of one hand, held at waist height. So each culture has
its own variant, but the fact that horns are such an obvious distinguishing feature of cattle
means that, despite local variations, the bovine Schematic Gesture is reasonably
understandable in most cultures.
SYMBOLIC GESTURES
www.seyfihoca.com
26
Gestures which represent moods and ideas
A Symbolic Gesture indicates an abstract quality that has no simple equivalent in the world of
objects and movements. Here we are one stage further away from the obviousness of the
enacted Mimic Gesture.
How, for instance, would you make a silent sign for stupidity? You might launch into a fullblooded Theatrical Mime of a drooling village idiot. But total idiocy is not a precise way of
indicating the momentary stupidity of a healthy adult. Instead, you might tap your forefinger
against your temple, but this also lacks accuracy, since you might do precisely the same thing
when indicating that someone is brainy. All the tap does is to point to the brain. To make the
meaning more clear, you might instead twist your forefinger against your temple, indicating ‘a
screw loose’. Alternatively, you might rotate your forefinger close to your temple, signalling
that the brain is going round and round and is not stable.
Many people would understand these temple-forefinger actions, but others would not. They
would have their own local, stupidity gestures, which we in our turn would find confusing,
such as tapping the elbow of the raised forearm, flapping the hand up and down in front of
half-closed eyes, rotating a raised hand, or laying one forefinger flat across the forehead.
The situation is further complicated by the fact that some stupidity signals mean totally
different things in different countries. To take one example, in Saudi Arabia stupidity can be
signalled by touching the lower eyelid with the tip of the forefinger. But this same action, in
various other countries, can mean disbelief, approval, agreement, mistrust, scepticism,
alertness, secrecy, craftiness, danger, or criminality. The reason for this apparent chaos of
meanings is simple enough. By pointing to the eye, the gesturer is doing no more than stress
the symbolic importance of the eye as a seeing organ. Beyond that, the action says nothing, so
that the message can become either: ‘Yes, I see’, or ‘I can’t believe my eyes’, or ‘Keep a
sharp look-out’, or ‘I like what I see’, or almost any other seeing signal you care to imagine.
In such a case it is essential to know the precise ‘seeing’ property being represented by the
symbolism of the gesture in any particular culture.
So we are faced with two basic problems where Symbolic Gestures are concerned: either one
meaning may be signalled by different actions, or several meanings may be signalled by the
same action, as we move from culture to culture. The only solution is to approach each culture
with an open mind and learn their Symbolic Gestures as one would their vocabulary.
As part of this process, it helps if a link can be found between the action and the meaning, but
this is not always possible. In some cases we simply do not know how certain Symbolic
Gestures arose. It is clear that they are symbolic because they now represent some abstract
quality, but how they first acquired the link between action and meaning has been lost
somewhere in their long history. A good instance of this is the ‘cuckold’ sign from Italy. This
consists of making a pair of horns, either with two forefingers held at the temples, or with a
forefinger and little finger of one hand held in front of the body. There is little doubt about
what the fingers are meant to be: they are the horns of a bull. As such, they would rate as part
of a Schematic Gesture. But they do not send out the simple message ‘bull’. Instead they now
indicate ‘sexual betrayal’. The action is therefore a Symbolic gesture and, in order to explain
it, it becomes necessary to find the link between bulls and sexual betrayal.
www.seyfihoca.com
27
Historically, the link appears to be lost, with the result that some rather wild speculations have
been made. A complication arises in the form of the ‘horned hand’, also common in Italy,
which has a totally different significance, even though it employs the same motif of bull’s
horns. The Y horned hand is essentially a protective gesture, made to ward off imagined
dangers. Here it is clear enough that it is the bull’s great power, ferocity and masculinity that
is being invoked as a symbolic aid to protect the gesturer. But this only makes it even more
difficult to explain the other use of the bull’s-horns gesture as a sign of a ‘pathetic’ cuckold.
A suggested explanation of this contradiction is that it is due to one gesture using as its
starting point the bull’s power, while the other - the cuckold sign - selects the bull’s frequent
castration. Since the domestication of cattle began, there have always been too many bulls in
relation to cows. A good, uncastrated bull can serve between 50 and 100 cows a year, so that
it is only necessary to retain a small proportion of intact bulls for breeding purposes. The rest
are rendered much more docile and easy to handle for beef production, by castration. In folklore, then, these impotent males must stand helplessly by, while the few sexually active bulls
‘steal their rightful females’; hence the symbolism of: bull = cuckold.
A completely different explanation once offered was that, when the cuckold discovers that his
wife has betrayed him, he becomes so enraged and jealous that he bellows and rushes
violently about like a ‘mad bull’.
A more classical interpretation involves Diana the Huntress, who made horns into a symbol of
male downfall. Actaeon, another hunter, is said to have sneaked a look at her naked body
when she was bathing. This so angered her that she turned him into a horned beast and set his
own hounds upon him, who promptly killed and ate him.
Alternatively, there is the version dealing with ancient religious prostitutes. These ladies
worshipped gods who wore ‘horns of honour’ - that is, horns in their other role as symbols of
power and masculinity - and the gods were so pleased with the wives who became sacred
whores that they transferred their godly horns on to the heads of the husbands who had
ordered their women to act in this role. In this way, the horns of honour became the horns of
ridicule.
As if this were not enough, it is also claimed elsewhere, and with equal conviction, that
because stags have horns (antlers were often called horns in earlier periods) and because most
stags in the rutting season lose their females to a few dominant males who round up large
harems, the majority of ‘horned’ deer are unhappy ‘cuckolds’.
Finally, there is the bizarre interpretation that bulls and deer have nothing to do with it.
Instead, it is thought that the ancient practice of grafting the spurs of a castrated cockrel on to
the root of its excised comb, where they apparently grew and became ‘horns’, is the origin of
the symbolic link between horns and cuckolds. This claim is backed up by the fact that the
German equivalent word for ‘cuckold’ (hahnrei) originally meant ‘capon’.
If, after reading these rival claims, you feel that all you have really learned is the meaning of
the phrase ‘cock-and-bull story’, you can be forgiven. Clearly, we are in the realm of fertile
imagination rather than historical record. But this example has been dealt with at length to
show how, in so many cases, the true story of the origin of a Symbolic Gesture is no longer
available to us. Many other similarly conflicting examples are known, but this one will suffice
to demonstrate the general principle.
www.seyfihoca.com
28
There are exceptions, of course, and certain of the Symbolic Gestures we make today, and
take for granted, can easily be traced to their origins. ‘Keeping your fingers crossed’ is a good
example of this. Although used by many non-Christians, this action of making the cross,
using only the first and second fingers, is an ancient protective device of the Christian church.
In earlier times it was commonplace to make a more conspicuous sign of the cross (to cross
oneself) by moving the whole arm, first downwards and then sideways, in front of the body,
tracing the shape of the cross in the air. This can still be seen in some countries today in a
non-religious context, acting as a ‘good luck’ protective device. In more trivial situations it
has been widely replaced, however, by the act of holding up one hand to show that the second
finger is tightly crossed over the first, with the crossing movement of the arm omitted.
Originally this was the secret version of ‘crossing oneself’ and was done with the hand in
question carefully hidden from view. It may still be done in this secret way, as when trying to
protect oneself from the consequences of lying, but as a ‘good luck’ sign it has now come out
into the open. This development is easily explained by the fact that crossing the fingers lacks
an obvious religious character. Symbolically, the finger-crossing may be calling on the
protection of the Christian God, but the small finger action performed is so far removed from
the priestly arm crossing action, that it can without difficulty slide into everyday life as a
casual wish for good fortune. Proof of this is that many people do not even realize that they
are demanding an act of Christian worship - historically speaking - when they shout out:
‘Keep your fingers crossed!’
TECHNICAL GESTURES
Gestures used by specialist minorities
Technical Gestures are invented by a specialist minority for use strictly within the limits of
their particular activity. They are meaningless to anyone outside the specialization and operate
in such a narrow field that they cannot be considered as playing a part in the mainstream of
visual communication of any culture.
Television-studio signals are a good example of Technical Gestures in use today. The studio
commentator we see on our screens at home is face to face with a ‘studio manager’. The
manager is linked to the programme director in the control room by means of headphones and
conveys the director’s instructions to the commentator by simple visual gestures. To warn the
commentator that he will have to start speaking at any moment, the manager raises a forearm
and holds it stiffly erect. To start him speaking, he brings the forearm swiftly down to point at
the commentator. To warn him that he must stop speaking in a few seconds, the manager
rotates his forearm, as if it were the hand of a clock going very fast - ‘Time is running out
fast.’ To ask him to lengthen the speaking time and say more, he holds his hands together in
front of his chest and pulls them slowly apart, as if stretching something – ‘stretch it out.’ To
tell the speaker to stop dead this instant, the manager makes a slashing action with his hand
across his throat - ‘Cut!’ There are no set rules laid down for these signals. They grew up in
the early days of television and, although the main ones listed here are fairly widespread
today, each studio may well have its own special variants, worked out to suit a particular
performer.
Other Technical Gestures are found wherever an activity prohibits verbal contact. Skindivers,
for instance, cannot speak to one another and need simple signals to deal with potentially
dangerous situations. In particular they need gestures for danger, cold, cramp and fatigue.
Other messages, such as yes, no, good, bad, up and down, are easily enough understood by
www.seyfihoca.com
29
the use of everyday actions and require no Technical Gestures to make sense. But how could
you signal to a companion that you had cramp? The answer is that you would open and close
one hand rhythmically - a simple gesture, but one that might nevertheless save a life.
Disaster can sometimes occur because a Technical Gesture is required from someone who is
not a specialist in a technical field. Suppose some holiday-makers take out a boat, and it sinks,
and they swim to the safety of a small, rocky island. Wet and frightened, they crouch there
wondering what to do next, when to their immense relief a small fishing-boat comes chugging
towards them. As it draws level with the island, they wave frantically at it. The people on
board wave back, and the boat chugs on and disappears. If the stranded holiday-makers had
been marine ‘specialists’, they would have known that, at sea, waving is only used as a
greeting. To signal distress, they should have raised and lowered their arms stiffly from their
sides. This is the accepted marine gesture for ‘Help!’
Ironically, if the shipwrecked signallers had been marine experts and had given the correct
distress signal, the potential rescue boat might well have been manned by holiday-makers,
who would have been completely nonplussed by the strange actions and would probably have
ignored them. When a technical sphere is invaded by the non-technical, gesture problems
always arise.
Firemen, crane-drivers, airport-tarmac signalmen, gambling-casino croupiers, dealers at
auctions, and restaurant staff, all have their own special Technical Gestures. Either because
they must keep quiet, must be discreet, or cannot be heard, they develop their own sets of
signals. The rest of us can ignore them, unless we, too, wish to enter their specialized spheres.
CODED GESTURES
Sign-language based on a formal system
Coded Gestures, unlike all others, are part of a formal system of signals. They interrelate with
one another in a complex and systematic way, so that they constitute a true language. The
special feature of this category is that the individual units are valueless without reference to
the other units in the code. Technical Gestures may be systematically planned, but, with them,
each signal can operate quite independently of the others. With Coded Gestures, by contrast,
all the units interlock with one another on rigidly formulated principles, like the letters and
words in a verbal language.
The most important example is the Deaf-and-dumb Sign Language of hand signals, of which
there is both a one-handed and a two-handed version. Also, there is the Semaphore Language
of arm signals, and the Tic-tac Language of the race course. These all require considerable
skill and training and belong in a totally different world from the familiar gestures we employ
in everyday life. They serve as a valuable reminder, though, of the incredibly sensitive
potential we all possess for visual communication. It makes it all the more plausible to argue
that we are all of us responding, with greater sensitivity than we may realize, to the ordinary
gestures we witness each day of our lives.
(From Manwatching by Desmond Morris.)
REGIONAL SIGNALS
www.seyfihoca.com
30
The way signals change from country to country and district to district
A Regional Signal is one that has a limited geographical range. If a Norwegian, a Korean and
a Masai were marooned together on a desert island, they would easily be able to communicate
their basic moods and intentions to one another by their actions. All humanity shares a large
repertoire of common movements, expressions and postures. But there would also be
misunderstandings. Each man would have acquired from his own culture a special set of
Regional Signals that would be meaningless to the others. If the Norwegian were shipwrecked
instead with a Swede and a Dane, he would find his task much easier, because their closer
origins would mean a greater share of these regional gestures, since localized actions, like
many words, do not follow precisely the present-day national boundaries.
This comparison of gestures with words is significant because it reveals immediately our state
of ignorance as regards gestural geography. We already know a great deal about linguistic
maps, but we know far too little about Gesture Maps. Ask a linguist to describe the
distribution of any language you like to name and he will be able to provide accurate, detailed
information for you. Take any word, and he will be able to demonstrate its spread from
country to country. He can even present you with local dialect maps for some parts of the
world and show you, like Professor Higgins in Pygmalion, how slang expressions are limited
to certain small areas of big cities. But ask anyone for a world-wide gesture atlas, and you will
be disappointed.
A start has already been made, however, and new field work is now beginning. Although this
research is only in its infancy, recent studies in Europe and around the Mediterranean are
providing some valuable clues about the way gestures change as one travels from locality to
locality. For example, there is a simple gesture in which the forefinger taps the side of the
nose. In England most people interpret this as meaning secrecy or conspiracy. The message is:
'Keep it dark, don't spread it around.' But as one moves down across Europe to central Italy,
the dominant meaning changes to become a helpful warning: 'Take care, there is danger-they
are crafty.' The two messages are related, because they are both concerned with cunning. In
England it is we who are cunning, by not divulging our secret. But in central Italy it is they
who are cunning, and we must be warned against them. The Nose Tap gesture symbolizes
cunning in both cases, but the source of the cunning has shifted.
This is an example of a gesture keeping the same form over a wide range, and also retaining
the same basic meaning, but nevertheless carrying a quite distinct message in two regions.
The more gestures that are mapped in the field, the more common this type of change is
proving to be. Another instance is found in the Eye Touch gesture, where the forefinger
touches the face just below the eye and pulls the skin downwards, opening the eye wider. In
England and France this has the dominant meaning: 'You can't fool me - I see what you are up
to.' But in Italy this shifts to: 'Keep your eyes peeled - pay attention, he's a crook.' In other
words the basic meaning remains one of alertness, but it changes from 'I am alert' to 'You be
alert'.
In both these cases, there is a small number of people in each region who interpret the gesture
in its other meaning. It is not an all-or-none situation, merely a shift in dominance of one
message over the other. This gives some idea of the subtlety of regional changes.
Occasionally there is a total switch as one moves from one district to the next, but more often
than not the change is only a matter of degree.
www.seyfihoca.com
31
Sometimes it is possible to relate the geography of modern Regional Signals to past historical
events. The Chin Flick gesture, in which the backs of the fingers are swept upwards and
forwards against the underside of the chin, is an insulting action in both France and northern
Italy. There it means 'Get lost-you are annoying me.' In southern Italy it also has a negative
meaning, but the message it carries is no longer insulting. It now says simply 'There is
nothing' or 'No' or 'I cannot' or 'I don't want any'. This switch takes place between Rome and
Naples and gives rise to the intriguing possibility that the difference is due to a surviving
influence of ancient Greece. The Greeks colonized southern Italy, but stopped their northern
movement between Rome and Naples. Greeks today use the Chin Flick in the same way as the
southern Italians. In fact, the distribution of this, and certain other gestures, follows
remarkably accurately the range of the Greek civilization at its zenith. Our words and our
buildings still display the mark of early Greek influence, so it should not be too surprising if
ancient Greek gestures are equally tenacious. What is interesting is why they did not spread
farther as time passed. Greek architecture and philosophy expanded farther and farther in their
influences, but for some reason, gestures like the Chin Flick did not travel so well. Many
countries, such as England, lack them altogether, and others, like France, know them only in a
different role.
Another historical influence becomes obvious when one moves to North Africa. There, in
Tunisia, the Chin Flick gesture once again becomes totally insulting: a Tunisian gives a
'French' Chin Flick, rather than a 'Southern Italian' Chin Flick, despite the fact that France is
more remote. The explanation, borne out by other gesture links between France and Tunisia,
is that the French colonial influence in Tunisia has left its imperial mark even on informal
body-language. The modern Tunisian is gesturally more French than any of his closer
neighbours who have not experienced the French presence.
This gives rise to the question as to whether gestures are generally rather conservative,
compared with other social patterns. One talks about the latest fashions in clothing, but one
never hears of 'this season's crop of new gestures'. There does seem to be a cultural tenacity
about them, similar to the persistence found in much folklore and in many children's games
and rhymes. Yet new gestures do occasionally manage to creep in and establish themselves.
Two thousand years ago it was apparently the Greeks who were the 'gesturally virile' nation.
Today it is the British, with their Victory-sign and their Thumbs-up, and the Americans with
their OK Circle-sign. These have spread right across Europe and much of the rest of the world
as well, making their first great advance during the turmoil of the Second World War, and
managing to cling on since then, even in the gesture-rich countries of southern Europe. But
these are exceptions. Most of the local signs made today are centuries old and steeped in
history.
(From Manwatching by Desmond Morris)
The Voices of Time
Time talks. It speaks more plainly than words. The message it conveys comes through loud
and clear. Because it is manipulated less consciously, it is subject to less distortion than the
spoken language. It can shout the truth where words lie.
I was once a member of a mayors’ committee on human relations in a large city. My
assignment was to estimate what the chances were of non-discriminatory practices being
adopted by the different city departments. The first step in this project was to interview the
www.seyfihoca.com
32
department heads, two of whom were themselves members of minority groups. If one were to
believe the words of these officials, it seemed that all of them were more than willing to adopt
non-discriminatory labour practices. Yet I felt that, despite what they said, in only one case
was there much chance for a change. Why? The answer lay in how they used the silent
language of time and space.
Special attention had been given to arranging each interview. Department heads were asked to
be prepared to spend an hour or more discussing their thoughts with me. Nevertheless,
appointments were forgotten; long waits in outer offices (fifteen to forty-five minutes) were
common, and the length of the interview was often cut down to ten or fifteen minutes. I was
usually kept at an impersonal distance during the interview. In only one case did the
department head come from behind his desk. These men had a position and they were literally
and figuratively sticking to it!
The implications of this experience (one which public-opinion pollsters might well heed) is
quite obvious. What people do is frequently more important than what they say. In this case
the way these municipal potentates handled time was eloquent testimony to what they
inwardly believed, for the structure and meaning of time systems, as well as the time
intervals, are easy to identify. In regard to being late there are: “mumble something" periods,
slight apology periods, mildly insulting periods requiring full apology, rude periods, and
downright insulting periods. The psychoanalyst has long been aware of the significance of
communication on this level. He can point to the way his patients handle time as evidence of
“resistances" and “transference."
Different parts of the day, for example, are highly significant in certain contexts. Time may
indicate the importance of the occasion as well as on what level an interaction between
persons is to take place. In the United States if you telephone somebody very early in the
morning, while he is shaving or having breakfast, the time of the call usually signals a matter
of utmost importance and extreme urgency. The same applies for calls after 11.00 p.m. A call
received during sleeping hours is apt to be taken as a matter of life and death, hence the rude
joke value of these calls among the young. Our realization that time talks is even reflected in
such common expressions as, "What time does the clock say?"
An example of how thoroughly these things are taken for granted was reported to me by John
Useem, an American social anthropologist, in an illuminating case from the South Pacific.
The natives of one of the islands had been having a difficult time getting their white
supervisors to hire them in a way consistent with their traditional status system. Through
ignorance the supervisors had hired too many of one group and by so doing had disrupted the
existing balance of power among the natives. The entire population of the island was seething
because of this error. Since the Americans continued in their ignorance and refused to hire
according to local practice, the head men of the two factions met one night to discuss an
acceptable reallocation of jobs. When they finally arrived at a solution, they went en masse to
see the plant manager and woke him up to tell him what had been decided. Unfortunately it
was then between two and three o’clock in the morning. They did not know that it is a sign of
extreme urgency to wake up Americans at this hour. As one might expect, the American plant
manager, who understood neither the local language nor the culture nor what the hullabaloo
was all about, thought he had a riot on his hands and called out the Marines. It simply never
occurred to him that the parts of the day have a different meaning for these people than they
have for us.
www.seyfihoca.com
33
On the other hand, plant managers in the United States are fully aware of the significance of a
communication made during the middle of the morning or afternoon that takes everyone away
from his work. Whenever they want to make an important announcement they will ask:
“When shall we let them know?" In the social world a girl feels insulted when she is asked for
a date at the last minute by someone she doesn’t know very well, and the person who extends
an invitation to a dinner party with only three or four days’ notice has to apologize. How
different from the people of the Middle East with whom it is pointless to make an
appointment too far in advance, because the informal structure of their time system places
everything beyond a week into a single category of “future" in which plans tend to “slip off
their minds.”
Advance notice is often referred to in America as “lead time," an expression which is
significant in a culture where schedules are important. While it is learned informally, most of
us are familiar with how it works in our own culture, even though we cannot state the rules
technically. The rules for lead time in other cultures, however, have rarely been analysed. At
the most they are known by experience to those who have lived abroad for some time. Yet
think how important it is to know how much time is required to prepare people, or for them to
prepare themselves, for things to come. Sometimes lead time would seem to be very extended.
At other times, in the Middle East, any period longer than a week may be too long.
How troublesome differing ways of handling time can be is well illustrated by the case of an
American agriculturalist assigned to duty as an attaché of our embassy in a Latin country.
After what seemed to him a suitable period he let it be known that he would like to call on the
minister who was his counterpart. For various reasons, the suggested time was not suitable; all
sorts of cues came back to the effect that the time was not yet ripe to visit the minister. Our
friend, however, persisted and forced an appointment which was reluctantly granted. Arriving
a little before the hour (the American respect pattern), he waited. The hour came and passed;
five minutes - ten minutes - fifteen minutes. At this point he suggested to the secretary that
perhaps the minister did not know he was waiting in the outer office. This gave him the
feeling that he had done something concrete and also helped to overcome the anxiety that was
stirring inside him. Twenty minutes - twenty-five minutes - thirty minutes - forty-five minutes
(the insult period)!
He jumped up and told the secretary that he had been “cooling his heels” in an outer office for
forty-five minutes and he was “damned sick and tired” of this type of treatment. The message
was relayed to the minister, who said, in effect, “Let him cool his heels.” The attaché’s stay in
the country was not a happy one.
The principal source of misunderstanding lay in the fact that in the country in question the
five-minute delay interval was not significant. Forty-five minutes, on the other hand, instead
of being at the tail end of the waiting scale, was just barely at the beginning. To suggest to an
American’s secretary that perhaps her boss didn’t know you were there after waiting sixty
seconds would seem absurd, as would raising a storm about “cooling your heels” for five
minutes. Yet this is precisely the way the minister registered the protestations of the American
in his outer office! He felt, as usual, that Americans were being totally unreasonable.
Throughout this unfortunate episode the attach? was acting according to the way he had been
brought up. At home in the United States his responses would have been normal ones and his
behaviour legitimate. Yet even if he had been told before he left home this sort of thing would
happen, he would have had difficulty not feeling insulted after he had been kept waiting for
www.seyfihoca.com
34
forty-five minutes. If, on the other hand, he had been taught the details of the local time
system just as he should have been taught the local spoken language, it would have been
possible for him to adjust himself accordingly.
What bothers people in situations of this sort is that they don’t realize they are being subjected
to another form of communication, one that works part of the time with language and part of
the time independently of it. The fact that the message conveyed is couched in no formal
vocabulary makes things doubly difficult, because neither party can get very explicit about
what is actually taking place. Each can only say what he thinks is happening and how he feels
about it. The thought of what is being communicated is what hurts.
AMERICAN TIME
People of the Western world, particularly Americans, tend to think of time as something fixed
in nature, something around us from which we cannot escape; an ever present part of the
environment, just like the air we breathe. That it might be experienced in any other way seems
unnatural and strange, a feeling which is rarely modified even when we begin to discover how
really different it is handled by some other people. Within the West itself certain cultures rank
time much lower in over-all importance than we do. In Latin America, for example, where
time is treated rather cavalierly, one commonly hears the expression, “Our time or your
time?” “Hora americana, hora mejicana?”
As a rule, Americans think of time as a road or a ribbon stretching into the future, along
which one progresses. The road has segments or compartments which are best kept discrete
(“one thing at a time”). People who cannot schedule time are looked down upon as
impractical. In at least some parts of Latin America, the North American (their term for us)
finds himself annoyed when he has made an appointment with somebody, only to find a lot of
other things going on at the same time. An old friend of mine of Spanish cultural heritage
used to run his business according to the “Latino” system. This meant that up to fifteen people
were in his office at the same time. Business which might have been finished in a quarter of
an hour sometimes took a whole day. He realized, of course, that the Anglo-Americans were
disturbed by this and used to make some allowance for them, a dispensation which meant that
they spent only an hour or so in his office when they had planned on a few minutes. The
American concept of the discreteness of time and the necessity for scheduling was at variance
with this amiable and seemingly confusing Latin system. However, if my friend had adhered
to the American system he would have destroyed a vital part of his prosperity.
People who came to do business with him also came to find out things and to visit each other.
The ten to fifteen Spanish-Americans and Indians who used to sit around the office (among
whom I later found myself after I had learned to relax a little) played their own part in a
particular type of communications network.
Not only do we Americans segment and schedule time, but we look ahead and are oriented
almost entirely toward the future. We like new things and are preoccupied with change. We
want to know how to overcome resistance to change. In fact, scientific theories and even some
pseudo-scientific ones, which incorporate a striking theory of change, are often given special
attention.
Time with us is handled much like a material; we earn it, spend it, save it, waste it. To us it is
somewhat immoral to have two things going on at the same time. In Latin America it is not
www.seyfihoca.com
35
uncommon for one man to have a number of simultaneous jobs which he either carries on
from one desk or which he moves between, spending a small amount of time on each.
While we look to the future, our view of it is limited. The future to us is foreseeable future,
not the future of the South Asian that involves many centuries. Indeed, our perspective is so
short as to inhibit the operation of a good many practical projects, such as sixty- and onehundred-year conservation works requiring public support and public funds. Anybody who
has worked in industry or in the government of the United States has heard the following:
"Gentlemen, this is for the long term! Five or ten years.”
For us a “long time” can be almost anything - ten or twenty years, two or three months, a few
weeks, or even a couple of days. The South Asian, however, feels that it is perfectly realistic
to think of a “long time” in terms of thousands of years or even an endless period. A colleague
once described their conceptualization of time as follows: “Time is like a museum with
endless corridors and alcoves. You, the viewer, are walking through the museum in the dark,
holding a light to each scene as you pass it. God is the curator of the museum, and only He
knows all that is in it. One lifetime represents one alcove.”
The American’s view of the future is linked to a view of the past, for tradition plays an
equally limited part in American culture. As a whole, we push it aside or leave it to a few
souls who are interested in the past for some very special reason.
There are, of course, a few pockets, such as New England and the South, where tradition is
emphasized. But in the realm of business, which is the dominant model of United States life,
tradition is equated with experience, and experience is thought of as being very close to if not
synonymous with know-how. Know-how is one of our prized possessions, so that when we
look backward it is rarely to take pleasure in the past itself but usually to calculate the knowhow, to assess the prognosis for success in the future.
Promptness is also valued highly in American life. If people are not prompt, it is often taken
either as an insult or as an indication that they are not quite responsible. There are those, of a
psychological bent, who would say that we are obsessed with time. They can point to
individuals in American culture who are literally time-ridden. And even the rest of us feel
very strongly about time because we have been taught to take it so seriously. We have
stressed this aspect of culture and developed it to a point unequalled anywhere in the world,
except, perhaps, in Switzerland and North Germany. Many people criticize our obsessional
handling of time. They attribute ulcers and hypertension to the pressure engendered by such a
system. Perhaps they are right.
SOME OTHER CONCEPTS OF TIME
Even within the very borders of the United States there are people who handle time in a way
which is almost incomprehensible to those who have not made a major effort to understand it.
The Pueblo Indians, for example, who live in the Southwest, have a sense of time which is at
complete variance with the clock-bound habits of the ordinary American citizen. For the
Pueblos events begin when the time is ripe and no sooner.
I can still remember a Christmas dance I attended some twenty-five years ago at one of the
pueblos near the Rio Grande. I had to travel over bumpy roads for forty-five miles to get
there. At seven thousand feet the ordeal of winter cold at one o’clock in the morning is almost
www.seyfihoca.com
36
unbearable. Shivering in the still darkness of the pueblo, I kept searching for a clue as to when
the dance would begin.
Outside everything was impenetrably quiet. Occasionally there was the muffled beat of a deep
pueblo drum, the opening of a door, or the piercing of the night’s darkness with a shaft of
light. In the church where the dance was to take place a few white towns-folk were huddled
together on a balcony, groping for some clue which would suggest how much longer they
were going to suffer. “Last year I heard they started at ten o’clock.” “They can’t start until the
priest comes.” “There is no way of telling when they will start.” All this punctuated by
chattering teeth and the stamping of feet to keep up circulation.
Suddenly an Indian opened the door, entered, and poked up the fire in the stove. Everyone
nudged his neighbour: “Maybe they are going to begin now." Another hour passed. Another
Indian came in from outside, walked across the nave of the church, and disappeared through
another door. “Certainly now they will begin. After all, it’s almost two o’clock.” Someone
guessed they were just being ornery in the hope that the white men would go away. Another
had a friend in the pueblo and went to his house to ask when the dance would begin. Nobody
knew. Suddenly, when the whites were almost exhausted, there burst upon the night the deep
sounds of the drums, rattles, and low male voices singing. Without warning the dance had
begun.
After years of performances such as this, no white man in his right mind will hazard a guess
as to when one of these ceremonial dances will begin. Those of us who have learned now
know that the dance doesn’t start at a particular time. It is geared to no schedule. It starts
when “things” are ready!
As I pointed out, the white civilized Westener has a shallow view of the future compared to
the Oriental. Yet set beside the Navajo Indians of northern Arizona, he seems a model of
long-term patience. The Navajo and the European-American have been trying to adjust their
concepts of time for almost a hundred years. So far they have not done too well. To the oldtime Navajo time is like space - only the here and now is quite real. The future has little
reality to it.
An old friend of mine reared with the Navajo expressed it this way: “You know how the
Navajo love horses and how much they love to gamble and bet on horse races. Well, if you
were to say to a Navajo, ‘My friend, you know my quarter horse that won all the races at
Flagstaff last Fourth of July?’ that Navajo would eagerly say ‘yes, yes,’ he knew the horse;
and if you were to say, ‘In the fall I am going to give you that horse,’ the Navajo’s face would
fall and he would turn round and walk away. On the other hand, if you were to say to him,
‘Do you see that old bag of bones I just rode up on? That old hay-bellied mare with the knock
knees and pigeon toes, with the bridle that’s falling apart and the saddle that’s worn out? You
can have that horse, my friend, it’s yours. Take it, ride it away now.’ Then the Navajo would
beam and shake your hand and jump on his new horse and ride away. Of the two, only the
immediate gift has reality; a promise of future benefits is not even worth thinking about.”
In the early days of the range control and soil conservation programs it was almost impossible
to convince the Navajo that there was anything to be gained from giving up their beloved
sheep for benefits which could be enjoyed ten or twenty years in the future. Once I was
engaged in the supervision of the construction of small earth dams and like everyone else had
little success at first in convincing Navajo workmen that they should work hard and build the
www.seyfihoca.com
37
dam quickly, so that there would be more dams and more water for the sheep. The argument
that they could have one dam or ten, depending on how hard they worked, conveyed nothing.
It wasn’t until I learned to translate our behaviour into their terms that they produced as we
knew they could.
The solution came about in this way. I had been discussing the problem with a friend,
Lorenzo Hubbell, who had lived on the reservation all his life. When there were difficulties I
used to find it helpful to unburden myself to him. Somewhere in his remarks there was always
a key to the underlying patterns of Navajo life. As we talked I learned that the Navajo
understood and respected a bargain. I had some inkling of this when I noticed how unsettled
the Indians became when they were permitted to fall down on the job they had agreed to do.
In particular they seemed to be apprehensive lest they be asked to repay an unfulfilled
obligation at some future time. I decided to sit down with the Navajo crew and talk to them
about the work. It was quite useless to argue about the future advantages which would accrue
from working hard; linear reasoning and logic were meaningless. They did respond, however,
when I indicated that the government was giving them money to get out of debt, providing
jobs near their families, and giving them water for their sheep. I stressed the fact that in
exchange for this, they must work eight hours every day. This was presented as a bargain.
Following my clarification the work progressed satisfactorily.
One of my Indian workmen inadvertently provided another example of the cultural conflict
centring around time. His name was “Little Sunday.” He was small, wiry, and winning. Since
it is not polite to ask the Navajo about their names or even to ask them what their name is, it
was necessary to inquire of others how he came to be named “Little Sunday.” The explanation
was a revealing one.
In the early days of the white traders the Indians had considerable difficulty getting used to
the fact that we Europeans divided time into strange and unnatural periods instead of having a
“natural” succession of days which began with the new moon and ended with the old. They
were particularly perplexed by the notion of the week introduced by the traders and
missionaries. Imagine a Navajo Indian living some forty or fifty miles from a trading store
that is a hundred miles north of the railroad deciding that he needs flour and maybe a little
lard for bread. He thinks about the flour and the lard, and he thinks about his friends and the
fun he will have trading, or maybe he wonders if the trader will give him credit or how much
money he can get for the hide he has. After riding horseback for a day and a half to two days
he reaches the store all ready to trade. The store is locked up tight. There are a couple of other
Navajo Indians camped in the hogan built by the trader. They say the trader is inside but he
won’t trade because it’s Sunday. They bang on his door and he tells them, “Go away, it’s
Sunday,” and the Navajo says, “But I came from way up on Black Mesa, and I am hungry. I
need some food.“What can the trader do? Soon he opens the store and then all the Navajo
pour in. One of the most frequent and insistent Sunday visitors was a man who earned for
himself the sobriquet “Big Sunday.” “Little Sunday,” it turns out, ran a close second.
The Sioux Indians provide us with another interesting example of the differing views toward
time. Not so long ago a man who was introduced as the superintendent of the Sioux came to
my office. I learned that he had been born on the reservation and was a product of both Indian
and white cultures, having earned his A.B. at one of the Ivy League colleges.
During a long and fascinating account of the many problems which his tribe was having in
adjusting to our way of life, he suddenly remarked: “What would you think of a people who
www.seyfihoca.com
38
had no word for time? My people have no word for ‘late’ or for ‘waiting’, for that matter.
They don’t know what it is to wait or to be late.” He then continued, “I decided that until they
could tell the time and knew what time was they could never adjust themselves to white
culture. So I set about to teach them time. There wasn’t a clock that was running in any of the
reservation classrooms. So I first bought some decent clocks. Then I made the school buses
start on time, and if an Indian was two minutes late that was just too bad. The bus started at
eight forty-two and he had to be there.”
He was right of course. The Sioux could not adjust to European ways until they had learned
the meaning of time. The superintendent’s methods may have sounded a bit extreme, but they
were the only ones that would work. The idea of starting the buses off and making the drivers
hold to a rigid schedule was a stroke of genius; much kinder to the Indian, who could better
afford to miss a bus on the reservation than lose a job in town because he was late.
There is, in fact, no other way to teach time to people who handle it as differently from us as
the Sioux. The quickest way is to get very technical about it and to make it mean something.
Later on these people can learn the informal variations, but until they have experienced and
then mastered our type of time they will never adjust to our culture.
Thousands of miles away from the reservations of the American Indian we come to another
way of handing time which is apt to be completely unsettling to the unprepared visitor. The
inhabitants of the atoll of Truk in the Southwest Pacific treat time in a fashion that has
complicated life for themselves as well as for others, since it poses special problems not only
for their civil and military governors and the anthropologists recording their life but for their
own chiefs as well.
Time does not heal on Truk! Past events stack up, placing an ever-increasing burden on the
Trukese and weighing heavily on the present. They are, in fact, treated as though they had just
occurred. This was borne out by something which happened shortly after the American
occupation of the atoll at the end of World War II.
A villager arrived all out of breath at the military government headquarters. He said that a
murder had been committed in the village and that the murderer was running around loose.
Quite naturally the military governor became alarmed. He was about to dispatch M.P.s to
arrest the culprit when he remembered that someone had warned him about acting
precipitously when dealing with “natives.” A little enquiry turned up the fact that the victim
had been “fooling around” with the murderer’s wife. Still more enquiry of a routine type,
designed to establish the place and date of the crime, revealed that the murder had not
occurred a few hours or even days ago, as one might expect, but seventeen years before. The
murderer had been running around loose in the village all this time.
A further example of how time does not heal on Truk is that of a land dispute that started with
the German occupation in the 1890s, was carried on down through the Japanese occupation,
and was still current and acrimonious when the Americans arrived in 1946.
Prior to Missionary Moses’ arrival on Uman in 1867 life on Truk was characterized by violent
and bloody warfare. Villages, instead of being built on the shore where life was a little easier,
were placed on the sides of mountains where they could be better protected. Attacks would
come without notice and often without apparent provocation. Or a fight might start if a man
stole a coconut from a tree that was not his or waylaid a woman and took advantage of her.
www.seyfihoca.com
39
Years later someone would start thinking about the wrong and decide that it had not been
righted. A village would be attacked again in the middle of the night.
When charges were brought against a chief for things he had done to his people, every little
slight, every minor graft would be listed; nothing would be forgotten. Damages would be
asked for everything. It seemed preposterous to us Americans, particularly when we looked at
the lists of charges. “How could a chief be so corrupt?” “How could the people remember so
much?”
Though the Truk islanders carry the accumulated burden of time past on their shoulders, they
show an almost total inability to grasp the notion that two events can take place at the same
time when they are any distance apart. When the Japanese occupied Truk at the end of World
War I they took Artie Moses, chief of the island of Uman to Tokyo. Artie was made to send a
wireless message back to his people as a demonstration of the wizardry of Japanese
technology. His family refused to believe that he had sent it, that he had said anything at all,
though they knew he was in Tokyo. Places at a distance are very real to them, but people who
are away are very much away, and any interaction with them is unthinkable.
An entirely different handling of time is reported by the anthropologist Paul Bohannan for the
Tiv, a primitive people who live in Nigeria. Like the Navajo, they point to the sun to indicate
a general time of day, and they also observe the movement of the moon as it waxes and
wanes. What is different is the way they use and experience time. For the Tiv, time is like a
capsule. There is time for visiting, for cooking, or for working; and when one is in one of
those times, one does not shift to another.
The Tiv equivalent of the week lasts five to seven days. It is not tied into periodic natural
events, such as the phases of the moon. The day of the week is named after the things which
are being sold in the nearest “market.” If we had the equivalent, Monday would be
“automobiles” in Washington, D.C., “furniture” in Baltimore, and “yard goods” in New York.
Each of these might be followed by the days for appliances, liquor and diamonds in the
respective cities. This would mean that as you travelled about the day of the week would keep
changing, depending on where you were.
A requisite of our own temporal system is that the components must add up: Sixty seconds
have to equal one minute, sixty minutes one hour. The American is perplexed by people who
do not do this. The African specialist Henri Alexandre Junod, reporting on the Thonga, tells
of a medicine man who had memorized a seventy-year chronology and could detail the events
of each and every year in sequence. Yet this same man spoke of the period he had memorized
as an “era” which he computed at “four months and eight hundred years’ duration.” The usual
reaction to this story and others like it is that the man was primitive, like a child, and did not
understand what he was saying, because how could seventy years possibly be the same as
eight hundred? As students of culture we can no longer dismiss other conceptualizations of
reality by saying that they are childlike. We must go much deeper. In the case of the Thonga,
it seemed that a “chronology” is one thing and an “era” something else quite different, and
there is no relation between the two in operational terms.
If these distinctions between European-American time and other conceptions of time seem to
draw too heavily on primitive peoples, let me mention two other examples - from cultures
which are as civilized, if not as industrialized, as our own. In comparing the United States
with Iran and Afghanistan very great differences in the handling of time appear. The
www.seyfihoca.com
40
American attitude toward appointments is an example. Once while in Tehran I had the
opportunity to observe some young Iranians making plans for a party. After plans were made
to pick up everyone at appointed times and places everything began to fall apart. People
would leave messages that they were unable to take so-and-so or were going somewhere else,
knowing full well that the person who had been given the message couldn’t possibly deliver
it. One girl was left stranded on a street corner, and no one seemed to be concerned about it.
One of my informants explained that he himself had had many similar experiences. Once he
had made eleven appointments to meet a friend. Each time one of them had failed to show up.
The twelfth time they swore that they would both be there, that nothing would interfere. The
friend failed to arrive. After waiting for forty-five minutes my informant phoned his friend
and found him still at home. The following conversation is an approximation of what took
place:
"Is that you, Abdul?” “Yes.” “Why aren’t you here? I thought we were to meet for sure.” “Oh,
but it was raining,” said Abdul with a sort of whining intonation that is very common in Parsi.
If present appointments are treated rather cavalierly, the past in Iran takes on a very great
importance. People look back on what they feel are the wonders of the past and the great ages
of Persian culture. Yet the future seems to have very little reality or certainty to it.
Businessmen have been known to invest hundreds of thousands of dollars in factories of
various sorts without making the slightest plan as to how to use them. A complete woollen
mill was bought and shipped to Tehran before the buyer had raised enough money to erect it,
to buy supplies, or even to train personnel. When American teams of technicians came to help
Iran’s economy they constantly had to cope with what seemed to them to be an almost total
lack of planning.
Moving east from Iran to Afghanistan, one gets farther afield from American time concepts.
A few years ago in Kabul a man appeared, looking for his brother. He asked all the merchants
of the market place if they had seen his brother and told him where he was staying in case his
brother arrived and wanted to find him. The next year he was back and repeated the
performance. By this time one of the members of the American embassy had heard about his
inquiries and asked if he had found his brother. The man answered that he and his brother had
agreed to meet in Kabul, but neither of them had said what year.
Strange as some of these stories about the ways in which people handle time may seem, they
become understandable when correctly analysed. To do this adequately requires an adequate
theory of culture. Before we return to the subject of time again - in a much later chapter of
this book - I hope that I will have provided just such a theory. It will not only shed light on the
way time is meshed with many other aspects of society but will provide a key to unlock some
of the secrets of the eloquent language of culture which speaks in so many different ways.
(Edward T. Hall: The Silent Language published by Doubleday & Company, New York in
1959)
BIOLOGY
Evolution and Natural Selection
www.seyfihoca.com
41
The idea of evolution was known to some of the Greek philosophers. By the time of Aristotle,
speculation had suggested that more perfect types had not only followed less perfect ones but
actually had developed from them. But all this was guessing; no real evidence was
forthcoming. When, in modern times, the idea of evolution was revived, it appeared in the
writings of the philosophers-Bacon, Descartes, Leibniz and Kant. Herbert Spencer was
preaching a full evolutionary doctrine in the years just before Darwin's book was published,
while most naturalists would have none of it. Nevertheless a few biologists ran counter to the
prevailing view, and pointed to such facts as the essential unity of structure in all warmblooded animals.
The first complete theory was that of Lamarck (1744-1829), who thought that modifications
due to environment, if constant and lasting, would be inherited and produce a new type.
Though no evidence for such inheritance was available, the theory gave a working hypothesis
for naturalists to use, and many of the social and philanthropic efforts of the nineteenth
century were framed on the tacit assumption that acquired improvements would be inherited.
But the man whose book gave both Darwin and Wallace the clue was the Reverend Robert
Malthus (1766-1834), sometime curate of Albury in Surrey. The English people were
increasing rapidly, and Malthus argued that the human race tends to outrun its means of
subsistence unless the redundant individuals are eliminated. This may not always be true, but
Darwin writes:
In October 1838, I happened to read for amusement Malthus on Population, and being well
prepared to appreciate the struggle for existence which everywhere goes on, from long
continued observation of the habits of animals and plants, it at once struck me that, under
these circumstances, favourable variations would tend to be preserved, and unfavourable ones
to be destroyed. The result of this would be the formation of new species. Here then I had a
theory by which to work.
Darwin spent twenty years collecting countless facts and making experiments on breeding and
variation in plants and animals. By 1844 he had convinced himself that species are not
immutable, but worked on to get further evidence. On 18 June 1858 he received from Alfred
Russell Wallace a paper written in Ternate, in the space of three days after reading Malthus's
book. Darwin saw at once that Wallace had hit upon the essence of his own theory. Lyell and
Hooker arranged with the Linnaean Society to read on July 1st 1858 Wallace's paper together
with a letter from Darwin and an abstract of his theory written in 1844, Then Darwin wrote
out an account of his labours, and on 24th November 1859 published his great book The
Origin of Species.
In any race of plants or animals, the individuals differ from each other in innate qualities.
Darwin offered no explanation of these variations, but merely accepted their existence. When
the pressure of numbers or the competition for mates is great, any variation in structure which
is of use in the struggle has 'survival value', and gives its possessor an improved chance of
prolonging life and leaving offspring. That variation therefore tends to spread through the race
by the elimination of those who do not possess it, and a new variety or even species may be
established. As Huxley said, this idea was wholly unknown till 1858. Huxley said the book
was like a flash of lightning in the darkness. He wrote
It did the immense service of freeing us from the dilemma - Refuse to accept the Creation
hypothesis, and what have you to propose that can be accepted by any cautious reasoner? In
www.seyfihoca.com
42
1857 I had no answer ready, and I do not think anyone else had. A year later we reproached
ourselves with dullness for being perplexed with such an enquiry. My reflection when I first
made myself master of the central idea of the Origin was 'How extremely stupid not to have
thought of that!'
The hypothesis of natural selection may not be a complete explanation, but it led to a greater
thing than itself-an acceptance of the theory of organic evolution, which the years have but
confirmed. Yet at first some naturalists joined the opposition. To the many, who were unable
to judge the biological evidence, the effect of the theory of evolution seemed incredible as
well as devastating, to run counter to common sense and to overwhelm all philosophic and
religious landmarks. Even educated man, choosing between the Book of Genesis and the
Origin of Species, proclaimed with Disraeli that he was 'on the side of the Angels'.
Darwin himself took a modest view. While thinking that natural selection was the chief cause
of evolution, he did not exclude Lamarck's idea that characters acquired by long use or disuse
might be inherited, though no evidence seemed to be forthcoming. But about 1890 Weismann
drew a sharp distinction between the body (or soma) and the germ cells which it contains.
Somatic cells can only reproduce cells like themselves, but germ cells give rise not only to the
germ cells of a new individual but to all the many types of cell in his body. Germ cells
descend from germ cells in a pure line of germ plasm, but somatic cells trace their origin to
germ cells. From this point of view, the body of each individual is an unimportant by-product
of his parents' germ cells. The body dies, leaving no offspring, but the germ plasms show an
unbroken continuity. The products of the germ cells are not likely to be affected by changes in
the body. So Weismann's doctrine offered an explanation of the apparent noninheritance of
acquired characters.
The supporters of pure Darwinism came to regard the minute variations as enough to explain
natural selection and natural selection enough to explain evolution. But animal breeders and
horticulturalists knew that sudden large mutations occur, especially after crossing, and that
new varieties might be established at once. Then in 1900 forgotten work by Mendel was
rediscovered and a new chapter opened.
In 1869 Darwin's cousin, Francis Galton, applied these principles to mental qualities. By
searching books of reference, Galton examined the inheritance of ability. For instance, he
found that the chance of the son of a judge showing great ability was about 500 times as high
as that of a man taken at random, and for the judge's father it was nearly as much. While no
prediction can be made about individuals, on the average of large numbers, the inheritance of
ability is certain.
(From Chapter VIII of A Shorter History of Science by Sir W. C. Dampier.)
Evolution and Natural Selection
The idea of evolution was known to some of the Greek philosophers. By the time of Aristotle,
speculation had suggested that more perfect types had not only followed less perfect ones but
actually had developed from them. But all this was guessing; no real evidence was
forthcoming. When, in modern times, the idea of evolution was revived, it appeared in the
writings of the philosophers-Bacon, Descartes, Leibniz and Kant. Herbert Spencer was
preaching a full evolutionary doctrine in the years just before Darwin's book was published,
while most naturalists would have none of it. Nevertheless a few biologists ran counter to the
www.seyfihoca.com
43
prevailing view, and pointed to such facts as the essential unity of structure in all warmblooded animals.
The first complete theory was that of Lamarck (1744-1829), who thought that modifications
due to environment, if constant and lasting, would be inherited and produce a new type.
Though no evidence for such inheritance was available, the theory gave a working hypothesis
for naturalists to use, and many of the social and philanthropic efforts of the nineteenth
century were framed on the tacit assumption that acquired improvements would be inherited.
But the man whose book gave both Darwin and Wallace the clue was the Reverend Robert
Malthus (1766-1834), sometime curate of Albury in Surrey. The English people were
increasing rapidly, and Malthus argued that the human race tends to outrun its means of
subsistence unless the redundant individuals are eliminated. This may not always be true, but
Darwin writes:
In October 1838, I happened to read for amusement Malthus on Population, and being well
prepared to appreciate the struggle for existence which everywhere goes on, from long
continued observation of the habits of animals and plants, it at once struck me that, under
these circumstances, favourable variations would tend to be preserved, and unfavourable ones
to be destroyed. The result of this would be the formation of new species. Here then I had a
theory by which to work.
Darwin spent twenty years collecting countless facts and making experiments on breeding and
variation in plants and animals. By 1844 he had convinced himself that species are not
immutable, but worked on to get further evidence. On 18 June 1858 he received from Alfred
Russell Wallace a paper written in Ternate, in the space of three days after reading Malthus's
book. Darwin saw at once that Wallace had hit upon the essence of his own theory. Lyell and
Hooker arranged with the Linnaean Society to read on July 1st 1858 Wallace's paper together
with a letter from Darwin and an abstract of his theory written in 1844, Then Darwin wrote
out an account of his labours, and on 24th November 1859 published his great book The
Origin of Species.
In any race of plants or animals, the individuals differ from each other in innate qualities.
Darwin offered no explanation of these variations, but merely accepted their existence. When
the pressure of numbers or the competition for mates is great, any variation in structure which
is of use in the struggle has 'survival value', and gives its possessor an improved chance of
prolonging life and leaving offspring. That variation therefore tends to spread through the race
by the elimination of those who do not possess it, and a new variety or even species may be
established. As Huxley said, this idea was wholly unknown till 1858. Huxley said the book
was like a flash of lightning in the darkness. He wrote
It did the immense service of freeing us from the dilemma - Refuse to accept the Creation
hypothesis, and what have you to propose that can be accepted by any cautious reasoner? In
1857 I had no answer ready, and I do not think anyone else had. A year later we reproached
ourselves with dullness for being perplexed with such an enquiry. My reflection when I first
made myself master of the central idea of the Origin was 'How extremely stupid not to have
thought of that!'
The hypothesis of natural selection may not be a complete explanation, but it led to a greater
thing than itself-an acceptance of the theory of organic evolution, which the years have but
www.seyfihoca.com
44
confirmed. Yet at first some naturalists joined the opposition. To the many, who were unable
to judge the biological evidence, the effect of the theory of evolution seemed incredible as
well as devastating, to run counter to common sense and to overwhelm all philosophic and
religious landmarks. Even educated man, choosing between the Book of Genesis and the
Origin of Species, proclaimed with Disraeli that he was 'on the side of the Angels'.
Darwin himself took a modest view. While thinking that natural selection was the chief cause
of evolution, he did not exclude Lamarck's idea that characters acquired by long use or disuse
might be inherited, though no evidence seemed to be forthcoming. But about 1890 Weismann
drew a sharp distinction between the body (or soma) and the germ cells which it contains.
Somatic cells can only reproduce cells like themselves, but germ cells give rise not only to the
germ cells of a new individual but to all the many types of cell in his body. Germ cells
descend from germ cells in a pure line of germ plasm, but somatic cells trace their origin to
germ cells. From this point of view, the body of each individual is an unimportant by-product
of his parents' germ cells. The body dies, leaving no offspring, but the germ plasms show an
unbroken continuity. The products of the germ cells are not likely to be affected by changes in
the body. So Weismann's doctrine offered an explanation of the apparent noninheritance of
acquired characters.
The supporters of pure Darwinism came to regard the minute variations as enough to explain
natural selection and natural selection enough to explain evolution. But animal breeders and
horticulturalists knew that sudden large mutations occur, especially after crossing, and that
new varieties might be established at once. Then in 1900 forgotten work by Mendel was
rediscovered and a new chapter opened.
In 1869 Darwin's cousin, Francis Galton, applied these principles to mental qualities. By
searching books of reference, Galton examined the inheritance of ability. For instance, he
found that the chance of the son of a judge showing great ability was about 500 times as high
as that of a man taken at random, and for the judge's father it was nearly as much. While no
prediction can be made about individuals, on the average of large numbers, the inheritance of
ability is certain.
(From Chapter VIII of A Shorter History of Science by Sir W. C. Dampier.)
Banting and the Discovery of Insulin
While at the Medical School, Banting went into the library and looked at the November issue
of Surgery, Gynaecology and Obstetrics. The first article was entitled 'The relation of the
Islets of Langerhans to Diabetes' by Dr. Moses Barron of Minneapolis. Banting had to talk to
his students next day on the functions of the pancreas, so he took the journal home with him.
One paragraph in Barron's review of previous literature on the subject referred to the
experiments on tying the pancreatic ducts of rabbits made by Arnozen and Vaillard thirty-six
years earlier. Banting had not heard of these experiments before, but he knew that attempts to
treat diabetes with extracts of the pancreas had failed; and he wondered why.
A possible answer that occurred to him was that the hormone from the islets of Langerhans
was destroyed during the extraction of the pancreas. The question then was what might
destroy it; and his thoughts turned to the digestive ferment that the pancreas produced. He
knew this was very powerful, so powerful that it could break up and dissolve all sorts of
www.seyfihoca.com
45
protein foods including the toughest meats. Perhaps, during the process of extraction, this
ferment destroyed the vital hormone.
If that were so, Banting reasoned, the extraction ought to be delayed until the pancreas was no
longer producing this ferment. According to the experiments of Arnozen and Vaillard, this
condition could be reached by tying the pancreatic ducts. It was two o'clock in the morning of
October 31, 1920, when he wrote in his small black notebook: 'Tie off pancreas ducts of dogs.
Wait six or eight weeks. Remove and extract.'
Although he did not know it, this was much the same idea that had come to Lydia de Witt
fourteen years earlier. But it was not for the idea alone that Banting deserves to be
remembered; his greatness lay in the way he put it into practice. He had to wait until the
spring of 1921 before he could start work, and he filled in the time by reading all the literature
on the subject he could find. He still missed Lydia de Witt's work. At last he was given his
laboratory-small, hot and rather primitive-and his ten dogs. His assistant, Charles Best, was a
recent graduate in physiology and biochemistry who had been working under Macleod on
sugars. They began work on May 16.
Banting began by tying off the pancreatic ducts of a number of dogs, which was quite easy.
Then he had to remove the pancreas from other dogs to give them diabetes. The operation was
not easy, and Banting's training and ability as a surgeon proved invaluable. Even so, several
dogs died before he evolved a suitable technique.
On July 6 Banting and Best chloroformed two of the dogs whose pancreatic ducts they had
tied seven weeks earlier, and were disappointed to find that the pancreas had not degenerated
as they had hoped. They had not tied the ducts with the correct degree of tension needed-the
margin of error was very small. And they had only one week left to complete their work.
Macleod was away in Europe, but an extension was granted by the authorities, and the
experiment was continued. On July 27 another duct-tied dog was chloroformed, and when
Banting operated he found that the pancreas had shrivelled to about one-third of its original
size. It was removed, chopped into pieces and mixed with saline; and a small amount of a
filtered extract was injected into one of the diabetic dogs. Within two hours its blood sugar
had fallen considerably, and before long the dog became conscious, rose, and wagged its tail.
The effect of the injection was so dramatic that Banting and Best could hardly believe it; but
further experiments made them sure that they had indeed found what they were looking for.
They had succeeded in extracting the anti-diabetic hormone secreted by the islets of
Langerhans. They called it 'isletin'. It was some time later that Macleod renamed it insulin, a
word that had been suggested in 1910. Insulin did not cure diabetes. After a while the dog
relapsed, and further injections were needed to revive it again. But with regular injections of
insulin a dog with diabetes could live.
Banting and Best next succeeded in obtaining insulin by injecting secretin to stimulate the
production of the digestive ferment from the pancreas and exhaust the cells from which it
came. This was a much quicker method than tying the ducts and waiting several weeks; and
although the practical results were disappointing, its importance to the theory was
considerable.
So far insulin had been extracted only in sufficient quantity for laboratory work, and already
Banting and Best were seeking means of getting larger supplies. They now obtained insulin
www.seyfihoca.com
46
from the pancreas of a foetal calf-that is, a calf that had not yet been born. Nature, ever
practical, does not supply digestive ferments until a calf starts eating, so there was nothing to
destroy the insulin during extraction. This new success enabled Banting and Best to keep up
an adequate supply of insulin for more extensive experiments. At the same time they realized
that if their work was to have practical results in medical treatment it would be necessary to
get much larger supplies. And they could only come from adult cattle in the slaughterhouse.
The problem was to find a means of extracting the insulin from the pancreas of an ordinary
adult animal.
The problem was solved well enough to provide insulin for the first injections on human
beings. Two patients in Toronto General Hospital were chosen-a fourteen-year-old boy and a
doctor, both very seriously ill with diabetes: 'hopeless' cases. When treated with insulinalthough still in a relatively impure form-they improved at once. The boy is alive and well today.
'Research in medicine is specialized,' Banting said later, 'and as in all organized walks of life,
a division of labour is necessary. In consequence, a division of labour in the field of insulin
took place.' Professor J. B. Collip, a biochemist, was called in to produce a purer extract. He
succeeded very quickly; and other workers made it possible to obtain insulin on a really large
scale. Before very long insulin injections became the standard treatment for diabetes all over
the world. They still are today.
Banting, only thirty-one years old, was suddenly famous. Although for some extraordinary
reason he was not knighted until 1934, he was awarded the Nobel Prize for Medicine in
1923jointly with Macleod.
(Fom Chapter VIII of Great Discoveries in Modern Science by Patrick Pringle.)
The Galapagos Islands
After Tahiti the Galapagos were the most famous of all the tropical islands in the Pacific.
They had been discovered in 1535 by Fray Tomas de Berlanga, Bishop of Panama, and were
now owned by Ecuador, 500 odd miles away. Already in the 1830s some sixty or seventy
whalers, mostly American, called there every year for 'refreshments', They replenished their
water tanks from the springs, they captured tortoises for meat, (galapagos is the Spanish word
for giant tortoises), and they called for mail at Post Office Bay where a box was set up on the
beach. Every whaling captain took from it any letters which he thought he might be able to
forward. Herman Melville called in at the Galapagos aboard the Acushnet not long after the
Beagle's visit, and the 'blighted Encantadas' are a part of the saga of the white whale. 'Little
but reptile life is here found', wrote Melville, 'the chief sound of life is a hiss'.
Apart from their practical uses there was nothing much to recommend the Galapagos; they
were not lush and beautiful islands like the Tahiti group, they were (and still are) far off the
usual maritime routes, circled by capricious currents, and nobody lived in them then except
for a handful of political prisoners who had been stranded there by the Ecuador government.
The fame of the islands was founded upon one thing; they were infinitely strange, unlike any
other islands in the world. No one who went there ever forgot theta. For the Beagle this was
just another port of call in a very long voyage, but for Darwin it was much more than that, for
it was here, in the most unexpected way-just as a man might have a sudden inspiration while
he is travelling in a car or a train-that he began to form a coherent view of the evolution of life
www.seyfihoca.com
47
on this planet. To put it into his own words: 'Here, both in space and time, we seem to be
brought somewhat near to that great fact-that mystery of mysteries-the first appearance of new
beings on this earth'.
The Beagle cruised for just over a month in the Galapagos, and whenever they reached an
interesting point FitzRoy dropped off a boatload of men to explore. On Narborough Island the
turtles were coming in at night to lay their eggs in the sand, thousands of them; they laid six
eggs in each hole. On Charles Island there was a penal settlement of two hundred convicts,
who cultivated sugar-cane, bananas and corn on the high ground. But the group that concerns
us is the one that was put ashore on James Island. Here Darwin, Covington, Bynoe and two
sailors were landed with a tent and provisions, and FitzRoy promised to come back and pick
them up at the end of a week. Darwin visited other islands as well, but they did not differ very
much from James Island, and so we can conveniently group all his experiences into this one
extraordinary week. They set up their tent on the beach, laid out their bedding and their stores,
and then began to look around them.
The marine lizards, on closer inspection, turned out to be miniature dragons, several feet in
length, and they had great gaping mouths with pouches under them and long flat tails; 'imps
of darkness', Darwin called them. They swarmed in thousands; everywhere Darwin went they
scuttled away before him, and they were even blacker than the forbidding black rocks on
which they lived. Everything about these iguanas was odd. They never went more than ten
yards inland; either they sunned themselves on the shore or dived into the sea where at once
they became expert swimmers, holding their webbed feet close to their sides and propelling
themselves along with strong swift strokes of their tails. Through the clear water one could
see them cruising close to the bottom, and they could stay submerged for a very long time; a
sailor threw one into the sea with a heavy weight attached to it, and when he fished it up an
hour later it was still alive and kicking. They fed on seaweed, a fact that Darwin and Bynoe
ascertained when with Bynoe's surgical instruments they opened one up and examined the
contents of its stomach. And yet, like some sailors, these marine beasts hated the sea. Darwin
took one by the tail and hurled it into a big pool that had been left in the rocks by the ebb-tide.
At once it swam back to the land. Again Darwin caught it and threw it back, and again it
returned. No matter what he did the animal simply would not stay in the sea, and Darwin was
forced to conclude that it feared the sharks there and instinctively, when threatened by
anything, came ashore where it had no enemies. Their breeding season was November, when
they put on their courting colours and surrounded themselves with their harems.
The other creatures on the coast were also strange in different ways; flightless cormorants,
penguins and seals, both cold-sea creatures, unpredictably living here in these tropical waters,
and a scarlet crab that scuttled over the lizards' backs, hunting for ticks. Walking inland with
Covington, Darwin arrived among some scattered cactuses, and here two enormous tortoises
were feeding. They were quite deaf and did not notice the two men until they had drawn level
with their eyes. Then they hissed loudly and drew in their heads. These animals were so big
and heavy that it was impossible to lift them or even turn them over on their sides Darwin and
Covington tried-and they could easily bear the weight of a man. Darwin got aboard and found
it a very wobbly seat, but he in no way impeded the tortoise's progress; he calculated that it
managed 60 yards in ten minutes, or 360 yards an hour, which would be roughly four miles a
day - 'allowing a little time for it to eat on the road'.
www.seyfihoca.com
48
The origin of species
INTRODUCTION
When on board H.M.S. Beagle as naturalist, I was much struck with certain facts in the
distribution of the organic beings inhabiting South America, and in the geological relations of
the present to the past inhabitants of that continent. These facts, as will be seen in the latter
chapters of this volume, seemed to throw some light on the origin of species- that mystery of
mysteries, as it has been called by one of our greatest philosophers. On my return home, it
occurred to me, in 1837, that something might perhaps be made out on this question by
patiently accumulating and reflecting on all sorts of facts which could possibly have any
bearing on it. After five years' work I allowed myself to speculate on the subject, and drew up
some short notes; these I enlarged in 1844 into a sketch of the conclusions, which then
seemed to me probable: from that period to the present day I have steadily pursued the same
object. I hope that I may be excused for entering on these personal details, as I give them to
show that I have not been hasty in coming to a decision.
My work is now (1859) nearly finished; but as it will take me many more years to complete it,
and as my health is far from strong, I have been urged to publish this abstract. I have more
especially been induced to do this, as Mr. Wallace, who is now studying the natural history of
the Malay Archipelago, has arrived at almost exactly the same general conclusions that I have
on the origin of species. In 1858 he sent me a memoir on this subject, with a request that I
would forward it to Sir Charles Lyell, who sent it to the Linnean Society, and it is published
in the third volume of the Journal of that society. Sir C. Lyell and Dr. Hooker, who both knew
of my work- the latter having read my sketch of 1844- honoured me by thinking it advisable
to publish, with Mr. Wallace's excellent memoir, some brief extracts from my manuscripts.
This abstract, which I now publish, must necessarily be imperfect. cannot here give references
and authorities for my several statements; and I must trust to the reader reposing some
confidence in my accuracy. No doubt errors will have crept in, though I hope I have always
been cautious in trusting to good authorities alone. I can here give only the general
conclusions at which I have arrived, with a few facts in illustration, but which, I hope, in most
cases will suffice. No one can feel more sensible than I do of the necessity of hereafter
publishing in detail all the facts, with references, on which my conclusions have been
grounded; and I hope in a future work to do this. For I am well aware that scarcely a single
point is discussed in this volume on which facts cannot be adduced, often apparently leading
to conclusions directly opposite to those at which I have arrived. A fair result can be obtained
only by fully stating and balancing the facts and arguments on both sides of each question;
and this is here impossible.
I much regret that want of space prevents my having the satisfaction of acknowledging the
generous assistance which I have received from very many naturalists, some of them
personally unknown to me. I cannot, however, let this opportunity pass without expressing my
deep obligations to Dr. Hooker, who, for the last fifteen years, has aided me in every possible
way by his large stores of knowledge and his excellent judgment.
In considering the Origin of Species, it is quite conceivable that a naturalist, reflecting on the
mutual affinities of organic beings, on their embryological relations, their geographical
www.seyfihoca.com
49
distribution, geological succession, and other such facts, might come to the conclusion that
species had not been independently created, but had descended, like varieties, from other
species. Nevertheless, such a conclusion, even if well founded, would be unsatisfactory, until
it could be shown how the innumerable species inhabiting this world have been modified, so
as to acquire that perfection of structure and coadaptation which justly excites our admiration.
Naturalists continually refer to external conditions, such as climate, food, &c., as the only
possible cause of variation. In one limited sense, as we shall hereafter see, this may be true;
but it is preposterous to attribute to mere external conditions, the structure, for instance, of the
woodpecker, with its feet, tail, beak, and tongue, so admirably adapted to catch insects under
the bark of trees. In the case of the mistletoe, which draws its nourishment from certain trees,
which has seeds that must be transported by certain birds, and which has flowers with
separate sexes absolutely requiring the agency of certain insects to bring pollen from one
flower to the other, it is equally preposterous to account for the structure of this parasite, with
its relations to several distinct organic beings, by the effects of external conditions, or of
habit, or of the volition of the plant itself.
It is, therefore, of the highest importance to gain a clear insight into the means of modification
and coadaptation. At the commencement of my observations it seemed to me probable that a
careful study of domesticated animals and of cultivated plants would offer the best chance of
making out this obscure problem. Nor have I been disappointed; in this and in all other
perplexing cases I have invariably found that our knowledge, imperfect though it be, of
variation under domestication, afforded the best and safest clue. I may venture to express my
conviction of the high value of such studies, although they have been very commonly
neglected by naturalists.
From these considerations, I shall devote the first chapter of this Abstract to Variation under
Domestication. We shall thus see that a large amount of hereditary modification is at least
possible; and, what is equally or more important, we shall see how great is the power of man
in accumulating by his Selection successive slight variations. I will then pass on to the
variability of species in a state of nature; but I shall, unfortunately, be compelled to treat this
subject far too briefly, as it can be treated properly only by giving long catalogues of facts.
We shall, however, be enabled to discuss what circumstances are most favourable to
variation. In the next chapter the Struggle for Existence amongst all organic beings
throughout the world, which inevitably follows from the high geometrical ratio of their
increase, will be considered. This is the doctrine of Malthus, applied to the whole animal and
vegetable kingdoms. As many more individuals of each species are born than can possibly
survive; and as, consequently, there is a frequently recurring struggle for existence, it follows
that any being, if it vary however slightly in any manner profitable to itself, under the
complex and sometimes varying conditions of life, will have a better chance of surviving, and
thus be naturally selected. From the strong principle of inheritance, any selected variety will
tend to propagate its new and modified form.
This fundamental subject of Natural Selection will be treated at some length in the fourth
chapter; and we shall then see how Natural Selection almost inevitably causes much
Extinction of the less improved forms of life, and leads to what I have called Divergence of
Character. In the next chapter I shall discuss the complex and little known laws of variation.
In the five succeeding chapters, the most apparent and gravest difficulties in accepting the
theory will be given: namely, first, the difficulties of transitions, or how a simple being or a
simple organ can be changed and perfected into a highly developed being or into an
elaborately constructed organ; secondly, the subject of Instinct, or the mental powers of
www.seyfihoca.com
50
animals; thirdly, Hybridism, or the infertility of species and the fertility of varieties when
intercrossed; and fourthly, the imperfection of the Geological Record. In the next chapter I
shall consider the geological succession of organic beings throughout time; in the twelfth and
thirteenth, their geographical distribution throughout space; in the fourteenth, their
classification or mutual affinities, both when mature and in an embryonic condition. In the
last chapter I shall give a brief recapitulation of the whole work, and a few concluding
remarks.
No one ought to feel surprise at much remaining as yet unexplained in regard to the origin of
species and varieties, if he make due allowance for our profound ignorance in regard to the
mutual relations of the many beings which live around us. Who can explain why one species
ranges widely and is very numerous, and why another allied species has a narrow range and is
rare? Yet these relations are of the highest importance, for they determine the present welfare
and, as I believe, the future success and modification of every inhabitant of this world. Still
less do we know of the mutual relations of the innumerable inhabitants of the world during
the many past geological epochs in its history. Although much remains obscure, and will long
remain obscure, I can entertain no doubt, after the most deliberate study and dispassionate
judgment of which I am capable, that the view which most naturalists until recently
entertained, and which I formerly entertained - namely, that each species has been
independently created- is erroneous. I am fully convinced that species are not immutable; but
that those belonging to what are called the same genera are lineal descendants of some other
and generally extinct species, in the same manner as the acknowledged varieties of any one
species are the descendants of that species. Furthermore, I am convinced that Natural
Selection has been the most important, but not the exclusive, means of modification.
(Introduction to On The Origin of Species by Charles Darwin, 1859)
INTRODUCTION
When on board H.M.S. Beagle as naturalist, I was much struck with certain facts in the
distribution of the organic beings inhabiting South America, and in the geological relations of
the present to the past inhabitants of that continent. These facts, as will be seen in the latter
chapters of this volume, seemed to throw some light on the origin of species- that mystery of
mysteries, as it has been called by one of our greatest philosophers. On my return home, it
occurred to me, in 1837, that something might perhaps be made out on this question by
patiently accumulating and reflecting on all sorts of facts which could possibly have any
bearing on it. After five years' work I allowed myself to speculate on the subject, and drew up
some short notes; these I enlarged in 1844 into a sketch of the conclusions, which then
seemed to me probable: from that period to the present day I have steadily pursued the same
object. I hope that I may be excused for entering on these personal details, as I give them to
show that I have not been hasty in coming to a decision.
My work is now (1859) nearly finished; but as it will take me many more years to complete it,
and as my health is far from strong, I have been urged to publish this abstract. I have more
especially been induced to do this, as Mr. Wallace, who is now studying the natural history of
the Malay Archipelago, has arrived at almost exactly the same general conclusions that I have
on the origin of species. In 1858 he sent me a memoir on this subject, with a request that I
would forward it to Sir Charles Lyell, who sent it to the Linnean Society, and it is published
in the third volume of the Journal of that society. Sir C. Lyell and Dr. Hooker, who both knew
www.seyfihoca.com
51
of my work- the latter having read my sketch of 1844- honoured me by thinking it advisable
to publish, with Mr. Wallace's excellent memoir, some brief extracts from my manuscripts.
This abstract, which I now publish, must necessarily be imperfect. cannot here give references
and authorities for my several statements; and I must trust to the reader reposing some
confidence in my accuracy. No doubt errors will have crept in, though I hope I have always
been cautious in trusting to good authorities alone. I can here give only the general
conclusions at which I have arrived, with a few facts in illustration, but which, I hope, in most
cases will suffice. No one can feel more sensible than I do of the necessity of hereafter
publishing in detail all the facts, with references, on which my conclusions have been
grounded; and I hope in a future work to do this. For I am well aware that scarcely a single
point is discussed in this volume on which facts cannot be adduced, often apparently leading
to conclusions directly opposite to those at which I have arrived. A fair result can be obtained
only by fully stating and balancing the facts and arguments on both sides of each question;
and this is here impossible.
I much regret that want of space prevents my having the satisfaction of acknowledging the
generous assistance which I have received from very many naturalists, some of them
personally unknown to me. I cannot, however, let this opportunity pass without expressing my
deep obligations to Dr. Hooker, who, for the last fifteen years, has aided me in every possible
way by his large stores of knowledge and his excellent judgment.
In considering the Origin of Species, it is quite conceivable that a naturalist, reflecting on the
mutual affinities of organic beings, on their embryological relations, their geographical
distribution, geological succession, and other such facts, might come to the conclusion that
species had not been independently created, but had descended, like varieties, from other
species. Nevertheless, such a conclusion, even if well founded, would be unsatisfactory, until
it could be shown how the innumerable species inhabiting this world have been modified, so
as to acquire that perfection of structure and coadaptation which justly excites our admiration.
Naturalists continually refer to external conditions, such as climate, food, &c., as the only
possible cause of variation. In one limited sense, as we shall hereafter see, this may be true;
but it is preposterous to attribute to mere external conditions, the structure, for instance, of the
woodpecker, with its feet, tail, beak, and tongue, so admirably adapted to catch insects under
the bark of trees. In the case of the mistletoe, which draws its nourishment from certain trees,
which has seeds that must be transported by certain birds, and which has flowers with
separate sexes absolutely requiring the agency of certain insects to bring pollen from one
flower to the other, it is equally preposterous to account for the structure of this parasite, with
its relations to several distinct organic beings, by the effects of external conditions, or of
habit, or of the volition of the plant itself.
It is, therefore, of the highest importance to gain a clear insight into the means of modification
and coadaptation. At the commencement of my observations it seemed to me probable that a
careful study of domesticated animals and of cultivated plants would offer the best chance of
making out this obscure problem. Nor have I been disappointed; in this and in all other
perplexing cases I have invariably found that our knowledge, imperfect though it be, of
variation under domestication, afforded the best and safest clue. I may venture to express my
conviction of the high value of such studies, although they have been very commonly
neglected by naturalists.
www.seyfihoca.com
52
From these considerations, I shall devote the first chapter of this Abstract to Variation under
Domestication. We shall thus see that a large amount of hereditary modification is at least
possible; and, what is equally or more important, we shall see how great is the power of man
in accumulating by his Selection successive slight variations. I will then pass on to the
variability of species in a state of nature; but I shall, unfortunately, be compelled to treat this
subject far too briefly, as it can be treated properly only by giving long catalogues of facts.
We shall, however, be enabled to discuss what circumstances are most favourable to
variation. In the next chapter the Struggle for Existence amongst all organic beings
throughout the world, which inevitably follows from the high geometrical ratio of their
increase, will be considered. This is the doctrine of Malthus, applied to the whole animal and
vegetable kingdoms. As many more individuals of each species are born than can possibly
survive; and as, consequently, there is a frequently recurring struggle for existence, it follows
that any being, if it vary however slightly in any manner profitable to itself, under the
complex and sometimes varying conditions of life, will have a better chance of surviving, and
thus be naturally selected. From the strong principle of inheritance, any selected variety will
tend to propagate its new and modified form.
This fundamental subject of Natural Selection will be treated at some length in the fourth
chapter; and we shall then see how Natural Selection almost inevitably causes much
Extinction of the less improved forms of life, and leads to what I have called Divergence of
Character. In the next chapter I shall discuss the complex and little known laws of variation.
In the five succeeding chapters, the most apparent and gravest difficulties in accepting the
theory will be given: namely, first, the difficulties of transitions, or how a simple being or a
simple organ can be changed and perfected into a highly developed being or into an
elaborately constructed organ; secondly, the subject of Instinct, or the mental powers of
animals; thirdly, Hybridism, or the infertility of species and the fertility of varieties when
intercrossed; and fourthly, the imperfection of the Geological Record. In the next chapter I
shall consider the geological succession of organic beings throughout time; in the twelfth and
thirteenth, their geographical distribution throughout space; in the fourteenth, their
classification or mutual affinities, both when mature and in an embryonic condition. In the
last chapter I shall give a brief recapitulation of the whole work, and a few concluding
remarks.
No one ought to feel surprise at much remaining as yet unexplained in regard to the origin of
species and varieties, if he make due allowance for our profound ignorance in regard to the
mutual relations of the many beings which live around us. Who can explain why one species
ranges widely and is very numerous, and why another allied species has a narrow range and is
rare? Yet these relations are of the highest importance, for they determine the present welfare
and, as I believe, the future success and modification of every inhabitant of this world. Still
less do we know of the mutual relations of the innumerable inhabitants of the world during
the many past geological epochs in its history. Although much remains obscure, and will long
remain obscure, I can entertain no doubt, after the most deliberate study and dispassionate
judgment of which I am capable, that the view which most naturalists until recently
entertained, and which I formerly entertained - namely, that each species has been
independently created- is erroneous. I am fully convinced that species are not immutable; but
that those belonging to what are called the same genera are lineal descendants of some other
and generally extinct species, in the same manner as the acknowledged varieties of any one
species are the descendants of that species. Furthermore, I am convinced that Natural
Selection has been the most important, but not the exclusive, means of modification.
www.seyfihoca.com
53
(Introduction to On The Origin of Species by Charles Darwin, 1859)
BUSINESS
Brands Up.
New Labour has proved a marketable package, but it may be that Tony Blair and his cabinet
colleagues should now go the whole hog and reinvent themselves as individual brands. A
survey recently found that consumers consider Heineken more trustworthy than the Prime
Minister, and brands like BT and BMW are better known than Gordon Brown and Jack Straw.
You know where you are with a brand. Invented in the 19th century to reassure consumers
they were getting the real McCoy, brands have long been the way shoppers navigate in a sea
of unknowns. They are beacons of consistency, badges of style. You know what you're
getting when you buy Marmite, Tennents lager or PG Tips. Or do you? Most of our bestknown brands have become baubles for multinationals. Brands are revenue streams, assets
which are made to sweat and sold between international corporations for millions. Marmite
may have been a national institution since 1902 when it was invented as a product of the
brewing industry in Burton on Trent. But it's now owned by the New Jersey based American
giant, CPC International - as are Bovril, Pot Noodles and Hellman's Mayonnaise. Lea &
Perrins is part of the French company, Danone, and PG Tips isn't owned by Brooke Bond; it's
part of the gigantic portfolio of margarine baron Unilever.
You think brands deliver consistency? Think again. Persil is owned he by Unilever in Britain
and by Henkel in Germany. Persil is Omo in Spain and the Netherlands and Skip in France
and Greece. Flora is Flora in the UK, but it's Becel in France and the Netherlands, and it's
Rama in Germany and Russia.
Names are misleading. Customers of the posh sounding Jeeves of Belgravia may be dismayed
to know it used to be owned by the down market chain Sketchley, until they sold it to a
German shoe company called Mr Minute. The QE2 sounds quintessentially British, but book
a cruise and you'll be swelling the coffers of Norwegian ship builders Kvaerner.
It's somehow equally disappointing to learn that the trendy new chain All Bar One is owned
by boring old Bass. All Bar One may be decently designed, but so it should be. Bass has a
long way to go to make amends for inflicting on the high street those fake Irish bars and
eyesores, O'Neills. Don't, however, think you can get away from Bass by drinking Grolsch,
Carling, Hoopers Hooch, Caffreys or Britvic Soft Drinks - Bass owns the lot. And if you
round off your evening by staying in the Holiday Inn, you'll bump up its profits nicely.
We like to think brands mean something. Consumers don't buy products; they make style
state-ments and are often prepared to pay a premium for the privilege. “The amount of
reliance placed on a brand is quite high, but it's not very well justified,” says Robert East. You
certainly may not have planned to benefit Barbie makers, Mattel, when you bought Scrabble,
nor profit the highly secretive Proc-ter and Gamble when you popped a Pringle or washed
your hair with Vidal Sassoon. You may have known P&G owned Ariel, Tampax and
Pampers, but did you know it also owns Sunny Delight orange drink, Crest, Clearasil, Pantene
Pro V, Cover Girl, Max Factor and Hugo Boss?
www.seyfihoca.com
54
You might think Virgin at least delivers on its promise. As brands go, it has more reason than
most to lay claim to certain values. It's inextricably associated with the founder, Richard
Branson. Yet even Virgin is not what it seems. In February, The Economist did an audit of the
empire: Virgin owns less than 50 per cent of Virgin Direct, Virgin Cola, Virgin Spirits, Virgin
Cinema, Virgin Vie and the Virgin Clothing Company. Hardly like a Virgin at all.
Does it matter? Should we worry about who owns what? “No,” says Nicholas Morgan,
marketing director for premium malt whiskies at United Distillers (owned by Diageo.)
“People buy a bottle of Bells or Johnnie Walker. They don't think about United Distillers and
we don't want them to. Information about the company gets in the way. It's not good for
anyone.
Consultants Inter-brand agree - but so they would. Inventors of brand names Hobnobs and
Mondeo, Interbrand were the first people to say you can put a price on a brand, and write it on
the balance sheet. “Most consumers are not that bothered about who owns the brand,
providing they get the service they want,” says its brand evaluation director, Alex Batchdor.
“It's getting the product right that matters. When ownership changes, no one gives a stuff.”
But shoppers may care much more than he would like to think. Angry consumers have in the
past inflicted major boycotts on the products of Barclays, Shell and Nestlé when they didn't
like what the companies were up to. Batchebr at Interbrand and Morgan at United Distillers
say no one is being misled - it's very easy to find out who owns what. But not from the
product it's not. Bylaw, all products need to have is an address you can write to. There's no
need to put any-thing about the parent company. So there's no way of knowing, when you're
buying a product of, say, Kraft Jacobs Suchard, that its parent company is the tobacco giant
Philip Morris. There's even less chance of knowing that the Chinese government owns part of
Midland Bank and First Direct (because the Hong Kong government bought an 8.9 per cent
stake in the banks' parent company, HSBC.)
Nor is it easy to unravel what ownership means at Rolls Royce after BMW bought the marque
and VW bought the factory. Ask either company exactly what is going on and both refer you
to Germany.
Not caring about ownership is also not an argument you could get past a Manchester United
supporter. Man U is the sports brand par excellence. The All Blacks are one of many clubs on
record as saying it's the brand on which they want to model themselves. It's the best-known
sports brand in the world - which is precisely why Rupert Murdoch would like to buy it, much
to the fans' dismay.
Andy Walsh, spokesman for the Manchester United Independent Supporters Association, will
fight a change of ownership tooth and nail. “We'd lose all our independence. It would no
longer be Man U. A football club generates a feeling of family, rather than a business, and in
terms of the emotional attachment, who owns it matters very much.”
There are other reasons why ownership matters. “Ownership is important in terms of public
policy and accountability,” says Robert East, professor of Marketing at Kingston University
Business School. It's where responsibility lies. Excluding own brands and fruit and veg, most
supermarket products are made by just three companies (four if you include booze). They are
Unilever, P&G, Nestlé and Diageo. It's a stranglehold even the supermarkets seem unaware
of.
www.seyfihoca.com
55
“Every week, we used to send a salesman to the supermarkets from each of our four main
companies: Van Den Berg Foods, Bird's Eye Walls, Lever Bros, and Elida Fabergé,” says a
spokesman for Unilever. “The supermarkets often had no idea they were all part of the same
company.”
Ownership certainly matters to the companies. This year Guinness (which already owned
United Distillers), merged with Grand Metropolitan. The newly formed group came up with
the unlovely name Diageo. The regulators made it sell Dewar's whisky and Bombay gin. The
two brands cost Bacardi-Martini £1.15 billion.
Brands are the life-blood of companies. You can buy a familiar but floundering brand, as
Unilever did with Colman's mustard, re-market it, then through brand extension flog the brand
to death. On the back of the mustard (made in Norwich for seven generations), Unilever has
now launched Colman's dry sauces, and Colman's and Oxo condiments.
But Unilever is beginning to change its brand strategy. “We think people want to know who is
controlling what, and who's behind the things they buy,” says a spokesman, Stephen Milton.
“They don't want some faceless conglomerate, and we think it's a trend that will continue.”
You can therefore expect to hear more about Unilever - which is just as well, as you're likely
to have plenty of its products in your home.
It's a high-risk strategy. Persil Power, accused of rotting your clothes, probably caused less of
a hiccup to Unilever's share price than New Coke's did to Coca-Cola. Being known to all your
customers - government/regulators, share-holders, trade and consumers - by the same name
means you have to take great care of it. A blip in one area, and the whole thing crashes down as Virgin may find now it has put its name on trains.
But it seems that the rewards can be greater. Whichever way they foster their brands, Nestlé,
Unilever, P&G, Diageo and the rest have a long way to go. Every time someone does a
survey, the super-brands that come out on top are the one-product or one-category companies,
known by the names under which they trade. You might not like them a great deal, but with
Coca-Cola, McDonald's, Sony and Microsoft you do at least know where you stand.
(From The Guardian November 5th 1998)
How to be a great manager
At the most general level, successful managers tend to have four characteristics:
ï‚·
ï‚·
ï‚·
ï‚·
they take enormous pleasure and pride in the growth of their people;
they are basically cheerful optimists - someone has to keep up morale when setbacks
occur;
they don't promise more than they can deliver;
when they move on from a job, they always leave the situation a little better than it
was when they arrived.
The following is a list of some essential tasks at which a manager must excel to be truly
effective.
www.seyfihoca.com
56
Great managers accept blame: When the big wheel from head office visits and expresses
displeasure, the great manager immediately accepts full responsibility. In everyday working
life, the best managers are constantly aware that they selected and should have developed
their people. Errors made by team members are in a very real sense their responsibility.
Great managers give praise: Praise is probably the most under-used management tool. Great
managers are forever trying to catch their people doing something right, and congratulating
them on it. And when praise comes from outside, they are swift not merely to publicise the
fact, 'but to make clear who has earned it. Managers who regularly give praise are in a much
stronger position to criticise or reprimand poor performance. If you simply comment when
you are dissatisfied with performance, it is all too common for your words to be taken as a
straightforward expression of personal dislike.
Great managers make blue sky: Very few people are comfortable with the idea that they will
be doing exactly what they are doing today in 10 years' time. Great managers anticipate
people's dissatisfaction.
Great managers put themselves about: Most managers now accept the need to find out not
merely what their team is thinking, but what the rest of the world, including their customers,
is saying. So MBWA (management by walking about) is an excellent thing, though it has to
be distinguished from MBWAWP (management by walking about - without purpose), where
senior management wander aimlessly, annoying customers, worrying staff and generally
making a nuisance of themselves.
Great managers judge on merit: A great deal more difficult than it sounds. It's virtually
impossible to divorce your feelings about someone - whether you like or dislike them - from
how you view their actions. But suspicions of discrimination or favouritism are fatal to the
smooth running of any team, so the great manager accepts this as an aspect of the game that
really needs to be worked on.
Great managers exploit strengths, not weaknesses, in themselves and in their people: Weak
managers feel threatened by other people's strengths. They also revel in the discovery of
weakness and regard it as something to be exploited rather than remedied. Great managers
have no truck with this destructive thinking. They see strengths, in themselves as well as in
other people, as things to be built on, and weakness as something to be accommodated,
worked around and, if possible, eliminated.
Great managers make things happen: The old-fashioned approach to management was rather
like the old-fashioned approach to child-rearing: 'Go and see what the children are doing and
tell them to stop it!' Great managers have confidence that their people will be working in their'
interests and do everything they can to create an environment in which people feel free to
express themselves.
Great managers make themselves redundant: Not as drastic as it sounds! What great managers
do is learn new skills and acquire useful information from the outside world, and then
immediately pass them on, to ensure that if they were to be run down by a bus, the team
would still have the benefit of the new information. No one in an organisation should be doing
work that could be accomplished equally effectively by someone less well paid than
themselves. So great managers are perpetually on the look-out for ' higher-level activities to
occupy their own time, while constantly passing on tasks that they have already mastered.
www.seyfihoca.com
57
(From The Independent )
Derivatives - The Beauty
Financial markets have grown more volatile since exchange rates were freed in 1973. Interest
rates and exchange rates now fluctuate more rapidly than at any time since the Crash of 1929.
At the same time, companies' profit margins have been squeezed by the lowering of trade
barriers and increased international competition. The result is that companies worldwide have
been forced to come to terms with their financial risks. No longer can managers stick their
heads in the sand and pretend that because their firms make cars, or sell soap powders, they
need only worry about this year's convertible or whether their new formula washes whiter
than Brand X. As many have found to their cost, ignoring interest-rate, currency or
commodity risks can hurt a company just as badly as the failure of a new product.
Derivatives offer companies the chance to reduce their financial risks - chiefly by transferring
them to someone (usually a bank) who is willing to assume and manage them. As they realize
this, more and more companies are using derivatives to hedge their exposures. America's
General Accounting Office reported that between 1989 and 1992 derivative volumes grew
145% to $12.1 trillion (in terms of the notional amount represented). This does not include
about $5.5 trillion of foreign-exchange forwards. Interest-rate risk was the main risk hedged at the end of 1992, interest-rate contracts accounted for 62% of total notionals, compared with
37% for foreign exchange.
In the US companies can now be sued for not hedging their exposures. In 1992, the Indiana
Court of Appeal held that the directors of a grain elevator co-operative had breached their
fiduciary duty by failing to sell forward the co-op's grain to hedge against a drop in prices.
Since 90% of the co-operative's operating income came from grain sales, its shareholders
argued that it was only prudent for the directors to have protected the co-op from the huge
losses it suffered (Brave v Roth, Indiana Court of Appeal). In another case, shareholders sued
Compaq Computers for violating securities laws by failing to disclose that it lacked adequate
mechanisms to hedge foreign-exchange risks.
Hedging does not necessarily remove all of a company's financial risk. When a firm hedges a
financial exposure, it is protecting itself against adverse market moves. If the markets move in
what would normally be the company's favour, the hedger could find itself in a position that
combined the worst of both hedged and unhedged worlds. For many firms, though, this is a
worthwhile price to pay for ensuring stability or certainty for some of their cashflows.
(From Managing derivative risks by Lillian Chew)
Motives
There is a multitude of psychological theories about what motivates man. Is the force inside
man, outside man, conditioned or not conditioned, goal directed or not goal directed? These
are all very controversial issues in academic research into what gets people to want to work.
Most people in organizations are not concerned with academic controversies and rely on their
commonsense view of behaviour.
The simplest motivation theory suggests that man is motivated by a series of needs or
motives. This theory argues that some of the motives are inherited and some learnt: that some
www.seyfihoca.com
58
are primarily physiological, while others are primarily psychological. Other theories deny the
existence of needs or motives. Therefore, at one extreme the behaviourists argue that
behaviour is a series of learned responses to stimuli, and at the other extreme systems
theorists talk about all systems - individuals, groups, and organizations - having needs.
Motivation can be either a conscious or an unconscious process: the allocation of time and
energy to work in return for rewards. Both internal and external stimuli lead to action.
Internalized values, hopes, expectations, and goals affect the decision process of the
individual, and thereby affect the resultant behaviour. Motivation is not an 'engine' built inside
an individual - as so many training managers believe. It is the individual responding to a
whole range of experiences, and responding as a totality, not as 'a need'. If we are threatened
by physical force, the stimulus for activity is external. If the hormone secretions in our bodies
operate effectively then we will wish to behave in physically satisfying ways. In both
examples, some of the force is inside the individual, while some of the stimulus is external.
How the individual will respond, how much energy he will expend, and how important are the
consequences (rewards) are all factors which moderate his motivation.
There have been many attempts to classify personal moderators in the decision process. The
most popular construct is the need, and categories of needs (e.g., body needs, safety needs,
social needs, achievement needs) dominate the literature. Goal categories, remarkably like
need categories, are also popular (e.g., money, status, power, friendship). Satisfaction theories
are a variation of goal theories, but have produced even 14 more controversial classifications
(e.g., implicit and explicit rewards).
There is no space here to go into what is primarily an academic debate on theories of
behaviour. I will contend that people are motivated to realize the outcome of ends or goals.
Where I use the term 'need', I do so in the sense of ends or goals desired by the individual. I
have difficulty in accepting a 'need' as a personality construct. However, desiring or wanting
an outcome does reveal something about a person, and 'need' can be used to refer to that
wanting. To many psychologists this view will be heresy, but I doubt if managers care what
the energy force is called (need, want, goal, etc.).
Organizational psychologists adopt hierarchies of goals or needs, along the lines suggested by
Maslow, McClelland, Ghiselli, and Likert. Maslow's need classifications are the most
extensively used, mainly because they seem to fit organizations rather than because they have
been empirically verified. We have little data to support the concept of a hierarchy of needs in
which lower order needs are satisfied before higher (hierarchically) order needs. However,
while need hierarchies may be difficult to accept, there is a great deal of data on the relevance
of these needs or ends or goals for individuals working in organizations, and it is these data
which are of value to managers.
The managers' dilemma is that, while they must accept the individual differences that exist
among their staff, organizational (and particularly personnel) practices assume that such
differences do not exist. The field of organization theory has been - and still is - plagued by
the conflict between the individual and the organization. As the orientation of this book is
towards organizations, it is important to deal with sameness or similarities between people,
while acknowledging differences within groups.
(From Managing people at work by John Hunt)
www.seyfihoca.com
59
Research and Development
There are two kinds of research: research and development, and basic research. The purpose
of research and development is to invent a product for sale. Edison invented the first
commercially successful light bulb, but he did not invent the underlying science that made the
light bulb possible. Edison at least understood the science, though, which was the primary
difference between inventing the light bulb and inventing fire. Basic research is something
else - ostensibly the search for knowledge for its own sake. Basic research provides the
scientific knowledge upon which R&D is later based. Sending telescopes into orbit or
building superconducting supercolliders is basic research. There is no way, for example, that
the $1.5 billion Hubble space telescope is going to lead directly to a new car or computer or
method of solid waste disposal. That is not what it is for. If a product ever results from basic
research, it usually does so fifteen to twenty years later, following a later period of research
and development.
Nearly all companies do research and development, but only a few do basic research. The
companies that can afford to do basic research (and cannot afford not to) are ones that
dominate their markets. Most basic research in industry is done by companies that have at
least a 50 percent market share. They have both the greatest resources to spare for this type of
activity and the most to lose if, by choosing not to do basic research, they eventually lose their
technical advantage over competitors. Such companies typically devote about 1 percent of
sales each year to research intended not to develop specific products but to ensure that the
company remains a dominant player in its industry twenty years from now. It is cheap
insurance, since failing to do basic research guarantees that the next major advance will be
owned by someone else.
The problem with industrial basic research, and what differentiates it from government basic
research, is this fact that its true product is insurance, not knowledge. If a researcher at the
government-sponsored Lawrence Livermore Lab comes up with some particularly clever new
way to kill millions of people, there is no doubt that his work will be exploited and that
weapons using the technology will eventually be built. The simple rule about weapons is that
if they can be built, they will be built. But basic researchers in industry find their work is at
the mercy of the marketplace and their captains-of-industry bosses. If a researcher at General
Motors comes up with a technology that will allow cars to be built for $100 each, GM
executives will quickly move to bury the technology, no matter how good it is, because it
threatens their current business, which is based on cars that cost thousands of dollars each to
build. Consumers would revolt if it became known that GM was still charging high prices for
cars that cost $100 each to build, so the better part of business valor is to stick with the old
technology since it results in more profit dollars per car produced.
In the business world, just because something can be built does not at all guarantee that it will
be built, which explains why RCA took a look at the work of George Heilmeier, a young
researcher working at the company's research center in New Jersey and quickly decided to
stop work on Heilmeier's invention, the liquid crystal display. RCA made this mid-1960s
decision because LCDs might have threatened its then-profitable business of building cathode
ray picture tubes. Twenty-five years later, of course, RCA is no longer a factor in the
television market, and LCD displays - nearly all made in Japan - are everywhere.
(From Accidental empires by Robert X Cringeley)
www.seyfihoca.com
60
SONY
I had decided during my first trip abroad in 1953 that our full name - Tokyo Tsushin Kogyo
Kabushiki Kaisha - was not a good name to put on a product. It was a tongue-twister. Even in
Japan, we shortened it sometimes to Totsuko, but when I was in the United States I learned
that nobody could pronounce either name. The English-language translation - Tokyo
Telecommunications Engineering Company - was too clumsy. We tried Tokyo Teletech for a
while, but then we learned there was an American company using the name Teletech.
It seemed to me that our company name didn't have a chance of being recognized unless we
came up with something ingenious. I also thought that whatever new name we came up with
should serve double duty- that is, it should be both our company name and our brand name.
That way we would not have to pay double the advertising cost to make both well known.
We tried a symbol for a while, an inverted pyramid inside a thin circle with small wedges cut
from the sides of the pyramid to give us a stylized letter "T." But for our first transistors and
for our first transistor radio, we wanted a brand name that was special and clever and that
people would remember. We decided our transistor radio would be the first consumer, product
available to the public with our new brand name on it.
I thought a lot about this when I was in the United States, where I noticed that many
companies were using three letter logotypes, such as ABC, NBC, RCA, and AT&T. Some
companies were also using just their full name as their logo. This looked like something new
to me. When I was a boy, I had learned to recognize the names of imported automobiles by
their symbols, the three-pointed star for Mercedes, the blue oval with Ford in it, the Cadillac
crown, the Pierce Arrow arrow, the Winged Victory of Rolls-Royce. Later, many car
companies began to use their names together with the symbol, like Chevrolet, Ford,
Buick, and others, and I could recognize their names even if I couldn't actually read them. I
pondered every possibility. Ibuka and I took a long time deciding on a name. We agreed we
didn't want a symbol. The name would be the symbol, and therefore it should be short, no
more than four or five characters. All Japanese companies have a company badge and a lapel
pin, usually in the shape of the company symbol, but except for a prominent few, such as the
three diamonds of Mitsubishi, for example, it would be impossible for an outsider to
recognize them. Like the automobile companies that began relying less and less on symbols
and more and more on their names, we felt we really needed a name to carry our message.
Every day we would write down possibilities and discuss them whenever we had the time. We
wanted a new name that could be recognized anywhere in the world, one that could be
pronounced the same in any language. We made dozens and dozens of tries. Ibuka and I went
through dictionaries looking for a bright name, and we came across the Latin word sonus,
meaning "sound." The word itself seemed to have sound in it. Our business was full of sound,
so we began to zero in on sonus. At that time in Japan borrowed English slang and nicknames
were becoming popular and some people referred to bright young and cute boys as "sonny,"
or "sonny-boys," and, of course, "sunny" and "sonny" both had an optimistic and bright sound
similar to the Latin root with which we were working. And we also thought of ourselves as
"sonny-boys" in those days. Unfortunately, the single word "sonny" by itself would give us
troubles in Japan because in the romanization of our language, the word "sonny" would be
pronounced "sohn-nee," which means to lose money. That was no way to launch a new
product. We pondered this problem for a little while and the answer struck me one day: why
not just drop one of the letters and make it "Sony"? That was it!
www.seyfihoca.com
61
The new name had the advantage of not meaning anything but "Sony" in any language; it was
easy to remember, and it carried the connotations we wanted. Furthermore, as I reminded
Ibuka, because it was written in roman letters, people in many countries could think of it as
being in their own language. All over the world governments were spending money to teach
people how to read English and use the roman alphabet, including Japan. And the more
people who learned English and the roman alphabet, the more people would recognize our
company and product name-at no cost to us.
We kept our old corporate name for some time after we began putting the Sony logotype on
our products. For our first product logo, we used a tall, thin sloping initial letter inside a
square box, but I soon realized that the best way to get name recognition would be to make
the name as legible and simple as possible, so we moved to the more traditional and simple
capital letters that remain today. The name itself is the logo.
We managed to produce our first transistorized radio in 1955 and our first tiny "pocketable"
transistor radio in 1957. It was the world's smallest, but actually it was a bit bigger than a
standard men's shirt pocket, and that gave us a problem for a while, even though we never
said which pocket we had in mind when we said "pocketable." We liked the idea of a
salesman being able to demonstrate how simple it would be to drop it into a shirt pocket. We
came up with a simple solution. We had some shirts made for our salesmen with slightly
larger than normal pockets, just big enough to slip the radio into.
The introduction of this proud achievement was tinged with disappointment that our first
transistorized radio was not the very first one on the market. An American company called
Regency, supported by Texas Instruments, and using TI transistors, put out a radio with the
Regency brand name a few months before ours, but the company gave up without putting
much effort into marketing it. As the first in the field, they might have capitalized on their
position and created a tremendous market for their product, as we did. But they apparently
judged mistakenly that there was no future in this business and gave it up.
Our fine little radio carried our company's new brand name, Sony, and we had big plans for
the future of transistorized electronics and hopes that the success of our small "pocketable"
radio would be a harbinger of successes to come.
In June 1957, we put up our first billboard carrying the Sony name opposite the entrance to
Tokyo's Haneda International Airport, and at the end of the year we put up another in the
heart of the Ginza district of Tokyo. In January 1958 we officially changed our company
name to Sony Corporation and were listed on the Tokyo Stock Exchange that December.
We had registered the name Sony in one hundred and seventy countries and territories and in
various categories, not just electronics, in order to protect it from being used by others on
products that would exploit the similarity. But we soon learned that we had failed to protect
ourselves from some entrepreneurs right at home in Japan. One day we learned that somebody
was selling "Sony" chocolate.
We were very proud of our new corporate name and I was really upset that someone would
try to capitalize on it. The company that picked up our name had used a completely different
name on their products before and only changed the name when ours became popular. They
registered the name "Sony" for a line of chocolates and snack foods and even changed their
company trade name to Sony Foods. In their logo they used the same type of letters we used.
www.seyfihoca.com
62
In those days we sometimes used a small cartoon character called "Sonny Boy" in our
advertising. The character was actually called "Atchan," and was created by cartoonist
Fuyuhiko Okabe of the Japanese newspaper Asahi Shimbun. The bogus Sony chocolate
merchants started using a similar cartoon. Seeing this stuff on sale in major department stores
made me sick with anger. We took the imposters to court and brought famous people such as
entertainers, newspapermen, and critics to confirm the damage that was being done to us. One
witness said he thought the appearance of Sony chocolate meant that the Sony Corporation
was in financial difficulty if it had to resort to selling chocolate instead of high-technology
electronics. Another witness said she had the impression that since Sony was really a
technical company, the chocolate must be some kind of synthetic. We were afraid that if these
chocolates continued to fill the marketplace, it would completely destroy the trust people had
in our company.
I have always believed that a trademark is the life of an enterprise and that it must be
protected boldly. A trademark and a company name are not just clever gimmicks-they carry
responsibility and guarantee the quality of the product. If someone tries to get a free ride on
the reputation and the ability of another who has worked to build up public trust, it is nothing
short of thievery. We were not flattered by this theft of our name.
Court cases take a long time in Japan, and the case dragged on for almost four years, but we
won. And for the first time in Japanese history, the court used the unfair competition law
rather than patent or trademark registration laws in granting us relief. The chocolate people
had registered the name, all right, but only after our name had become popular. In trying to
prove that the name was open for anyone to use, their lawyers went to the major libraries of
the country to show that the name was in the public domain, but they were in for a shock.
They came away empty-handed because no matter what dictionaries they went to they could
not find the word Sony. We knew they would discover that; we had done it ourselves long
before. The name is unique, and it is ours.
On our thirty-fifth anniversary, we thought we should consider revising our trademark. Styles
and fashions were changing in clothing, in product design, and in virtually everything, so we
thought that perhaps we should consider changing the style of the letters of our name. We
held an international competition, and we received hundreds of suggestions, along with
hundreds of pleas from our dealers not to change. After reviewing all the suggestions, we
decided not to make any changes. S O N Y still looked very good to us, and we decided, as
they say today, that there was no point in fixing something that was far from broken.
(From Made in Japan by Akio Morita)
American And Japanese Styles
Japanese attitudes toward work seem to be critically different from American attitudes.
Japanese people tend to be much better adjusted to the notion of work, any kind of work, as
honourable. Nobody would look down on a man who retires at age fifty-five or sixty and then
to keep earning money takes a more menial job than the one he left. I should mention that toplevel executives usually have no mandatory retirement age, and many stay on into their
seventies and even their eighties.
At Sony we have mandatory retirement from the presidency at sixty-five, but to utilize their
experience and knowledge we keep former executives who have retired as consultants. We
www.seyfihoca.com
63
provide them with office space and staff, so that they can work apart from the day-to-day
affairs of the company, at Ibuka Hall, a building located five minutes away from the
headquarters building. From time to time, we ask them for advice and they attend conferences
and other events as representatives of Sony. Many of those people who retire from managerial
jobs find executive positions in smaller companies or subsidiary companies of Sony where
their managerial experience and skill are needed and valued.
Workers generally are willing to learn new skills. Japan has never devised a system like the
American, in which a person is trained to do one thing and then refuses to take a job doing
anything else-and is even supported by government funds while he looks for a job that suits
his specific tastes. Because of Japan's special situation, our people do not have that luxury.
And our unemployment rate lately has not reached 3 percent.
One old style of management that is still being practiced by many companies in the United
States and by some in Japan is based on the idea that the company that is successful is the one
that can produce the conventional product most efficiently at cheaper cost. Efficiency, in this
system, becomes a god. Ultimately, it means that machinery is everything, and the ideal
factory is a perfectly automated one, perhaps one that is unmanned. This machinelike
management is a management of dehumanization.
But technology has accelerated at an unparalleled pace in the past few decades and it has
entailed digesting new knowledge, new information, and different technologies. Today,
management must be able to establish new business ahead of its competitors, rather than
pursue higher efficiency in manufacturing conventional products. In the U.S. and Europe
today, old-fashioned low-level jobs are being protected while the new technologies are being
neglected.
More important, an employee today is no longer a slave to machinery who is expected to
repeat simple mechanical operations like Charlie Chaplin in the film Modern Times. He is no
longer a beast of burden who works under the carrot-and stick rule and sells his labour. After
all, manual labour can be taken over by machine or computer. Modern industry has to be
brain-intensive and so does the employee. Neither machinery nor animals can carry out brainintensive tasks. In the late sixties, when integrated circuits had to be assembled by hand, the
deft fingers of Asian women were greatly in demand by U.S. companies. As the design of
these devices became more and more complicated, along came more sophisticated machinery,
such as laser trimmers, which required not deft fingers but agile minds and intelligence. And
so this upgrading of the workers is something that every country will have to be concerned
about, and the idea of preserving old-fashioned jobs in the modern era does not make sense.
This means educating new employees and re-educating older employees for new challenges.
That is not all. At Sony we at times have scientists participate in sales for a while because we
don't want our scientists to live in ivory towers. I have always felt they should know that we
are in a very competitive business and should have some experience in the front lines of the
business. Part of the training program for graduates who enter Sony as recruits fresh out of
university includes a program where non-technical persons undergo a month of training at a
factory and technical persons work as salespeople in a Sony shop or department store, selling
our products.
Japanese labour practices are often called old-fashioned in today's world, and some say the
old work ethic is eroding in Japan as it has elsewhere, but I do not think this is inevitable. As I
www.seyfihoca.com
64
see it, the desire to work and to perform well is not something unnatural that has to be
imposed on people. I think all people get a sense of satisfaction from accomplishing work that
is challenging, when their work and role in the company are being recognized. Managers
abroad seem to overlook this. People in America, for example, have been conditioned to a
system in which a person sells his labour for a price. In a way, that's good because people
cannot coast; they know they have to work to earn their money or be fired. (I also think the
way Americans make their children do work to earn their allowance is a fine idea; in Japan we
often just give the money without requiring anything of our children.) In Japan we do take the
risk of promising people job security, and then we have to keep motivating them. Yet I
believe it is a big mistake to think that money is the only way to compensate a person for his
work.
People need money, but they also want to be happy in their work and proud of it. So if we
give a lot of responsibility to a younger man, even if he doesn't have a title, he will believe he
has a good future and will be happy to work hard. In the United States, title and job and
monetary incentives are all tied together. That is why, if a young person has a big job,
management thinks he has to have a big salary. But in Japan we customarily give raises each
year as employees get older and more experienced in the company. If we give an unusually
high salary to one person, we cannot continue to give him annual increases indefinitely. At
some point, his salary will have to level off, and at that point, he is likely to get discouraged.
So we like to give the same sort of raise to all. I think this keeps our people well motivated.
This may be a Japanese trait, but I do not think so.
I believe people work for satisfaction. I know that advertisements and commercials in the U.S.
seem to hold up leisure as the most satisfying goal in life, but it is not that way in Japan yet. I
really believe there is such a thing as company patriotism and job satisfaction - and that it is
as important as money. It goes without saying that you must pay good wages. But that also
means, of course, that the company must not throw money away on huge bonuses for
executives or other frivolities but must share its fate with the workers. Japanese workers seem
to feel better about themselves if they get raises as they age, on an expectable curve. We have
tried other ways.
When we started our research laboratory, we had to go out and find researchers, and because
these people had more education and were, naturally, older than our normal new employees
we decided they should have higher wages, equivalent to U.S. salary levels. One suggested
plan was to put them under short-term contract, say three years, after which we would decide
whether to renew or not. But before we decided on this new pay scheme, I asked the new
employees whether they would prefer the more common system of lower pay to start, but with
yearly increases, or the three-year contract at a much higher wage.
Not one of them asked for the American-level salary. Everyone opted for long-range security.
That is why I tell the Americans I meet that people don't work only for money. But often
when I say it, they respond, "Yes, I see, but how much do you pay the ones who really work
hard?" Now this is an important point. When a worker knows he will be getting a raise each
year, he can feel so secure that he thinks there is no need to work hard. Workers must be
motivated to want to do a good job. We Japanese are, after all, human beings, with much in
common with people everywhere. Our evaluation system is complex and is designed to find
really capable persons, give them challenging jobs, and let them excel. It isn't the pay we give
that makes the difference-it is the challenge and the recognition they get on the job.
www.seyfihoca.com
65
My eldest son, Hideo, may not be the best example of the typical Japanese worker, but he has
an interesting and, I think, typical view of work in Japan. He has studied in Britain and the
United States, and all his life he wanted to work for Sony. He went to work as an Artists and
Repertory man at the CBS-Sony record company on the urging of Norio Ohga. He and I felt
that for him to come directly into Sony headquarters would be wrong, because of the family
connection and the overtones of nepotism. So he was proving himself at CBS-Sony. He
worked with foreign and local artists and became famous and successful in the record industry
in Japan. He worked very hard, from about noon until three or four o'clock in the morning,
doing his regular office business during the day and then dealing with musicians after they
finished their work. Hideo doesn't drink, and so it was hard for him to sit around the Tokyo
discos and bars with these rock stars, drinking Coca-Cola while they relaxed with whiskey in
the wee small hours of the morning. But it was important for him to do this, and although he
could have gone on a long time resting on his laurels, he took stock of himself on his thirtieth
birthday and made a decision.
As he put it, "In the record business, there are many people in their late thirties and early
forties wearing jogging shoes and white socks and jeans and T-shirts to the office. I looked at
those guys and said; I don't want to be like that when I am forty or forty-five. This business is
fine and I have been successful, and I have no reason to leave it. If I keep this job, I thought, I
might end up being a top officer of CBS-Sony, but I didn't want to see myself at fifty coming
into the office at one o'clock in the afternoon in jogging shoes and white socks saying 'Good
morning.' I felt I had to prove to myself after seven years in the record business that I could
work from nine to five, like ordinary people."
He was assigned to the Sony accounting division-quite a change, you might think, from the
artistic side of the record business-and some might have wondered whether he could make it
or not, but I believed he could. His attitude is very Japanese, despite his international
upbringing:
"All jobs are basically the same. You have to apply yourself, whether you are a record A&R
man, a salesman on the street, or an accounting clerk. You get paid and you work one hundred
percent to do the job at hand. As an A&R man, I was interested and excited and happy, but
naturally as long as you are satisfied with your work and are using your energy, you will be
happy. I was also very excited about the accounting division. I found out something new
every day, struggling with a whole bunch of invoices and the payment sheets, the balance
sheet, the profit and loss statement, and working with all those numbers. I began to get a
broad picture of the company, its financial position and what is happening day to day and
which way the company is heading. I discovered that that excitement and making music at the
studio are the same thing."
•
In the late sixties a European Commission internal memo on Japan was leaked, and a great stir
was created because it referred to the Japanese as "workaholics" who live in "rabbit hutches."
There is no doubt that inadequate housing is a major problem in Japan, and nobody could
deny that the Japanese are probably the hardest working people in the world. We have many
holidays in Japan, but only about the same number as the United States. We do not give long
summer vacations, even to our schoolchildren.
www.seyfihoca.com
66
At Sony we were one of the first Japanese companies to close down our factory for one week
in the summer, so that everybody could take off at the same time. And we long ago instituted
the five-day, forty-hour week. The Japan Labour Standards Act still provides for a maximum
forty-eight-hour workweek, though it is soon to be revised downward, and the average
workweek in manufacturing is now forty-three hours. But even with up to twenty days of paid
vacation a year, Japanese workers managed to take fewer days off and spend more days on the
job than workers in the United States and Europe.
It was only in 1983 that banks and financial institutions began to experiment with the five-day
week, closing one Saturday a month, and eventually the whole nation will move closer to the
five-day week. Still, International Labour Organization data show that Japanese work longer
weeks and have fewer labour disputes than workers in the U.S., the U.K., France, or West
Germany. What I think this shows is that the Japanese worker appears to be satisfied with a
system that is not designed only to reward people with high pay and leisure.
At Sony we learned that the problem with an employee who is accustomed to work only for
the sake of money is that he often forgets that he is expected to work for the group entity, and
this self-cantered attitude of working for himself and his family to the exclusion of the goals
of his co-workers and the company is not healthy. It is management's responsibility to keep
challenging each employee to do important work that he will find satisfying and to work
within the family. To do this, we often reorganize the work at Sony to suit the talents and
abilities of the workers.
I have sometimes referred to American companies as being structures like brick walls while
Japanese companies are more like stone walls. By that I mean that in an American company,
the company's plans are all made up in advance, and the framework for each job is decided
upon. Then, as a glance at the classified section of any American newspaper will show, the
company sets out to find a person to fit each job. When an applicant is examined, if he is
found to be oversized or undersized for the framework, he will usually be rejected. So this
structure is like a wall built of bricks: the shape of each employee must fit in perfectly, or not
at all.
In Japan recruits are hired, and then we have to learn how to make use of them. They are a
highly educated but irregular lot. The manager takes a good long look at these rough stones,
and he has to build a wall by combining them in the best possible way, just as a master mason
builds a stone wall. The stones are sometimes round, sometimes square, long, large, or small,
but somehow the management must figure out how to put them together. People also mature,
and Japanese managers must also think of the shapes of these stones as changing from time to
time. As the business changes, it becomes necessary to refit the stones into different places. I
do not want to carry this analogy too far, but it is a fact that adaptability of workers and
managements has become a hallmark of Japanese enterprise.
When Japanese companies in declining or sunset industries change their line of business or
add to it, workers are offered retraining and, for the most part, they accept it eagerly. This
sometimes requires a family move to the new job, and Japanese families are, again, generally
disposed to do this.
(From Made in Japan by Akio Morita)
CHEMISTRY
www.seyfihoca.com
67
Metallurgy: Making Alloys
The majority of alloys are prepared by mixing metals in the molten state; then the mixture is
poured into metal or sand moulds and allowed to solidify. Generally the major ingredient is
melted first; then the others are added to it and should completely dissolve. For instance, if a
plumber makes solder he may melt his lead, add tin, stir, and cast the alloy into stick form.
Some pairs of metals do not dissolve in this way. When this is so it is unlikely that a useful
alloy will be formed. Thus if the plumber were to add aluminium, instead of tin, to the lead,
the two metals would not dissolve - they would behave like oil and water. When cast, the
metals would separate into two layers, the heavy lead below and aluminium above.
One difficulty in making alloys is that metals have different melting points. Thus copper melts
at 1,083°C, while zinc melts at 419°C and boils at 907°C So, in making brass, if we just put
pieces of copper and zinc in a crucible and heated them above 1,083°C, both the metals would
certainly melt. But at that high temperature the liquid zinc would also boil away and the
vapour would oxidize in the air. The method adopted in this case is to heat first the metal
having the higher melting point, namely the copper. When this is molten, the solid zinc is
added and is quickly dissolved in the liquid copper before very much zinc has boiled away.
Even so, in the making of brass, allowance has to be made for unavoidable zinc loss which
amounts to about one part in twenty of the zinc. Consequently, in weighing out the metals
previous to alloying, an extra quantity of zinc has to be added.
Sometimes the making of alloys is complicated because the higher melting point metal is in
the smaller proportion. For example, one light alloy contains 92 per cent aluminium (melting
point 660°C) with 8 per cent copper (melting point 1,083°C). To manufacture this alloy it
would be undesirable to melt the few pounds of copper and add nearly twelve times the
weight of aluminium. The metal would have to be heated so much to persuade the large bulk
of aluminium to dissolve that gases would be absorbed, leading to unsoundness. In this, as in
many other cases, the alloying is done in two stages. First an intermediate 'hardener alloy' is
made, containing 50 per cent copper and 50 per cent aluminium, which alloy has a melting
point considerably lower than that of copper and, in fact, below that of aluminium. Then the
aluminium is melted and the correct amount of the hardener alloy added; thus, to make l00lb
of the aluminium-copper alloy we should require 84lb. of aluminium to be melted first and
16lb of hardener alloy to be added to it.
In a few cases, the melting point of the alloy can be worked out approximately by arithmetic.
For instance, if copper (melting point 1,083°C) is alloyed with nickel (melting point 1,454°C)
a fifty-fifty alloy will melt at about halfway between the two temperatures. Even in this case
the behaviour of the alloy on melting is not simple. A copper-nickel alloy does not melt or
freeze at one fixed and definite temperature, but progressively solidifies over a range of
temperature. Thus, if a fifty-fifty copper-nickel alloy is liquefied and then gradually cooled, it
starts freezing at 1,312°C, and as the temperature falls, more and more of the alloy becomes
solid until finally at 1,248°C it has completely solidified. Except in certain special cases this
'freezing range' occurs in all alloys, but it is not found in pure metals, metallic, or chemical
compounds, and in some special alloy compositions, referred to below, all of which melt and
freeze at one definite temperature.
The alloying of tin and lead furnishes an example of one of these special cases. Lead melts at
327°C and tin at 232°C. If lead is added to molten tin and the alloy is then cooled, the
freezing point of the alloy is found to be lower than the freezing points of both lead and tin
www.seyfihoca.com
68
(see figure 1). For instance, if a molten alloy containing 90 per cent tin and 10 per cent lead is
cooled, the mixture reaches a temperature of 217°C before it begins to solidify. Then, as the
alloy cools further, it gradually changes from a completely fluid condition, through a stage
when it is like gruel, until it becomes as thick as porridge, and finally, at a temperature as low
as 183°C, the whole alloy has become completely solid. By referring to figure 1, it can be
seen that with 80 per cent tin, the alloy starts solidifying at 203°C, and finishes only when the
temperature has fallen to 183°C (note the recurrence of the 183°C).
What happens at the other end of the series, when tin is added to lead? Once again the
freezing point is lowered. An alloy with only 20 per cent tin and the remainder lead starts to
freeze at 279°C and completes solidification at the now familiar temperature of 183°C. One
particular alloy, containing 62 per cent tin and 38 per cent lead, melts and solidifies entirely at
183°C. Obviously this temperature of 183°C and the 62/38 per cent composition are important
in the tin-lead alloy system. Similar effects occur in many other alloy systems and the special
composition which has the lowest freezing point of the series and which entirely freezes at
that temperature has been given a special name. The particular alloy is known as the 'eutectic'
alloy and the freezing temperature (183°C in the case of the tin-lead alloys) is called the
eutectic temperature.
By a careful choice of constituents, it is possible to make alloys with unusually low melting
points. Such a fusible alloy is a complex eutectic of four or five metals, mixed so that the
www.seyfihoca.com
69
melting point is depressed until the lowest melting point possible from any mixture of the
selected metals is obtained. A familiar fusible alloy, known as Wood's metal, has a
composition:
Bismuth
Lead
Tin
Cadmium
4 parts
2 parts
1 part
1 part
and its melting point is about 70°C; that is, less than the boiling point of water. Practical
jokers have frequently amused themselves by casting this fusible alloy into the shape of a
teaspoon, which will melt when used to stir a cup of hot tea.
These low melting point alloys are regularly in use for more serious purposes, as for example,
in automatic anti-fire sprinklers installed in the ceilings of buildings. Each jet of the water
sprinkler system contains a piece of fusible alloy, so that if a fire occurs and the temperature
rises sufficiently high, the alloy melts and the water is released through the jets of the
sprinkler.
(From Metals in the Service of Man by W. Alexander & A. Street.)
Electricity Helps Chemistry: Electro-plating
A liquid which is decomposed when an electric current passes through it is called an
electrolyte. The process is called electrolysis, and the two wires or plates dipping into the
electrolyte are called electrodes. The electrode which is connected to the positive terminal of
the cell or battery is called the anode. The electrode which is connected to the negative
terminal of the battery is called the cathode.
Let us examine what happens when two copper electrodes are used in a solution of copper
sulphate. The circuit is shown in the diagram. The right-hand diagram shows the two copper
electrodes dipping into the copper sulphate solution contained in a glass jar. The current
enters by the anode (+), passes through the solution, enters the cathode (-), and then leaves the
cathode as shown by the arrow. In the left-hand diagram, V represents the glass vessel
containing the copper sulphate (electrolyte), and the two electrodes are marked + for the
www.seyfihoca.com
70
anode and - for the cathode. When the switch S is closed, the current flows from the - terminal
of the battery B in the direction of the arrow to the anode (+) of V, through the solution to the
cathode (-), then round the circuit through S back to the negative terminal of the battery B.
Before starting this experiment the weights of the two copper plates which are to be used for
the anode and cathode must be written down carefully for future reference. Next, place the
anode and cathode in the copper sulphate solution and connect them up to the battery B and
switch S. The switch is then placed in the 'on' position and the current is allowed to flow
through the circuit for about half an hour. The anode and cathode are then removed and dried
carefully in blotting paper before being weighed a second time.
You will find that a surprising thing has happened. The anode now weighs a few milligrams
less than before and the cathode weighs a few milligrams more than before. The weight lost
by the anode is exactly equal to the gain in weight by the cathode. In some strange way a few
milligrams of copper have been removed from the anode and carried through the electrolyte
by the current and have finally become firmly attached to the cathode. This is a most exciting
discovery, for we have learned how to use an electric current to transfer tiny particles of
copper from the anode to the cathode.
Nineteenth-century industry soon found out how to apply this exciting discovery to our
everyday lives. Scientists found that many other metals could be transferred from anode to
cathode. The anode had to be made of the metal which it was desired to transfer to the
cathode, and the electrolyte had to be a suitable solution or salt of the metal. Then the cathode
always became plated with metal from the anode. Copper, silver, gold, nickel, zinc and
chromium can all be used in this process, which is called electro-plating. Electro-plating is
used widely in industry for a number of reasons. Firstly, it is used for decoration. Coatings of
nickel, gold, silver or chromium give a nice shiny appearance to articles and make them look
much more expensive. Watch-cases and cutlery are often plated with silver or gold to give
them a smart appearance so that they become attractive to intending buyers. Handlebars of
bicycles and the shiny fittings of cars are also made attractive by means of nickel and
chromium plating.
This leads us to the second reason for electro-plating - as a protection against rust or
corrosion. Iron and steel corrode easily when exposed to the atmosphere. Car fittings and the
shiny parts of bicycles are electro-plated chiefly for this reason, so that they may stand up to
the hard wear and tear of daily use. Zinc is formed into a protective layer for iron sheets by
the electroplating process which we now call galvanizing. Galvanized iron sheets resist the
effects of wind and weather much better than sheets made of iron. Tin is also used as a
protective agent. Sheets of thin iron are plated with tin and used for canning fruit and jam, and
for all kinds of 'tin' cans used in industry and trade. We may sum up by saying that industry
has used the process of electro-plating first to protect metal surfaces which would otherwise
corrode; and secondly to provide a beautiful and attractive finish to useful articles. As a result,
our bicycles and cars, our watches and cutlery, our building and manufacturing materials last
much longer and are much more pleasant to look at.
The process of electrolysis is used for the production of very pure specimens of metal. Most
metals in industrial use contain many impurities. About 1 million tons of refined copper are
produced each year by electrolysis. In this case the anode consists of crude copper and the
cathode of thin sheets of pure copper. As the current passes, pure copper from the anode
passes over to the cathode, and all impurities fall off the anode as a kind of mud. In this way
www.seyfihoca.com
71
pure copper is collected at one electrode and the muddy residue, which falls off the cathode,
sinks to the bottom of the vat and is periodically removed.
Aluminium is so widely used today that we can scarcely think of times when it was not
available. Yet a few years back it was a costly metal because no satisfactory method had been
found of producing it commercially. Aluminium ores are so common in nature that scientists
and engineers made many attempts to find a cheap and convenient method of refining them.
The problem was finally solved by electrolysis, using a carbon anode and aluminium ores,
which had been melted at a temperature of about 1,000°C, as the electrolyte. Aluminium is
now plentiful and it is being put to fresh uses every day.
Electrolysis has an important industrial application in the printing trade, for it is often used to
make the 'blocks' from which pictures and type are printed. A wax mould is first made of the
printing block which is to be reproduced. Since wax is a non-conductor of electricity it is
dusted over with graphite so that the surface becomes a conductor and can act as a cathode.
This mould then becomes the cathode upon which copper or chromium is deposited from the
anode. When the wax is taken out of the electrolyte it is coated with a fine shell of metal. The
wax is removed by heating and the metal shell acts as a mould into which molten type metal
can be poured. Plates made in this way are very hard-wearing and can be used to print many
thousands of copies of newspapers, journals and magazines.
(From General Science by N. Ahmad, W. F. Hawkins And W. M. Zaki.)
ECONOMICS
THE CONVENTIONAL WISDOM
The first requirement for an understanding of contemporary economic and social life is a clear
view of the relation between events and the ideas which interpret them. For each of these has
a life of its own, and much as it may seem a contradiction in terms each is capable for a
considerable period of pursuing an independent course.
The reason is not difficult to discover. Economic, like other social life, does not conform to a
simple and coherent pattern. On the contrary it often seems incoherent, inchoate, and
intellectually frustrating. But one must have an explanation or interpretation of economic
behaviour. Neither man's curiosity nor his inherent ego allows him to remain contentedly
oblivious to anything that is so close to his life.
Because economic and social phenomena are so forbidding, or at least so seem, and because
they yield few hard tests of what exists and what does not, they afford to the individual a
luxury not given by physical phenomena. Within a considerable range he is permitted to
believe what he pleases, he may hold whatever view of the world he finds most agreeable or
otherwise to his taste.
As a consequence, in the interpretation of all social life there is a persistent and never-ending
competition between what is relevant and what is merely acceptable. In this competition,
while a strategic advantage lies with what exists, all tactical advantage is with the acceptable.
Audiences of all kinds most applaud what they like best. And in social comment the test of
audience approval, far more than the test of truth, comes to influence comment. The speaker
www.seyfihoca.com
72
or writer who addresses his audience with the proclaimed intent of telling the hard, shocking
facts invariably goes on to expound what the audience most wants to hear.
Just as truth ultimately serves to create a consensus, so in the short run does acceptability.
Ideas come to be organized around what the community as a whole or particular audiences
find acceptable. And as the laboratory worker devotes himself to discovering scientific
verities, so the ghost writer and the public relations man concern themselves with identifying
the acceptable. If their clients are rewarded with applause, these artisans are qualified in their
craft. If not they have failed. However, by sampling audience reaction in advance, or by
pretesting speeches, articles, and other communications, the risk of failure can now be greatly
minimized.
Numerous factors contribute to the acceptability of ideas. To a very large extent, of course,
we associate truth with convenience - with what most closely accords with self-interest and
individual well-being or promises best to avoid awkward effort or unwelcome dislocation of
life. We also find highly acceptable what contributes most to self-esteem. Speakers before the
United States Chamber of Commerce rarely denigrate the business man as an economic force.
Those who appear before the AFL-CIO are prone to identify social progress with a strong
trade union movement. But perhaps most important of all, people approve most of what they
best understand. As just noted, economic and social behaviour are complex and mentally
tiring. Therefore we adhere, as though to a raft, to those ideas which represent our
understanding. This is a prime manifestation of vested interest. For a vested interest in
understanding is more preciously guarded than any other treasure. It is why men react, not
infrequently with something akin to religious passion, to the defence of what they have so
laboriously learned. Familiarity may breed contempt in some areas of human behaviour, but
in the field of social ideas it is the touchstone of acceptability.
Because familiarity is such an important test of acceptability, the acceptable ideas have great
stability. They are highly predictable. It will be convenient to have a name for the ideas which
are esteemed at any time for their acceptability, and it should be a term that emphasized this
predictability. I shall refer to those ideas henceforth as the conventional wisdom.
(From The Affluent Society by J. K. Galbraith)
MARKETS
A market is commonly thought of as a place where commodities are bought and sold. Thus
fruit and vegetables are sold wholesale at Covent Garden Market and meat is sold wholesale
at Smithfield Market. But there are markets for things other than commodities, in the usual
sense. There are real estate markets, foreign exchange markets, labour markets, short-term
capital markets, and so on; there may be a market for anything which has a price. And there
may be no particular place to which dealings are confined. Buyers and sellers may be
scattered over the whole world and instead of actually meeting together in a market-place they
may deal with one another by telephone, telegram, cable or letter. Even if dealings are
restricted to a particular place, the dealers may consist wholly or in part of agents acting on
instructions from clients far away. Thus agents buy meat at Smithfield on behalf of retail
butchers all over England; and brokers on the London Stock Exchange buy and sell securities
on instructions from clients all over the world. We must therefore define a market as any area
over which buyers and sellers are in such close touch with one another, either directly or
www.seyfihoca.com
73
through dealers, that the prices obtainable in one part of the market affect the prices paid in
other parts.
Modern means of communication are so rapid that a buyer can discover what price a seller is
asking, and can accept it if he wishes, although he may be thousands of miles away. Thus the
market for anything is, potentially, the whole world. But in fact things have, normally, only a
local or national market.
This may be because nearly the whole demand is concentrated in one locality. These special
local demands, however, are of quite minor importance. The main reason why many things
have not a world market is that they are costly or difficult to transport.
The lower the value per ton of a good, the greater is the percentage addition made to its price
by a fixed charge per ton-mile for transport. Thus, if coal is £2 a ton and tin £200 a ton at the
place of production, a given transport charge forms a percentage of the price of coal a hundred
times greater than of the price of tin. Hence transport costs may restrict the market for goods
with a low value per ton, even if, as is often the case, they are carried at relatively low rates. It
may be cheaper to produce, say, coal or iron ore at A than at B, but the cost of transporting it
from A to B may outweigh the difference in production costs, so that it is produced for local
consumption at B, and B does not normally form part of the market output of A. For example,
coal is produced much more cheaply in the United States than in Europe, but, owing to the
cost of transporting coal by rail from the inland mines to the Atlantic seaboard of the United
States, American coal seldom finds its way to Europe.
Sea transport, however, is very much cheaper than land transport. Hence commodities of this
type produced near a port can often be sent profitably quite long distances by sea. Thus
Swedish iron ore comes by sea from Narvik to the Ruhr, and British coal is exported to
Canada and South America.
The markets for real estate are local. Soil has been transported from French vineyards to
California, and historic mansions have been demolished in Europe to be re-erected in the
United States, but as a rule land and buildings are not transported.
Some goods, like new bread and fresh cream and strawberries, must be consumed very soon
after they have been produced, and this restricts their sale to local markets. Other goods do
not travel well. Thus many local wines which cannot stand transport can be bought in the
district more cheaply than similar wines which have a wider market. The development of
refrigeration, and of other devices which enable foodstuffs to be preserved and transported,
has greatly widened the market for such things as meat and fish and some kinds of fruit. But
such devices often transform the articles, from the standpoint of consumers, into a different
commodity. Condensed milk is not the same as fresh milk, and chilled meat or frozen butter
has not the same taste as fresh.
Many workers are reluctant to move to a different country, or even to a different part of their
own country, to get a higher wage. This should not be exaggerated. Before the war of 1914,
over a million persons a year emigrated overseas from Europe. Following it, there were
considerable movements of population within Great Britain away from the depressed areas
towards the more prosperous South. Employers may take the initiative. Thus girl textile
workers have been engaged in Yorkshire to work in Australia, and during the inter-war years
www.seyfihoca.com
74
French employers engaged groups of Poles and Italians to work in the coal-mines and steelworks of France. Nevertheless labour markets are mainly local, or at any rate national.
Transport services by rail or tram are obviously local in that passengers or goods must travel
between points on the fixed track. A firm may charter, for example, a Greek ship rather than
an English ship, if it is cheaper, but low railway rates in Belgium are no help to the firm
which wishes to send goods across Canada. In the same way, such things as gas, water, and
electricity, supplied by means of pipes or wires, cannot be sold to places not connected with
the system of pipes or wires.
(From Economics by Frederick Benham)
INVESTMENT
Of the various purposes which money serves, some essentially depend upon the assumption
that its real value is really constant over a period of time. The chief of these are those
connected, in a wide sense, with contracts for the investment of money. Such contracts namely those which provide for the payment of fixed sums of money over a long period of
time - are the characteristic of what it is convenient to call the Investment System, as distinct
from the property system generally. Under this phase of capitalism, as developed during the
nineteenth century, many arrangements were devised for separating the management of
property from its ownership. These arrangements were of three leading types: (i) those in
which the proprietor, while parting with the management of his property, retained his
ownership of it - i.e. of the actual land, buildings, and machinery, or of whatever else it
consisted in, this mode of tenure being typified by a holding of ordinary shares in a joint-stock
company; (ii) those in which he parted with the property temporarily, receiving a fixed sum of
money annually in the meantime, but regained his property eventually, as typified by a lease;
and (iii) those in which he parted with his real property permanently, in return either for a
perpetual annuity fixed in terms of money, or for a terminable annuity and the repayment of
the principal in money at the end of the term, as typified by mortgages, bonds, debentures,
and preference shares. This third type represents the full development of Investment.
Contracts to receive fixed sums of money at future dates (made without provision for possible
changes in the real value of money at those dates) must have existed as long as money has
been lent and borrowed. In the form of leases and mortgages, and also of permanent loans to
Governments and to a few private bodies, such as the East India Company, they were already
frequent in the eighteenth century. But during the nineteenth century they developed a new
and increasing importance, and had, by the beginning of the twentieth, divided the propertied
classes into two groups-the 'business men' and the 'investors' - with partly divergent interests.
The division was not sharp as between individuals; for business men might be investors also,
and investors might hold ordinary shares; but the division was nevertheless real, and not the
less important because it was seldom noticed.
By the aid of this system the active business class could call to the aid of their enterprises not
only their own wealth but the savings of the whole community; and the professional and
propertied classes, on the other hand, could find an employment for their resources, which
involved them in little trouble, no responsibility, and (it was believed) small risk.
For a hundred years the system worked, throughout Europe, with an extraordinary success and
facilitated the growth of wealth on an unprecedented scale. To save and to invest became at
www.seyfihoca.com
75
once the duty and the delight of a large class. The savings were seldom drawn on, and,
accumulating at compound interest, made possible the material triumphs which we now all
take for granted. The morals, the politics, the literature, and the religion of the age joined in a
grand conspiracy for the promotion of saving. God and Mammon were reconciled. Peace on
earth to men of good means. A rich man could, after all, enter into the Kingdom of Heaven-if
only he saved. A new harmony sounded from the celestial spheres. 'It is curious to observe
how, through the wise and beneficent arrangement of Providence, men thus do the greatest
service to the public, when they are thinking of nothing but their own gain': so sang the
angels.
The atmosphere thus created well harmonized the demands of expanding business and the
needs of an expanding population with the growth of a comfortable non-business class. But
amidst the general enjoyment of ease and progress, the extent to which the system depended
on the stability of the money to which the investing classes had committed their fortunes was
generally overlooked; and an unquestioning confidence was
I apparently felt that this matter would look after itself. Investments spread and multiplied,
until, for the middle classes of the world, the gilt-edged bond came to typify all that was most
permanent and most secure. So rooted in our day has been the conventional belief in the
stability and safety of a money contract that, according to English law, trustees have been
encouraged to embark their trust funds exclusively in such transactions, and are indeed
forbidden, except in the case of real estate (an exception which is itself a survival of the
conditions of an earlier age) to employ them otherwise.
As in other respects, so also in this, the nineteenth century relied on the future permanence of
its own happy experiences and disregarded the warning of past misfortunes. It chose to forget
that there is no historical warrant for expecting money to be represented even by a constant
quantity of a particular metal, far less by a constant purchasing power. Yet Money is simply
that which the State declares from time to time to be a good legal discharge of money
contracts. In 1914 gold had not been the English standard for a century or the sole standard of
any other country for half a century. There is no record of a prolonged war or a great social
upheaval which has not been accompanied by a change in the legal tender, but an almost
unbroken chronicle in every country which has a history, back to the earliest dawn of
economic record, of a progressive deterioration in the real value of the successive legal
tenders which have represented money.
(From A Tract on Monetary Reform by J M. Keynes)
BARTER
The existence of pure barter does not necessarily indicate a very primitive form of
civilization. Often the system survives long after the community has progressed considerably
in other respects. This may be due to conservatism, since primitive peoples are reluctant to
change their trading methods, even though they be sufficiently intelligent and advanced to
adopt more convenient methods. In some cases there is prejudice against the adoption of a
monetary economy, though such prejudice is usually directed against the use of coins rather
than against primitive money. In many cases barter continues to be the principal method of
trading long after the adoption of some form of money, for the simple reason that there is not
enough money to go round. And a decline in the supply of money often causes a relapse into
www.seyfihoca.com
76
barter. Distrust in money has also been responsible for reversion to the barter system; such
distrust may have been caused by debasement or inflation.
In the light of the stock phrases used by many economists about the inconvenience of barter it
may appear puzzling to the student that any community which was sufficiently advanced to
realize the possibilities of a monetary system should continue to practise such an inconvenient
method. The explanation is that in a primitive community barter is not nearly so inconvenient
as it appears through modern eyes. Economists are inclined to exaggerate its inconvenience
because they look at it from the point of view of modern man. The instances-real or
imaginary-they quote are calculated to make their readers wonder how any community could
possibly have existed under barter except in extremely primitive conditions. Some of them
seek to demonstrate the absurdity of barter by describing the difficulties that would arise if
our modern communities were to attempt to practise it. It is, of course, easy for a lecturer to
earn the laughter of his audience by telling them about the pathetic efforts of some market
gardener who has to find a barber in need of radishes before he can have his hair cut. What
the lecturer and his audience do not realize is that in a primitive community the grower of
radishes usually cuts his own hair, or has it cut by a member of his family or household; and
that even in primitive communities with barbers as an independent profession the barber and
the gardener have a fair idea about each other's requirements, and have no difficulty in suiting
each other. If the barber does not happen to require to-day any of the products the gardener is
in a position to offer, he simply performs his service in return for the future delivery of
products he is expected to need sooner or later.
Even the genuine instances quoted by economists to illustrate the absurdity of barter are apt to
be misleading in their implication. There is, for instance, the well-known experience of Mlle.
Zelie, singer at the Théâtre Lyrique in Paris, who, in the course of a tour round the world,
gave a concert on one of the Society Islands, and received the fee of three pigs, twenty-three
turkeys, forty-four chickens, five thousand coconuts and considerable quantities of bananas,
lemons and oranges, representing one-third of the box office takings. In a letter published by
Wolowski and quoted to boredom by economists ever since, she says that, although this
amount of livestock and vegetables would have been worth about four thousand francs in
Paris, in the Society Islands it was of very little use to her. Another much-quoted experience
is that of Cameron in Tanganyika, when in order to buy an urgently needed boat he first had
to swap brass wire against cloth, then cloth against ivory and finally ivory against the boat.
What the economists quoting these and other similar instances do not appear to realize is that
the difficulties complained of are not inherent in the system of barter. They are largely
anomalies arising from sudden contact between two different civilizations. A native singer in
the Society Islands would not have been embarrassed at receiving payment in kind, since she
would have known ways in which to dispose of her takings, or store them for future use. Nor
would a native of Tanganyika have found the system of barter prevailing there at the time of
Cameron's visit nearly so difficult as Cameron did. Knowing local conditions, he would have
been prepared for the difficulties, and, before embarking on a major capital transaction such
as the purchase of a boat, he would have made arrangements accordingly. In any case, the fact
that the goods required could not be obtained by a single transaction would not have worried
him unduly. The majority of primitive peoples enjoy bartering and bargaining, and the time
lost in putting through three transactions instead of one would not matter to them nearly as
much as to modern man living at high speed, especially to an explorer in a hurry to proceed
on his journey. And while Cameron must have suffered a loss in each of the three
transactions, a local man with adequate time at his disposal and with a thorough knowledge of
www.seyfihoca.com
77
his market would have chosen the right moment for effecting the necessary exchanges on
terms reasonably advantageous to him.
(From Primitive Money by Paul Einzig)
PRODUCTIVITY AS A GUIDE TO WAGES
Through defective presentation the Government has allowed the wages pause to be interpreted
as involving a substantial sacrifice by all concerned, and especially by those least able to
afford it. The very reverse is true. In fact, substantial benefits would accrue to everybody by
maintaining existing wages levels, because if this were done for a reasonable time, real
earnings, or purchasing power, would improve. And if thereafter reasonable restraint were
exercised in asking for, or giving, wages increases, there would be a real hope of recapturing
the habit of price reduction and so of still further improving purchasing power.
It is not true that prosperity must be equated with 'gentle' inflation-whatever that means. And
while lip service is paid to the concept that inflation is caused by wage increases outstripping
increases in productivity, few people are willing to do something positive about it. There is
already active opposition to the wages pause, and no discernible determination by any
sections of the community concerned with wage negotiations to devise a logical wages policy.
It was, of course, unfortunate for the climate of industrial relations that the wages pause was
presented to both managements and workers without the most careful preparation and
education. Too much emphasis was, and still is, placed on actual wage rates and earnings,
whereas what matters most is real earnings, or purchasing power. Charts comparing, for
example, wage rates, earnings, and profits are unrealistic unless we add the line for real
earnings. When this is added, and also the line for productivity, it becomes startlingly clear
that however great the total wage claim and the eventual settlement, real earnings will always
follow much the same course as productivity. When substantial claims are substantially met
real earnings may for a little time rise above the line of productivity. But eventually they
come together again.
In spite of some sizeable variations profits have followed much the same course as weekly
wage rates though they have been almost consistently below them. Thus all the tough
bargaining by the unions on the basis of rising profits has only effectively raised real
purchasing power by about the same factor as rising productivity would in any case have
achieved had wages merely kept pace with it.
Of course, it is easy to see what is wrong. Put in its simplest terms it is that we have lost the
habit of price reduction. Everyone wants it but few achieve it in the face of ever rising costs.
At present we are importing too much and exporting too little. The reasons for this are
complex. Design, salesmanship, after-sales service, and many other factors enter into it. What
is certain is that it will be even more difficult to sell abroad after the next round of wage
claims have been met and incorporated into the price of the product. What would the picture
look like now if wage increases had more nearly been associated with the increase in
productivity? Certainly our products would have cost overseas and home customers fewer
pounds, shillings and pence without the purchasing power of the workers having been at all
impaired.
www.seyfihoca.com
78
It has long seemed to me that the alternative is a rationalized wage policy for industry,
maintaining the existing bargaining machinery, but avoiding the bitterness and acrimony that
is engendered at present, and also avoiding the time-wasting strikes and deflection of
management from its principal purpose. It may be argued that to base wage rates on a
productivity index would weaken the trade unions. I do not think so. If there were bargaining
committees of trade unions and employers for each industry and they based their negotiations
annually on whatever was the national increase in productivity, then I feel quite sure that
much more time would be left for the more important things - such as raising productivity
itself and improving working conditions and the training of craftsmen to adequate standards.
If workers knew that every year wage increases would be automatically considered without
their having to demand them, then this would remove the present acrimonious preliminaries to
negotiations and would guarantee that the result could never be inflationary - that is, wage
increases would be real increases in terms of purchasing power. It would also give workers a
more direct interest in raising productivity by relating it to their wage packets. As wages
would never exceed productivity, the cost of production could not rise from this cause; in all
probability it would fall. This would encourage price reduction and tend to give a lower retail
price index, so that real wages would increase, and any increase in the money rate would be
worth still more. This would solve the problem of the pensioners and other people with fixed
incomes. Furthermore, I am convinced that such a wage policy, when clearly understood,
would create a new spirit of collaboration between management, supervision, and
workpeople.
(J. J. Gracie From an article in The Guardian, December 7th, 1961)
THE FAILURE OF THE CLASSICAL THEORY OF COMMERCIAL POLICY
Let us try to sort out in appropriate groups the various influences which have been responsible
for these developments.
If we are to preserve a proper sense of proportion, I have no doubt whatever that right at the
top of the list we must put the direct impact and influence of war. This is a matter which is so
obvious that we are very apt to forget it. Yet whatever may be the importance of the political
and ideological tendencies which I shall shortly proceed to discuss, we get the perspective
wrong if we regard them as more important than the brute disruptive effects of the military
convulsions of our age. It was these convulsions which, by bursting the cake of custom and
compelling the supersession of the normal institutions of peace, created the states of mind in
which restrictive and disintegrating policies seemed legitimate. It may be said that if adequate
measures had been taken, the difficulties of disequilibrium would have been less; and that if
fundamental attitudes had not been disturbed by illiberal ideologies, the chances of applying
appropriate measures would have been greater. Doubtless there is truth in this. But we are not
dealing with communities of angels whose errors are always deliberate sins against the light.
We must not expect too much of the human spirit under strain; and we simplify history
unduly if, in the explanation of the policies of our time, we do not allot to the shock of war
something like autonomous status.
For somewhat similar reasons I am disposed to list separately the influence of mass
unemployment or imminent financial crisis. Of course, unemployment and financial crises are
not to be regarded as acts of God: there are often occasions when they are to be attributed to
wrong economic policies, in some cases perhaps springing from the same ideologies as the
www.seyfihoca.com
79
overt resistance to liberal commercial policies. But here again, I think we oversimplify if we
make our history monistic. In the explanation of how this or that community came to adopt
policies of commercial restriction, we do well to treat unemployment and financial crisis as at
least semi-independent causes. After all, we know that, in such circumstances, commercial
restrictions may actually have a favourable influence for a time: unemployment may be
diminished, a drain of gold or dollars arrested. And experience shows that it is just at such
times that liberal commercial policies are most in danger. Take, for instance, the final
abandonment of free trade by Great Britain in the early thirties. No one who lived through the
crisis of those days will be disposed to deny the influence of ideological factors. The
advocacy of tariff protection by Keynes, hitherto an outstanding free trader, had an impact
which should not be underestimated. But perhaps Keynes himself would not have gone that
way had there not been a depression. And certainly his influence would have been less if
people had not felt themselves to be in a sort of earthquake in which all the old guide posts
and landmarks were irrelevant.
Having thus recognized the catastrophic elements in the evolution of policy, we may now go
on to examine the more persistent and slow-moving forces. And since we are proceeding all
the time from the simpler to the more complex, we may put next on our list the influence of
producer interest. This is an influence which I am sure should be disentangled from those
which we have already examined. I know that it is sometimes argued that it is only because of
under-employment or financial dislocation that the pressure groups are effective; and I
willingly concede that in such situations they have, so to speak, very powerful allies. But I am
not willing to admit that it is only in such situations that they are successful. Producer interest
is ceaselessly active, seeking to protect itself against competition and the incidence of
disagreeable change. The influence of the agrarian interest in Europe which, while tending to
keep down the real incomes of European consumers, has wrought such havoc among
agricultural producers overseas, has certainly not been confined to times of general
unemployment. Nor - to allot blame evenly all round - have the many abuses of the infant
industry argument on the part of manufacturing interests. Much attention nowadays is given
to the alleged influence on history of the struggles between different classes, conceived on a
social basis. In my judgment, a more realistic view would pay more attention to the struggles
of different groups organized on a producer basis. These were the first foes of Classical
liberalism and they may very well be the last.
(From The Economist in the Twentieth Century, by Lionel Robbins)
EDUCATION
The Personal Qualities of a Teacher
Here I want to try to give you an answer to the question: What personal qualities are desirable
in a teacher? Probably no two people would draw up exactly similar lists, but I think the
following would be generally accepted.
First, the teacher's personality should be pleasantly live and attractive. This does not rule out
people who are physically plain, or even ugly, because many such have great personal charm.
But it does rule out such types as the over-excitable, melancholy, frigid, sarcastic, cynical,
frustrated, and over-bearing: I would say too, that it excludes all of dull or purely negative
personality. I still stick to what I said in my earlier book: that school children probably 'suffer
more from bores than from brutes'.
www.seyfihoca.com
80
Secondly, it is not merely desirable but essential for a teacher to have a genuine capacity for
sympathy-in the literal meaning of that word; a capacity to tune in to the minds and feelings
of other people, especially, since most teachers are school teachers, to the minds and feelings
of children. Closely related with this is the capacity to be tolerant-not, indeed, of what is
wrong, but of the frailty and immaturity of human nature which induce people, and again
especially children, to make mistakes.
Thirdly, I hold it essential for a teacher to be both intellectually and morally honest. This does
not mean being a plaster saint. It means that he will be aware of his intellectual strengths, and
limitations, and will have thought about and decided upon the moral principles by which his
life shall be guided. There is no contradiction in my going on to say that a teacher should be a
bit of an actor. That is part of the technique of teaching, which demands that every now and
then a teacher should be able to put on an act -to enliven a lesson, correct a fault, or award
praise. Children, especially young children, live in a world that is rather larger than life.
A teacher must remain mentally alert. He will not get into the profession if of low
intelligence, but it is all too easy, even for people of above-average intelligence, to stagnate
intellectually and that means to deteriorate intellectually. A teacher must be quick to adapt
himself to any situation, however improbable (they happen!) and able to improvise, if
necessary at less than a moment's notice. (Here I should stress that I use 'he' and 'his'
throughout the book simply as a matter of convention and convenience.)
On the other hand, a teacher must be capable of infinite patience. This, I may say, is largely a
matter of self-discipline and self-training; we are none of us born like that. He must be pretty
resilient; teaching makes great demands on nervous energy. And he should be able to take in
his stride the innumerable petty irritations any adult dealing with children has to endure.
Finally, I think a teacher should have the kind of mind which always wants to go on learning.
Teaching is a job at which one will never be perfect; there is always something more to learn
about it. There are three principal objects of study: the subject, or subjects, which the teacher
is teaching; the methods by which they can best be taught to the particular pupils in the
classes he is teaching; and -by far the most important-the children, young people, or adults to
whom they are to be taught. The two cardinal principles of British education today are that
education is education of the whole person, and that it is best acquired through full and active
co-operation between two persons, the teacher and the learner.
(From Teaching as a Career, by H . C . Dent.)
Rousseau's Emile
It is not intended, even if it were desirable, to give a running commentary on the Emile. The
reader is advised to study it for himself and to read in conjunction the account of the
education of Julie's children described in Part V of The New Heloise. All that space permits is
a summary of some of the more important ideas that Rousseau contributed to educational
theory. Speaking of the Emile, Lord Morley described it as 'one of the seminal books in the
history of literature, and of such books the worth resides less in the parts than in the whole. It
touched the deeper things of character. It filled parents with a sense of the dignity and
moment of their task. It cleared away the accumulation of clogging prejudices and obscure
inveterate usage, which made education one of the dark formalistic arts. It admitted floods of
www.seyfihoca.com
81
light and air into the tightly closed nurseries and schoolrooms. It effected the substitution of
growth for mechanism . . . it was the charter of youthful deliverance.'
It is the last sentence of this passage which expresses the most important influence that
Rousseau exercised upon education. One may justly hail him as the discoverer of the child.
This is not to forget that educational thinkers of the ancient world, the mediaeval period, and
the Renaissance, had kindly, sympathetic, and helpful ideas about the training of children. In
their view the child did not come first. They fixed their eyes upon what he was to be in the
future and the curriculum they approved and the methods they recommended were coloured
by this attitude. Rousseau, on the other hand, emphasised that the prime factor to be
considered in education is the child and his present nature as a child. He wrote: 'Nature wants
children to be children before they are men. If we deliberately pervert this order, we shall get
premature fruits which are neither ripe nor well flavoured, and which soon decay. ...
Childhood has ways of seeing, thinking, and feeling, peculiar to itself; nothing can be more
foolish than to substitute our ways for them.'
So important did this principle seem to him that he repeated it in the Emile. In the Preface to
the Emile, he wrote: 'We know nothing of childhood; and with our mistaken notions, the
further we advance, the further we go astray. The wisest writers devote themselves to what a
man ought to know, without asking what a child is capable of learning. They are always
looking for the man in the child without considering what he is before he becomes a man....
Begin thus by making a more careful study of your scholars, for it is clear that you know
nothing about them.' He carried out this advice in the Emile by expressing the view that
education is a continuous process which begins at birth and which can be divided into four
consecutive stages. Thus Book I deals with the infant; Book II with childhood; Book III with
the pre-adolescent between the ages of twelve and fifteen; and Book IV with adolescence.
These stages of development correspond to the progress made in the history of the human
race. To all intents and purposes, the infant is living at the animal level. The child can be
compared with primitive man. Boyhood is a period of self-sufficiency, whilst at adolescence
the sex impulses ripen, social behaviour becomes possible, and the young person is able to
conduct his life according to moral principles.
It is an easy matter to criticise Rousseau's account of the child's development from the
standpoint of present-day child study, but it is essential to bear in mind that in the eighteenth
century he was a pioneer in this field. His exposition was rendered less effective by his
adherence to a faculty psychology, but in his defence, one could urge that this was the
dominant view of his time. The continuity of the child's development may seem to be broken
by the emphasis placed upon the emergence of new faculties, for example, at adolescence, the
pupil appears to make a complete break with his former life. One may dispute Rousseau's
statement that the most dangerous period of human life lies between birth and the age of
twelve, but when these partial criticisms have been made, the fact remains that he
concentrated attention upon the child and his nature rather than upon the subject and the
pupil's future occupation. This was one of the most revolutionary steps that educational theory
had so far taken.
(From A Short History of Educational Ideas, by S. J. Curtis & M. E. A. Boultwood.)
The Beginnings of Scientific and Technical Education
www.seyfihoca.com
82
The higher instruction given to workers was mainly concerned at first with science. As early
as 1760 a professor at Glasgow, named Anderson, had begun to hold evening classes in
science, which working men were encouraged to attend. In his will he left an endowment for a
chair of natural philosophy at the University. Its first occupant was George Birkbeck (17761841), who held a degree in medicine. When he started his lectures in 1799 he found it
necessary to have a good deal of apparatus, and while this was being made under his
instructions he became acquainted with a number of Glasgow artisans. He found them so
intelligent and so eager to learn that he resolved to start a course of lectures and experiments
in mechanics 'solely for persons engaged in the practical exercise of the mechanical arts, men
whose situation in early life has precluded the possibility of acquiring even the smallest
portion of scientific knowledge.' The lectures proved a great success. After Birkbeck removed
to London in 1804, the lectures were continued by the next occupant of the chair; and finally,
in 1823, the members of the class organised it into a 'Mechanics' Institute'. Its purpose was
defined as 'instructing artisans in the scientific principles of arts and manufactures'.
Mechanics' institutes soon sprang up in many parts of the country. They were supported by
subscriptions from the members and by donations from sympathisers. By 1845 there were 610
institutions, with 102,050 members. They were naturally most popular in the manufacturing
districts, such as London, Lancashire, and Yorkshire; but there were a few successful
institutes also in such rural centres as Lewes, Basingstoke, Chichester and Lincoln. Each
Institute usually included a library, reading-room, and museum of models and apparatus.
Lectures were provided on mathematics and its applications, and on natural and experimental
science and drawing. Sometimes literary subjects, such as English and foreign languages,
were included. Travelling lecturers and circulating boxes of books helped to keep the smaller
institutes in touch with one another.
The mechanics' institutes played an important part in English education, and yet they were
only partially successful. By 180 two changes had become noticeable. Their membership
consisted more of clerks and apprentices and middle-class people than of working men, for
whose benefit they had been founded; and, as a corollary of this, their syllabuses had tended
to change. There was less purely technical instruction and more recreational activities and
popular lectures. Discussions, debates, and even social courses tended to take the place of ad
hoc courses designed to help artisans. There were several reasons for this change. The artisans
and working classes had not yet received an elementary education, which would form an
adequate foundation on which to build a superstructure of technical education. Reference has
already been made to the meagre limits of education provided by the monitorial schools and
other elementary schools. It must also be remembered that some of the children of the poor
hardly went to school at all and that the average length of school life was in any case only one
and a half or two years. Moreover, a great obstacle to the spread of knowledge at this period
was the high cost of newspapers, owing to the Government duty: from 1819 to 1836 there was
a stamp duty of 4d. a copy. In a Poor Law Commissioners' Report of 1834 there occurs this
passage: 'The dearness of newspapers in this country is an insurmountable obstacle to the
education of the poor. I could name twenty villages within a circuit of a few miles in which a
newspaper is never seen from one year's end to another.' Again, the fees for membership and
classes in mechanics' institutes tended to be too high for those for whom they were originally
designed. At the London Mechanics' Institute in 1823 the annual subscription was fixed at £1,
and this seems to have been a fairly usual charge. In 1826 1,477 workmen paid this fee at the
London Institute: but it would be a rather high fee for people of that type even today, and it
must have been much more onerous in the Corn Law days, after the Napoleonic Wars, when
wages generally were low. Thus the mechanics' institutes tended to decline in importance and
www.seyfihoca.com
83
change in character. But some of them retained much of their original character and were
stimulated into new life by the development of technical education during the second half of
the nineteenth century. For example, the London Mechanics' Institute was the forerunner of
the present Birkbeck College, which caters for evening students but is a constituent part of the
University of London. In the broadest sense, the mechanics' institutes have laid the foundation
for the development of our modern technical schools and colleges.
(From A Short History of English Education, by H. C. Barnard.)
Supposed Mental Faculties and Their Training
The mind is commonly thought to consist of a number of 'Faculties'-such as memory,
observation, perception, reasoning, will, judgment, and so on, pretty much the same as those
described by the phrenologists who feel the 'bumps' on a man's head and describe his
capacities according to the chart of the skull which they have mapped out. It is supposed that
a man has a good (or bad) memory in general, and that the exercise of the memory, say on
history dates or Latin verbs, will strengthen the memory as a whole; and that a training in the
observation of wild flowers will sharpen the whole 'faculty of observation' whatever is to be
observed; or again that mathematics, since it exercises the 'faculty of reasoning', equally
improves reasoning about politics, social problems, and religion.
This view as to mental faculties is still widely held by most educated people who have not
studied Psychology, and it still has a harmful influence on education in some ways. Let us
consider this 'popular' psychology carefully, discussing in this chapter the supposed
intellectual 'Faculties'.
An example of error in popular views about the mind appears in the idea of a faculty of
observation. One often hears it said that we should train the observation of our pupils; and it is
imagined that by training them to observe certain things we are training them to observe
anything and everything. A method of instruction used in the Army, in which a man had to
observe quickly and remember a score of various objects in a tray, seems to have been based
on this idea.
One of my students once gave a lesson on Botany in the presence of an inspector of schools.
After the lesson, the inspector said to her: 'Yes: that was an interesting lesson, but what I want
to know is, are you training the pupil's powers of observation? would they, for example, be
able to tell you the colour of the tie I was wearing?' The inspector overlooked the fact that the
more the pupils had concentrated their attention on the lesson, the less would they be likely to
notice the colour of his tie; and that the more interested they were in the flowers studied, the
less would they be likely to attend to him or his personal appearance. (I should like to add that
this incident occurred a good many years ago. Inspectors are better informed nowadays on
psychological matters.)
Observation, in fact, depends on interest and knowledge. If three friends travel abroad, one an
architect, another a botanist, and the third a stockbroker travelling with them only to take a
'cure' abroad, and interested only in his health and moneymaking, then the architect is likely
to notice the style of houses and other buildings more than his friends do, because he is
specially interested in them. The botanist will observe especially the flowers and trees of the
country more than his friends; and he will actually see more details because he knows what to
look for. Observation is guided by knowledge, and prompted by interest. We have, however,
www.seyfihoca.com
84
no reason to suppose that the botanist, trained in such observation, or the architect, keenly
observant of the buildings, will be more observant than the stockbroker of the faces of the
foreign people they meet, or the dress of the women. Indeed, they are more likely to have
their attention diverted by the objects of their special interests. So training in the careful
observation of the varied endings of Latin words, or of the changes in chemical substances in
experiments, will have no effect on the observation of pictures or the movements of the stars.
These popular ideas about the mind and its faculties sometimes have the element of truth in
them which makes it all the harder for the psychologist to eliminate their exaggeration. For
example, as to observation: a careful training in observing plants under the microscope
includes a training in method, in the value of the precise description of what is actually seen
(and not merely what one thinks should be there), and so on: and a student with such
experience will gain something from it if he turns to a similar study, e.g. zoology or geology,
especially when he also uses the microscope. Here we see that the adoption of an ideal of
truth and exactitude in such work or a training in a method of procedure in one kind of work,
may result in its application in a similar kind of work though it does not always do so.
(From Psychology and its Bearing on Education, by C. W. Valentine.)
The Concept of Number
The long struggle forward in children's thinking comes out very clearly in the development of
their number ideas-the part of Piaget's work which is now best known in this country. It offers
a striking illustration both of the nature of his discoveries and of the basic pattern of mental
growth. We can watch how the child starts from a level of utter confusion, without a notion of
what number really means, even though he may be able to count up to ten or twenty; a level
where number is completely mixed up with size, shape, and arrangement, or constantly shifts
according to the way it is subdivided or added up. And we can see how, on an average two
years later, children declare of their own accord that a number must stay the same, whatever
you do with it, so long as you do not actually add to it or take away from it; or that whatever
you have done with it, you can always reverse this and get back to where you started from; or
that you can always show it to be the same by counting; and so on.
The following are a few examples of the ways in which Piaget's experiments bring out this
pattern of growth:
1. Each child was presented with two vessels of the same shape and size containing equal
quantities of coloured liquid. Then the contents of one of them was poured into (a) two similar
but smaller vessels, (b) several such, (c) a tall but narrow vessel, (d) a broad but shallow one.
In each case the child was asked whether the quantity of liquid was still the same as in the
untouched vessel.
Piaget found that at a first stage, around 4-5 years, children took it for granted that the
quantity of liquid was now different - either more because the level was higher, or more
because there were more glasses, or less because the new vessel was narrower, or less because
the levels in the two or more glasses were lower. In other words, there was no idea of a
constant quantity, independent of its changing forms; if its appearance changed, the quantity
changed and could become either more or less according to what aspect of the new
appearance caught the child's eye. At a second stage, at about 5½-6, children had reached a
transitional phase, in which they wavered uncertainly between the visual appearance and the
www.seyfihoca.com
85
dawning idea of conservation in their minds. Thus the quantity of liquid might be regarded as
still the same when it was poured into two smaller glasses, but as greater when it was poured
into three. Or as remaining the same if the difference in level or cross-section in the new
vessel was small, but not if it was larger. Or the child might try to allow for the relation
between cross section and level, and experiment uncertainly without reaching any clear
conclusion. In the third stage, 6½-8, children give the correct answers right away, either by
reference to the height width relation, or by pointing out that the quantity has not been
changed: 'It's only been poured out.'
2. As a check on these results, Piaget carried out a similar set of experiments, with beads
instead of liquids. In this way something closer to counting could be introduced (e.g. the child
putting beads into a container one by one as the experimenter did the same into another
vessel). Also he could be asked to imagine that the beads inside each vessel were arranged
into the familiar shape of a necklace. The outcome was entirely the same. At the first stage,
the children thought that the quantity of beads would be either more or less, according as the
level looked higher, or the width greater, or there were more vessels, and this happened even
when a child had put one bead into his vessel for each one that the experimenter placed in his.
At stage 2 there is a struggle in the child's mind as before. This may show itself for example
by his first going wrong when comparing levels between a wider and a taller vessel; then
correcting himself if asked to think in terms of the necklaces; but when the beads are spread
over two or more containers, still thinking that the necklace will be longer. At stage 3 once
more the children reply correctly and cannot be shaken, however the questions or the
experiments may be varied.
(From The Growth of Understanding in the Young Child, by Nathan Isaacs.)
English in the Primary School
The writing of their own stories was started from the very first, long before the children could
read. The thought came from them, and was interpreted by means of a picture. When the
picture was finished, each one told me what it was about, and I wrote for them a few words
about their picture. I have kept a set of books all by one child, between the ages of four and a
half and seven years, which make a most valuable study of infant progress.
Jeffery came to me at four, a bright little fellow with a mop of wonderful auburn curls and a
gift of romancing which one morning caused him to rush into school declaring that a sow had
just bitten his head off 'down by the church'. He showed skill in handling crayons and writing
tools straight away, so I made him a sewn book and began on an illustrated version of 'The
House that Jack Built', a rhyme which he knew. He drew all the pictures, and after each one
copied, in his fashion, the appropriate text. Next came a book called 'The Farmer', in which he
himself thought up a situation every day and drew a picture of it; then I came along, wrote for
him what he wanted to say, and he copied it. As soon as this was finished he demanded a new
book in which to write the 'story of my donkey-the one that ran away'. As he had never had a
donkey, I recognized his romantic imagination at work, and was delighted at the thought of it
being put to such good use. This book was full of the most wonderful, lively pictures and it
was evident that Jeff was often considerably irked by the necessity of having to wait for me to
write down what he wanted to say. He was by this time only a month or two over five years
old, but was already beginning to read and I used to find him struggling valiantly on bits of
odd paper to write for himself the next part of the tale.
www.seyfihoca.com
86
Then we reached the long summer holiday. After it, Jeff returned to me a much more mature
little boy, and though still only five and a half, was reading very well when we had been a
month or so into the new term. His new book was called 'Jeff the Brave Cowboy', and this
time he wrote for himself on a piece of paper. I corrected it, and the fair copy was made into
his book. This book was really remarkable for the sudden leap forward he made in his use of
English. It was vigorous and economical, with adjectives and adverbs appearing quite
naturally in the right places, but only where they were needed for extra emphasis. The
undoubted success of 'The Brave Cowboy' put Jeff's feet firmly on the road to authorship. He
was away on 'The Dog Fight' before the previous one was really finished, because, having
seen a dog fight on his way to school one morning, he simply could not wait to start putting
down his memory of it. This was a change from the purely imaginative work of his previous
creations and very good as English practice, for he discovered not only the value of
possessing powers of keen observation, but of knowing the language that allowed one to
record what one had noticed in detail.
It appeared that Jeff's mother, a glorious red-head like himself, had thrown a bucket of water
over the fighting dogs. While he was drawing a picture of this, giving his mother
unmistakable hair, he was thinking deeply about it, for he must have been wondering how to
record the mood of the moment in words. He came to me for advice. Although he was still
only five years old, I explained to him the use of conversation marks, thinking that that was
what he had asked. He was not satisfied.
'Does that make it say how Mummy said it?' he asked. 'She didn't just say "Stop it". She said
(miming the throwing of the water and stamping his foot) "STOP IT!" '
After Christmas Jeff, now six, wanted to write a pirate story. After reading some pirate books,
he started on 'Jeff and the Flying Bird', this time writing his own book without any correction
or other interference from me.
One day (Jeff) stood on the prow of his ship, he lookted in his telscope.
What did he see ?
He saw a ship sailing across the sea to Jeffs ship.
Attack shouted Jeff.
At seven he wrote 'Jeff the Space Man', text in pencil but pictures in coloured inks. At the end
of his first session with his new book, he came to me to show it. I was sitting at my desk, he
standing by the side of it. I took the book from him, admired the picture, and read aloud:
Jeff pulled the lever and pressed the button. The space ship, swished up into the air
and was gene
I happened to glance at the real Jeff, who was pressing with all his might an imaginary button
on the side of my desk, after which he held the pit of his stomach and gazed at the school
ceiling.
'Gosh,' he said, in an awed voice, 'can't you just hear it?' I could.
www.seyfihoca.com
87
'The ssspace ssship ssswissshed ....' Of course! After that Jeff went on exercising his
imagination, his observation, and his increasing command of English, in book after book after
book. His last was a treatise on freshwater fish and how to catch them. At eleven-plus his I.Q.
showed him to be only of good average intelligence. I mention this just to prove that he was
not the brilliantly academic type of child from whom one would expect this lively interest in
English as a matter of course.
(From An Experiment in Education, by Sybil Marshall.)
Britain and Japan: Two roads to higher education
Britain and Japan are the two great pioneers of industrialism and therefore of the modern
world. Britain was the pioneer industrial nation of the Western, European-dominated world,
Japan of the Eastern, non-European and, to many eyes, the hope of the Third World. The
countries have always had much in common. Both groups of islands off an initially more
civilized and powerful continent, they had to defend themselves from military and cultural
conquest and to that end developed a powerful navy and an independence of mind which
made them increasingly different from their continental neighbours. Behind their sea defences
they were able to pursue their own ideals and ambitions which enabled them in the end to
originate changes in industry and society which, because they brought wealth and power,
others tried to imitate. The British at the height of their imperial power and economic
domination recognized in the emerging Japanese a fellow pioneer and an ally. They called her
‘the Britain of the East’ and in the 1902 Treaty were the first to recognize Japan as a world
power.
Yet the two countries took utterly different roads to industrialism, power and wealth. Britain,
the first industrial nation, evolved slowly without knowing - because nobody then knew what
an Industrial Revolution was - where she was going or what the end of modernization would
be. Japan, coming late to economic growth in a world which saw and felt only too clearly
what the gains and dangers of industrialism were, adopted it self-consciously and with
explosive and revolutionary speed. And they still bear the marks of these differences of
origin, timing and approach. Britain had the first Industrial Revolution because she had the
right kind of society to generate it; but for that very reason she was riot forced to change her
society as much as later developing countries, and she now has the wrong kind of society for
sustaining a high and continuing rate of economic growth. That does not mean that she has the
wrong kind of society to provide a highly civilized and comfortable life for her people. On the
contrary, just as the British were pioneers of the industrial society dominated by a gospel of
work so they may now be the pioneers of the post-industrial society dedicated to the gospels
of leisure and welfare.
Japan on the other hand has astonished the world by the degree to which she was prepared to
change her society in order to industrialize, and the speed at which, in less than a hundred
years, she transformed herself from a feudal society of samurai, artisans and peasants into one
of the most efficient industrial and egalitarian meritocracies in the world. However, it must be
said that Tokugawa Japan was no ordinary feudal society, and had hidden advantages for
industrial development which most feudal societies lack: one of the most urbanized
populations in the world, with a highly educated ruling class of efficient bureaucrats, large
numbers of skilled craftsmen and sophisticated merchants, and a more literate populace than
most other countries, even in the West. But the fact remains that the leaders of the Meiji
Restoration were prepared to abolish feudalism, establish equality before the law, and make
www.seyfihoca.com
88
everyone, rich or poor, samurai, worker or peasant, contribute to the survival and
development of the country.
If the British and Japanese roads to industrialism were different, their two roads to higher
education were even more different. The British educational road was far slower, more
indirect and evolutionary even than their road to industrial development and, indeed, in the
early stages had little connection with economic growth. The ancient universities, particularly
in England as distinct from Scotland, had become little more than finishing schools for young
gentlemen, chiefly landowners’ sons and young clergymen. They did not conduct research
and not one of the major inventions of the early Industrial Revolution originated in a
university. Oxford and Cambridge were even less important to English society when the
Industrial Revolution began than they were over a century earlier - many of the ruling elite
came to find little to interest them in the repetition of classical Greek and Latin texts.
When Japan was beginning its great transformation under the Meiji, the main contribution of
the British universities to economic growth was still in the future. It may seem surprising that,
in relation to industrial development and modernization, British higher education in the late
19th century was no more advanced than the new Japanese system. By 1900 university
students in both Britain and Japan were less than one per cent of the student age group. In
both countries higher education was exclusively for the elite, but whereas in Britain the elite
graduates went predominantly into the home civil service, colonial government and the
traditional professions, in Japan they went not only into these but still more into industry and
commerce and the newer technological professions.
This was because Japanese higher education, like the whole modern education system, was
created by the Meiji reformers for the express purpose of modernizing Japan. Japan, contrary
to popular belief in the West, did not start from scratch. Under the Tokugawa there were
higher schools and colleges based on Confucian learning, no more out of touch with the needs
of a traditional ruling elite than were Oxford and Cambridge. But the difference was that the
Meiji elite knew that higher education had to be changed, and changed radically if Japan was
to be transformed into a modern nation able to expel the barbarians and become a strong and
wealthy country. Under the Fundamental Code of Education of 1872 they set out to establish
a modern system of education with an elementary school within reach of every child, a
secondary school in every school district, and a university in each of eight academic regions.
In the next forty years, Japanese higher education expanded explosively. By 1922 there were
6 imperial and 20 non-imperial universities and 235 other higher institutions. Moreover, the
whole system was geared to industrialization and economic growth, to the production of
bureaucrats, managers, technologists and technicians. Whereas in Britain the sons of the elite
at this stage avoided industry, which was left to the largely self-educated, trained in industry
itself, in Japan the sons of the Shizoku, the ex-samurai who formed the majority of students in
the universities, went indiscriminately into the service of the state and of private industry.
Britain too began a remarkable expansion of higher education in the late 19th century. New
universities, more responsive to the scientific and industrial needs of their regions, came into
existence in nearly all the great cities which did not already have them: Manchester, Leeds,
Liverpool, Birmingham, Bristol, Newcastle, Nottingham, Sheffield, and so on. These new
civic universities were much more dedicated to scientific and technological research and had a
provocative and stimulating effect on the older universities too, and Oxford and Cambridge
came to develop science and engineering and other modern subjects. Thus at the time Japan
was using higher education as an instrument of industrialization, Britain began to do the same.
www.seyfihoca.com
89
The road remained substantially different, however. Unlike the Japanese, the great majority of
British managers never went to university. Some went to a variety of technical colleges which
grew up to meet the demand which the universities had so long neglected, but the great
majority were trained on the job with the help of evening schools where they learned to pass
the examinations of the new professional bodies like the Institution of Mechanical Engineers
or the Institute of Chemistry.
Thus the British road to industrial higher education was largely a part-time road, Most modern
universities began as technical or other colleges, mostly for part-time students. This helps to
explain why Britain, with one of the smallest university systems amongst the advanced
countries, could sustain a competitive industrial economy, and even remain the world’s
largest exporter of manufactured goods down to the First World War.
During the 1960s the number of British universities nearly doubled, from 25 to 45 and in
addition 30 polytechnics were formed from existing technical colleges. But British industry
still depends to a larger extent than any other advanced country on part-time education and
training on the job.
Japan by contrast believes in full-time higher education, and has far larger numbers in
universities and colleges. Since the Second World War, initially under the stimulus of the
American Occupation, the system has grown from 64 universities and 352 other colleges with
about 250,000 students in l948 to 43l universities and 580 other colleges with nearly 2 million
students in 1977, equal to 38 per cent of the age group. In terms of full-time students Britain
is still only on the threshold of mass higher education; Japan is already moving towards
universal higher education.
Most educationists still believe that if only the British would spend as much on education as
the Japanese they could achieve the same rate of economic growth. But perhaps too much
influence is claimed for it by educationists. Could it not be said that education is an effect
rather than a cause - or rather, that it is an effect before it can become a cause. It is an effect of
the social values and social structure of the society which creates and provides it.
In other words, the British and the Japanese took two different roads to higher education and
to modern industrialism because they were two different kinds of society; with different aims
and ambitions, different moral and social values, different principles of social connexion and
of social structure.
In one sense the aims and objectives of the two societies were very similar. They both
harnessed personal ambition to the drive to wealth and power. The key to the British
Industrial Revolution was social ambition, ‘keeping up with the ‘Joneses,’ the desire to rise in
society by making money and to confirm that rise by spending money on conspicuous
consumer goods. In a similar way, the Japanese of the Meiji Restoration strove to become rich
and powerful in order to expel the barbarians and restore the country’s independence. The two
kinds of ambition were fundamentally different. The British landlords, farmers, industrialists
and workers were self-seeking and individualistic in their ambition, and national economic
growth was a by-product of their individual success. The Japanese samurai, merchants,
artisans, and peasants strove to succeed, but success was not measured as much by personal
wealth, status and power, as by the admiration and applause of one’s family, one’s colleagues
and one’s fellow citizens. Individual prosperity was a by-product of the group’s. The British
www.seyfihoca.com
90
(and Western) kind of ambition may be called ‘acquisitive individualism’ and the Japanese
kind ‘participative self-fulfilment’.
Acquisitive individualism in Britain has deep roots in English society. Even now the British
are more concerned with their individual share of the national cake than with increasing its
size.
The Japanese by contrast have never been individualists in this sense. They have always put
the group - the family, the village, the feudal han, the nation - before the individual and his
material gain. The individual has found his reward in and through the group and in loyalty to
its leader, who represents the group to the outside world.
This ideal of participative self-fulfilment has deep roots in Japanese society and goes back to
the nature of the Japanese family, the ie. In Western terms, ie is best translated as ‘household’
rather than ‘family’, since it was more open to newcomers such as sons-in-law than the
Western family, and outgoing members who married into another ie ceased to belong. Its
major feature was that every member, however new, strove for respect in the eyes or the
household and received protection and loyalty in return. This was the origin of that
participative self-fulfilment, that striving for success in and through whichever group one
came to belong to, which is the secret of Japanese unselfish ambition and co-operative
loyalty.
Yet there are limits to the group responsibility produced by the ie tradition. Because it was
rooted in a system of group rivalries which drew a sharp distinction between one’s own group
and all the others - which is why it is difficult, for example, to unite workers from different
companies in the same trade union - there is less sense of responsibility in Japan for the
welfare of those who do not belong to the same group. That is why welfare, social security,
pensions, medical services and leisure facilities are mainly organized by the large
corporations for their own workers, and the state welfare system is still undeveloped
compared with Britain and Europe.
Britain, despite its acquisitive individualism, always had another tradition, an aristocratic
tradition of paternalism or ‘noblesse oblige’ which, oddly enough, remained enshrined in the
older, aristocratic universities of Oxford and Cambridge while acquisitive individualism was
creating or capturing the newer universities of the industrial society. This tradition found its
way into the university settlement movement in the slums of London and other great cities,
into housing improvement and factory reform, into adult education for the working class, into
social work, even into British non- Marxist socialism, and into the welfare state. It was a
tradition which went beyond all groups, whether of family, trade, profession or class. It asked
in effect, ‘who is my neighbour?’ and it answered zany member of society who needs my
help’. This is the hidden principle which has saved Britain from the excesses of acquisitive
individualism.
Although British trade unions, employers and professional bodies today fight each other for a
bigger share of the cake regardless of what happens to the cake as a whole, there is a
gentleman’s agreement, stemming from that other, more gentlemanly tradition, that the
welfare of the poor, the disabled, the elderly, the sick and the unemployed comes first. For the
same reason, economic growth comes second. Welfare, leisure, a clean environment and a
civilized social life are now more important acquisitions to the British than a larger post-tax
income. Acquisitive individualism has shifted its ground, from material possessions to the
www.seyfihoca.com
91
non-material goods of health, education for pleasure, an enjoyable environment and a more
leisurely and pleasurable life.
Britain and Japan took two different roads to higher education and to industrialism because
they were two very different societies with different social structures and ideals. If the British
could borrow some of their unselfish self-fulfilment and co-operative efficiency from the
Japanese and the Japanese some of their concern for social welfare and public amenity from
the British, perhaps East and West could at last meet in mutual under- standing and each have
the best of both worlds.
Enrolments in universities and colleges in Britain and Japan as percentage of the student age
group
Britain
Japan
1885
(1.0)
0.5
1900
1.2
0.9
1920
1.8
1924
2.7
1930
3.1
1938
2.7
1940
4.0
1955
6.1
8.7
1960
8.3
10.2
1965
8.9
15.9
1970
13.8
18.7
1974
14.0
27.9
1977
33.9
1979
13.9
Note:
The British figures include full-time advanced students in universities, teacher training
colleges and further education; the Japanese figures those in universities, two-year colleges
and higher technical colleges (and excluding higher vocational schools).
What type of student do you have to teach?
Most lecturers try to help students develop their understanding. But understanding a foreign
language is not the same as understanding why someone is upset or understanding
electromagnetism or understanding history. It is not to be expected therefore that the same
teaching methods will be appropriate to these different kinds of understanding.
Most forms of understanding are expressed by concepts which differ from everyday ones. For
example, we all know that suitcases get heavier the longer you carry them, but in science this
is described in terms of constant weight plus increasing fatigue. The concept “weight” is
introduced and laid alongside the commonsense concept of “heaviness’. Similarly we all
know that time passes quickly when we are absorbed and slowly when we are bored, but
science tells us that this is an illusion; time really ticks away at a steady rate. Note that
conceptual change should not be the aim, as is sometimes suggested, since people still also
www.seyfihoca.com
92
need their common sense. The aim is to add new sets of concepts and to explain when to use
which set.
But “understanding” is not the only kind of learning which students need to master.
Instruction, demonstration and error-correction are the key teaching activities - which are
quite different from those needed to reach understanding - while practice is the main learning
activity.
Students also have to memorise information and be able to recall it when required, as well as
acquire several other kinds of learning (such as know-how and attitudes and values) each of
which calls for different teaching methods. So learning-centred teaching includes a conscious
matching of teaching methods to the intended kind of learning.
While good teaching involves, among other things, helping students to achieve their chosen
learning goals, the picture is further complicated by the different learning styles adopted by
different groups of students.
Many ways of categorisation and modelling students as learners have been suggested, of
which the following are as useful as any, particularly in connection with understanding.
(Differences between learners’ natural learning styles are not so significant when skills are
being taught, since the appropriate style is determined more by the activity involved than by
students’ natural capabilities.)
Some students are “holists”: which means they like to take an overview of a subject first and
then fill in the details and concepts in their own way.
Others are “serialists” who like to follow a logical progression of a subject, beginning at the
beginning. Educational researcher Gordon Pask structured some teaching materials in both a
holist and a serialist manner, and then tested previously-sorted cohorts of students using them.
He found that the best performance of those who were mismatched (i.e. holist students with
serialist material, and vice versa) was worse than the worst performance of those who were
matched to the learning materials.
This seems to imply, for example, that educational textbooks - which are naturally serialist in
character - should include signposts, summaries, alternative explanations of difficult concepts,
explanatory figure captions, a glossary of terms, a good index, etc, to help holist students find
their own way through them. Similarly projects, which are naturally holist in character, since
they are usually specified in terms of a final goal, can cause problems for serialists, who may
therefore need step-by-step guidance.
Another group of students are “visualisers” whose learning is helped by the inclusion of
diagrams, pictures, flow-charts, films, etc. Others are “verbalisers” and prefer to listen, read,
discuss, argue, attend tutorials and write during their conceptual development. And some are
“doers” and find that overt practical activity is best. The saying that “to hear is to forget, to
see is to remember, but to do is to understand” is only true for “doers”. With a typical mix of
students, attempts should be made to cater for each preferred style.
It is well known nowadays that for the development of “understanding” and for the
memorisation of information it is important that students adopt a “deep approach” to their
learning, rather than a “surface approach’. The deep approach refers to an intention to develop
www.seyfihoca.com
93
their understanding and to challenge ideas, while the “surface approach” is the intention to
memorise information and to follow instructions. Although students are naturally inclined
towards one approach rather than the other - often with a new subject the inclination is
towards the surface approach - this can vary from subject to subject and can usually be
changed by the teaching they receive. Overloading, for example, will encourage the surface
approach; stimulating interest may encourage the deep approach. Given the deep approach,
even good lectures can make a considerable contribution to students’ “understanding”.
Recently the need to encourage the deep approach in students has been allowed to dominate
the choice of teaching method, sometimes at the expense of effective teaching.
Constructivism in science teaching, for example, in which students are encouraged to devise
their own explanations of phenomena, certainly tends to encourage the deep approach, but it
can also leave students with misconceptions. Similarly, though problem-based learning is
usually popular with students, it teaches “know-how” rather than “understanding”: unless
explicit conceptual guidance is also given.
The fact that students have different preferred learning styles also has important implications
for course evaluation through feedback. It often seems to be assumed that students are a
homogeneous bunch and that therefore a majority opinion condemning a certain aspect of a
course justifies changing it for the future. But this can well be a mistake. If a course is well
matched, say, to “holist verbalisers” it is unlikely to be found very helpful to “serialist
visualisers”. In other words, feedback is likely to reveal as much about the students as about
the course or lecturer, and can be quite misleading unless it is properly analysed in terms of
the preferred learning styles of the particular cohort of students.
Indeed, student feedback about the teaching of “understanding” can, in any case, be quite
misleading, since students cannot be expected to judge what has been helpful to them until
much of the necessary conceptual development has occurred. Only after “the penny has
dropped” is such feedback likely to be reliable. Similarly, favourable feedback about the
necessary but tedious practising of important “skills” cannot normally be expected.
These considerations are all aspects of learning-centred teaching, with which all lecturers
should, in due course, become familiar. Innovation in education without taking these matters
into consideration is at best cavalier, at worst irresponsible, for it is the students who suffer
from teachers’ ill-founded experiments.
John Sparkes, Times Higher Education Supplement. February 6th, 1998.
Spoon-fed feel lost at the cutting edge
Before arriving at university students will have been powerfully influenced by their school’s
approach to learning particular subjects. Yet this is only rarely taken into account by teachers
in higher education, according to new research carried out at Nottingham University, which
could explain why so many students experience problems making the transition.
Historian Alan Booth says there is a growing feeling on both sides of the Atlantic that the
shift from school to university-style learning could be vastly improved. But little consensus
exists about who or what is at fault when the students cannot cope. “School teachers
commonly blame the poor quality of university teaching, citing factors such as large first-year
lectures, the widespread use of inexperienced postgraduate tutors and the general lack of
www.seyfihoca.com
94
concern for students in an environment where research is dominant in career progression,” Dr
Booth said.
Many university tutors on the other hand claim that the school system is failing to prepare
students for what will be expected of them at university. A-level history in particular is seen
to be teacher-dominated, creating a passive dependency culture.
But while both sides are bent on attacking each other, little is heard during such exchanges
from the students themselves, according to Dr Booth, who has devised a questionnaire to test
the views of more than 200 first-year history students at Nottingham over a three-year period.
The students were asked about their experience of how history is taught at the outset of their
degree programme. It quickly became clear that teaching methods in school were pretty staid.
About 30 per cent of respondents claimed to have made significant use of primary sources
(few felt very confident in handling them) and this had mostly been in connection with project
work. Only 16 per cent had used video/audio; 2 per cent had experienced field trips and less
than 1 per cent had engaged in role-play.
Dr Booth found students and teachers were frequently restricted by the assessment style
which remains dominated by exams. These put obstacles in the way of more adventurous
teaching and active learning, he said. Of the students in the survey just 13 per cent felt their
A-level course had prepared them very well for work at university. Three-quarters felt it had
prepared them fairly well.
One typical comment sums up the contrasting approach: “At A-level we tended to be spoonfed with dictated notes and if we were told to do any background reading (which was rare) we
were told exactly which pages to read out of the book”.
To test this further the students were asked how well they were prepared in specific skills
central to degree level history study. The answers reveal that the students felt most confident
at taking notes from lectures and organising their notes. They were least able to give an oral
presentation and there was no great confidence in contributing to seminars, knowing how
much to read, using primary sources and searching for texts. Even reading and taking notes
from a book were often problematic. Just 6 per cent of the sample said they felt competent at
writing essays, the staple A level assessment activity.
The personal influence of the teacher was paramount. In fact individual teachers were the
centre of students’ learning at A level with some 86 per cent of respondents reporting that
their teachers had been more influential in their development as historians than the students’
own reading and thinking.
The ideal teacher turned out to be someone who was enthusiastic about the subject; a good
clear communicator who encouraged discussion. The ideal teacher was able to develop
students involvement and independence. He or she was approachable and willing to help. The
bad teacher, according to the survey, dictates notes and allows no room for discussion. He or
she makes students learn strings of facts; appears uninterested in the subject and fails to listen
to other points of view.
www.seyfihoca.com
95
No matter how poor the students judged their preparedness for degree-level study, however,
there was a fairly widespread optimism that the experience would change them significantly,
particularly in terms of their open mindedness and ability to cope with people.
But it was clear, Dr Booth said, that the importance attached by many departments to thirdyear teaching could be misplaced. “Very often tutors regard the third year as the crucial time,
allowing postgraduates to do a lot of the earlier teaching. But I am coming to the conclusion
that the first year at university is the critical point of intervention”.
Alison Utley, Times Higher Education Supplement. February 6th, 1998.
GEOLOGY/GEOGRAPHY
The Age of the Earth
The age of the earth has aroused the interest of scientists, clergy, and laymen. The first
scientists to attack the problem were physicists, basing their estimates on assumptions that are
not now generally accepted. G. H. Darwin calculated that 57 million years had elapsed since
the moon was separated from the earth, and Lord Kelvin estimated that 20 - 40 million years
were needed for the earth to cool from a molten condition to its present temperature. Although
these estimates were much greater than the 6,000 years decided upon some two hundred years
earlier from a Biblical study, geologists thought the earth was much older than 50 or 60
million years. In 1899 the physicist Joly calculated the age of the ocean from the amount of
sodium contained in its waters. Sodium is dissolved from rocks during weathering and carried
by streams to the ocean. Multiplying the volume of water in the ocean by the percentage of
sodium in solution, the total amount of sodium in the ocean is determined as 16 quadrillion
tons. Dividing this enormous quantity by the annual load of sodium contributed by streams
gives the number of years required to deposit the sodium at the present rate. This calculation
has been checked by Clark and by Knopi with the resulting figure in round numbers of
1,000,000,000 years for the age of the ocean. This is to be regarded as a minimum age for the
earth, because all the sodium carried by streams is not now in the ocean and the rate of
deposition has not been constant. The great beds of rock salt (sodium chloride), now stored as
sedimentary rocks on land, were derived by evaporation of salt once in the ocean. The annual
contribution of sodium by streams is higher at present than it was in past geological periods,
for sodium is now released from sedimentary rocks more easily than it was from the silicates
of igneous rocks before sedimentary beds of salt were common. Also, man mines and uses
tons of salt that are added annually to the streams. These considerations indicate that the
ocean and the earth have been in existence much longer than 1,000,000,000 years, but there is
no quantitative method of deciding how much the figure should be increased.
Geologists have attempted to estimate the length of geologic time from the deposition of
sedimentary rocks. This method of measuring time was recognized about 450 B.C. by the
Greek historian Herodotus after observing deposition by the Nile and realizing that its delta
was the result of repetitions of that process. Schuchert has assembled fifteen such estimates of
the age of the earth ranging from 3 to 1,584 million years with the majority falling near 100
million years. These are based upon the known thicknesses of sedimentary rocks and the
average time required to deposit one foot of sediment. The thicknesses as well as the rates of
deposition used by geologists in making these estimates vary. Recently Schuchert has
compiled for North America the known maximum thicknesses of sedimentary rocks deposited
since the beginning of Cambrian time and found them to be 259,000 feet, about 50 miles. This
www.seyfihoca.com
96
thickness may be increased as other information accumulates, but the real difficulty with the
method is to decide on a representative rate of deposition, because modern streams vary
considerably in the amount of sediment deposited. In past geological periods the amount
deposited may have varied even more, depending on the height of the continents above sea
level, the kind of sediment transported, and other factors. But even if we knew exact values
for the thickness of PreCambrian and PostCambrian rocks and for the average rate of
deposition, the figure so obtained would not give us the full length of time involved. At many
localities the rocks are separated by periods of erosion called unconformities, during which
the continents stood so high that the products of erosion were carried beyond the limits of the
present continents and “lost intervals” of unknown duration were recorded in the depositional
record. It is also recognized that underwater breaks or diastems caused by solution due to
acids in sea water and erosion by submarine currents may have reduced the original thickness
of some formations. Geologists appreciated these limitations and hoped that a method would
be discovered which would yield convincing evidence of the vast time recorded in rocks.
Unexpected help came from physicists studying the radioactive behavior of certain heavy
elements such as uranium, thorium, and actinium. These elements disintegrate with the
evolution of heat and give off particles at a constant rate that is not affected by high
temperatures and great pressures. Helium gas is liberated, radium is one of the intermediate
products, and the stable end product is lead with an atomic weight different from ordinary
lead. Eight stages have been established in the radium disintegration series, in which elements
of lower atomic weights are formed at a rate which has been carefully measured. Thus,
uranium with an atomic weight of 238 is progressively changed by the loss of positively
charged helium atoms each having an atomic weight of 4 until there is formed a stable
product, uranium lead with an atomic weight of 206. Knowing the uranium-lead ratio and the
rate at which atomic disintegration proceeds, it is possible to determine the time when the
uranium mineral crystallized and the age of the rock containing it. By this method the oldest
rock, which is of Archeozoic age, is 1,850,000,000 years old, while those of the latest
Cambrian are 450,000,000 years old. Allowing time for the deposition of the early Cambrian
formations, the beginning of the Paleozoic is estimated in round numbers at 500,000,000
years ago. This method dates the oldest intrusive rock thus far found to contain radioactive
minerals. But even older rocks occur on the earth’s surface, for they existed when these
intrusions penetrated them. How much time should be assigned to them, we have no accurate
way of judging. Recently attention has centered upon the radio activity of the isotopes of
potassium, which disintegrate into calcium with an atomic weight of 40 instead of 40.08 of
ordinary calcium. On this basis A. K. Brewer has calculated the age of the earth at not more
than 2,500,000,000 years, but there is some question that this method has the same order of
accuracy as the uranium-lead method. Geologists are satisfied with the time values now
allotted by physicists for the long intervals of mountain-making, erosion, and deposition by
which the earth gradually reached its present condition.
DIVISIONS OF GEOLOGICAL TIME
The rocks of the accessible part of the earth are divided into five major divisions or eras,
which are in the order of decreasing age, Archeozoic, Proterozoic, Paleozoic, Mesozoic, and
Cenozoic. Superposition is the criterion of age. Each rock is considered younger than the one
on which it rests, provided there is no structural evidence to the contrary, such as overturning
or thrust faulting. As one looks at a tall building there is no doubt in the mind of the observer
that the top story was erected after the one on which it rests and is younger than it in order of
time. So it is in stratigraphy in which strata are arranged in an orderly sequence based upon
www.seyfihoca.com
97
their relative positions. Certainly the igneous and metamorphic rocks at the bottom of the
Grand Canyon are the oldest rocks exposed along the Colorado River in Arizona and each
successively higher formation is relatively younger than the one beneath it. The rocks of the
Mississippi Valley are inclined at various angles so that successively younger rocks overlap
from Minnesota to the Gulf of Mexico. Strata are arranged in recognizable groups by
geologists utilizing a principle announced by William Smith in 1799. While surveying in
England Smith discovered that fossil shells of one geological formation were different from
those above and below. Once the vertical range and sequence of fossils are established the
relative position of each formation can be determined by its fossil content. By examining the
succession of rocks in various parts of the world it was found that the restriction of certain life
forms to definite intervals of deposition was world wide and occurred always in the same
order. Apparently certain organisms lived in the ocean or on the land for a time, then became
extinct and were succeeded by new forms that were usually higher in their development than
the ones whose places they inherited. Thus, the name assigned to each era implies the stage of
development of life on the earth during the interval in which the rocks accumulated. The eras
are subdivided into periods, which are grouped together in to indicate the highest forms of life
during that interval. As the rocks of increasingly younger periods are examined higher types
of life appear in the proper order, invertebrates, fish, amphibians, reptiles, mammals, man.
From this it is evident that certain fossil forms limited to a definite vertical range may be used
as index fossils of that division of geological time. Also, in this table are given for each era
estimates of the beginning, duration, and thickness of sediments, based largely upon a report
of a Committee of the National Research Council on the Age of the Earth. At the close of and
within each era widespread mountain-making disturbances or revolutions took place, which
changed the distribution of land and sea and affected directly or indirectly the life of the sea
and the land. The close of the Paleozoic era brought with it the rise of the Appalachian
Mountains. It has been estimated that only 3 per cent of the Paleozoic forms of life survived
and lived on into the Mesozoic era. The birth of the Rocky Mountains at the close of the
Mesozoic was accompanied by widespread destruction of reptilian life. Faunal successions
responded noticeably to crustal disturbances.
UNCONFORMITIES. In subdividing rocks geologists have been guided by the periods of
erosion resulting from extensive mountain construction. Uplift of the continents causes the
shallow seas to withdraw from land thereby deepening the ocean and allowing erosion to start
on the evacuated land areas. Since all the oceans are connected, sea level throughout the
world was affected in many instances, leaving a record of crustal movements in the
depositional history of each of the continents. At many places the rocks of one era are
separated from those of another by unconformities or erosion intervals, in which miles of
rocks were eroded from the crests of folds before sedimentation was resumed on the truncated
edges of the mountain structure. There are four stages in the development of an angular
unconformity, so named because there is an angular difference between the bedding of the
lower series and that of the overlying series. If the series above and below an unconformity
consist of marine formations, four movements of the area relative to sea level took place. In
stage 1 the sandstones and shales comprise a conformable marine series, which was laid down
by continuous deposition with the bedding of one formation conforming to the next. We have
seen that the deposition of 24,000 feet of sediment requires repeated sinking of the area below
sea level. In stage 2 the region was folded and elevated above sea level, so that erosion could
take place. Since erosion starts as soon as the land develops an effective slope for corrosion,
there is no proof that this structure ever stood 24,000 feet high. But, the evidence is clear that
24,000 feet were eroded to produce the flat surface, shown in stage 3. In order that the over
lying marine series could be deposited the area had to be again submerged below sea level.
www.seyfihoca.com
98
Since the region now stands above sea level, a fourth movement is necessary. In some cases
crustal movement does not tilt or fold the beds, but merely elevates horizontal strata so that
erosion removes material and leaves an irregular surface on which sedimentation may be
resumed with the deposition of an overlying formation parallel to the first. An erosion interval
between parallel formations is a disconformity. But not all unconformities and
disconformities are confined to the close of eras. Local deformation and uplift caused erosion
between formations within the same era and within the same period. In the Grand Canyon
region Devonian rocks rest on the eroded surface of Cambrian formations. At other North
American and European localities Ordovician and Silurian rocks occupy this interval, so that
the disconformity within the Paleozoic era at this locality represents two whole periods. It is
only by carefully tracing the sequence of rocks of one region into another that the immensity
of geological time can be appreciated from stratigraphy.
Oils
There are three main groups of oils: animal, vegetable and mineral. Great quantities of animal
oil come from whales, those enormous creatures of the sea which are the largest remaining
animals in the world. To protect the whale from the cold of the Arctic seas, nature has
provided it with a thick covering of fat called blubber. When the whale is killed, the blubber is
stripped off and boiled down, either on board ship or on shore. It produces a great quantity of
oil which can be made into food foör human consumption. A few other creatures yield oil, but
none so much as the whale. The livers of the cod and the halibut, two kinds of fish, yield
nourishing oil. Both cod liver oil and halibut liver oil are given to sick children and other
invalids who need certain vitamins. These oils may be bought at any chemist’s.
Vegetable oil has been known from antiquity. No household can get on without it, for it is
used in cooking. Perfumes may be made from the oils of certain flowers. Soaps are made
from vegetable and animal oils.
To the ordinary man, one kind of oil may be as important as another. But when the politician
or the engineer refers to oil, he almost always means mineral oil, the oil that drives tanks,
aeroplanes and warships, motor-cars and diesel locomotives; the oil that is used to lubricate
all kinds of machinery. This is the oil that has changed the life of the common man. When it is
refined into petrol it is used to drive the internal combustion engine. To it we owe the
existence of the motorcar, which has replaced the private carriage drawn by the horse. To it
we owe the possibility of flying. It has changed the methods of warfare on land and sea. This
kind of oil comes out of the earth. Because it burns well, it is used as fuel and in some ways it
is superior to coal in this respect. Many big ships now burn oil instead of coal. Because it
burns brightly, it is used for illumination; countless homes are still illuminated with oilburning lamps. Because it is very slippery, it is used for lubrication. Two metal surfaces
rubbing together cause friction and heat; but if they are separated by a thin film of oil, the
friction and heat are reduced. No machine would work for long if it were not properly
lubricated. The oil used for this purpose must be of the correct thickness; if it is too thin it will
not give sufficient lubrication, and if it is too thick it will not reach all parts that must be
lubricated.
The existence of oil wells has been known for a long time. Some of the Indians of North
America used to collect and sell the oil from the wells of Pennsylvania. No one, however,
seems to have realised the importance of this oil until it was found that paraffin-oil could be
www.seyfihoca.com
99
made from it; this led to the development of the wells and to the making of enormous profits.
When the internal combustion engine was invented, oil became of worldwide importance.
What was the origin of the oil which now drives our motor-cars and air-craft? Scientists are
confident about the formation of coal, but they do not seem so sure when asked about oil.
They think that the oil under the surface of the earth originated in the distant past, and was
formed from living things in the sea. Countless billions of minute sea creatures and plants
lived and sank to the sea bed. They were covered with huge deposits of mud; and by
processes of chemistry, pressure and temperature were changed through long ages into what
we know as oil. For these creatures to become oil, it was necessary that they should be
imprisoned between layers of rock for an enormous length of time. The statement that oil
originated in the sea is confirmed by a glance at a map showing the chief oilfields of the
world; very few of them are far distant from the oceans of today. In some places gas and oil
come up to the surface of the sea from its bed. The rocks in which oil is found are of marine
origin too. They are sedimentary rocks, rocks which were laid down by the action of water on
the bed of the ocean. Almost always the remains of shells, and other proofs of sea life, are
found close to the oil. A very common sedimentary rock is called shale, which is a soft rock
and was obviously formed by being deposited on the sea bed. And where there is shale there
is likely to be oil.
Geologists, scientists who study rocks, indicate the likely places to the oil drillers. In some
cases oil comes out of the ground without any drilling at all and has been used for hundreds of
years. In the island of Trinidad the oil is in the form of asphalt, a substance used for making
roads. Sir Walter Raleigh visited the famous pitch lake of Trinidad in 1595; it is said to
contain nine thousand million tons of asphalt. There are probably huge quantities of crude oil
beneath the surface.
The king of the oilfield is the driller. He is a very skilled man. Sometimes he sends his drill
more than a mile into the earth. During the process of drilling, gas and oil at great pressure
may suddenly be met, and if this rushes out and catches fire the oil well may never be brought
into operation at all. This danger is well known and steps are always taken to prevent it.
There is a lot of luck in drilling for oil. The drill may just miss the oil although it is near; on
the other hand, it may strike oil at a fairly high level. When the drill goes down, it brings up
soil. The samples of soil from various depths are examined for traces of oil. If they are
disappointed at one place, the drillers go to another. Great sums of money have been spent,
for example in the deserts of Egypt, in ‘prospecting’ for oil. Sometimes little is found. When
we buy a few gallons of petrol for our cars, we pay not only the cost of the petrol, but also
part of the cost of the search that is always going on.
When the crude oil is obtained from the field, it is taken to the refineries to be treated. The
commonest form of treatment is heating. When the oil is heated, the first vapours to rise are
cooled and become the finest petrol. Petrol has a low boiling point; if a little is poured into the
hand, it soon vaporizes. Gas that comes off the oil later is condensed into paraffin. Last of all
the lubricating oils of various grades are produced. What remains is heavy oil that is used as
fuel.
There are four main areas of the world where deposits of oil appear. The first is that of the
Middle East, and includes the regions near the Caspian Sea, the Black Sea, the Red Sea and
www.seyfihoca.com
100
the Persian Gulf. Another is the area between North and South America, and the third,
between Asia and Australia, includes the islands of Sumatra, Borneo and Java.
The fourth area is the part near the North Pole. When all the present oil-fields are exhausted, it
is possible that this cold region may become the scene of oil activity. Yet the difficulties will
be great, and the costs may be so high that no company will undertake the work. If progress in
using atomic power to drive machines is fast enough, it is possible that oil-driven engines may
give place to the new kind of engine. In that case the demand for oil will fall, the oilfields will
gradually disappear, and the deposits at the North Pole may rest where they are for ever.
(From Power and Progress by G. C. Thornley (Longman))
The use of land
A very important world problem - in fact, I am inclined to say it is the most important of all
the great world problems which face us at the present time - is the rapidly increasing pressure
of population on land and on land resources.
It is not so much the actual population of the world but its rate of increase which is important.
It works out to be about 1.6 per cent per annum net increase. In terms of numbers this means
something like forty to fifty-five million additional people every year. Canada has a
population of twenty million - rather less than six months' climb in world population. Take
Australia. There are ten million people in Australia. So, it takes the world less than three
months to add to itself a population which peoples that vast country. Let us take our own
crowded country - England and Wales: forty-five to fifty million people - just about a year's
supply.
By this time tomorrow, and every day, there will be added to the earth about 120,000 extra
people - just about the population of the city of York.
I am not talking about birth rate. This is net increase. To give you some idea of birth rate, look
at the seconds hand of your watch. Every second three babies are born somewhere in the
world. Another baby! Another baby! Another baby! You cannot speak quickly enough to keep
pace with the birth rate.
This enormous increase of population will create immense problems. By A.D. 2000, unless
something desperate happens, there will be as many as 7,000,000,000 people on the surface of
this earth! So this is a problem which you are going to see in your lifetime.
Why is this enormous increase in population taking place? It is really due to the spread of the
knowledge and the practice of what is coming to be called Death Control. You have heard of
Birth Control? Death Control is something rather different. Death Control recognizes the
work of the doctors and the nurses and the hospitals and the health services in keeping alive
people who, a few years ago, would have died of some of the incredibly serious killing
diseases, as they used to be. Squalid conditions, which we can remedy by an improved
standard of living, caused a lot of disease and dirt. Medical examinations at school catch
diseases early and ensure healthier school children. Scientists are at work stamping out
malaria and other more deadly diseases. If you are seriously ill there is an ambulance to take
you to a modern hospital. Medical care helps to keep people alive longer. We used to think
seventy was a good age; now eighty, ninety, it may be, are coming to be recognized as a
www.seyfihoca.com
101
normal age for human beings. People are living longer because of this Death Control, and
fewer children are dying, so the population of the world is shooting up.
Imagine the position if you and I and everyone else living on earth shared the surface between
us. How much should we have each? It would be just over twelve acres - the sort of size of a
small holding. But not all that is useful land which is going to produce food. We can cut out
one-fifth of it, for example, as being too cold. That is land which is covered with ice and snow
- Antarctica and Greenland and the great frozen areas of northern Canada. Then we can cut
out another fifth as being too dry - the great deserts of the world like the Sahara and the heart
of Australia and other areas where there is no known water supply to feed crops and so to
produce food. Then we can cut out another fifth as being too mountainous or with too great an
elevation above sea level. Then we can cut out another tenth as land which has insufficient
soil, probably just rock at the surface. Now, out of the twelve acres only about four are left as
suitable for producing food. But not all that is used. It includes land with enough soil and
enough rainfall or water, and enough heat which, at present, we are not using, such as, for
example, the great Amazon forests and the Congo forest and the grasslands of Africa. How
much are we actually using? Only a little over one acre is what is required to support one
human being on an average at the present time.
Now we come to the next point, and that is, the haves and the have-nots amongst the countries
of the world. The standard share per person for the world is a little over twelve acres per head;
potentially usable, about four acres; and actually used about 1.1 acre. We are very often told
in Britain to take the United States as an example of what is done or what might be done.
Every little American is born into this world with a heritage of the home country, the
continental United States, of just about the world average - about twelve acres. We can
estimate that probably some six acres of the total of twelve of the American homeland is
cultivable in the sense I have just given you. But the amount actually used - what the
Americans call 'improved land' in crops and pasture on farms - is three and a half acres. So the
Americans have over three times the world average of land on which to produce food for
themselves. On that land they produce more food than they actually require, so they have a
surplus for export.
Now suppose we take the United States' great neighbour to the north, Canada. Every
Canadian has 140 acres to roam around in. A lot of it is away in the frozen north, but there is
still an enormous area of land in Canada waiting to be settled and developed. The official
figure is twenty-two acres. The Canadians use at the moment four acres, and they too have a
large food surplus available for export.
Now turn to our own country. Including land of all sorts, there is just over one acre per head
in the United Kingdom of Great Britain and Northern Ireland. That is why we have to be so
very careful with it. How much do we actually use? Just over half an acre to produce food that is as farm land. The story is much the same if you separate off Northern Ireland and
Scotland and just take England and Wales. In this very crowded country, we have only 0.8
acres per head of land of all sorts to do everything with which we need. That is why we have
to think so very carefully of this problem.
India, with 2.5 acres per head, has considerably more land than we have in this country. Not
all of it is usable for food production. But there is land which could be reclaimed by modern
methods, that is being tackled at the present time. The crucial figure is the actual area in
agricultural use - three-quarters of an acre! The yields from this land are low, methods of
www.seyfihoca.com
102
production are primitive, and that is why the Indians are so very near the starvation level for
almost every year of their lives. But they are not as badly off where land is concerned as
Japan.
The Japanese figures are the same as our own country in overall land - 1.1 acres per person but it is a very mountainous country with volcanoes, and so much less is cultivable. Less than
a fifth of an acre - 0.17 of an acre - is under cultivation. You see at once the tremendous land
problem which there is in Japan.
There is a great variation, of course, in the intensity with which land is used. In the United
States they are extravagant in the use of land and take, perhaps, twenty times as much to feed
one person as in Japan. You may talk about the Japanese agriculture being twenty times as
efficient as the American, but that raises a lot of questions.
The intensive cultivation characteristic of Japan uses every little bit of land and only the
barren hillsides are not required. Much of the agriculture is based on rice. The farm workers
plant by hand every individual rice plant, and this kind of intensive cultivation enables the
Japanese to support seven persons per acre.
By contrast, think of the ranch lands in North and South America, with animals ranging over
immense tracts of land. A diet of beef and of milk is extravagant of land ; in other words, it
takes a lot of land for the number of calories produced. In this sense it is less efficient than the
Japanese rice-growing agriculture. But not everyone likes eating rice.
Where the sea is concerned, we are scarcely, at the present time, out of the old Stone Age. In
the Stone Age, the people simply went out, killed wild animals - if they were lucky - and had
a good meal; if they were unlucky they just went hungry. At the present day, we do almost the
same thing in the sea, hunting wild fish from boats. In the future, perhaps, we shall cultivate
the sea; we shall grow small fish and fish spawn in tanks, take them to the part of the ocean
where we want them, let them grow to the right size, and harvest them. This is not fantasy,
because, at the present time, fish are being cultivated like that in ponds and tanks in India, and
various parts of the Far East so that the people there have a supply of protein. There is a great
development possible.
A lot of things are going to happen in the next fifty years. It is enormously important to
increase the yield of grain plants and a great deal has happened through the work of the
geneticists in the last few years. For instance, there has been an enormous world increase in
the production of what Americans call corn (maize to us) due to the development of new
strains. Throughout agriculture geneticists are improving plants to get higher yields.
From 'Using Land Wisely' Discovery (Granada Television, 1961)
HISTORY
THE NATURE, OBJECT AND PURPOSE OF HISTORY
I shall therefore propound answers to my four questions such as I think any present-day
historian would accept. Here they will be rough and ready answers, but they will serve for a
provisional definition of our subject-matter and they will be defended and elaborated as the
argument proceeds.
www.seyfihoca.com
103
(a) What is history? Every historian would agree, I think, that history is a kind of research or
inquiry. What kind of inquiry it is I do not yet ask. The point is that generically it belongs to
what we call the sciences: that is, the forms of thought whereby we ask questions and try to
answer them. Science in general, it is important to realize, does not consist in collecting what
we already know and arranging it in this or that kind of pattern. It consists in fastening upon
something we do not know, and trying to discover it. Playing patience with things we already
know may be a useful means towards this end, but it is not the end itself. It is at best only the
means. It is scientifically valuable only in so far as the new arrangement gives us the answer
to a question we have already decided to ask. That is why all science begins from the
knowledge of our own ignorance: not our ignorance of everything, but our ignorance of some
definite thing-the origin of parliament, the cause of cancer, the chemical composition of the
sun, the way to make a pump work without muscular exertion on the part of a man or a horse
or some other docile animal. Science is finding things out: and in that sense history is a
science.
(b) What is the object of history? One science differs from another in that it finds out things of
a different kind. What kinds of things does history find out? I answer, res gestae: actions of
human beings that have been done in the past. Although this answer raises all kinds of further
questions
many of which are controversial, still, however they may be answered, the answers do not
discredit the proposition that history is the science of res gestae, the attempt to answer
questions about human actions done in the past.
(c) How does history proceed? History proceeds by the interpretation of evidence: where
evidence is a collective name for things which singly are called documents, and a document is
a thing existing here and now, of such a kind that the historian, by thinking about it, can get
answers to the questions he asks about past events. Here again there are plenty of difficult
questions to ask as to what the characteristics of evidence are and how it is interpreted. But
there is no need for us to raise them at this stage. However they are answered, historians will
agree that historical procedure, or method, consists essentially of interpreting evidence.
(d) Lastly, what is history for? This is perhaps a harder question than the others; a man who
answers it will have to reflect rather more widely than a man who answers the three we have
answered already. He must reflect not only on historical thinking but on other things as well,
because to say that something is `for' something implies a distinction between A and B, where
A is good for something and B is that for which something is good. But I will suggest an
answer, and express the opinion that no historian would reject it, although the further
questions to which it gives rise are numerous and difficult.
My answer is that history is `for' human self-knowledge. It is generally thought to be of
importance to man that he should know himself: where knowing himself means knowing not
his merely personal peculiarities, the things that distinguish him from other men, but his
nature as man. Knowing yourself means knowing, first, what it is to be a man; secondly,
knowing what it is to be the kind of man you are; and thirdly, knowing what it is to be the
man you are and nobody else is. Knowing yourself means knowing what you can do; and
since nobody knows what he can do until he tries, the only clue to what man can do is what
man has done. The value of history, then, is that it teaches us what man has done and thus
what man is.
www.seyfihoca.com
104
(From The idea of history, by R. G. Collingwood.)
THE EXPANSION OF WESTERN CIVILIZATION
The predominance of the Western civilization throughout he world on the eve of the fateful
year 194 was, indeed, both recent and unprecedented. It was unprecedented in this sense-that,
though many civilizations before that of Europe had radiated their influence far beyond their
original homelands, none had previously cast its net right round the globe.
The civilization of Eastern orthodox Christendom, which grew up in mediaeval Byzantium,
had been carried by the Russians to the Pacific; but, so far from spreading westwards, it had
itself succumbed to Western influence since the close of the seventeenth century. The
civilization of Islam had expanded from the Middle East to Central Asia and Central Africa,
to the Atlantic coast of Morocco and the Pacific coasts of the East Indies, but it had obtained
no permanent foothold in Europe and had never crossed the Atlantic into the New World. The
civilization of ancient Greece and Rome had extended its political dominion into NorthWestern Europe under the Roman Empire and its artistic inspiration into India and the Far
East, where the Graeco-Roman models had stimulated the development of Buddhist art. Yet
the Roman Empire and the Chinese Empire had co-existed on the face of the same planet for
two centuries with scarcely any direct intercourse, either political or economic. It was the
same with the other ancient civilizations. Ancient India radiated her religion, her art, her
commerce and her colonists into the Far East and the East Indies, but never penetrated the
West. As far as we know for certain, the only civilization that has ever yet become worldwide
is ours.
Moreover, this is a very recent event. Nowadays we are apt to forget that Western Europe
made two unsuccessful attempts to expand before she eventually succeeded.
The first of these attempts was the mediaeval movement in the Mediterranean for which the
most convenient general name is the Crusades. In the Crusades, the attempt to impose the
political and economic dominion of West Europeans upon other peoples ended in a complete
failure, while, in the interchange of culture, the West Europeans received a greater impress
from the Muslims and Byzantines than they imparted to them. The second attempt was that of
the Spaniards and Portuguese in the sixteenth century of our era. This was more or less
successful in the New World, but, elsewhere, Western civilization, as propagated by the
Spaniards and Portuguese, was rejected after about a century's trial.
The third attempt was begun in the seventeenth century by the Dutch, French and English, and
these three West European nations were the principal authors of the world-wide ascendancy
that our Western civilization was enjoying in 1914. By 1914 the network of European trade
and European means of communication had become world-wide. On the plane of politics, the
European nations had not only colonized the New World, but had conquered India and
tropical Africa.
The political ascendancy of Europe, however, though outwardly even more imposing than her
economic ascendancy, was really more precarious. The daughter-nations overseas had already
set their feet firmly on the road towards independent nationhood. The United States and the
Latin American Republics had long since established their independence by revolutionary
wars; and the self-governing British Dominions were in the process of establishing theirs by
peaceful evolution. In India and tropical Africa, European domination was being maintained
www.seyfihoca.com
105
by a handful of Europeans who lived there as pilgrims and sojourners. They had not found it
possible to acclimatize themselves sufficiently to bring up their children in the tropics; this
meant that the hold of Europeans upon the tropics had not been made independent of a
European base of operations. Finally, the cultural influence of the West European civilization
upon Russians, Muslims, Hindus, Chinese, Japanese, and tropical Africans was so recent a
ferment that it was not yet possible to predict whether it would evaporate without permanent
effect, or whether it would turn the dough sour, or whether it would successfully leaven the
lump.
This then, in very rough outline, was the position of Europe in the world on the eve of the
War of 1914-1918. She was in the enjoyment of an undisputed ascendancy, and the peculiar
civilization which she had built up for herself was in process of becoming world-wide. Yet
this position, brilliant though it was, was not merely unprecedented and recent; it was also
insecure. It was insecure chiefly because, at the very time when European expansion was
approaching its climax, the foundations of West European civilization had been broken up
and the great deeps loosed by the release and emergence of two elemental forces in European
social life-the forces of industrialism and democracy, which were brought into a merely
temporary and unstable equilibrium by the formula of nationalism. It is evident that a Europe
which was undergoing a terrific double strain of this inward transformation and outward
expansion-both on the heroic scale-could not with impunity squander her resources, spend her
material wealth and man-power unproductively, or exhaust her muscular and nervous energy.
If her total command of resources was considerably greater than that which any other
civilization had ever enjoyed, these resources were relative to the calls upon them; and the
liabilities of Europe on the eve of 1914, as well as her assets, were of an unprecedented
magnitude. Europe could not afford to wage even one World War; and when we take stock of
her position in the world after a Second World War and compare it with her position before
1914, we are confronted with a contrast that is staggering to the imagination.
(From Civilization on Trial, by Arnold Toynbee)
THE CAREER OF JENGHIS KHAN
We have now to tell of the last and greatest of all the raids of nomadism upon the civilizations
of the East and West. We have traced in this history the development side by side of these two
ways of living, and we have pointed out that as the civilizations grew more extensive and
better organized, the arms, the mobility, and the intelligence of the nomads also improved.
The nomad was not simply an uncivilized man, he was a man specialized and specializing
along his own line. From the very beginning of history the nomad and the settled people have
been in reaction. Whenever civilization seems to be choking amidst its weeds of wealth and
debt and servitude, when its faiths seem rotting into cynicism and its powers of further growth
are hopelessly entangled in effete formulae, the nomad drives in like a plough to break up the
festering stagnation and release the world to new beginnings. The Mongol aggression, which
began with the thirteenth century, was the greatest, and so far it has been the last, of all these
destructive ploughings of human association.
From entire obscurity the Mongols came very suddenly into history towards the close of the
twelfth century. They appeared in the country to the north of China, in the land of origin of
the Huns and Turks, and they were manifestly of the same strain as these peoples. They were
gathered together under a chief; under his son Jenghis Khan their power grew with
extraordinary swiftness.
www.seyfihoca.com
106
The career of Jenghis Khan and his immediate successors astounded the world, and probably
astounded no one more than these Mongol Khans themselves. The Mongols were in the
twelfth century a tribe subject to those Kin who had conquered North-east China. They were a
horde of nomadic horsemen living in tents, and subsisting mainly upon mare's milk products
and meat. Their occupations were pasturage and hunting, varied by war. They drifted
northward as the snows melted for summer pasture, and southward to winter pasture after the
custom of the steppes. Their military education began with a successful insurrection against
the Kin. The empire of Kin had the resources of half' China behind it and in the struggle the
Mongols learnt very much of the military science of the Chinese. By the end of the twelfth
century they were already a fighting tribe of exceptional quality.
The opening years of the career of Jenghis were spent in developing his military machine, in
assimilating the Mongols and the associated tribes about them into an organised army. His
first considerable extension of power was westward, when the Tartar Kirghis and the Uigurs
(who were the Tartar people of the Tarim basin) were not so much conquered as induced to
join his organisation. He then attacked the Kin empire and took Pekin (1214). The Khitan
people, who had been so recently subdued by the Kin, threw in their fortunes with his, and
were of very great help to him. The settled Chinese population went on sowing and reaping
and trading during this change of masters without lending its weight to either side.
We have already mentioned the very recent Kharismian empire of Turkestan, Persia and
North India. This empire extended eastward to Kashgar, and it must have seemed one of the
most progressive and hopeful empires of the time. Jenghis Khan, while still engaged in this
war with the Kin empire, sent envoys to Kharismia. They were put to death, an almost
incredible stupidity. The Kharismian government, to use the political jargon of today, had
decided not to `recognize' Jenghis Khan and took this spirited course with him. Thereupon
(1218) the great host of horsemen that Jenghis Khan had consolidated swept over the Pamirs
and down into Turkestan. It was well armed, and probably it had some guns and gunpowder
for siege work-for the Chinese were certainly using gunpowder at this time, and the Mongols
learnt its use from them. Kashgar, Khokand, Bokhara fell and then Samarkand, the capital of
the Kharismian empire. Thereafter nothing held the Mongols in the Kharismian territories.
They swept westward to the Caspian, and southward as far as Lahore. To the north of the
Caspian a Mongol army encountered a Russian force from Kieff. There was a series of battles,
in which the Russian armies were finally defeated and the Grand Duke of Kieff was taken
prisoner. So it was that the Mongols appeared on the northern shores of the Black Sea. A
panic swept Constantinople, which set itself to reconstruct its fortifications. Meanwhile other
armies were engaged in the conquest of the empire of the Hsia in China. This was annexed,
and only the southern part of the Kin empire remained unsubdued. In 1227 Jenghis Khan died
in the midst of a career of triumph. His empire reached already from the Pacific to the
Dnieper. And it was an empire still vigorously expanding.
Like all empires founded by nomads, it was, to begin with, purely a military and
administrative empire, a framework rather than a rule. It centred on the personality of the
monarch, and its relations with the mass of the population over which it ruled was simply one
of taxation for the maintenance of the horde. But Jenghis Khan had called to his aid a very
able and experienced administrator of the Kin empire, who was learned in all the traditions
and science of the Chinese. This statesman, Yeliu Chutsai, was able to carry on the affairs of
the Mongols long after the death of Jenghis Khan, and there can be little doubt that he is one
of the great political heroes of history. He tempered the barbaric ferocity of his masters, and
saved innumerable cities and works of art from destruction. He collected archives and
www.seyfihoca.com
107
inscriptions, and when he was accused of corruption, his sole wealth was found to consist of
documents and a few musical instruments. To him perhaps quite as much as to Jenghis is the
efficiency of the Mongol military machine to be ascribed. Under Jenghis, we may note
further, we find the completest religious toleration established across the entire breadth of
Asia.
(From The Outline of History, by H G Wells)
THE TRIAL AND EXECUTION OF CHARLES I
JANUARY, 1649
Officers and men had come home from the war in a fierce mood, prepared to commit any
violence. Colonel Pride and his musqueteers, stationed at the door of the Com-mons, excluded
some hundred members and carried off nearly fifty more to prison. On January 4th, 1649, the
Rump that remained, not one hundred strong, passed this resolution:
That the people are, under God, the original of all just power: and that the Commons
of England, in Parliament assembled, being chosen by and representing the people,
have the supreme power in this nation; that whatsoever is enacted or declared for law
by the Commons in Parliament assembled, hath the force of law, and all the people of
this nation are concluded thereby, although the consent of the King or House of Peers
be not had thereunto.
In logical accordance with these new principles, it was voted a few weeks later that the House
of Lords was 'useless and dangerous and ought to be abolished'. The King disappeared from
the constitution by a yet more signal and symbolic act. 'We will cut off his head', said
Cromwell, 'with the crown upon it.'
The Commons appointed a Commission to try Charles Stuart. Unless the theoretic declaration
of the omnipotence of the Lower House held good by 'the Law of Nature', this Commission
had no legal power. Furthermore, Charles had committed no legal crime. Law and pity, would
both plead on his side. His own outward dignity and patience made his appeal to pity and law
the most effective appeal in that sort that is recorded in the history of England. Law and pity
were known of old to Englishmen; the new ideas ranged against these two powerful pleaders
were new and strange. The sovereignty of the people and the equality of man with man in the
scales of justice were first ushered into the world of English politics by this deed. Against
them stood the old-world ideals, as Charles proclaimed them from the scaffold with his dying
breath:
For the people (he said, as he looked out into eternity) truly I desire their liberty and freedom
as much as anybody whatsoever; but I must tell you, their liberty and freedom consists in
having government, those laws by which their lives and goods may be most their own. It is
not their having a share in the government; that is nothing appertaining to them. A subject and
a sovereign are clear different things.
That, indeed, was the issue. Those who sought to substitute a better theory of government for
that proclaimed by Charles as his legacy to England did not pause to consider that the people,
for whose new-born sovereignty they struck, was heart and soul against the deed. By the side
of this objection, the charge that they acted 'by no law' shrinks to insignificance. They were
www.seyfihoca.com
108
striving to make a law of democracy, and could not be expected to effect this by observing the
old laws. But their fault lay in this, that the new law by which they condemned Charles, while
it claimed to derive from the people of England, did not, save in theory, derive from that
source at all. When the bleeding head was held up; the shout of the soldiers was drowned in
the groans of the vast multitude.
If there was any chance that the establishment of a more democratic form of government
could gradually win the sup-port of the people at large, that chance was thrown away by the
execution of the King. The deed was done against the wish of many even of the Independents
and Republicans; it outraged beyond hope of reconciliation the two parties in the state who
were strong in numbers and conservative tradition, the Presbyterians and the Cavaliers; and it
alienated the great mass of men who had no party at all. Thus the Republicans, at the outset of
their career, made it impossible for themselves ever to appeal in free election to the people
whom they had called to sovereignty.
It is much easier to show that the execution was a mistake and a crime (and it certainly was
both) than to show what else should have been done. Any other course, if considered in the
light of the actual circumstances, seems open to the gravest objection. The situation at the end
of 1648 was this - that any sort of government by consent had been rendered impossible for
years to come, mainly by the untrustworthy character of the King, and by the intolerant action
of Parliament after the victory won for it by the Army. Cromwell had advocated a real
settlement by consent, only to have it rejected by King, Parliament, and Army alike. The
situation had thereby been rendered impossible, through no fault of his.
(From England under the Stuarts by G. M. Trevelyan)
PETER THE GREAT
The impact of Peter the Great upon Muscovy was like that of a peasant hitting a horse with
his fist. Muscovy bore many of the marks permanently and henceforward she became known
as Russia. Yet his reforms, for all their importance, did not create a new form of state: they
were to a large extent the very rapid and energetic extension of ideas and practices already to
be found in the generation before him. The old and the new continued to live on, in part
conflicting, in part coalescing. Tsarism in his lifetime received a new stamp owing to his
wholly exceptional character and abilities, but its functioning remained as before dependent
upon different sec-tions of the landed class and the machinery of government. Peter did not
introduce the idea of the service state; he pushed it to extremes, enforced compulsory service
in the army, navy, and government on the landowners and himself set the example as the first
servant of the state. He wrote of himself with full justification: 'I have not spared and I do not
spare my life for my fatherland and people.'
Peter was repellent in his brutality, coarseness and utter disregard of human life; but he was
mightily propellent, through his ever-flaming energy and will-power, his insatiable curiosity
and love of experiment, his refusal to accept any defeat and his capacity to launch, however
crudely and over-hastily at first, great schemes on an immense, however wasteful, scale. From
earliest childhood his over-riding personal interest was the art of war, by sea as well as by
land; but he understood it in the widest sense as involving the full utilization of the human and
material resources of his country. He was at war for twenty-eight consecutive years, from
1695. He began when he was twenty-four; when he finally had peace he was fifty-two and
had little more than a year to live.
www.seyfihoca.com
109
Almost all Peter's reforms were born of military and naval
requirements. Russia must be westernized in order to ensure the 'two necessary things in
government, namely order and defence'. His task was, again, as he himself put it, to convert
his subjects from 'beasts into human beings', from 'children into adults'. His strongest modern
critics allow that he was inspired by devotion to Russia, not by personal ambition, and that he
aimed at inculcating by example and precept rational ideas of citizenship in terms of efficient,
and therefore educated, service to the state, in contrast with blind service to a sacrosanct ruler
throned far away on high in the hallowed veneration of Muscovite ceremonial.
His reforms until about 1715 were imposed piecemeal, chaotically and (as he himself
admitted) far too hastily, in dependence on the urgent pressure of the moment. He was
constantly scrapping or remodelling this or that portion of his own handiwork in frantic search
for more recruits, more labour, more revenue, more munitions. In his last dozen years, when
war was less onerous and his contacts with the West were closer, the peremptory edicts, often
conflicting with each other, gave way to systematic, carefully elaborated legislation that
remoulded state and church alike. His brutal violence, the enormous demands that he exacted
and his external flouting of national ways and prejudices supplied fertile ground for
opposition. He had to crush in blood four serious risings, and he condemned to death his own
son and heir, Alexis, on the ground of his being the ringleader of reaction (Q18). In actuality
Alexis was a passive creature who was only hoping for his father's death in terrified fear of
his own. The opposition was leaderless; and as well it was too heterogeneous; almost all
interests in Russia were divided between supporters and opponents of Peter.
He aimed at transforming tsarism into a European kind of absolute monarchy, and to a
considerable extent he succeeded. Russia was never the same again, even though the pace was
too hot and there was regress after his death. He declared himself to be 'an absolute monarch
who does not have to answer for any of his actions to anyone in the world; but he has power
and authority for the purpose of governing his states and lands according to his will and wise
decision as a Christian sovereign'. This version of enlightened despotism, typically enough,
appeared in Peter's new code for the army (1716). The creation of a national standing army on
Western models was one of the most fundamental of his legacies, and the links of tsarism with
military power and the military spirit were henceforth knitted even more closely than before.
One external sign is significant. Peter himself almost always appeared as a soldier or sailor
(when not dressed as a mechanic) and all succeeding emperors did likewise; his predecessors
(when not hunting) had usually appeared in hieratic pomp, half tsar, half high-priest.
No tsar has made such a lasting impression on Russia as Peter, whether in his works or his
personality. He was an unheard-of tsar-for some no tsar at all, but Antichrist. He brought the
tsar to earth and entwined himself in the hopes and fears and groans of his subjects as a dark
and terrible force, rooting up the past, putting to rout the Swedes; as a ruler such as they had
never conceived before, to be seen throughout the length and breadth of the land, immense in
stature, with his tireless stride that was more of a run and his huge calloused hands of which
he was so proud; a ruler who went into battle as a bombardier, who wielded an axe as well as
any of his subjects, who could kill a man with a single blow of his fist-and on occasion did.
He made Russia conscious of great destiny, and ever since Europe and Asia have had to
reckon with her.
(From Survey of Russian History, by B. H. Sumner)
www.seyfihoca.com
110
THE UNITED STATES IN 1790
In the physiographic features of the settled part of America there was a certain uniformity.
The coast-line was low and un-inviting, except in northern New England, where it had
something of the rugged picturesqueness of the western coast of Britain. South of New York
stretched a long succession of barrier beaches in flattish curves, parting at the entrances of the
great bays of Chesapeake and Delaware, and enclosing the shallow lagoons of Albemarle and
Pamlico. A vast forest of conifers and hardwood swept up from the coast over the crest of the
Appalachians, and down to the Great Lakes, the prairies of Illinois, the savannahs of the lower
Mississippi, and the Gulf of Mexico. Except for natural meadows along the rivers, open
country did not exist in the land that the English colonists wrested from the Indians. Their
farms had been cleared from the forest; and it was still too early to aver that the colonists had
conquered the forest. Volney wrote that during his journey in 1796 through the length and
breadth of the United States he scarcely travelled for more than three miles together on open
and cleared land. 'Compared with France, the entire country is one vast wood.' Only in
southern New England, and the eastern portion of the Middle States, did the cultivated area
exceed the woodland; and the clearings became less frequent as one approached the
Appalachians.
Like western Europe, the United States lies wholly within the northern temperate zone, and
the belt of prevailing westerly winds. The earliest European explorers had passed it by for the
Caribbean and the St. Lawrence because they were seeking tropical plantations, fur-trading
posts, and fishing stations. Their successors, in search of farm-lands, found the greater part of
the Thirteen Colonies suitable for life and labour as are few portions of the non-European
world. Yet the climate of the area settled by 1790 is in many respects unlike that of Europe.
Westerly winds reach it across a continent, without the moisture and the tempering of the
Atlantic. North-west is the prevailing wind in winter, and south-west in summer.
Consequently the summers are everywhere hotter than in the British Isles, and the winters,
north of Virginia, colder; the extremes of heat and cold in the same season are greater; the
rainfall less, although adequate for animal and plant life. Near the sea-coast a sea-turn in the
wind may soften outlines, but inland the dry air, clear sky, and brilliant sunlight foreshorten
distant prospects, and make the landscape sharp and hard.
In most parts of the United States the weather is either fair or foul. It rains or shines with a
businesslike intensity; in comparison, the weather of the British Isles is perpetually unsettled.
In the coastal plain of the Carolinas and the Gulf, there is a soft gradation between the
seasons, and a languor in the air; else-where, the transition from a winter of ice and snow to a
summer of almost tropical heat is abrupt.
Our spring gits everythin' in tune
An' gives one leap from April into June
wrote Lowell. Except where the boreal forest of conifers maintains its sombre green, the sharp
dry frosts of October turn the forest to a tapestry of scarlet and gold, crimson and russet. High
winds strip the leaves in November, and by New Year's day the country north of Baltimore,
along the Appalachians, and east of the Sierras, should be tucked into a blanket of snow.
(From History of the United States by S. E. Morison)
www.seyfihoca.com
111
Civilisation and History
Most of the people who appear most often and most gloriously in the history books are great
conquerors and generals and soldiers, whereas the people who really helped civilization
forward are often never mentioned at all. We do not know who first set a broken leg, or
launched a seaworthy boat, or calculated the length of the year, or manured a field; but we
know all about the killers and destroyers. People think a great deal of them, so much so that
on all the highest pillars in the great cities of the world you will find the figure of a conqueror
or a general or a soldier. And I think most people believe that the greatest countries are those
that have beaten in battle the greatest number of other countries and ruled over them as
conquerors. It is just possible they are, but they are not the most civilized. Animals fight; so
do savages; hence to be good at fighting is to be good in the way in which an animal or a
savage is good, but it is not to be civilized. Even being good at getting other people to fight
for you and telling them how to do it most efficiently - this, after all, is what conquerors and
generals have done - is not being civilized. People fight to settle quarrels. Fighting means
killing, and civilized peoples ought to be able to find some way of settling their disputes other
than by seeing which side can kill off the greater number of the other side, and then saying
that that side which has killed most has won. And not only has won, but, because it has won,
has been in the right. For that is what going to war means; it means saying that might is right.
That is what the story of mankind has on the whole been like. Even our own age has fought
the two greatest wars in history, in which millions of people were killed or mutilated. And
while today it is true that people do not fight and kill each other in the streets - while, that is to
say, we have got to the stage of keeping the rules and behaving properly to each other in daily
life - nations and countries have not learnt to do this yet, and still behave like savages.
But we must not expect too much. After all, the race of men has only just started. From the
point of view of evolution, human beings are very young children indeed, babies, in fact, of a
few months old. Scientists reckon that there has been life of some sort on the earth in the form
of jellyfish and that kind of creature for about twelve hundred million years; but there have
been men for only one million years, and there have been civilized men for about eight
thousand years at the outside. These figures are difficult to grasp; so let us scale them down.
Suppose that we reckon the whole past of living creatures on the earth as one hundred years;
then the whole past of man works out at about one month, and during that month there have
been civilizations for between seven and eight hours. So you see there has been little time to
learn in, but there will be oceans of time in which to learn better. Taking man's civilized past
at about seven or eight hours, we may estimate his future, that is to say, the whole period
between now and when the sun grows too cold to maintain life any longer on the earth, at
about one hundred thousand years. Thus mankind is only at the beginning of its civilized life,
and as I say, we must not expect too much. The past of man has been on the whole a pretty
beastly business, a business of fighting and bullying and gorging and grabbing and hurting.
We must not expect even civilized peoples not to have done these things. All we can ask is
that they will sometimes have done something else.
From The Story of Civilization by C. E. M. Joad (A. D. Peters & Co. 1962)
Coal
The rapid growth of steam power relied directly on large supplies of its only fuel: coal. There
was a great increase in the amount of coal mined in Britain.
www.seyfihoca.com
112
Coal was still cut by hand and the pits introduced few really new techniques. The increased
demand was met by employing more miners and by making them dig deeper. This was made
possible by more efficient steam pumps and steam-driven winding engines which used wire
ropes to raise the coal to the surface, by better ventilation and by the miner's safety lamp
which detected some dangerous gases. By the 1830s scams well below 1000 feet were being
worked in south Durham and the inland coalfields of Lancashire and Staffordshire. Central
Scotland and south Wales were mined more intensively later. By the end of the nineteenth
century the best and most accessible seams were worked out. As the miners followed the
eastern scams of the south Yorkshire-north Midlands area, which dipped further below the
surface, shafts of 3,000 feet were not uncommon.
Some of the work in mines was done by women and children. Boys and girls were often put in
charge of the winding engines or of opening and shutting the trap doors which controlled the
ventilation of the mines. Then they had to crouch all day in the same spot by themselves in the
dark. When these evils were at last publicized in 1842 by a Royal Commission, many mines
no longer employed women, but Parliament made it illegal for them all. It also forbade them
to employ boys under the age of ten. The limit, which was very difficult to enforce, was
increased to twelve in the 1870s. Subsequently it rose with the school leaving age.
Mining was very dangerous. Loose rocks were easily dislodged and the risk of being killed or
injured by one was always greater in the tall scams where they had further to fall. In the north
of England fatal accidents were not even followed by inquests to discover why they had
happened until after 1815. Few safety precautions were taken before the mid-nineteenth
century. The mine owners insisted that they were not responsible. The men were most
reluctant to put up enough props to prevent the roof from falling in and to inspect the winding
gem: and other machinery on which their lives depended. If they did, they spent less time
mining and so earned less money because the miners' pay was based not on how long they
worked but on how much coal they extracted. They preferred to take risks.
The deeper seams contained a dangerous gas called 'fire-damp' which could be exploded by
the miners' candles. The safety lamp, which was invented in the early nineteenth century, did
not really solve this problem, but it was often used to detect gas and so made the mining of
deeper seams possible. There the air was more foul, the temperature higher (one pit paid the
men an extra 6d a day for working in 130°F) and the risk of fire-damp even greater. In the
1840s a series of terrible explosions in the deeper mines led to stricter regulations, which
inspectors helped enforce. The inspectors were particularly keen on proper ventilating
machines and, although deeper shafts were sunk, they did not become more dangerous.
However, many serious accidents still occurred.
(From Britain Transformed, Penguin Books)
LANGUAGE
'Primitiveness' in Language
'Primitive' is a word that is often used ill-advisedly in discussions of language. Many people
think that 'primitive' is indeed a term to be applied to languages, though only to some
languages, and not usually to the language they themselves speak. They might agree in calling
'primitive' those uses of language that concern greetings, grumbles and commands, but they
www.seyfihoca.com
113
would probably insist that these were especially common in the so-called 'primitive
languages'. These are misconceptions that we must quickly clear from our minds.
So far as we can tell, all human languages are equally complete and perfect as instruments of
communication: that is, every language appears to be as well equipped as any other to say the
things its speakers want to say. It may or may not be appropriate to talk about primitive
peoples or cultures, but that is another matter. Certainly, not all groups of people are equally
competent in nuclear physics or psychology or the cultivation of rice or the engraving of
Benares brass. But this is not the fault of their language. The Eskimos can speak about snow
with a great deal more precision and subtlety than we can in English, but this is not because
the Eskimo language (one of those sometimes miscalled 'primitive') is inherently more precise
and subtle than English. This example does not bring to light a defect in English, a show of
unexpected 'primitiveness'. The position is simply and obviously that the Eskimos and the
English live in different environments. The English language would be just as rich in terms
for different kinds of snow, presumably, if the environments in which English was habitually
used made such distinction important.
Similarly, we have no reason to doubt that the Eskimo language could be as precise and subtle
on the subject of motor manufacture or cricket if these topics formed part of the Eskimos' life.
For obvious historical reasons, Englishmen in the nineteenth century could not talk about
motorcars with the minute discrimination which is possible today: cars were not a part of their
culture. But they had a host of terms for horse-drawn vehicles which send us, puzzled, to a
historical dictionary when we are reading Scott or Dickens. How many of us could distinguish
between a chaise, a landau, a victoria, a brougham, a coupe, a gig, a diligence, a whisky, a
calash, a tilbury, a carriole, a phaeton, and a clarence ?
The discussion of 'primitiveness', incidentally, provides us with a good reason for sharply and
absolutely distinguishing human language from animal communication, because there is no
sign of any intermediate stage between the two. Whether we examine the earliest records of
any language, or the present-day language of some small tribe in a far-away place, we come
no nearer to finding a stage of human language more resembling animal communication and
more 'primitive' than our own. In general, as has been said, any language is as good as any
other to express what its speakers want to say. An East African finds Swahili as convenient,
natural and complete as an East Londoner finds English. In general the Yorkshire Dalesman's
dialect is neither more nor less primitive or ill-fitted to its speaker's wants than Cockney is for
the Londoner's. We must always beware the temptation to adopt a naive parochialism which
makes us feel that someone else's language is less pleasant or less effective an instrument than
our own.
This is not to say that an individual necessarily sounds as pleasant or as effective as he might
be, when using his language, but we must not confuse a language with an individual's ability
to use it. Nor are we saying that one language has no deficiencies as compared with another.
The English words 'home' and 'gentleman' have no exact counterparts in French, for example.
These are tiny details in which English may well be thought to have the advantage over
French, but a large-scale comparison would not lead to the conclusion that English was the
superior language, since it would reveal other details in which the converse was true. Some
years ago it came as something of a shock to us that we had no exact word for translating the
name that General de Gaulle had given to his party - Rassemblement du Peuple Francais. The
B.B.C. for some time used the word 'rally', and although this scarcely answers the purpose it
www.seyfihoca.com
114
is a rather better translation of 'rassemblement' than either of the alternatives offered by one
well-known French-English dictionary, 'muster' and 'mob'.
The more we consider the question, then, the less reasonable does it seem to call any language
'inferior', let alone 'primitive'.
The Sanskrit of the Rig-Veda four thousand years ago was as perfect an instrument for what
its users wanted to say as its modern descendant, Hindi, or as English.
(From The Use of English, by Randolph Quirk )
English in the Fifteenth Century
Soon after Chaucer's death, in the fifteenth century, there was a renewed drift towards
simplification in English. Final unaccented vowels, by 1400 already reduced to a very slight
number, were entirely lost. Still more nouns were shifted to the majority declension (with
plurals in -s) out of the small group left in the minority declensions. More and more verbs
shifted to the weak conjugation from those still retaining the internal vowel change. For a
time, of course, there was a choice of forms: for example, between 'he clomb' and 'he
climbed'; 'he halp' and 'he helped'. Some of the quaint surviving constructions out of Old
English, such as impersonal verbs with the dative, the inflected genitive case for nouns
denoting things, and the double negative, began to fall into disuse. They persist in the
fifteenth century, indeed even into the sixteenth, but they are increasingly felt to be archaic
survivals.
Another important usage became increasingly prevalent in the fifteenth and early sixteenth:
the bolstering of verbs with a number of auxiliaries derived from 'do' and 'be'. In Middle
English a question was asked with the simple form of the verb in inverted position: 'What say
you? What think you?' For a couple of centuries after 1400 this was still done habitually, but
more and more people fell into the habit of saying 'What do you say? What do you think?' The
'do' was colourless and merely brought about a deferment of the main verb. In effect it makes
our English usage somewhat like Russian, which says 'What you say? What you think?'
without any inversion of the verb before the subject. In simple statements the 'do' forms were
used for situations where we no longer feel the need for them. An Elizabethan would say 'I do
greatly fear it' (an unrestricted statement). We should use the less emphatic 'I fear it greatly.'
During the same period there began the gradual spread of the so-called progressive
conjugation, with forms of 'to be': ' I am coming; he is sitting down.' These two special forms
of English conjugation have developed an intricate etiquette, with many modifications of
usage, which cause great trouble to the foreign student. One of the last distinctions he masters
is the one between 'I eat breakfast every morning' and 'I am eating breakfast now'; between 'I
believe that, and 'I do indeed believe that.'
One of the most fateful innovations in English culture, the use of the printing press, had its
effects on the language in many ways. The dialect of London, which for over a century had
been gaining in currency and prestige, took an enormous spurt when it was more or less
codified as the language of the press. As Caxton and his successors normalized it, roughly
speaking, it became the language of officialdom, of polite letters, of the spreading commerce
centres at the capital. The local dialects competed with it even less successfully than formerly.
The art of reading, though still a privilege of the favoured few, was extended lower into the
www.seyfihoca.com
115
ranks of the middle classes. With the secularizing of education later on, the mastery of the
printed page was extended to still humbler folk. Boys who, like William Shakespeare, were
sons of small-town merchants and craftsmen, could learn to read Latin literature and the Bible
even if they had no intention of entering the Church. Times had distinctly changed since the
thirteenth century. It may be added that changes in society-the gradual emergence of a
mercantile civilization-gave scope to printing which it would never have had in the earlier
Middle Ages. The invention was timely in more than one sense.
All this may have been anticipated by the early printers. Their technological innovations may
have been expected to facilitate the spread of culture. But they could not have foreseen that
the spelling which they standardized, more or less, as the record of contemporary
pronunciation, would have been perpetuated for centuries afterwards. Today, when our
pronunciation has become quite different, we are still teaching our unhappy children to spell
as Caxton did. Respect for the printed page has become something like fetish-worship. A few
idiosyncrasies have been carefully preserved, although the reason for them is no longer
understood. When Caxton first set up the new business in London he brought with him
Flemish workers from the Low Countries, where he himself had learned it. Now the Flemish
used the spelling 'gh' to represent their own voiced guttural continuant, a long-rolled-out
sound (y) unlike our English (g). English had no such sound at the time, but the employees in
Caxton's shop were accustomed to
combining the two letters, and continued to do so in setting up certain English words. In
words like 'ghost' and 'ghastly' it has persisted, one of the many mute witnesses to
orthographical conservatism.
(From The Gift of Tongues, by Margaret Schlauch.)
An International Language
Some languages are spoken by quite small communities and they are hardly likely to survive.
Before the end of the twentieth century many languages in Africa, Asia, and America will
have passed into complete oblivion unless some competent linguist has found time to record
them. The languages that remain are constantly changing with the changing needs and
circumstances of the people who speak them. Change is the manifestation of life in language.
The great languages of the world, such as English, Russian, Chinese, Japanese, German,
Spanish, French, Italian, Portuguese, Dutch, and Arabic, are just as liable to change as
Swahili, Tamil, or Choctaw. Change may, it is true, be artificially retarded in special
conditions: for example in the great liturgical languages of mankind, such as Sanskrit, the
language of the orthodox Hindu religion of India; or Pali, the sacred language of Buddhism;
or Latin, the liturgical language of the Roman Church. By arduous schooling a man may train
himself to read, write, and converse in these crystallized forms of speech. Sanskrit, Pali, and
Latin are magnificent and awe-inspiring exceptions to the otherwise universal principle of
change. Their immutability depends upon two main factors or conditions: first, that they are
not normally used in everyday conversation, but are entrusted instead to the care of a
privileged class of priests and scholars; and secondly, that they possess memorable recorded
literatures and liturgies which are constantly read and recited in acts of religious devotion and
worship.
It is just because these two conditions do not apply to artificial languages like Volapuk,
Esperanto, and Ido, that they, however carefully devised and constructed, cannot come alive
www.seyfihoca.com
116
and then escape from the law of change. Over one hundred artificial languages have been
framed by men in recent times, but the three just named are far more widely known and used
than any others. Volapuk, or 'World's Speech', was created by a Bavarian pastor named Johan
Martin Schleyer, in 1879, when it was acclaimed with enthusiasm as the future universal
speech of mankind. Only eight years later, however, many of Volapuk's most ardent
supporters abandoned it in favour of the system invented by the 'hopeful doctor', Doktoro
Esperanto, a Polish Jew named Lazarus Zamenhof (1859-1917). Esperanto is certainly an
improvement upon Volapuk in that it is both more flexible and more regular. Even within
Zamenhof's lifetime, however, the mechanism of Esperanto was improved in various ways,
and in 1907 Ido (a made-up name consisting of the initials of International Delegation
substantive suffix -o) was formulated. This Delegation included scholars prominent in various
branches of learning, but its recommendations were not accepted by the main body of
Esperantists who were reluctant to admit that all their well-established textbooks might now
be out of date. Today Esperanto, and not its more advanced form Ido, is easily the first
constructed language in the world and it has proved its worth at numerous international
gatherings. It no longer aspires to supplant ethnic languages. Like those other artificial
languages created in the twentieth century - Edgar de Wahl's Occidental (1922), Otto
Jespersen's Novial (1928), Guiseppe Peano's Interlingua, or Latino sine Flexione (1908), and
Lancelot Hogben's Interglossa (1943), and many more - Esperanto can be regarded as a
valuable bridge-language which any man may find unexpectedly useful in unforeseen
contingencies. Learning Esperanto is a pleasant pastime, and manipulating its regularized
affixes and inflections may become a healthy form of mental gymnastics. Nevertheless, even
loyal esperantists have been known to chafe and strain under the necessary bonds of
orthodoxy. However much society may desire and demand that it should remain constant,
'language changes always and everywhere'. In the New World, where opportunities are
limitless and enthusiasm boundless, and where whole families have been reputed to adopt
Esperanto as their everyday language, it has become modified considerably within the space
of one year to suit the special circumstances and way of life of that particular community. The
worlds in which different social communities live are separate worlds, not just one world with
different linguistic labels attached. An American and a Russian may converse pleasantly in
Esperanto about travel, food, dress, and sport, but they may be quite incapable of talking
seriously in Esperanto about religion, science, or philosophy. 'Men imagine', as Francis Bacon
said long ago, 'that their minds have command over language: but it often happens that
language bears rule over their minds.' Whether we like it or not, we are all very much under
the spell of that particular form of speech which has become the medium of discourse for our
society.
(From Language in the Modern World, by Simeon Potter.)
Language as Symbolism
Animals struggle with each other for food or for leadership, but they do not, like human
beings, struggle with each other for things that stand for food or leadership: such things as our
paper symbols of wealth (money, bonds, titles), badges of rank to wear on our clothes, or lownumber licence plates, supposed by some people to stand for social precedence. For animals,
the relationship in which one thing stands for something else does not appear to exist except
in very rudimentary form.
The process by means of which human beings can arbitrarily make certain things stand for
other things may be called the symbolic process. Whenever two or more human beings can
www.seyfihoca.com
117
communicate with each other, they can, by agreement, make anything stand for anything. For
example, here are two symbols:
X
Y
We can agree to let X stand for buttons and Y stand for bows: then we can freely change our
agreement and let X stand for Chaucer and Y for Shakespeare, X for the CIO and Y for the
AFL. We are, as human beings, uniquely free to manufacture and manipulate and assign
values to our symbols as we please. Indeed, we can go further by making symbols that stand
for symbols. If necessary we can, for instance, let the symbol M stand for all the X's in the
above example (buttons, Chaucer, CIO) and let N stand for all the Y's (bows, Shakespeare,
AFL). Then we can make another symbol, T, stand for M and N, which would be an instance
of a symbol of a symbol of symbols. This freedom to create symbols of any assigned value
and to create symbols that stand for symbols is essential to what we call the symbolic process.
Everywhere we turn, we see the symbolic process at work. Feathers worn on the head or
stripes on the sleeve can be made to stand for military leadership; cowrie shells or rings of
brass or pieces of paper can stand for wealth; crossed sticks can stand for a set of religious
beliefs; buttons, elks' teeth, ribbons, special styles of ornamental haircutting or tattooing, can
stand for social affiliations. The symbolic process permeates human life at the most primitive
as well as at the most civilized levels.
Of all forms of symbolism, language is the most highly developed, most subtle, and most
complicated. It has been pointed out that human beings, by agreement, can make anything
stand for anything. Now human beings have agreed, in the course of centuries of mutual
dependency, to let the various noises that they can produce with their lungs, throats, tongues,
teeth, and lips systematically stand for specified happenings in their nervous systems. We call
that system of agreements language. For example, we who speak English have been so trained
that, when our nervous systems register the presence of a certain kind of animal, we may
make the following noise: 'There's a cat.' Anyone hearing us expects to find that, by looking
in the same direction, he will experience a similar event in his nervous system-one that will
lead him to make an almost identical noise. Again, we have been so trained that when we are
conscious of wanting food we make the noise 'I'm hungry.'
There is, as has been said, no necessary connection between the symbol and that which is
symbolized. Just as men can wear yachting costumes without ever having been near a yacht,
so they can make the noise, 'I'm hungry', without being hungry. Furthermore, just as social
rank can be symbolized by feathers in the hair, by tattooing on the breast, by gold ornaments
on the watch chain, or by a thousand different devices according to the culture we live in, so
the fact of being hungry can be symbolized by a thousand different noises according to the
culture we live in: J'ai faim', or 'Es hungert mich', or 'Ho appetito', or 'Hara ga hetta', and so
on.
However obvious these facts may appear at first glance, they are actually not so obvious as
they seem except when we take special pains to think about the subject. Symbols and things
symbolized are independent of each other: nevertheless, we all have a way of feeling as if,
and sometimes acting as if, there were necessary connections. For example, there is the vague
sense we all have that foreign languages are inherently absurd: foreigners have such funny
names for things, and why can't they call things by their right names? This feeling exhibits
itself most strongly in those English and American tourists who seem to believe that they can
www.seyfihoca.com
118
make the natives of any country understand English if they shout loud enough. Like the little
boy who is reported to have said: 'Pigs are called pigs because they are such dirty animals',
they feel that the symbol is inherently connected in some way with the things symbolized.
Then there are the people who feel that since snakes are 'nasty, slimy creatures' (incidentally,
snakes are not slimy), the word 'snake' is a nasty, slimy word.
(From Language in Thought and Action, by S. Hayakawa)
From Word Symbol to Phoneme Symbol
The second process, by which pictures cease to be ideograms and come to stand for specific
linguistic forms, is even more important. The earliest Sumerian writings are just lists of
objects with symbols for numbers against them: for example, four semicircles and the picture
of an ox's head would read 'four oxen'. It seems that writing arose to meet the needs of the
highly centralized city state, and the first writings are records of payments to the temple or
city treasury, and similar transactions.
In this way, pictorial symbols come to stand for various words, which are the names of
concrete objects like sheep, oxen, the sun, houses and so on. Next, by a process of extension,
the same symbols are made to stand for more abstract words related to the original word. Thus
a picture of the sun may come to stand for the words for 'bright', or 'white', and later for the
words 'day' and 'time', and a picture of a whip for words like 'power' or 'authority'.
Perhaps the really crucial development, however, is 'phonetization', the association of a
symbol with a particular sound (or group o£ sounds). First, a symbol for a concrete object is
transferred to some more abstract object which is denoted by the same or a similar word. For
example, the Sumerian word ti meant 'arrow', and so was represented by an arrow in the
script; but there was also a Sumerian word ti which meant 'life', so the arrow symbol came to
be used for this too. The arrow symbol was then felt to stand for the sound of the word ti, and
was used for the syllable ti in longer words. In this way, the original word symbols developed
into syllable symbols, which could be grouped together to spell out a word.
An analogous process in English can be imagined on these lines. A picture of a tavern is used
to represent the word inn. Because of the identity of sound, the same symbol then becomes
used for the word in. At the same time a picture of an eye is used for the word eye, and then
by extension is used for the word sight. Finally the tavern symbol and the eye symbol are
combined to write the words incite and insight, and have now become syllabic symbols. If we
wanted to distinguish between insight and incite in our syllabic script, we could add a third
symbol to show which of the two was intended: we could draw a picture of an orator to show
that we meant incite, or add a symbol for some word like 'wisdom' to show that we meant
insight. When we used the eye symbol by itself, we might wish to indicate whether it stood
for the word eye or the word sight; one way of doing this would be to add a symbol after it
suggesting one of the sounds used in the word intended: for example, if we had a symbol for
the words sow, sew, so, we could add this after our eye symbol to indicate that the required
word began with s. These and similar methods are used in ancient Egyptian and Sumerian
writing.
Sumerian writing is very mixed, using ideograms, word symbols, syllable symbols, and
various kinds of indicators of the types mentioned. Out of it, however, developed the almost
purely syllabic system of cuneiform writing which was used for Akkadian (the language of
www.seyfihoca.com
119
the ancient Babylonians and Assyrians), and which for centuries dominated the writing of the
Near East.
Ancient Egyptian writing also developed into a syllabic system, and was particularly
important for the development of true alphabetic writing (i.e. a script that has symbols
representing phonemes). The important thing about the Egyptian system was that the vowels
were not indicated. Most of the signs (about eighty) stood for a group of two consonants, plus
any vowels whatever. For example, the symbol for a house (par) stood for the group pr, and
this could mean par, per, apr, epr, epra, and so on. But there were twenty-four signs which
stood for only one consonant plus any vowel; for example, the symbol representing a mouth
(ra) stood for the consonant r, and could mean ra, ar, re, er, and so on. When the West Semitic
peoples living round the eastern shores of the Mediterranean developed a script, they did so
by taking over from the Egyptians just these twenty-four signs. Originally, this must have
been a syllable system, in which each of the signs stood for a number of possible syllables,
like the Egyptian ra, ar, re, er, etc.: but in fact it is formally identical with a purely alphabetic
system in which only the consonants are written and the vowels are left out.
The final step, of having fixed and regular symbols for the vowels, was made by the Greeks
when they took over this Semitic alphabet. Some of the consonant sounds of Phoenician did
not exist in Greek, and the Greeks used the corresponding symbols for vowels. For example,
the first letter of the West Semitic alphabet, derived from the picture of an ox, was 'aleph', and
stood for a kind of h sound (represented in the spelling by'); the Greeks of the period did not
use this sound, and took the letter over as alpha, representing an a sound. Thus was reached at
last a system of writing where symbols stand for phonemes, and all later alphabetic systems
are ultimately derived from this Greek achievement. The great advantage of the system is the
relatively small number of symbols needed, which makes universal literacy possible.
(From The Story of Language, by A. L. Barber.)
LAW
MODERN CONSTITUTIONS
It is natural to ask, in the light of this discussion, why it is that countries have Constitutions,
and why most of them make the Constitution superior to the ordinary law.
If we investigate the origins of modern Constitutions, we find that, practically without
exception, they were drawn up and adopted because people wished to make a fresh start, so
far as the statement of their system of government was concerned. The desire or need for a
fresh start arose either because, as in the United States, some neighbouring communities
wished to unite together under a new government, or because, as in Austria or Hungary or
Czechoslovakia after 1918, communities had been released from an Empire as the result of a
war and were now free to govern themselves; or because, as in France in 1789 or the U.S.S.R.
in 1917, a revolution had made a break with the past and a new form of government on new
principles was desired; or because, as in Germany after 1918 or in France in 1875 or in 1946,
defeat in war had broken the continuity of government and a fresh start was needed after the
war. The circumstances in which a break with the past and the need for a fresh start come
about vary from country to country, but in almost every case in modern times, countries have
a Constitution for the very simple and elementary reason that they wanted, for some reason, to
begin again and so they put down in writing the main outline, at least, of their proposed
www.seyfihoca.com
120
system of government. This has been the practice certainly since 1781 when the American
Constitution was drafted, and as the years passed no doubt imitation and the force of example
have led all countries to think it necessary to have a Constitution.
This does not explain, however, why many countries think it necessary to give the
Constitution a higher status in law than other rules of law. The short explanation of this
phenomenon is that in many countries a Constitution is thought of as an instrument by which
government can be controlled. Constitutions spring from a belief in limited government.
Countries differ however in the extent to which they wish to impose limitations. Sometimes
the Constitution limits the executive or subordinate local bodies; sometimes it limits the
legislature also, but only so far as amendment of the Constitution itself is concerned; and
sometimes it imposes restrictions upon the legislature which go far beyond this point and
forbid it to make laws upon certain subjects or in a certain way or with certain effects.
Whatever the nature and the extent of the restrictions, however, they are based upon a
common belief in limited government and in the use of a Constitution to impose these
limitations.
The nature of the limitations to be imposed on a government, and therefore the degree to
which a Constitution will be supreme over a government, depends upon the objects which the
framers of the Constitution wish to safeguard. In the first place they may want to do no more
than ensure that the Constitution is not altered casually or carelessly or by subterfuge or by
implication; they may want to secure that this important document is not lightly tampered
with, but solemnly, with due notice and deliberation, consciously amended. In that case it is
legitimate to require some special process of constitutional amendment - say, that the
legislature may amend the Constitution only by a two-thirds majority or after a general
election or perhaps upon three months notice.
The framers of Constitutions have more than this in mind. They may feel that a certain kind of
relationship between legislature and the executive is important, or that the judicature should
have a certain guaranteed degree of independence of the legislature and executive. They may
feel that there are certain rights which citizens have and which the legislature or the executive
must not invade or remove. They may feel that certain laws should not be made at all. The
framers of the American Constitution, for example, forbade Congress to pass any ex
post,facto law, that is, a law made after the occurrence of the action or the situation which it
seeks to regulate-a type of law which may render a man guilty of an offence through an action
which, when he committed it, was innocent. The framers of the Irish Constitution of 1937
forbade the legislature to pass any law permitting divorce.
Further safeguards may be called for when distinct and different communities decide to join
together under a common government but are anxious to retain certain rights for themselves.
If these communities differ in language, race, and religion, safeguards may be needed to
guarantee to them a free exercise of these national characteristics. Those who framed the
Swiss, the Canadian, and the South African Constitutions, to name a few only, had to consider
these questions. Even when communities do not differ in language, race, or religion, they may
still be unwilling to unite unless they are guaranteed a measure of independence inside the
union. To meet this demand the Constitution must not only divide powers between the
government of the Union and the governments of the individual, component parts, but it must
also be supreme in so far at any rate as it enshrines and safeguards this division of powers.
www.seyfihoca.com
121
In some countries only one of the considerations mentioned above may operate, in others
some, and in some, all. Thus, in the Irish Constitution, the framers were anxious that
amendment should be a deliberate process, that the rights of citizens should be safeguarded
and that certain types of laws should not be passed at all, and therefore they made the
Constitution supreme and imposed restrictions upon the legislature to achieve these ends. The
framers of the American Constitution also had these objects in mind, but on top of that they
had to provide for the desire of the thirteen colonies to be united for some purposes only and
to remain independent for others. This was an additional reason for giving supremacy to the
Constitution and for introducing certain extra safeguards into it.
(From Modern constitutions by K. C. Wheare)
THE FUNCTIONS OF GOVERNMENT
England had become a political unit before the Norman Conquest. By the middle of the
fourteenth century England had a general or common law, and the main functions of
government, those relating to the agricultural economy of the village apart, were vested in the
central authorities that surrounded the king and in the king's local representatives. From this
time we can speak of England as a 'State'.
The growth of the division of labour, which may also be described as the disappearance of
subsistence agriculture, and which became marked in England with the development of the
wool trade, did not of itself imply a corresponding increase in the functions of government.
Some increase there was, of necessity. A more extensive system for maintaining order, some
organization for the repair of roads and bridges, a development of the instruments for the
settling of disputes and the imposition of taxation, were obviously required by the gradual
breakdown of village self-sufficiency. Trade and commerce are, however, matters of private
arrangement, and the State itself need do no more than provide the laws to regulate disputes,
the judicial institutions to administer the laws, and the currency to serve as the instrument of
exchange. It happens that in England the State went somewhat further, and was compelled to
make some attempt to control the movement of labour. So long as labour was provided within
the manor by labourers who themselves had interests in the land of the manor, the problem
was one for the manor alone; but hired labour became more and more the practice as
specialization developed, in ancillary trades as well as in agriculture itself, and labourers in
search of work left their manors, with the result that the State interfered in the interests of
public order and the needs of employers. This was especially so after the Black Death created
a dearth of labour. Justices of labour were created to regulate labour, and were subsequently
merged with the justices of the peace, who were originally concerned only with the
apprehension of criminals.
By the sixteenth century the State had to concern itself far more than in the past with external
relations, and at the same time it had not only to deal internally with 'wandering beggars' who
were 'rogues and vagabonds', but also to provide poor relief; and the requirements of transport
compelled more effective provision for the maintenance of roads and bridges. The result of
further economic development was that in the eighteenth century the State had additional
functions relating to the regulation of foreign trade, and the government of colonies.
Moreover, though the foundation of the nation's economic life remained in the village, the
towns became increasingly important.
www.seyfihoca.com
122
The Industrial Revolution - falsely so named, because there was no clear break in the chain of
development, and the use of steam-power merely accelerated the speed of change - altered the
emphasis. Mining, iron smelting, and other industries became as important as agriculture. The
domestic wool industry was superseded by weaving factories, and foreign trade developed the
cotton industry. New methods of transport were developed. These changes did not in
themselves require any new functions of government; indeed, the ideas of the time, especially
after Adam Smith, were in favour of easing such governmental restrictions on trade and
industry as already existed and of developing a system of free competition. There were,
however, repercussions in the field of government. New roads, canals, and, in due course,
railways were required. New ports and harbours were opened. Though some or all of these
might be provided by private enterprise, the intervention of the State was necessary to enable
land to be acquired, traffic to be regulated, roads crossed, bridges built, and so on.
The most important effect from a governmental aspect, however, was the congregation of vast
numbers of inhabitants into towns, some of which were ancient market towns, some mere
villages, and some entirely new growths where before agriculture had reigned supreme. The
old methods by which the family had obtained its own water and disposed of its own sewage
and refuse were dangerous to health in these great urban centres. New police forces had to be
created, streets paved and lighted, refuse collected, and water supplies and main drainage
provided. The existence of unemployment on a large scale created a new problem for the poor
law. In particular it was recognized that ill-health was responsible for much unemployment
and, therefore, for much expenditure on poor relief. On the one hand medical services and
hospitals had to be provided, and on the other hand there was new agitation for preventive
remedies-pure water, clean streets, good sewerage, and sewage disposal. Also, these remedies
were not enough if a large part of the population spent most of its time in unhealthy factories
and mines. Hours and other conditions of employment had to be regulated; and though this
involved in the earlier experiments only restrictions on private enterprise, they were soon.
found to be ineffective without State inspection.
Thus, while the State in the nineteenth century was freeing trade from the cumbersome
restrictions of the eighteenth century, it proceeded to regulate industry both directly in the
interests of public health and indirectly by providing services out of the produce of taxation.
After 1867 sections of the working class had the vote, and the individualist principles which
appealed to the middle class of the industrial towns became much less strong. Accordingly,
the provision of services became part of State policy. It recognized, for instance, its
responsibility for the education of the young and provided schools, not merely by subsidizing
ecclesiastical bodies, as it had done since 1833, nor merely for the benefit of 'pauper children',
but directly through the local authorities and for the benefit of all children. There was, too, a
general and progressive reform of all the public services, and new services, such as housing,
transport, and electricity, were provided. In the present century the pace of development has
accelerated; and while on the one hand existing services have been expanded, new services
like pensions, insurance, and broadcasting have developed; and since 1945 some of the
services formerly provided by private enterprise have been taken over by the State.
(From The law and the constitution by Ivor Jennings)
INITIATIVE AND REFERENDUM
It is sometimes argued that a democratic system requires the embodiment of the initiative and
the referendum in the constitution. A people, it is said, does not really control its own life if its
www.seyfihoca.com
123
only direct participation in the business of making legal imperatives is confined to choosing
the persons responsible for their substance. By the initiative, the popular will can take positive
form; and by the referendum, the people can prevent action by its representatives with which
it is not in agreement. Direct government, it is claimed, provides a necessary supplement to a
representative system; otherwise, as Rousseau said of the English people, it is only free at
election time.
But this is, it may be suggested, both to mistake the nature of the problems which have to be
decided, and the place at which popular opinion can obtain the most valuable results in action.
In all modern states, the size of the electorate is necessarily so large that the people can hardly
do more, as a whole people, than give a direct negative or affirmative to the questions direct
government would place before them. Legislation, however, is a matter not less of detail than
of principle; and no electorate can deal with the details of a measure submitted to it for
consideration. Direct government, in fact, is too crude an instrument for the purposes of
modern government. It fails to make discussion effective at the point where discussion is
required; and it leaves no room for the process of amendment. One might, it is true, leave
certain broad questions of principle to popular vote, whether, for instance, the supply of
electricity should be a national or a private service. But all other questions are so delicate and
complex that the electorate would have neither the interest nor the knowledge, when taken as
an undifferentiated electorate, to arrive at adequate decisions.
Nor is this all. Not only can most questions not be framed in a way which can make direct
government effective; the secondary results of the system are also unsatisfactory. It is hardly
compatible, for instance, with the parliamentary system since it places the essential
responsibility for measures outside the legislature. Such a division of responsibility destroys
that coherence of effort which enables a people adequately to judge the work of its
representatives. It assumes, further, that public opinion exists about the process of legislation,
as well as about its results. But the real problem of government is not forcibly to extract from
the electorate an undifferentiated and uninterested opinion upon measures about which it is
not likely to be informed. It is rather to relate to the law-making process that part of public
opinion which is relevant to, and competent about, its substance before that substance is made
a legal imperative. 'This involves not direct government, but a method of associating the
relevant interest-units of the community with the making of measures that will affect their
lives. A referendum, for example, on a national scheme of health insurance would give far
less valuable results than a technique of consultation in which the opinions of doctors, tradeunions, and similar associations were given a full opportunity to state their view before the
scheme was debated in the legislative assembly. Effective opinion for the purpose of
government, in a word, is almost always opinion which is organized and differentiated from
that of the multitude by the possession of special knowledge. Popular opinion, as such, will
rarely give other than negative results; and it seems to be the lesson of experience, very
notably on the record of Switzerland, that it is so firmly encased in traditional habit, as to
make social experiment difficult when it is a reserve-power-outside.
(From An Introduction to Politics by H. J. Laski)
S v S (FINANCIAL PROVISION: DEPARTING FROM EQUALITY) [2001] 2 FLR 246
[2001] 2 FLR 246
Family Division
www.seyfihoca.com
124
Peter Collier QC
(sitting as a Deputy Judge of the High Court)
2 February 2001
Financial provision – Factors to be considered – Departure from equality
The parties were married in 1967 when the wife was 25 and the husband 27. There were two
children, a son aged 37 who was financially independent, and a daughter, aged 27, who lived
with her mother and was supported by her financially to some extent. The wife worked
outside the home until her son was born and thereafter ran the house and worked in the
husband’s business. The marriage broke down in 1998 when the husband left the matrimonial
home and disclosed that he had been having an affair with another woman since 1992, and
that she had a child by him. At the hearing of the wife’s application for ancillary relief the
judge found that the husband’s share of the total assets of the parties amounted to £2,102,890,
consisting of his interests in various businesses, his pension fund, some properties, his halfinterest in the matrimonial home and investments, a yacht, and a number of vintage cars. The
wife had her half-share in the house, her car, and a portfolio of investments, all amounting to
£629,166. It was argued on the wife’s behalf that this was a case in which to follow the
principles set out in White v White, so as to achieve a clean break and an equal division of the
assets, by the transfer by the husband to the wife of his share of the matrimonial home and of
the joint investments, together with a lump sum of £400,000. Counsel for the husband
contended that in the circumstances that approach would not be fair having regard to the
criteria in s 25 of the Matrimonial Causes Act 1973.
Held – the assets of the parties had been valued on a net value basis, ie taking account of sale
costs and tax liabilities, and did not fall to be reduced in the husband’s case because some of
the capital was not readily available or was not producing an income. However, having regard
to all the matters in s 25, the award of £400,000 to the wife, required to achieve equality,
would discriminate against the husband. On the facts, such an award would mean that the wife
lived in luxury, while for the husband with a new family to support, things would be much
tighter. If an agreement was harder on one party than the other, then there was good reason to
depart from equality. The court should aim to provide both parties with a comfortable house
and sufficient money to discharge their needs and obligations. On that basis there would be an
order for the husband to transfer to the wife his half-share of the matrimonial home and of the
joint investments, and to pay to her a lump sum of £300,000.
Statutory provision considered
Matrimonial Causes Act 1973, s 25
Cases referred to in judgment
Leadbeater v Leadbeater [1985] FLR 789, FD
Piglowska v Piglowski [1999] 2 FLR 763, [1999] 1 WLR 1360, [1999] 3 All ER 632, CA and
HL
White v White [2000] 2 FLR 981, [2000] 3 WLR 1571, [2001] 1 All ER 43, HL
www.seyfihoca.com
125
[2001] 2 FLR 247
Martin Pointer QC for the petitioner
Rodger Hayward-Smith QC and Andrew Marsden for the respondent
PETER COLLIER QC: This is an application for ancillary relief by Mrs S, following the
breakdown of her marriage to her husband Mr S.
The marriage
The petitioner (who I will hereafter refer to as ‘the wife’) is Mrs S. She is 58 years of age,
having been born on 14 June 1942. The respondent, her husband, is Mr S. He is now 60 years
of age, having been born on 30 April 1940. They were married on 20 September 1967, when
she was 25 and he was 27.
The children
S, born on 27 July 1969, he is now 31. He lives in America and is, of course, financially
independent. He has not featured in the evidence, unlike his sister M, who is now 27, having
been born on 15 July 1973. She has completed her tertiary education; in fact she has two MA
degrees, but at the present time has no permanent employment. Her father thinks she should
be more independent than does her mother, who continues to support her, having provided a
car for her, by way of paying the hire-purchase instalments on the car, which is said to be a
loan to her. She has also assisted her, towards the end of last year, in her attempts to set up a
small retail business, by buying goods for her; again that is a loan of some £5000 which has
been shown in the schedules.
The husband helped M to buy a property in Colchester some while ago now. She currently
rents that out, receiving about £500 per month, the mortgage costing her some £340 a month;
and at the present time she chooses to live with her mother.
There was a Trust Fund set up for the children in 1976, for tax and inheritance reasons. Since
December 1998, M has received a total of £31,556 from the Trust Funds.
Employment
At the time of the marriage the husband was working with his brother, H, in a business that
they owned and ran, known as ‘[SCL]’, making car alarms and battery chargers. The business
had been started by them in 1960.
When they married, the wife was working as a PA/secretary to the Managing Director of
MGN. In 1969, when S was expected, she gave up work and, thereafter, she has not worked
outside the family or the family businesses. It is accepted that she worked in the home,
bringing up the children and running the home. When S was about 2, she began to work for
the husband’s company. She was an employee, she was on the books; she did secretarial and
what I described as ‘general duties’, but they covered quite a lot of matters, including hosting
people who came to visit.
www.seyfihoca.com
126
He says that she has exaggerated her role in the business, and she says that he has minimised
it. I find it is not necessary to define it with precision, suffice to say that in my judgment she
made a significant contribution to the business.
[2001] 2 FLR 248
Homes
The first matrimonial home was in Tomswood Road, Chigwell; it was bought for £12,000,
with the assistance of a gift of £3000 from the wife’s father, A. The husband says that he does
not recall exactly how much, but he believes they also had some help from his family. Again,
greater precision will not affect the fact that these two people began to share their lives, with
support from their families, it was a real partnership, each bringing what they could to the
partnership, supported by both their families and they intended to work hard to make their
family and business something of which they could both be proud.
In 1976 or 1977, Tomswood Road was sold for £43,000 and they purchased Hunters,
Chigwell, for £58,750 with the aid of a mortgage of £25,000. The mortgage was paid off out
of the proceeds of the subsequent sale of the business. The wife remains there to this day, and
it is accepted that she should stay there. I am told that the house is not, itself, attractive, being
a chalet-bungalow that has been extended, and at the time these proceedings commenced it
was in need of substantial redecoration. Over the years that she has lived there, the wife has
put time and money into the development of the garden, which is clearly her pride and joy.
She envisages that in due course, if necessary, she could put in a ramp and push her zimmerframe up and down between house and garden.
In October 1991, the husband and his brother H sold SCL to RG, and some of the proceeds
have been used to purchase investment portfolios: there is a joint portfolio; there are separate
ones for each of the husband and wife; there is one creating a settlement trust for the children;
there is also a pension fund, which now provides an income for the husband.
Some time before 1992, the husband formed a relationship with a Miss D who bore his
daughter, B, in that year. The existence of both that relationship and that child were kept
secret from the wife until April of 1998. I simply record that the husband had had a brief
relationship with another woman in the mid 1980s, he had left the matrimonial home for a
short time but had returned.
Businesses
In addition to SCL, there are other businesses in which the husband has been and, in some
instances, remains involved. PPEL, was set up by the parties to receive commission on the
sale of battery chargers for RG after the sale of SCL to RG. PPEL commenced business in
1993. The company has had an income of as high as £19,000, but in recent years it has
deteriorated to £2000–£3000 a year. I have been told that the company was used as a vehicle
to buy equipment for and fund travel by the wife; the husband says that he anticipates no
future profits from this company.
SL and TFL were also companies run by the husband and his brother. I shall say more of
those later when I come to deal with valuations. AVIL – I need say little about that: it was a
www.seyfihoca.com
127
business in which the parties invested money and which failed. Some moneys had been repaid
to the husband, and I understand there will be no further income from that source.
Properties
There are other properties that have come within the family at different times. Lower Clapton
Road, these are commercial premises which are let.
[2001] 2 FLR 249
They are premises left to the husband by his mother, and comprise a shop, a flat and an
advertising hoarding. They currently produce an income: the shop and flat, some £11,000
plus; the hoarding, £7500.
There is an issue regarding the hoarding. At an earlier stage in the proceedings it was stated,
in relation to its value, that ‘the income from the advertising panel, which in our view is high,
is relatively insecure’. Since then, it has been disclosed that the local council, Hackney, had
written to the husband on 18 November 1997 stating that in their view there was neither
'deemed nor 'express' consent for this board which is in a conservation area. Since then, the
advertiser, Postmobile, has written to the council stating that the board has been in its present
situation for so long that it must have deemed consent. That letter has been acknowledged and
there has been no further action on the part of the local authority. The husband told me that he
thought the board had been there for over 45 years. On that evidence, I am not able to
discount the rent received in relation to these premises.
There are factory premises in Springfield, Chelmsford. These premises were purchased
through a Business Expansion Scheme and are rented out, producing an income in the region
of some £10,000 a year.
Aveley Way, Malden, is the husband’s present home where he lives with Miss D, B and
another daughter of Miss D’s by a former relationship.
For completeness, I should perhaps add that prior to the husband’s purchase of that property,
into which he moved with his new family, Miss D had been living in a council house which
she has now purchased with the aid of a loan of £33,331 from the husband. That property she
now lets out at £5400 pa, and I am told that the property is now worth some £55,000.
Cars
The parties have an interest in classic cars. The husband rebuilt several of these and they
would spend a significant amount of time touring and rallying in these cars. The family’s
present collection of cars include several classic cars, an E-Type Jaguar, a Jowett Jupiter and a
Jowett Javelin. They are all with the husband, who has built a garage at his new home to
house them. He also owns an Audi Quatro and a Peugeot 205. The agreed value of all those
cars is £41,200. For everyday use the husband has an imported Toyota Estate car worth some
£1500; the wife has a Volkswagen Golf, bought with the help of a loan from her sister of
£17,000 in November 1998. The Golf is now said to be worth £10,000.
The boat
www.seyfihoca.com
128
There is a boat, an Oyster 435, known as ‘Star Charger 4’, which is currently moored in
Antibes in the South of France. There is no agreed valuation, but I am asked to come to a
view about its value based on a variety of material that has been put before me and to which I
shall turn in due course.
Bank accounts
There are on both sides various bank accounts.
Breakdown of the marriage
The breakdown of the marriage happened in April 1998. The husband does not dispute the
wife’s account that he left the matrimonial home on 27 April 1998, having disclosed his affair
2 days earlier on 25 April 1998.
[2001] 2 FLR 250
Present circumstances
The wife continues to live in the former matrimonial home, she has her own investment
income and the £350 per month ordered by way of interim periodical payments.
The husband lives with Miss D, who is 41 years old, B, who is now aged 8, and Miss D’s
other child, E, who is aged 15, at Aveley Way, Malden. It has been described to me as a
typical small modern estate house on the outskirts of the town. The husband says he bought it
only intending to stay there until he knows the outcome of these proceedings, when he hopes
to buy something more suitable where he can keep his cars and spend his remaining years. He
is retired but clearly he remains an active man.
Valuation of the assets
These are mostly agreed, but the outstanding issues remain in relation to the husband’s
interest in the companies SL and TFL, and the value of Star Charger 4, the boat.
I deal first with SL and TFL. SL manufactures tubular furniture, and TFL does the same, and
there is a trading relationship between the two companies. TFL owes SL in excess of
£300,000, and TFL has a deficiency of some £63,000. I am told it is to SL’s advantage to
continue the present situation as it is able to use up tax losses, but to run the two companies
involves the doubling of some of the costs. The management of both companies is the same;
as I understand it, they have the same directors and similar shareholders. Mr P is the only nonfamily member who is a director, otherwise, the husband, plus his sister-in-law, PS, and her
father, DL, are the directors.
I am asked to resolve the dispute as to the valuation of the husband’s interest in these two
companies. In order to assist me, I heard from two accountants, ME, called on behalf of the
wife, and PM on behalf of the husband.
In relation to SL, they agree that the basis of the valuation should be the company’s net asset
value. There is a small difference between them as to whether one should discount SL’s net
assets by the amount of TFL’s deficiency, which is £63,539, since TFL still owes SL
www.seyfihoca.com
129
£306,563. The net asset value of the shares at face value, on ME’s valuation, is £92,582, and
PM’s valuation produces £102,478.
They also differ in the degree to which the appropriate figure should be discounted. I am told
that the scale accepted by the Inland Revenue is between 40 and 50%. PM says the proper
discount is 40%, ME says it is 50%. Which is chosen is dependent upon the extent of the
control that the husband has over the company. The Articles of Association provide that the
sale should be, in the first instance, to its present shareholders who are all in some way
connected with the family, and, of course, Mr P. If they decide not to buy, then the company
can purchase the shares if it has sufficient reserves. Finally, if neither of those routes produce
a sale, then a third party sale can take place.
What is the extent of the control that the husband has? It is argued that he does not involve
himself in the day-to-day management of the business; that the other directors, as I have said,
are Mr P who manages the company, DL and his daughter, PS, who is the widow of H, the
husband’s brother.
ME is the company auditor and occasionally attends board meetings. He
[2001] 2 FLR 251
did not disagree when I said to him that I thought it would be suggested that the husband was
a strong character. He told me that the husband does attend board meetings where he
expresses his views and that he is respected by the other directors. I am satisfied that in key
decisions of the sort we are contemplating, the husband would have substantial influence; and
so, given the agreed scale of 40–50%, I have assessed the discount at 40% in relation to SL.
Also given the relationship of the companies, I do not propose to make a reduction in value
because of TFL’s deficiency. Consequently, I determine the value of the husband’s interest in
SL at £62,000.
That leaves TFL. Unlike SL, where ME and PM agree about the basis of the valuation, that is
not the position with TFL. There are no assets, so I cannot value the company on a net asset
basis. The company has paid no dividends, so I cannot value it on a dividend yield basis. PM
says that nevertheless a trade sale basis can be arrived at on the basis of the profit. ME says
that there is no track record and the profit is not reliable.
Having explored with ME how these companies are related in their trading, it became clear
that TFL was intended to have a genuinely separate existence, resulting from a joint venture
between SL and another business. However, the other party gradually withdrew, leaving the
owners of SL to run TFL. Their two directors, Messrs. C and J, were replaced by the husband
and DL, and the company is now run solely to gain certain tax advantages arising from past
tax losses. The result is that TFL is, realistically, just a division of SL. ME said that once
someone was told the story of TFL, they would not want to buy TFL’s shares. I agree. In my
judgment, the shares have no saleable value, certainly not one that could be assessed on a
trade sale basis (as PM attempted to do), and no other basis has been put forward to me.
So I assess the value of the husband’s interest in SL at £62,000, but the value of his interest in
TFL as nil. Mr Hayward-Smith QC suggests that that figure should be further discounted as it
is said by the accountants not to be the equivalent of cash. I have considered carefully his
submissions, but in the end I have decided that although they are not realisable as
www.seyfihoca.com
130
immediately as if it was a publicly quoted company, I should not make any allowance for that,
otherwise I would have to go through the same exercise with all the properties, each of which
may take some considerable time to realise. I must assume that everything has a value and, in
the absence of firm evidence about a difference between market value and forced sale value,
and also a reason for preferring the latter in some particular instance, I should take the market
value of all the assets, including this one, as their true value.
The boat
That leaves the other issue, the value of the Star Charger boat. This was built between 1983
and 1985. The husband has rebuilt part of the cabin area. He jointly owned the boat with his
brother, but bought out his brother’s share for £50,000 when his brother bought another boat
in March 1998. Attempts had been made to sell the boat in 1996 through Oyster Brokerage of
Ipswich. They advertised the boat at an asking price of £125,000; no sale was achieved.
Oysters have since said, in the summer of 1999, that they had hoped to achieve a price of
£110,000–£100,000. Apparently it would sell better in the UK, so any sale here for a better
price would necessitate transporting the boat to the UK at a cost of about £4000, in addition to
the
[2001] 2 FLR 252
10% commission charged by the selling agent.
The self-same agents were asked for a valuation in 1999. They said that their insurance policy
precluded them from giving valuations, but said that they estimated the value at £82,500–
£87,500 in June of 1999. They were again asked in January 2001. Making the same
reservation regarding valuation/estimation, they said it would have depreciated by a further 5–
10% and so now would be worth £75,000
The insured sum throughout has been £112,900, which is a replacement value. The boat
suffers from osmosis, which has been taken into account in the estimations, but it will effect
its saleability. The wife, going on the original advert for sale, and discounting, says that the
value should be assessed at £95,000 less 10% commission.
Doing the best I can, it seems to me that if I take the mean between £82,500 and £87,500,
which is £85,000, and then deduct the commission and transport costs, and then also make
some small allowance for depreciation, I come to a figure of about £70,000, which is the
amount, in the absence of better evidence, I shall take as the value of the boat.
As the value of all the other assets is now agreed, those findings of mine produce a total asset
value of £2,732,055.
The arguments as to how those assets should be divided which were presented (in the broadest
outline) to me are as follows.
The wife’s case
On behalf of the wife it is argued that this is a classic case in which to follow the principles
set out by Lord Nicholls of Birkenhead in his speech in the House of Lords case of White v
White [2000] 2 FLR 981, [2000] 3 WLR 1571. It is said that there are sufficient funds for a
www.seyfihoca.com
131
clean break, with an equal division of all the assets. That can be achieved by transferring the
husband’s half-share of the house, and of the joint funds, to her absolutely, and adding to that
a lump sum to make up half the value of the assets; that sum is just over £400,000. It is said
that the husband has not displaced that as a starting point.
The husband’s case
For the husband, it is said that that is not the correct approach; that this is only marginally a
White case. It is said that when I look at the statutory criteria set out in s 25 of the
Matrimonial Causes Act 1973, my tentative conclusions will be that equality is not possible
here if I am to be fair and do justice to those criteria. He particularly argues that given the fact
that significant parts of the assets are, on the one hand, the home, which will go to the wife,
and, on the other hand, his pension fund, then I am restricted to the extent that I can balance
the figures equally.
Section 25 of the Matrimonial Causes Act 1973
It seems to me that I must begin by examining the statute. Lord Nicholls of Birkenhead said in
White v White [2000] 2 FLR 981, 992, [2000] 3 WLR 1571, 1581, quoting from Lord
Hoffmann’s speech in Piglowska v Piglowski [1999] 2 FLR 763, 782, [1999] 1 WLR 1360,
1379:
‘ section 25(2) does not rank the matters listed in that subsection in any kind of hierarchy.
[2001] 2 FLR 253
I think it would be helpful, therefore, if I deal first with those matters relative to background
and history. Subsection (d): ‘the age of each party to the marriage and the duration of the
marriage;’. I recited the history earlier: he 60, she is 58; the marriage was a long one, lasting
almost 31 years.
Subsection (f): ‘the contributions which each of the parties is or is likely in the foreseeable
future to make to the welfare of the family, including any contribution by looking after the
home or caring for the family;’. As I indicated earlier, I take the view that for many years this
marriage was a partnership, each contributing equally but in different ways, and no one has
argued otherwise.
Subsection (c): ‘the standard of living enjoyed by the family before the breakdown of the
marriage;’. It is part of the wife’s evidence that in the final year of marriage they got through
£120,000, and she would have me believe that this was typical expenditure. I find that the
£120,000 is not proved. I consider it was an unlikely sum to have been spent, given that it is
the sum of their two proposed future budgets, which includes a £24,500 mortgage for the
husband. Clearly they were very comfortably off, whatever they wanted to do they could and
did do. She told me they ‘lived the life of Riley’. She also said that her experience before the
separation was that there was plenty for whatever she wanted to do, he begrudged it but there
were adequate funds. They enjoyed classic cars that the husband rebuilt; they had a yacht in
the south of France; they took holidays as and when they wished. Her expenditure on the
garden was unrestrained. She made regular major shopping expeditions to the West End. In
latter years, the husband had a child he was supporting to the tune of £350 per month, without
his wife, who monitored their finances, suspecting anything.
www.seyfihoca.com
132
Subsection (e): ‘any physical or mental disability of either of the parties to the marriage;’.
There are none of significance. The wife has been depressed, but no one suggests that she
should go out to work or that she has special needs that should be provided for.
Subsection (g): ‘the conduct of each of the parties, if that conduct is such that it would in the
opinion of the court be inequitable to disregard it;’.
Subsection (h): ‘ the value to each of the parties to the marriage of any benefit which, by
reason of the dissolution or annulment of the marriage, that party will lose the chance of
acquiring.
Neither of those subsections has any relevance to my determination.
I turn to the two main subsections, those dealing with assets and resources on the one hand,
and with needs, obligations and responsibilities on the other.
Subsection (a):
‘ the income, earning capacity, property and other financial resources which each of the
parties to the marriage has or is likely to have in the foreseeable future, including in the case
of earning capacity any increase in that capacity which it would in the opinion of the court be
reasonable to expect a party to the marriage to take steps to acquire.
Income
Apart from her interim maintenance of £350 month, the wife has investment income only.
Currently there is that produced by her own portfolio and her half of the joint investments,
producing something in the region, I have been
[2001] 2 FLR 254
told, of £25,000. It is not suggested that she will have any income other than what is produced
by those investments, plus any produced by the lump sum that I order the husband to pay her.
The husband has income from a number of sources. His pension fund currently produces
£52,470 per year. His rental from properties produces £28,500 per year, though I note that
something less than that was shown in his tax return for the year ending April 1999. He has
investment income of his own, of about £12,500, and his half-share of the joint investments,
producing something over £5000. Some other earnings have been produced from time to time,
such as bonuses from SL. It is clear that he has more cards available to him in his hand than
she has in hers, and that he has some more flexibility than she does.
Earning capacity
She has none. I accept that he wants to remain retired, but I note that he is a company director,
having his say in those companies. I do not consider that either of them should take steps to
increase their earning capacity at this stage of their lives, they have both reached the age when
they can expect to live now off the fruits of their past labours.
Property and other financial resources
www.seyfihoca.com
133
The matrimonial home and the portfolio of joint investments are the only jointly owned items;
each has other property in their own name and other expectations as well as obligations.
The wife
She has her half-interest in the home, which is £252,200, and her half-share of the joint
investments at £81,050. She owns a car said to be worth £10,000, but she owes her sister
£17,000 which was loaned to her to buy that car. She is owed £5000 by M, which are moneys
advanced to help her set up a small business venture at the end of last year. She has some
small amounts in her bank accounts. Her main asset is her own portfolio of investments, said
to be worth £286,608. She has to date paid legal costs of £12,553 (relevant by reason of the
decision in Leadbeater v Leadbeater [1985] FLR 789). Her total pot, therefore, amounts to
£629,166.
The husband
He has his half-interest in the home, £252,200 net, and his half-share of the joint investments,
£81,050. He owns a collection of cars valued at £41,200; his everyday car is worth £1500. He
owns the boat which I have assessed at £70,000. He has lent £33,331 to Miss D. He expects to
receive £75,000 from his brother H’s estate. He has an investment portfolio worth £356,313
net. He has two PEPs worth £12,526. He has SL shares, which I have valued at £62,000. He
has over £115,000 in various bank accounts. His pension fund is valued at 689,037. He has
paid costs to date of £38,369. His total pot, therefore, comes to £2,102,890. The total joint
assets amount to £2,732,055. I come to subs (b):
‘ the financial needs, obligations and responsibilities which each of the parties to the marriage
has or is likely to have in the foreseeable future.
[2001] 2 FLR 255
The section refers to ‘needs, obligations and responsibilities’.
Lord Nicholls of Birkenhead was critical of the phrase ‘reasonable requirements which has, in
recent years, been used in these courts to comprehend the assessment that is made under this
subsection in the context of the rest of the section.
I shall consider her ‘needs, obligations and responsibilities in the context that is set by the
other factors to which I have already referred. If this is different from ‘reasonable
requirements’, then so be it.
Housing
The wife needs a house; she wants to live in Hunters. She tells me it is the ugliest house she
knows, but it has the nicest garden she knows. She wants to stay there; there are no dependent
children, so there is no particular need to do so. The house is worth over £500,000, which is a
significant part of the assets. One of the difficulties in this case is created by her desire to stay
in that house.
The husband has bought a house; it has an equity of £80,000. He hopes to improve his lot in
relation to housing, dependent on the outcome of these proceedings. Mr Pointer QC says he
www.seyfihoca.com
134
does not plan to move, he is settled where he is, but I accept the husband’s evidence that he
would like to move, and he will if he is able.
Obligations and responsibilities
The husband has a new family, Miss D, B (aged 8) and E (aged 15), and there has been
argument about the effect that I should give to that. The wife has no obligations, though she
supports her daughter M to an extent, as I have already described. M’s father says that her
mother indulges her. If she does, that is her choice, perhaps understandable in the
circumstances.
General standard of living
I find that the parties have a similar standard of living, as one might expect from people who
have lived together for over 30 years. They are not critical of each other’s proposed budgets
overmuch. She questions his need for a boat costing £10,000 pa, and some expenditure on his
cars; and he inquires as to her proposed expenditure of £7000 pa in relation to clothing.
In relation to housekeeping figures, he has a budget of £7800, which is £150 a week; she has a
budget of £4633, which is £90. That seems to me to be perfectly reasonable, relative to and
consistent with the way they have lived in the past.
As for hobbies, he has his boat and cars, and she has her garden.
Those are the criteria I have to consider. The aim, or, in modern parlance, ‘the overriding
objective’, is now said to be to achieve a fair outcome. Well, what are the arguments? Mr
Pointer says this is a classic White case and that all I need to do, having arrived at a grand
total value for the assets, is to make an order that will result in them each having a half share
of that total.
Mr Hayward-Smith says: ‘Not so, that will result in unfairness and that there are reasons why
I should not go down that route. First, he says this is not really a White case: White assumes a
‘clean break case where the children are grown up and independent, and where the assets
exceed the amounts required by the parties for their financial needs in terms of a home and
income for each of them. In those circumstances, equality should be the
[2001] 2 FLR 256
yardstick against which any division should be checked.
Mr Hayward-Smith says that if this is a White case, it is ‘on the cusp’, and he has sought to
persuade me that for various reasons the capital is not all readily available and sufficiently
free to make that equal division. There is, he says, nothing binding on H’s estate to pay, and
no timescale within which they will pay; the estate is not straightforward and the husband will
not be paid until it is finalised.
I am not told that he has any charge on Miss D’s property; and, as she doesn’t earn, she will
not be in a position to raise a mortgage on it, so that loan could only be repaid if she sold her
property and he cannot force her to sell. The boat has been for sale previously and did not sell.
There is also the argument about the saleability of the SL shares. I must also note that whilst
www.seyfihoca.com
135
they remain in his hands pending sale, none of these assets produces any income. So, for all
these reasons, it is said that the assets are not such as enable an immediate fair division by the
ordering of a payment of a lump sum so as to even out the shares of assets held by each.
Secondly, it is argued that to simply divide the assets is to take no notice of the fact that the
wife unnecessarily diminished the capital available after the separation. It is said that she
removed and spent moneys from the accounts as if it was ‘going out of fashion’. It is said that
she liquidated some of her own capital unnecessarily.
Finally, under this head, it is said that it was quite unnecessary for her to borrow £17,000
from her sister to purchase a car, whilst at the same time purchasing for her daughter, on hirepurchase, a Polo motorcar.
It is said that I should take all that into account in deciding whether this is a White case; and,
if it is, that those sums should be reflected in the final calculations.
Thirdly, it is argued that the husband’s pension fund takes this case outside White: it is the
largest single part of the assets. It is said that this fund is now set up and he cannot vary it, he
cannot draw capital from it, and his flexibility in relation to how his income is delivered is
therefore significantly affected.
Finally, it is said that he has different and additional needs: he has responsibilities for Miss D,
B and E; he wants B to have a private education, as did his other two children; he wants also
to provide properly for his new family. It is said that it is accepted that they had come to him
knowing that he has a former wife to be provided for; but, nevertheless, they must be
considered and they do stretch the limited resources.
Mr Hayward-Smith invites me to perform an exercise whereby, leaving aside the house, I
should enhance the wife’s present asset position by an additional £143,475, made up of the
various sums she is said to have expended unnecessarily, so as to provide a current total of
£388,788. To that should be added the joint portfolio, a further £162,099. He says that if she
were then to be given a lump sum of £100,000 she would have a total free capital of
£650,887, plus a house worth £504,400, giving her assets worth £1,155,287. He says that with
the £650,887 she would, on a Duxbury basis, have some £45,000 pa net, not far short of what
she is looking for in her proposed budget of £48,611.
In relation to those arguments, Mr Pointer says that first, ‘assets are assets and I should not
over-concern myself with how and when they may be realised.
[2001] 2 FLR 257
In relation to alleged diminution of capital, he says that the figures are wrong and the
arguments are flawed: the proper date to take as the start point must be the date when the
husband left, which is 27 April 1998, not the start date on the page of the bank statement
which was 6 April 1998. He says that if we go from the date of separation, the amount drawn
or paid from the bank is not £38,700 but £22,500, and that she has accounted for £18,000, and
more, of that in her schedule at A/57.
As to the withdrawals of capital, amounting to £45,000 in 3 years, that has only enabled her to
bring her income up to about £45,000 pa that she needs to spend.
www.seyfihoca.com
136
In relation to the third point regarding the pension fund, he says that this is not different in
principle from the wife’s position: she will be entirely dependent upon investment income,
she will have to invest in a Duxbury fund and will have no more flexibility than him.
Fourthly, it is said that the position of his paramour and her children, only one of whom is his,
cannot reduce the rights of his wife as there are ample resources to provide for her and to
enable the husband to look after his new family.
I do not consider that I can differentiate between assets in relation to their value. The courts
have always approached these matters on a net value basis, ie, taking account of sale costs and
tax liabilities. White affirmed that approach.
I dealt earlier with the difficulty of applying forced sale values to particular items and said
that there would have to be very good reasons for doing so. I am not satisfied that there are
any such reasons here; and, indeed, no forced sale values have been provided. It seems to me
that the place that this might have some impact is on the time allowed for paying any lump
sum ordered.
I must therefore, in this case, take the values that have been agreed, or that I have found, and
work on the basis that that value is realisable within a reasonable time.
As regards unnecessary diminution of the value of the assets by the wife, I have no doubt that
she was devastated by the revelations that her husband had been involved in a ongoing
deception of her for many years and that he had had a child by a woman some 6 years
previously, which woman he was still seeing and which child he was supporting. I have no
doubt that she did spend as she saw fit. She changed the locks and the alarm system. She
bought a new car, which is included in the schedules, both the car and the loan; and she
continued in the lifestyle to which she had grown accustomed. I am equally sure that she felt
drawn closer to her daughter as a result of her husband’s revelations.
I am not prepared to find that she was unnecessarily extravagant or that she diminished the
resources available to any significant extent. I accept Mr Pointer’s arguments about the dates
and the amounts of her bank withdrawals. I also accept his argument that the pension fund is,
in reality, no different from the Duxbury fund in which she will have to invest.
As to the position of the second family, this is really a question of seeing what happens when
the arithmetic is done on different bases.
The proposed equation that Mr Hayward-Smith puts forward seems to me to fail for several
reasons: first, I do not accept his argument about the diminution of capital by which I should
now notionally enhance the wife’s
[2001] 2 FLR 258
fund. Even if I did do that, then, as such capital as she has liquidated has been spent, she will
not be able to buy a sufficient fund to provide for her reasonable needs.
The other side of that equation would be that if the husband only had to raise £100,000 for a
lump sum payment, that would leave him with over £500,000 free funds, ie bank accounts and
investments, which could purchase an income of some £35,000 pa. That would need to be
www.seyfihoca.com
137
added to his present income of £80,000 from pension fund and rental properties, and he would
still have his boat, cars and money, in due course, to come from H’s estate and Miss D. Even
given his needs, that does not seem to me to be a fair outcome.
I must test it by comparing it to the outcome of an equal division. What happens if I go down
the equality route? It requires the payment of a lump some to the wife of £400,000, in addition
to the transfer of the house and the joint investments. That would put £835,000 into the wife’s
hands. A Duxbury calculation says that that gives her over £50,000 pa; she only asks for
£48,000. As Mr Pointer says, it is not for me, in some patriarchal way, to say what she should
and should not do with her money, to allow some of her items and to disallow others.
However, I am permitted to observe that she would, on her own figures, be living in relative
luxury. By that I mean that she will remain in the family house; she will be able to enjoy her
garden, spending, as she plans to do, over £7500 per year on it; also spending £7000 a year on
clothes and £7000 a year on miscellaneous items not included in her very detailed budget.
If the husband was ordered to pay that sum, and chose to do so by liquidating his investments,
then they would produce just under £400,000. He would then have just under £100,000 in
bank accounts, plus his expectation from H’s estate, the money due from Miss D, his boat,
car, SL shares and his properties. His income from his pension fund and rental properties
would be the £80,000 gross, which is £51,000 net, plus whatever came from the balance of his
assets as he chose to invest them.
He still has several choices open to him: his current home is subject to a mortgage which
expires in September 2004; he could pay that off as it was only a 5-year mortgage and is
heavily weighted to capital repayments. He has the next 15 years to decide when to convert
his pension fund into an annuity. He can make choices about his leisure activity. But is he not
as entitled, if at all possible, to maintain those in the same way that the wife will have her
garden?
Undoubtedly things will be very much tighter for him than for the wife. His budget contains
no provision for education for B; and, apart from his expenditure on his boat, is, on any view,
modest, for example, £150 a week housekeeping for a family of four.
So it seems to me that equality may not be the answer here. Is there another course that would
be fair? It seems to me that I must consider the various subsections of s 25 and aim to provide
as fair a distribution as possible. I am driven to conclude, against the background of this case,
that I should therefore aim to provide them each with a comfortable house and with sufficient
money to discharge their needs, obligations and responsibilities. If, in addition to the transfer
of the house and the joint assets, I order the husband to pay a lump sum of £300,000, then that
will provide the wife with a fund of £735,000 which should produce a net
[2001] 2 FLR 259
income that will meet her stated needs. She will have an unencumbered house worth over
£500,000. Her life will, so far as I can judge, continue in the style that she says she enjoyed
during the marriage.
The husband will have, immediately, somewhere in the region of £180,000 in liquid funds to
top up the income that comes from his pension fund and his rental properties. He will also
have free capital of £275,000 in property, including his home, and about £240,000 to come
www.seyfihoca.com
138
from the boat, the SL shares, H’s estate and Miss D’s loan. His net income will be just about
sufficient to meet his proposed budget.
I must then check my tentative conclusions against the yardstick of ‘equality’. My proposals
would share the joint pot as to £1,469,640 for the husband, and £1,262,416 for the wife. The
question for me is whether ‘if the shoe pinches (to use a phrase of Ward LJ in Piglowska v
Piglowski [1999] 2 FLR 763, [1999] 1 WLR 1360), and pinches more on one party than the
other, as a result of an equal distribution; is that sufficient demonstration that there is good
reason, in that case, to depart from equality? I believe that it is. I have looked at my proposals
to see if there is any discrimination in them. I do not believe that there is. It seems to me that
if I were to divide the assets equally, then I would be discriminating against the husband as a
result of his responsibility for his new family. I am satisfied that to deal with the matter in the
way I have proposed will result in as fair an outcome to both parties as is possible in this case.
I shall therefore order that the husband shall transfer to the wife his half-share in the
matrimonial home, Hunters, his half-share in the joint investments, and shall pay a lump sum
of £300,000 to the wife.
Order accordingly.
Solicitors:
Stokoe Partnership for the petitioner
Ellison & Co for the respondent
PATRICIA HARGROVE
Barrister
THE LEGAL CHARACTER OF INTERNATIONAL LAW
It has often been said that international law ought to be classified as a branch of ethics rather
than of law. The question is partly one of words, because its solution will clearly depend on
the definition of law which we choose to adopt; in any case it does not affect the value of the
subject one way or the other, though those who deny the legal character of international law
often speak as though 'ethical' were a depreciatory epithet. But in fact it is both practically
inconvenient and also contrary to the best juristic thought to deny its legal character. It is
inconvenient because if international law is nothing but international morality, it is certainly
not the whole of international morality, and it is difficult to see how we are to distinguish it
from those other admittedly moral standards which we apply in forming our judgements on
the conduct of states. Ordinary usage certainly uses two tests in judging the 'rightness' of a
state's act, a moral test and one which is somehow felt to be independent of morality. Every
state habitually commits acts of selfishness which are often gravely injurious to other states,
and yet are not contrary to international law; but we do not on that account necessarily judge
them to have been 'right'. It is confusing and pedantic to say that both these tests are moral.
Moreover, it is the pedantry of the theorist and not of the practical man; for questions of
international law are invariably treated as legal questions by the foreign offices which conduct
our international business, and in the courts, national or international, before which they are
brought; legal forms and methods are used in diplomatic controversies and in judicial and
arbitral proceedings, and authorities and precedents are cited in argument as a matter of
course. It is significant too that when a breach of international law is alleged by one party to a
controversy, the act impugned is practically never defended by claiming the right of private
www.seyfihoca.com
139
judgement, which would be the natural defence if the issue concerned the morality of the act,
but always by attempting to prove that no rule has been violated. This was true of the
defences put forward even for such palpable breaches of international law as the invasion of
Belgium in 194, or the bombardment of Corfu in 1923.
But if international law is not the same thing as international morality, and if in some
important respects at least it certainly resembles law, why should we hesitate to accept its
definitely legal character? The objection comes in the main from the followers of writers such
as Hobbes and Austin, who regard nothing as law which is not the will of a political superior.
But this is a misleading and inadequate analysis even of the law of a modern state; it cannot,
for instance, unless we distort the facts so as to fit them into the definition, account for the
existence of the English Common Law. In any case, even if such an analysis gave an adequate
explanation of law in the modern state, it would require us to assume that that law is the only
true law, and not merely law at a particular stage of growth or one species of a wider genus.
Such an assumption is historically unsound. Most of the characteristics which differentiate
international law from the law of the state and are often thought to throw doubt on its legal
character, such, for instance, as its basis in custom, the fact that the submission of parties to
the jurisdiction of courts is voluntary, the absence of regular processes either for creating or
enforcing it, are familiar features of early legal systems; and it is only in quite modern times,
when we have come to regard it as natural that the state should be constantly making new
laws and enforcing existing ones, that to identify law with the will of the state has become
even a plausible theory. If, as Sir Frederick Pollock writes, and as probably most competent
jurists would today agree, the only essential conditions for the existence of law are the
existence of a political community, and the recognition by its members of settled rules
binding upon them in that capacity, international law seems on the whole to satisfy these
conditions.
(From The Law of Nations by J. L. Brierly)
THE LAW OF NEGLIGENCE
Have you ever wondered how, with the infinite complexity of life, the law can provide for
every situation by a rule laid down in advance? Take the driving of a motor car. How can you
have a law that will govern every turn of the steering wheel? The answer found in the law
books is that cars must be driven, and almost everything else must be done, with reasonable
care and without negligence. This at least seems to state a rule. But is it a real rule, or just an
appearance of one? Unless one can give a content to the notion of reasonableness, and to its
opposite, negligence, the problem will be raised afresh: how can the law provide for every
eventuality?
In practice the negligence rule means that the judge or jury decides whether the defendant is
to be blamed or not, and this involves fixing upon something that he could and should have
done (or omitted to do) in order to avoid the mischief.
During the nineteenth century the judges attempted to clarify the law by inventing the notion
of the reasonable man. Instead of asking, as before, whether the defendant was guilty of
imprudence, the judges required the jury to consider whether the defendant had behaved like a
reasonable man, or a reasonably careful man, or a reasonably prudent man.
www.seyfihoca.com
140
Having invented the reasonable man, the judges had to make an effort to describe him.
According to some, he was the ordinary reasonable man. This was not the same as saying that
he was the ordinary man, but came rather near it. Indeed, Lord Bowen, with his gift for a
phrase, equated the reasonable man with 'the man on the Clapham omnibus', and this had led
to the widespread supposition that the standard of care required by the common law is that of
the average man. The supposition is certainly untrue.
'The ordinary reasonable man' of the judges' imagining is a meticulously careful person, so
careful that very few gentlemen come up to his standard. Every form of average is a measure
of central tendency, but the standard required by law is not a central human tendency of any
kind.
Take one obvious point. A defendant in an action for negligence would not be allowed to put
the passengers of a Clapham omnibus into the witness-box to say that they would have done
the same as he did. The evidence would not be listened to. One reason for this is that if
ordinary standards were conclusive, the courts could not use their influence to improve these
standards.
Again, ordinary people, even though normally they are circumspect in their behaviour, lapse
into carelessness now and then. As Elbert Hubbard expressed it, with humorous exaggeration,
'Every man is a damn fool for at least five minutes every day. Wisdom consists in not
exceeding the limit'. But the idealized stereotype of the law, so far from being given five
minutes' indulgence a day, is never allowed an off-moment. The defendant may be a good
driver with twenty-five years' clean record, yet he will be held liable in damages for
negligence if he injures someone during a momentary lapse.
That the judges set the standard by the nearly perfect man, rather than the average man, is
lavishly illustrated in the law reports. The master of a small motor vessel, when off
Greenhithe in the Thames, fainted at the wheel, as a result of eating bad tinned salmon, and a
collision followed. Up to the moment of fainting the master had felt quite well, and he was
obviously not negligent in fainting. A pure accident, you might say. But the judge held that
the master was negligent because he had failed to foresee that he might lose consciousness,
and had omitted to provide against this by having some other person on deck who might be
able to get to the bridge in time to prevent an accident if fainting occurred. As a landlubber I
am in no position to comment upon this decision, but obviously, it could not be applied to
land vehicles without absurdity. No one would suggest that the Clapham or any other omnibus
should carry a reserve driver, ready to seize the wheel in case the other driver faints.
Where, then, is our elusive standard to be found? If the reasonable man is not to be discovered
on the Clapham omnibus, can he be identified with the judge or juryman who has to decide
the issue? Technically, at least, the answer is again in the negative. A judge must not tell the
members of the jury to determine what they would have done, because that is not the
question. The individual jurors might not have acted as prudently as they now think, on
reflection, they ought to have acted in the situation. jurors are expected to follow Hume's
precept that, in considering the moral character of an act, we should adopt the role of
impartial spectator, seeking reactions as to what we would approve or disapprove, not as to
what we ourselves would have felt or done.
www.seyfihoca.com
141
Perhaps this is the key to the puzzle. The reasonable man is a phantom reflecting a certain
ideal on the part of the tribunal, whether judge or jury as the case may be. He is a
personification of the court's social judgment.
If this is the right conclusion, I think one is driven to admit that the device of trying to decide
cases in terms of the reasonable man is not such a good idea after all. Why not address oneself
directly to the problem of negligence? Values are important, but there is no point in
personifying values, or attributing them to a fictitious person. In speaking of what a
reasonable man does or does not do, we appear to be stating a fact, whereas in truth the
question is how people ought to behave. It is merely misleading to use an expression that
seems to indicate the behaviour of real people, when we are evaluating behaviour by reference
to an ideal standard.
(Glanville Williams from an article in The Listener, February 2nd, 1961)
THE DEATH PENALTY
I want to organize under five simple verbs my own reasons for thinking that the death penalty
is a bad thing. If we catch a man who has committed a murder, try him and convict him, we
have to do something more with him than punish him, because, although he must be punished,
there are several other things that ought to happen to him. I think that the whole theory of
what ought to be done to a convicted murderer can be summed up in the five verbs: prevent,
reform, research, deter and avenge. Let me take these five things in turn and see how the
death penalty now looks as a means of achieving them.
The first is 'prevent'. By this I mean preventing the same man from doing it again, to check
him in his career-though, of course, nobody makes a career of being a murderer, except the
insane, who are not at issue in the question of the death penalty. I believe that I am right in
saying that in the course of a century there is only one doubtful case of a convicted murderer,
after his release at the end of a normal life sentence, committing another murder. I think that
that means, statistically, that the released murderer is no more likely to murder again than
anybody else is. The question of long sentences comes in here. If the sane convicted murderer
is not to be hanged, should he be imprisoned, and should the length of his service be
determined in a way not the usual one for the actual sentence served? I think this question can
be answered only by looking at the statistics of how likely a man is to do it again. In other
words, how likely a prison sentence for a given number of years, 15, 20 or 30 years, is to
prevent him from doing it again. There is a wealth of statistics available to us on that. I do not
think they suggest that the convicted murderer who is not hanged should have his prison
sentence dealt with in any way differently from that in which prison sentences are usually
dealt with.
To turn to the second verb on my list, 'reform'. That is rather a nineteenth century word, and
perhaps we should now say 'rehabilitate', stressing more the helping of a man with his social
functions rather than adjusting his internal character; but that is a minor point. It is clear that,
whatever we may think about what is able to be achieved in our prison system by treatment in
the reformatory and rehabilitatory way - and it is open to criticism for lack of funds and so onit is obvious that less can be achieved if you hang a man. One man who is utterly
unreformable is a corpse; and hanging is out of the question, because you cannot achieve any
form of reform or rehabilitation by it.
www.seyfihoca.com
142
The next word is 'research'. This is not part of the traditional idea of what to do with a
convicted murderer. It is rather a new notion that it may be an appropriate purpose in
detaining a criminal and inflicting punishment and other things upon him that research should
be conducted into the criminal personality and the causes of crime. At the moment we hang
only the sanest criminals. We can get all the research we want into the motives, characters and
personality structures of those with diminished responsibility, the insane and those under an
age to be hanged. But the one we cannot research into is the man who is sane and who
commits capital murder in cold blood on purpose. It might be that if we were to keep this man
alive and turn psychiatrists and other qualified persons on to talking to him for twenty years
during his prison sentence we should find things that would enable us to take measures which
would reduce the murder rate and save the lives of the victims. But in hanging these men we
cut ourselves off from this possible source of knowledge of help to the victims of murder.
The fourth word, 'deter', is the crux of the whole thing. Abolitionists, as we all know, have
held for many years that evidence from abroad has for long been conclusive that the capital
penalty is not a uniquely effective deterrent against murder. Retentionists of the death penalty
have been saying for years that we are not like those abroad; we are a different country
economically; our national temperament is different; and there is this and that about us which
is not so about those in Italy, Norway or certain States of the United States, New Zealand,
India, or wherever it may be. Now we have this remarkable pamphlet which in effect closes
that gap in the abolitionists' argument. It shows within mortal certitude that we are exactly
like those abroad, and that in this country the death penalty is not a uniquely effective
deterrent against murder.
The last on the list of my five verbs is 'avenge'. Here the death penalty is uniquely effective. If
a man has taken life, the most effective, obvious and satisfying form of vengeance is to take
his life. I have no argument against that. I think it is true that if one accepts vengeance as a
purpose proper for the State in its handling of convicted criminals, then the death penalty
should stay for convicted murderers. For myself-and it is only a personal matter-I utterly
reject the idea that vengeance is a proper motive for the State in dealing with convicted
criminals; and I hope that, from the date of the publication of this pamphlet onwards, those
who wish to retain the death penalty will admit that its only merit is precisely that of
vengeance.
(Lord Kennet from a Speech in the House of Lords, November 9th, 1961)
MATHEMATICS
On Different Degrees of Smallness
We shall find that in our processes of calculation we have to deal with small quantities of
various degrees of smallness. We shall have also to learn under what circumstances we may
consider small quantities to be so minute that we may omit them from consideration.
Everything depends upon relative minuteness. Before we fix any rules let us think of some
familiar cases. There are 60 minutes in the hour, 24 hours in the day, 7 days in the week.
There are therefore 1,440 minutes in the day and 10,080 minutes in the week.
Obviously 1 minute is a very small quantity of time compared with a whole week. Indeed, our
forefathers considered it small as compared with an hour, and called it 'one minute', meaning
a minute fraction - namely one-sixtieth-of an hour. When they came to require still smaller
www.seyfihoca.com
143
subdivisions of time, they divided each minute into 60 still smaller parts, which in Queen
Elizabeth's days, they called 'second minutes' (i.e. small quantities of the second order of
minuteness). Nowadays we call these small quantities of the second order of smallness
'seconds'. But few people know why they are so called. Now if one minute is so small as
compared with a whole day, how much smaller by comparison is one second!
Again, think of a farthing as compared with a sovereign; it is worth only a little more than
1/1,000 part. A farthing more or less is of precious little importance compared with a
sovereign; it may certainly be regarded as a small quantity. But compare a farthing with
£1,000: relatively to this greater sum, the farthing is of no more importance than 1/1,000 of a
farthing would be to a sovereign. Even a golden sovereign is relatively a negligible quantity in
the wealth of a millionaire.
Now if we fix upon any numerical fraction as constituting the proportion which for any
purpose we call relatively small, we can easily state other fractions of a higher degree of
smallness.
Thus if, for the purpose of time, x/6o be called a small fraction, then 1/60 of 1/60 (being a
small fraction of a small fraction) may be regarded as a small quantity of the second order of
smallness. Or, if for any purpose we were to take 1 per cent (i.e. 1/100) as a small fraction,
then 1 per cent of 1 per cent (i.e. 1/10,000) would be a small fraction of the second order of
smallness; and 1/1,000,000 would be a small fraction of the third order of smallness, being 1
per cent of 1 per cent of 1 per cent.
Lastly, suppose that for some very precise purpose we should regard 1/1,000,000 as 'small'.
Thus, if a first-rate chronometer is not to lose or gain more than half a minute in a year, it
must keep time with an accuracy of 1 part in 1,051,200. Now if, for such a purpose, we regard
1/1,000,000 (or one millionth) as a small quantity, then 1/1,000,000 of 1/1,000,000, that is
1/1,000,000,000,000 (or one billionth) will be a small quantity of the second order of
smallness, and may be utterly disregarded, by comparison.
Then we see that the smaller a small quantity itself is, the more negligible does the
corresponding small quantity of the second order become. Hence we know that in all cases we
are justified in neglecting the small quantities of the second - or third (or higher) - orders, if
only we take the small quantity of the first order small enough in itself. But it must be
remembered that small quantities if they occur in our expressions as factors multiplied by
some other factor, may become important if the other factor is itself large. Even a farthing
becomes important if only it is multiplied by a few hundred.
Now in the calculus we write dx for a little bit of x. These things such as dx and du, and dy,
are called 'differentials', the differential of x, or of u, or of y, as the case may be. (You read
them as dee-eks, or dee-you, or dee-wy). If dx be a small bit of x, and relatively small, it does
not follow that such quantities as x.dx or a2.dx or ax.dx are negligible. But dx times dx would
be negligible, being a small quantity of the second order.
A very simple example will serve as illustration. Let us think of x as a quantity that can grow
by a small amount so as to become x + dx, where dx is the small increment added by growth.
The square of this is x2 + 2x.dx + (dx)2. The second term is not negligible because it is a firstorder quantity; while the third term is of the second order of smallness, being a bit of a bit of
x. Thus if we took dx to mean numerically, say 1/60 of x, then the second term would be 2/60
www.seyfihoca.com
144
of x2, whereas the third term would be 1/3,600 of x2. This last term is clearly less important
than the second. But if we go further and take dx to mean only 1/1000 of x, then the second
term will be 2/1,000 of x2, while the third term will be only 1/1,000,000 Of x2.
Geometrically this may be depicted as follows: draw a square the side of which we will take
to represent x. Now suppose the square to grow by having a bit dx added to its size each way.
The enlarged square is made up of the original square x2, the two rectangles at the top and on
the right, each of which is of area x. dx (or together 2x.dx), and the little square at the top
right-hand corner which is (dx)2. We have taken dx as quite a big fraction of x - about 1/5. But
suppose we had taken it as only 1/1000 - about the thickness of an inked line drawn with a
fine pen. Then the little corner square will have an area of only 1/10,000 of x2, and be
practically invisible. Clearly (dx)2 is negligible if only we consider the increment dx to be
itself small enough.
(From Calculus Made Easy by Silvanus P. Thompson.)
Chance or Probability
The aim of science is to describe the world in orderly language, in such a way that we can, if
possible, foresee the results of those alternative courses of action between which we are
always choosing. The kind of order which our description has is entirely one of convenience.
Our purpose is always to predict. Of course, it is most convenient if we can find an order by
cause and effect; it makes our choice simple; but it is not essential.
There is of course nothing sacred about the causal form of natural laws. We are accustomed to
this form, until it has become our standard of what every natural law ought to look like. If you
halve the space which a gas fills, and keep other things constant, then you will double the
pressure, we say. If you do such and such, the result will be so and so; and it will always be so
and so. And we feel by long habit that it is this 'always' which turns the prediction into a law.
But of course there is no reason why laws should have this always, all-or-nothing form. If you
self-cross the offspring of a pure white and a pure pink sweet pea, said Mendel, then on an
average one-quarter of these grandchildren will be white, and three-quarters will be pink. This
is as good a law as any other; it says what will happen, in good quantitative terms, and what it
says turns out to be true. It is not any less respectable for not making that parade of every time
certainty which the law of gases makes.
It is important to seize this point. If I say that after a fine week, it always rains on Sunday,
then this is recognized and respected as law. But if I say that after a fine week, it rains on
Sunday more often than not, then this somehow is felt to be an unsatisfactory statement; and it
is taken for granted that I have not really got down to some underlying law which would
chime with our habit of wanting science to say decisively either 'always' or 'never'. Somehow
it seems to lack the force of law.
Yet this is a mere prejudice. It is nice to have laws which say, 'This configuration of facts will
always be followed by event A, ten times out of ten.' But neither taste nor convenience really
make this a more essential form of law than one which says, 'This configuration of facts will
be followed by event A seven times out of ten, and by event B three times out of ten.' In form
the first is a causal law and the second a statistical law. But in content and in application,
there is no reason to prefer one to the other.
www.seyfihoca.com
145
There is, however, a limitation within every law which does not contain the word 'always'.
Bluntly, when I say that a configuration of facts will be followed sometimes by event A and at
other times by B, I cannot be certain whether at the next trial A or B will turn up. I may know
that A is to turn up seven times and B three times out of ten; but that brings me no nearer at
all to knowing which is to turn up on the one occasion I have my eye on next time. Mendel's
law is all very fine when you grow sweet peas by the acre; but it does not tell you, and cannot,
whether the single second generation seed in your window box will flower white or pink.
But this limitation carries with it a less obvious one. If we are not sure whether A or B will
turn up next time, then neither can we be sure which will turn up the time after, or the time
after that. We know that A is to turn up seven times and B three; but this can never mean that
every set of ten trials will give us exactly seven As and three Bs.
Then what do I mean by saying that we expect A to turn up seven times to every three times
which B turns up? I mean that among all the sets of ten trials which we can choose from an
extended series, picking as we like, the greatest number will contain seven As and three Bs.
This is the same thing as saying that if we have enough trials, the proportion of As to Bs will
tend to the ratio of seven to three. But of course, no run of trials, however extended, is
necessarily long enough. In no run of trials can we be sure of reaching precisely the balance
of seven to three.
Then how do I know that the law is in fact seven As and three Bs? What do I mean by saying
that the ratio tends to this in a long trial, when I never know if the trial is long enough? And
more, when I know that at the very moment when we have reached precisely this ratio, the
next single trial must upset it because it must add either a whole A or a whole B, and cannot
add seven-tenths of one and three-tenths of the other. I mean this. After ten trials, we may
have eight As and only two Bs; it is not at all improbable. But it is very improbable that, after
a hundred trials, we shall have as many as eighty As. It is excessively improbable that after a
thousand trials we shall have as many as eight hundred As; indeed it is highly improbable that
at this stage the ratio of As and Bs departs from seven to three by as much as five per cent.
And if after a hundred thousand trials we should get a ratio which differs from our law by as
much as one per cent, then we should have to face the fact that the law itself is almost
certainly in error.
Let me quote a practical example. The great naturalist Bufhon was a man of wide interests.
His interest in the laws of chance prompted him to ask an interesting question. If a needle is
thrown at random on a sheet of paper ruled with lines whose distance apart is exactly equal to
the length of the needle, how often can it be expected to fall on a line and how often into a
blank space? The answer is rather odd: it should fall on a line a little less than two times out
of three - precisely, it should fall on a line two times out of pi where pi is the familiar ratio of
the circumference of a circle to its diameter, which has the value 3.14159265.... How near can
we get to this answer in actual trials? This depends of course on the care with which we rule
the lines and do the throwing; but, after that, it depends only on our patience. In 1901 an
Italian mathematician, having taken due care, demonstrated his patience by making well over
3,000 throws. The value he got for pi was right to the sixth place of decimals, which is an
error of only a hundred thousandth part of one per cent.
This is the method to which modern science is moving. It uses no principle but that of
forecasting with as much assurance as possible, but with no more than is possible. That is, it
www.seyfihoca.com
146
idealizes the future from the outset, not as completely determined, but as determined within a
defined area of uncertainty.
(From The Common Sense of Science by J. Bronowski.)
PHILOSOPHY
Definition and Some of its Difficulties
If our thought is to be clear and we are to succeed in communicating it to other people, we
must have some method of fixing the meanings of the words we use. When we use a word
whose meaning is not certain, we may well be asked to define it. There is a useful traditional
device for doing this by indicating the class to which whatever is indicated by the term
belongs, and also the particular property which distinguishes it from all other members of the
same class. Thus we may define a whale as a 'marine animal that spouts'. 'Marine animal' in
this definition indicates the general class to which the whale belongs, and 'spouts' indicates
the particular property that distinguishes whales from other such marine animals as fishes,
seals, jelly-fish, lobsters, etc. In the same way we can define an even number as a finite
integer divisible by two, or a democracy as a system of government in which the people
themselves rule.
There are other ways, of course, of indicating the meanings of words. We may, for example,
find it hard to make a suitable definition of the word 'animal', so we say that an animal is such
a thing as a rabbit, dog, fish, etc. Similarly we may say that religion is such a system as
Christianity, Islam, Judaism, Christian Science, etc. This way of indicating the meaning of a
term by enumerating examples of what it includes is obviously of limited usefulness. If we
indicated our use of the word 'animal' as above, our hearers might, for example, be doubtful
whether a sea-anemone or a slug was to be included in the class of animals. It is, however, a
useful way of supplementing a definition if the definition itself is definite without being easily
understandable. If, for example, we explain what we mean by religion by saying: 'A religion
is a system of beliefs and practices connected with a spiritual world, such as Christianity,
Islam, Judaism, Christian Science, and so on', we may succeed in making our meaning more
clear than it would be if we had been given the definition alone.
Failure of an attempt at definition to serve its purpose may result from giving as
distinguishing mark one which either does not belong to all the things the definition is
intended to include, or does belong to some members of the same general class which the
definition is intended to exclude. Thinking, for example, of the most obvious difference
between a rabbit and a cabbage, we might be tempted to define an animal as a living organism
which is able to move about. This would combine both faults mentioned above, since some
animals (e.g. some shell-fish such as the oyster) are not able to move about for the whole or
part of their lives, while some vegetables (such as the fresh-water alga Volvox) do swim
about. Of course, anyone who used the above definition might claim to be defining 'animal' in
a new and original way to include Volvox and exclude oysters, but he would have failed to
produce a definition which defined the ordinary use of the word 'animal'.
More commonly an attempt at definition fails by not indicating correctly the general class to
which the thing defined belongs. One meets, for example, in psychological writings such
definitions as: 'Intelligence is a state of mind characterized by the ability to learn and solve
problems.' The second part of the definition is all right, but the word 'intelligence' is not used
www.seyfihoca.com
147
for a state of mind; and the person who defines 'intelligence' like this does not in his actual use
of the word make it stand for a state of mind. Such conditions as despair, concentration,
alertness, and hope can be called 'states of mind'. 'Intelligence' is used for a quality of mind,
not for a state. If the word 'quality' replaced the word 'state' in the above definition it would
indicate very well the current use of the word 'intelligence'.
(From Straight and Crooked Thinking, by R. H. Thouless.)
The Subject Matter of Philosophy
It seems clear that subjects or fields of study are determined by the kind of questions to which
they have been invented to provide the answer. The questions themselves are intelligible if,
and only if, we know where to look for the answers.
If you ask someone an ordinary question, say, 'Where is my coat?', 'Why was Mr Kennedy
elected President of the United States?', 'What is the Soviet system of criminal law?' he would
normally know how to set about finding an answer. We may not know the answers ourselves,
but we know that in the case of the question about the coat, the proper procedure is to look on
the chair, in the cupboard, etc. In the case of Mr Kennedy's election or the Soviet system of
law we consult writings of specialists for the kind of empirical evidence which leads to the
relevant conclusions, and renders them, if not certain, at any rate probable.
In other words, we know where to look for the answer: we know what makes some answers
plausible and others not. What makes this type of question intelligible in the first place is that
we think that the answer can be discovered by empirical means, that is, by orderly observation
or experiment or methods compounded of these, namely those of common sense or the natural
sciences. There is another class of questions where we are clear about the proper route by
which the answers are to be sought, namely the formal disciplines: mathematics, for example,
or logic, or grammar, or chess, where there are certain fixed axioms, certain accepted rules of
deduction, and the answer to problems is to be found by applying these rules in the manner
prescribed as correct.
The hallmark of these provinces of human thought is that once the question is put we know
which way to proceed to try to obtain the answer. The history of systematic human thought is
largely a sustained effort to formulate all the questions that occur to mankind in such a way
that they will fall into one or other of these two great baskets: the empirical, that is, questions
whose answers depend, in the end, on the data of observation; and the formal, that is,
questions whose answers depend on pure calculation, untrammelled by factual knowledge.
But there are certain questions that do not easily fit into this classification. 'What is an okapi?'
is answered easily enough by an act of empirical observation. Similarly 'What is the cube root
of 729?' is settled by a piece of calculation in accordance with accepted rules. But if I ask
'What is a number?', 'What is the purpose of human life on earth', 'Are you sure that all men
are brothers?' how do I set about looking for the answer?
There seems to be something queer about all these questions as wide apart as those about
number, or the brotherhood of man, or purposes of life; they differ from the questions in the
other baskets in that the question itself does not seem to contain a pointer to the way in which
the answer to it is to be found. The other, more ordinary, questions contain precisely such
pointers built-in techniques for finding the answers to them. The questions about number and
www.seyfihoca.com
148
so on reduce the questioner to perplexity, and annoy practical people precisely because they
do not seem to lead to clear answers or useful knowledge of any kind.
This shows that between the two original baskets, the empirical and the formal, there is an
intermediate basket, in which all those questions live which cannot easily be fitted into the
other two. These questions are of the most diverse nature; some appear to be questions of fact,
others of value; some are questions about words, others are about methods pursued by
scientists, artists, critics, common men in the ordinary affairs of life; still others are about the
relations between the various provinces of knowledge; some deal with the presuppositions of
thinking, some with the correct ends of moral or social or political action.
The only common characteristic which all these questions appear to have is that they cannot
be answered either by observation or calculation, either by inductive methods or deductive;
and as a crucial corollary of this, that those who ask them are faced with a perplexity from the
very beginning-they do not know where to look for the answer; there are no dictionaries,
encyclopaedias, compendia of knowledge, no experts, no orthodoxies, which can be referred
to with confidence as possessing unquestionable authority or knowledge in these matters.
Such questions tend to be called philosophical.
(From an article by Sir Isaiah Berlin in The Sunday Times, 14th November, 1962.)
What can we Communicate?
The obvious answer to the question how we know about the experiences of others is that they
are communicated to us, either through their natural manifestations in the form of gestures,
tears, laughter, play of feature and so forth, or by the use of language. A very good way to
find out what another person is thinking or feeling is to ask him. He may not answer, or if he
does answer he may not answer truly, but very often he will. The fact that the information
which people give about themselves can be deceptive does not entail that it is never to be
trusted. We do not depend on it alone; it may be, indeed, that the inferences which we draw
from people's non-verbal behaviour are more secure than those that we base upon what they
say about themselves, that actions speak more honestly than words. But were it not that we
can rely a great deal upon words, we should know very much less about each other than we
do.
At this point, however, a difficulty arises. If I am to acquire information in this way about
another person's experiences, I must understand what he says about them. And this would
seem to imply that I attach the same meaning to his words as he does. But how, it may be
asked, can I ever be sure that this is so? He tells me that he is in pain, but may it not be that
what he understands by pain is something quite different from anything that I should call by
that name? He tells me that something looks red to him, but how do I know that what he calls
'red' is not what I should call 'blue', or that it is not a colour unlike any that I have ever seen,
or that it does not differ from anything that I should even take to be a colour? All these things
would seem to be possible. Yet how are such questions ever to be decided?
In face of this difficulty, some philosophers have maintained that experiences as such are
uncommunicable. They have held that in so far as one uses words to refer to the content of
one's experiences, they can be intelligible only to oneself. No one else can understand them,
because no one else can get into one's mind to verify the statements which they express. What
can be communicated, on this view, is structures. I have no means of knowing that other
www.seyfihoca.com
149
people have sensations or feelings which are in any way like my own. I cannot tell even that
they mean the same by the words which they use to refer to physical objects, since the
perceptions which they take as establishing the existence of these objects may be utterly
different from any that I have ever had myself. If I could get into my neighbour's head to see
what it is that he refers to as a table, I might fail to recognize it altogether, just as I might fail
to recognize anything that he is disposed to call a colour or a pain. On the other hand,
however different the content of his experience may be from mine, I do know that its structure
is the same. The proof that it is the same is that his use of words corresponds with mine, in so
far as he applies them in a corresponding way. However different the table that he perceives
may be from the table that I perceive, he agrees with me in saying of certain things that they
are tables and of others that they are not. No matter what he actually sees when he refers to
colour, his classification of things according to their colour is the same as mine. Even if his
conception of pain is quite different from my own, his behaviour when he is in pain is such as
I consider to be appropriate. Thus the possible differences of content can, and indeed must be
disregarded. What we can establish is that experiences are similarly ordered. It is this
similarity of structure that provides us with our common world: and it is only descriptions of
this common world, that is, descriptions of structure, that we are able to communicate.
On this view, the language which different people seem to share consists, as it were, of flesh
and bones. The bones represent its public aspect; they serve alike for all. But each of us puts
flesh upon them in accordance with the character of his experience. Whether one person's way
of clothing the skeleton is or is not the same as another's is an unanswerable question. The
only thing that we can be satisfied about is the identity of the bones.
(From The Problem of Knowledge, By A. J. Ayer.)
Ethics
Ethics is traditionally a department of philosophy, and that is my reason for discussing it. I
hardly think myself that it ought to be included in the domain of philosophy, but to prove this
would take as long as to discuss the subject itself, and would be less interesting.
As a provisional definition, we may take ethics to consist of general principles which help to
determine rules of conduct. It is not the business of ethics to say how a person should act in
such and such specific circumstances; that is the province of casuistry. The word 'casuistry'
has acquired bad connotations, as a result of the Protestant and Jansenist attacks on the
Jesuits. But in its old and proper sense it represents a perfectly legitimate study. Take, say, the
question: In what circumstances is it right to tell a lie? Some people, unthinkingly, would say:
Never! But this answer cannot be seriously defended. Everybody admits that you should lie if
you meet a homicidal maniac pursuing a man with a view to murdering him, and he asks you
whether the man has passed your way. It is admitted that lying is a legitimate branch of the art
of warfare; also that priests may lie to guard the secrets of the confessional, and doctors to
protect the professional confidences of their patients. All such questions belong to casuistry in
the old sense, and it is evident that they are questions deserving to be asked and answered. But
they do not belong to ethics in the sense in which this study has been included in philosophy.
It is not the business of ethics to arrive at actual rules of conduct, such as: 'Thou shalt not
steal.' This is the province of morals. Ethics is expected to provide a basis from which such
rules can be deduced. The rules of morals differ according to the age, the race, and the creed
of the community concerned, to an extent that is hardly realized by those who have neither
www.seyfihoca.com
150
travelled nor studied anthropology. Even within a homogeneous community differences of
opinion arise. Should a man kill his wife's lover? The Church says no, the law says no, and
common sense says no; yet many people would say yes, and juries often refuse to condemn.
These doubtful cases arise when a moral rule is in process of changing. But ethics is
concerned with something more general than moral rules, and less subject to change. It is true
that, in a given community, an ethic which does not lead to the moral rules accepted by that
community is considered immoral. It does not, of course, follow that such an ethic is in fact
false since the moral rules of that community may be undesirable. Some tribes of headhunters hold that no man should marry until he can bring to the wedding the head of an enemy
slain by himself. Those who question this moral rule are held to be encouraging licence and
lowering the standard of manliness. Nevertheless, we should not demand of an ethic that it
should justify the moral rules of head-hunters.
Perhaps the best way to approach the subject of ethics is to ask what is meant when a person
says: 'You ought to do so-and-so' or 'I ought to do so-and-so.' Primarily, a sentence of this sort
has an emotional content; it means 'this is the act towards which I feel the emotion of
approval'. But we do not wish to leave the matter there; we want to find something more
objective and systematic and constant. The ethical teacher says: 'You ought to approve acts of
such-and-such kinds.' He generally gives reasons for this view, and we have to examine what
sorts of reasons are possible. We are here on very ancient ground. Socrates was concerned
mainly with ethics; Plato and Aristotle both discussed the subject at length; before their time,
Confucius and Buddha had each founded a religion consisting almost entirely of ethical
teaching, though in the case of Buddhism there was afterwards a growth of theological
doctrine. The views of the ancients on ethics are better worth studying than their views on
(say) physical science; the subject has not yet proved amenable to exact reasoning, and we
cannot boast that the moderns have as yet rendered their predecessors obsolete.
(From An Outline of Philosophy, by Bertrand Russell.)
Aristotle's Ethics
As to goodness of character in general, Aristotle says that we start by having a capacity for it,
but that it has to be developed by practice. How is it developed ? By doing virtuous acts. At
first sight this looks like a vicious circle. Aristotle tells us that we become virtuous by doing
virtuous acts, but how can we do virtuous acts unless we are already virtuous? Aristotle
answers that we begin by doing acts which are objectively virtuous, without having a reflex
knowledge of the acts and a deliberate choice of the acts as good, a choice resulting from an
habitual disposition. For instance, a child may be told by its parents not to lie. It obeys
without realizing perhaps the inherent goodness of telling the truth, and without having yet
formed a habit of telling the truth; but the acts of truth-telling gradually form the habit, and as
the process of education goes on, the child comes to realize that truth-telling is right in itself,
and to choose to tell the truth for its own sake, as being the right thing to do. It is then virtuous
in this respect. The accusation of the vicious circle is thus answered by the distinction
between the acts which create the good disposition and the acts which flow from the good
disposition once it has been created. Virtue itself is a disposition which has been developed
out of a capacity by the proper exercise of that capacity. (Further difficulties might arise, of
course, concerning the relation between the development of moral valuations and the
influence of social environment, suggestion of parents and teachers, etc., but with these
Aristotle does not deal.)
www.seyfihoca.com
151
How does virtue stand to vice? It is a common characteristic of all good actions that they have
a certain order and proportion, and virtue, in Aristotle's eyes, is a mean between two
extremes, the extremes being vices, one being a vice through excess, the other being a vice
through defect. Through excess or defect of what? Either in regard to a feeling or in regard to
an action. Thus, in regard to the feeling of confidence, the excess of this feeling constitutes
rashness - at least when the feeling issues in action, and it is with human actions that ethics
are concerned - while the defect is cowardice. The mean, then, will be a mean between
rashness on the one hand and cowardice on the other hand: this mean is courage and is the
virtue in respect to the feeling of confidence. Again, if we take the action of giving of money,
excess in regard to this action is prodigality - and this is a vice - while defect in regard to this
action is illiberality. The virtue, liberality, is the mean between the two vices, that of excess
and that of defect. Aristotle, therefore, describes or defines moral virtue as 'a disposition to
choose, consisting essentially in a mean relatively to us determined by a rule, i.e. the rule by
which a practically wise man would determine it.' Virtue, then, is a disposition, a disposition
to choose according to a rule, namely, the rule by which a truly virtuous man possessed of
moral insight would choose. Aristotle regarded the possession of practical wisdom, the ability
to see what is the right thing to do in the circumstances, as essential to the truly virtuous man,
and he attaches much more value to the moral judgments of the enlightened conscience than
to any a priori and merely theoretical conclusions. This may seem somewhat naive, but it
must be remembered that for Aristotle the prudent man will be the man who sees what is truly
good for a man in any set of circumstances: he is not required to enter upon any academic
preserve, but to see what truly befits human nature in those circumstances.
When Aristotle speaks of virtue as a mean, he is not thinking of a mean that has to be
calculated arithmetically: that is why he says in his definition 'relatively to us'. We cannot
determine what is excess, what mean and what defect by hard-and-fast, mathematical rules: so
much depends on the character of the feeling or action in question: in some cases it may be
preferable to err on the side of excess rather than on that of defect, while in other cases the
reverse may be true. Nor, of course, should the Aristotelian doctrine of the mean be taken as
equivalent to an exaltation of mediocrity in the moral life, for as far as excellence is
concerned virtue is an extreme: it is in respect of its essence and its definition that it is a
mean.
(From A History of Philosophy, by Frederick Copleston.)
The Road to Happiness
It is a commonplace among moralists that you cannot get happiness by pursuing it. This is
only true if you pursue it unwisely. Gamblers at Monte Carlo are pursuing money, and most
of them lose it instead, but there are other ways of pursuing money which often succeed. So it
is with happiness. If you pursue it by means of drink, you are forgetting the hang-over.
Epicurus pursued it by living only in congenial society and eating only dry bread,
supplemented by a little cheese on feast days. His method proved successful in his case, but
he was a valetudinarian, and most people would need something more vigorous. For most
people, the pursuit of happiness, unless supplemented in various ways, is too abstract and
theoretical to be adequate as a personal rule of life. But I think that whatever personal rule of
life you may choose it should not, except in rare and heroic cases, be incompatible with
happiness.
www.seyfihoca.com
152
There are a great many people who have all the material conditions of happiness, i.e. health
and a sufficient income, and who, nevertheless, are profoundly unhappy. In such cases it
would seem as if the fault must lie with a wrong theory as to how to live. In one sense, we
may say that any theory as to how to live is wrong. We imagine ourselves more different from
the animals than we are. Animals live on impulse, and are happy as long as external
conditions are favourable. If you have a cat it will enjoy life if it has food and warmth and
opportunities for an occasional night on the tiles. Your needs are more complex than those of
your cat, but they still have their basis in instinct. In civilized societies, especially in Englishspeaking societies, this is too apt to be forgotten. People propose to themselves some one
paramount objective, and restrain all impulses that do not minister to it. A businessman may
be so anxious to grow rich that to this end he sacrifices health and private affections. When at
last he has become rich, no pleasure remains to him except harrying other people by
exhortations to imitate his noble example. Many rich ladies, although nature has not endowed
them with any spontaneous pleasure in literature or art, decide to be thought cultured, and
spend boring hours learning the right thing to say about fashionable new books that are
written to give delight, not to afford opportunities for dusty snobbism.
If you look around you at the men and women whom you can call happy, you will see that
they all have certain things in common. The most important of these things is an activity
which at most times is enjoyable on its own account, and which, in addition, gradually builds
up something that you are glad to see coming into existence. Women who take an instinctive
pleasure in their children can get this kind of satisfaction out of bringing up a family. Artists
and authors and men of science get happiness in this way if their own work seems good to
them. But there are many humbler forms of the same kind of pleasure. Many men who spend
their working life in the City devote their week-ends to voluntary and unremunerated toil in
their gardens, and when the spring comes they experience all the joys of having created
beauty.
The whole subject of happiness has, in my opinion, been treated too solemnly. It had been
thought that man cannot be happy without a theory of life or a religion. Perhaps those who
have been rendered unhappy by a bad theory may need a better theory to help them to
recovery, just as you may need a tonic when you have been ill. But when things are normal a
man should be healthy without a tonic and happy without a theory. It is the simple things that
really matter. If a man delights in his wife and children, has success in work, and finds
pleasure in the alternation of day and night, spring and autumn, he will be happy whatever his
philosophy may be. If, on the other hand, he finds his wife hateful, his children's noise
unendurable, and the office a nightmare; if in the daytime he longs for night, and at night
sighs for the light of day, then what he needs is not a new philosophy but a new regimen - a
different diet, or more exercise, or what not.
Man is an animal, and his happiness depends on his physiology more than he likes to think.
This is a humble conclusion, but I cannot make myself disbelieve it. Unhappy businessmen, I
am convinced, would increase their happiness more by walking six miles every day than by
any conceivable change of philosophy. This, incidentally, was the opinion of Jefferson, who
on this ground deplored the horse. Language would have failed him if he could have foreseen
the motor-car.
(From an article in The Listener, by Bertrand Russell.)
Logic
www.seyfihoca.com
153
Discourse is connected thought, expressed in words. It moves this way and that, like the
shuttle in the loom (as Plato said) weaving the fabric of reasoned argument. In discourse with
others opinion is formed, knowledge is acquired, and truth attained. What is said by one
speaker, combined with what is said by another speaker, may yield a truth, not previously
known to either speaker. In that discourse with oneself which is called reflection or
meditation, a truth learned today links up with a truth learned yesterday, and the two truths
may point the way to some advance or even discovery of tomorrow. From what others have
said or from what we ourselves have thought, conclusions and inferences are drawn and they
are the special concern of Logic. It is all too easy to draw wrong conclusions and false
inferences; and discourse without the discipline of Logic is a fruitful source of false opinion,
ignorance and error.
Logic trains the mind to draw the right conclusion, and to avoid the wrong, to make the true
inference and not the false. It has formulated rules of inference to govern and guide debate
and to promote discovery. Logic has to deal as well with other important elements of
discourse, but its main province has always been, and still is, inference.
Idle talk and trivial conversation do not rank as discourse for our purpose. Logic has little to
do with the frivolous; its business is with the serious statement which admits truth or falsity.
Logic promotes truth; yet we can go far in Logic without knowing or caring much whether a
particular statement is true or false, in the ordinary acceptation of those words. By true in
ordinary speech we mean true to fact, and by false we mean the opposite. Now a statement,
true to fact, may in its context infringe a rule of Logic; and a statement, false in fact, may in
its context conform to the rules of Logic. The logician, as such, is not directly concerned with
fact, but is much more concerned with the observance of the rules of Logic, and therefore he
uses a pair of technical terms, valid and invalid, to express, respectively, what conforms to the
rules of logic and what does not conform thereto. By the aid of these terms he can set out the
rules of reasoning without committing himself as to whether a particular statement is true to
fact, or not. Valid comes from the Latin, validus, strong. A valid passport may make mistakes
in fact, but if duly signed and not out of date, it may do its work and get you through the
barrier. On the other hand, it may give the colour of the eyes and all the other facts correctly,
but if it is out of date, it will not do its work; it is invalid. The distinction between truth and
validity must be carefully observed. It is illogical and therefore incorrect to speak of a true
syllogism, if you mean a valid syllogism, or of a valid conclusion, if you mean a true
conclusion.
This distinction is of special importance here; for in the study of the syllogism, as such, Logic
concentrates on the form of the reasoning, for the most part, and is not directly concerned
with the truth of its contents. If the syllogism complies with the formal rules, it is valid, if not,
not. If the conclusion follows from the premisses, the conclusion is valid, even though
premiss and conclusion may not be true to fact. Example:
All fish are cold-blooded.
Whales are fish.
Whales are cold-blooded.
www.seyfihoca.com
154
The first premiss is true; the second is false; the conclusion is false; but the conclusion is
correctly drawn from the premisses, and therefore valid in its syllogism, even though it is not
true to fact.
The reverse can happen too. A proposition, true to fact, may appear as the conclusion of an
invalid syllogism. Example:
The industrious are prudent.
Ants are prudent.
Ants are industrious.
These examples are warnings against the habit of judging the validity or invalidity of a
syllogism by the truth or falsity of the conclusion. As students of Logic our first duty is to
look at the working of the syllogism, and to judge its validity, or otherwise, by the rules. No
good comes of confusing the two sets of terms, as is sometimes done. Truth is truth, and
validity is validity, and neither can do duty for the other. The lazy habit of styling a valid
conclusion true, or a true conclusion valid, weakens both our sense of truth and our feeling for
Logic.
(From Teach Yourself Logic, by A. A. Luce.)
Inductive and deductive logic
Now I want to talk about methods of finding one's way through these hierarchies - logic.
Two kinds of logic are used, inductive and deductive. Inductive inferences start with
observations of the machine and arrive at general conclusions. For example, if the cycle goes
over a bump and the engine misfires, and then goes over another bump and the engine
misfires, and then goes over another bump and the engine misfires, and then goes over a long
smooth stretch of road and there is no misfiring, and then goes over a fourth bump and the
engine misfires again, one can logically conclude that the misfiring is caused by the bumps.
That is induction: reasoning from particular experiences to general truths.
Deductive inferences do the reverse. They start with general knowledge and predict a specific
observation. For example, if, from reading the hierarchy of facts about the machine, the
mechanic knows the horn of the cycle is powered exclusively by electricity from the battery,
then he can logically infer that if the battery is dead the horn will not work. That is deduction.
Solution of problems too complicated for common sense to solve is achieved by long strings
of mixed inductive and deductive inferences that weave back and forth between the observed
machine and the mental hierarchy of the machine found in the manuals. The correct program
for this interweaving is formalized as scientific method.
Actually I've never seen a cycle-maintenance problem complex enough really to require fullscale formal scientific method. Repair problems are not that hard. When I think of formal
scientific method an image sometimes comes to mind of an enormous juggernaut, a huge
bulldozer-slow, tedious, lumbering, laborious, but invincible. It takes twice as long, five times
as long, maybe a dozen times as long as informal mechanic's techniques, but you know in the
www.seyfihoca.com
155
end you're going to get it. There's no fault isolation problem in motorcycle maintenance that
can stand up to it. When you've hit a really tough one, tried everything, racked your brain and
nothing works, and you know that this time Nature has really decided to be difficult, you say,
"Okay, Nature, that's the end of the nice guy," and you crank up the formal scientific method.
For this you keep a lab notebook. Everything gets written down, formally, so that you know at
all times where you are, where you've been, where you're going and where you want to get. In
scientific work and electronics technology this is necessary because otherwise the problems
get so complex you get lost in them and confused and forget what you know and what you
don't know and have to give up. In cycle maintenance things are not that involved, but when
confusion starts it's a good idea to hold it down by making everything formal and exact.
Sometimes just the act of writing down the problems straightens out your head as to what they
really are.
The logical statements entered into the notebook are broken down into six categories: (1)
statement of the problem, (2) hypotheses as to the cause of the problem, (3) experiments
designed to test each hypothesis, (4) predicted results of the experiments, (5) observed results
of the experiments and (6) conclusions from the results of the experiments. This is not
different from the formal arrangement of many college and high-school lab notebooks but the
purpose here is no longer just busywork. The purpose now is precise guidance of thoughts
that will fail if they are not accurate.
The real purpose of scientific method is to make sure Nature hasn't misled you into thinking
you know something you don't actually know. There's not a mechanic or scientist or
technician alive who hasn't suffered from that one so much that he's not instinctively on
guard. That's the main reason why so much scientific and mechanical information sounds so
dull and so cautious. If you get careless or go romanticizing scientific information, giving it a
flourish here and there, Nature will soon make a complete fool out of you. It does it often
enough anyway even when you don't give it opportunities. One must be extremely careful and
rigidly logical when dealing with Nature: one logical slip and an entire scientific edifice
comes tumbling down. One false deduction about the machine and you can get hung up
indefinitely.
In Part One of formal scientific method, which is the statement of the problem, the main skill
is in stating absolutely no more than you are positive you know. It is much better to enter a
statement "Solve Problem: Why doesn't cycle work?" which sounds dumb but is correct, than
it is to enter a statement "Solve Problem: What is wrong with the electrical system?" when
you don't absolutely know the trouble is in the electrical system. What you should state is
"Solve Problem: What is wrong with cycle?" and then state as the first entry of Part Two:
"Hypothesis Number One: The trouble is in the electrical system." You think of as many
hypotheses as you can, then you design experiments to test them to see which are true and
which are false.
This careful approach to the beginning questions keeps you from taking a major wrong turn
which might cause you weeks of extra work or can even hang you up completely. Scientific
questions often have a surface appearance of dumbness for this reason. They are asked in
order to prevent dumb mistakes later on.
Part Three, that part of formal scientific method called experimentation, is sometimes thought
of by romantics as all of science itself because that's the only part with much visual surface.
www.seyfihoca.com
156
They see lots of test tubes and bizarre equipment and people running around making
discoveries. They do not see the experiment as part of a larger intellectual process and so they
often confuse experiments with demonstrations, which look the same. A man conducting a
gee-whiz science show with fifty thousand dollars' worth of Frankenstein equipment is not
doing anything scientific if he knows beforehand what the results of his efforts are going to
be. A motorcycle mechanic, on the other hand, who honks the horn to see if the battery works
is informally conducting a true scientific experiment. He is testing a hypothesis by putting the
question to nature. The TV scientist who mutters sadly, "The experiment is a failure; we have
failed to achieve what we had hoped for," is suffering mainly from a bad scriptwriter. An
experiment is never a failure solely because it fails to achieve predicted results. An
experiment is a failure only when it also fails adequately to test the hypothesis in question,
when the data it produces don't prove anything one way or another.
Skill at this point consists of using experiments that test only the hypothesis in question,
nothing less, nothing more. If the horn honks, and the mechanic concludes that the whole
electrical system is working, he is in deep trouble. He has reached an illogical conclusion. The
honking horn only tells him that the battery and horn are working. To design an experiment
properly he has to think very rigidly in terms of what directly causes what. This you know
from the hierarchy. The horn doesn't make the cycle go. Neither does the battery, except in a
very indirect way. The point at which the electrical system directly causes the engine to fire is
at the spark plugs, and if you don't test here, at the output of the electrical system, you will
never really know whether the failure is electrical or not.
To test properly the mechanic removes the plug and lays it against the engine so that the base
around the plug is electrically grounded, kicks the starter lever and watches the spark-plug
gap for a blue spark. If there isn't any he can conclude one of two things: (a) there is an
electrical' failure or (b) his experiment is sloppy. If he is experienced he will try it a few more
times, checking connections, trying every way he can think of to get that plug to fire. Then, if
he can't get it to fire, he finally concludes that a is correct, there's an electrical failure, and the
experiment is over. He has proved that his hypothesis is correct.
In the final category, conclusions, skill comes in stating no more than the experiment has
proved. It hasn't proved that when he fixes the electrical system the motorcycle will start.
There may be other things wrong. But he does know that the motorcycle isn't going to run
until the electrical system is working and he sets up the next formal question: "Solve problem:
what is wrong with the electrical system?"
He then sets up hypotheses for these and tests them. By asking the right questions and
choosing the right tests and drawing the right conclusions the mechanic works his way down
the echelons of the motorcycle hierarchy until he has found the exact specific cause or causes
of the engine failure, and then he changes them so that they no longer cause the failure.
An untrained observer will see only physical labour and often get the idea that physical labour
is mainly what the mechanic does. Actually the physical labour is the smallest and easiest part
of what the mechanic does. By far the greatest part of his work is careful observation and
precise thinking. That is why mechanics sometimes seem so taciturn and withdrawn when
performing tests. They don't like it when you talk to them because they are concentrating on
mental images, hierarchies, and not really looking at you or the physical motorcycle at all.
They are using the experiment as part of a program to expand their hierarchy of knowledge of
www.seyfihoca.com
157
the faulty motorcycle and compare it to the correct hierarchy in their mind. They are looking
at underlying form.
(From Zen and the art of motorcycle maintenance by Robert Pirsig
PHYSICS
The Origin of the Sun and Planets
The suggestion that the material of the earth was indeed derived from an exploding star - a
supernova, is supported by strong evidence. The shower of stars must have been surrounded
by a cloud of gas - the cloud from which the stars had just condensed. A supernova,
undergoing violent disintegration, must have expelled gases that went to join this cloud, the
material from the supernova thereby getting mixed with the large quantity of hydrogen of
which the cloud was mainly composed. Our problem is then to explain how both the sun and
the planets were formed out of this mixture of materials.
It is a characteristic of a good detective story that one vital clue should reveal the solution to
the mystery, but that the clue and its significance should be far from obvious. Such a clue
exists in the present problem. It turns on the simple fact that the sun takes some 26 days to
spin once round on its axis-the axis being nearly perpendicular to the orbits of the planets,
which lie in nearly the same plane. The importance of this fact is that the sun has no business
to be rotating in 26 days. It ought to be rotating in a fraction of a day, several hundred times
faster than it is actually doing. Something has slowed the spin of the sun. It is this something
that yields the key to the mystery.
Stars are the products of condensations that occur in the dense inter-stellar gas clouds. A
notable cloud is the well-known Orion Nebula whose presence in the 'sword' of Onion can
easily be seen with binoculars. Stars forming out of the gas in such clouds must undergo a
very great degree of condensation. To begin with, the material of a star must occupy a very
large volume, because of the extremely small density of the inter-stellar gas. In order to
contain as much material as the sun does, a sphere of gas in the Orion Nebula must have a
diameter of some 10,000,000,000,000 miles. Contrast this with the present diameter of the
sun, which is only about a million miles. Evidently in order to produce a star like the sun a
blob of gas with an initial diameter of some 10 million million miles must be shrunk down in
some way to a mere million miles. This implies a shrinkage to one ten millionth of the
original size.
Now it is a consequence of the laws of dynamics that, unless some external process acts on it,
a blob of gas must spin more and more rapidly as it shrinks. The size of a condensation and
the speed of its spin keep an inverse proportion with each other. A decrease of size to one tenmillionth of the original dimensions leads to an increase in the speed of spin by 10 million.
But the rotation speed of the sun is only about 2 kilometres per second. At a speed of 100
kilometres per second the sun would spin round once in about half a day, instead of in the
observed time of 26 days.
Only one loophole remains. We must appeal to some external process to slow down the spin
of the solar condensation. Our problem is to discover how such an external process operates.
First we must decide at what stage of the condensation the external process acts. Does it act
while the condensing blob still has very large dimensions? Or does it operate only in the later
www.seyfihoca.com
158
stages, as the condensation reaches the compact stellar state ? Or does it operate more or less
equally throughout the whole shrinkage?
A strong hint that the process must act mainly in the late stages of the condensation comes
from observations of the rates of spin of stars. It is found that the rates of spin have a very
curious dependence on surface temperature. Stars like the sun, with surface temperatures less
than 6,000° C, rotate slowly like the sun. But stars with surface temperatures greater than
7,000° C rotate considerably more rapidly, their equatorial speeds of rotation being usually
greater than 5o kilometres per second. Although this is still much less than what we should
expect if no external process were operative, it is considerably greater than the equatorial
rotation speed possessed by the sun.
This shows that while the external process must be operative in all cases, it is operative to
different degrees that depend on the surface temperature of the final star. Now the difference
between one star and another can scarcely show at all during the early stages of the shrinkage.
Certainly the difference between two condensations, one yielding a star of surface
temperature 6,000° C and the other yielding a star of surface temperature 7,000° C, must be
very small indeed during the early stages: much too small for the stars to come to have
markedly different rotation speeds if the external process were of main effect during the early
stages. The inference is that the process operates mainly during the late stages of
condensation.
8. Now what was the external process? We have mentioned that rotary forces must have
become important during the late stages of condensation. The effect of these forces was to
cause the condensation to become more and more flattened at its poles. Eventually the
flattening became sufficient for an external rotating disc to begin growing out of the equator.
The sequence of events is illustrated in figure 1.
Once the sun had thus grown a disc the external process was able to come into operation. The
process consisted of a steady transference of 'rotational momentum from the sun to the disc.
Two birds were thereby killed with one stone. The sun was slowed down to its present slow
rate of spin and the disc, containing the material out of which the planets were subsequently to
condense, was pushed farther and farther from the sun. The solar condensation probably first
grew its disc when it had shrunk to a size somewhat less than the orbit of the innermost
planet, Mercury. The pushing outwards of the main bulk of the disc explains why the larger
planets now lie so far from the sun.
www.seyfihoca.com
159
It may be wondered why such an obvious theory was not put forward long ago. The answer is
that there seemed to be such grave objections to it that not until very recently has it been
examined at all seriously. And now it turns out that the objections are not so grave as was
previously believed.
(From Chapter VI of Frontiers of Astronomy by Fred Hoyle.)
Can Life Exist on the Planets?
The old view that every point of light in the sky represented a possible home for life is quite
foreign to modern astronomy. The stars have surface-temperatures of anything from 1,650
degrees to 60,000 degrees or more and are at far higher temperatures inside. A large part of
the matter of the universe consists of stellar matter at a temperature of millions of degrees, its
molecules being broken up into atoms, and the atoms broken up, partially or wholly, into their
constituent parts. The rest consists, for the most part, of nebular gas or dust. Now the very
concept of life implies duration in time; there can be no life-or at least no life at all similar to
that we know on earth-where atoms change their make up millions of times a second and no
pair of atoms can ever stay joined together. It also implies a certain mobility in space, and
these two implications restrict life to the small range of physical conditions in which the
liquid state is possible. Our survey of the universe has shown how small this range is in
comparison with that exhibited by the universe as a whole. It is not to be found in the stars nor
in the nebulae out of which the stars are born. Indeed, probably only an infinitesimal fraction
of the matter of the universe is in the liquid state.
Actually we know of no type of astronomical body in which the conditions can be favourable
to life except planets like our own revolving round a sun. Even these may be too hot or too
cold for life to obtain a footing. In the solar system, for instance, it is hard to imagine life
existing on Mercury or Neptune since liquids boil on the former and freeze hard on the latter.
www.seyfihoca.com
160
Even when all the requisite conditions are satisfied, will life come or will it not? We must
probably discard the at one time widely accepted view that if once life had come into the
universe in any way whatsoever, it would rapidly spread from planet to planet and from one
planetary system to another until the whole universe teemed with life; space now seems too
cold, and planetary systems too far apart. Our terrestrial life must in all probability have
originated on the earth itself. What we should like to know is whether it originated as the
result of some amazing accident or succession of coincidences, or whether it is the normal
event for inanimate matter to produce life in due course, when the physical environment is
suitable. We look to the biologist for the answer, which so far he has not been able to produce.
The astronomer might be able to give a partial answer if he could find evidence of life on
some other planet, for we should then at least know that life had occurred more than once in
the history of the universe, but so far no convincing evidence has been forthcoming. There is
no definite evidence of life anywhere in the universe, except on our own planet.
Apart from the certain knowledge that life exists on earth, our only definite knowledge is that,
at the best, life must be limited to a tiny fraction of the universe. Millions of millions of stars
exist which support no life, which have never done so and never will do so. Of the planetary
systems in the sky, many must be entirely lifeless, and in others life, if it exists at all, is
probably limited to a few of the planets.
Let us leave these rather abstract speculations and come down to earth. The earth, which
started life as a hot mass of gas, has gradually cooled, until it has now about touched bottom,
and has almost no heat beyond that which it receives from the sun. This just about balances
the amount it radiates away into space, so that it would stay at its present temperature for ever
if external conditions did not change, and any change in its condition will be forced on it by
changes occurring outside. These changes may be either gradual or catastrophic.
Of the gradual changes which are possible, the most obvious is a diminution in the light and
heat received from the sun. We have seen that if the sun consisted of pure hydrogen, it could
lose one part in 150 of its whole mass through the transformation of hydrogen into helium.
The energy thus set free would provide for radiation at the present rate through a period of
1000,000 million years.
The sun does not consist of pure hydrogen, and has never done so, but a fair proportion of its
present substance is probably hydrogen, and this ought to provide radiation for at least several
thousands of millions of years, at the present rate. After all the available supplies of hydrogen
are used up, the sun will, so far as we can guess, proceed to contract to the white dwarf state,
probably to a condition resembling that of the faint companion of Sirius. The shrinkage of the
sun to this state would transform our oceans into ice and our atmosphere into liquid air; it
seems impossible that terrestrial life could survive.
Such at least would be the normal course of events, the tragedy we have described happening
after a time of the order of perhaps 10,000 millions of years. But a variety of accidents may
intervene to bring the human race to an end long before any such interval has elapsed. To
mention only possible astronomical occurrences, the sun may run into another star, any
asteroid may hit any other asteroid and, as a result, be so deflected from its path as to strike
the earth, any of the stars in space may wander into the solar system and, in so doing, upset all
the planetary orbits to such an extent that the earth becomes impossible as an abode of life. It
www.seyfihoca.com
161
is difficult to estimate the likelihood of any of these events happening, but rough calculations
suggest that none of them is at all likely to happen within the next 10,000 million years or so.
A more serious possibility is that the sun's light and heat may increase so much as to shrivel
up all terrestrial life. We have seen how 'novae' occasionally appear in the sky, temporarily
emitting anything up to 25,000 times the radiation of the sun. It seems fairly certain that if our
sun were suddenly to become a nova, its emission of light and heat would so increase as to
scorch all life off the earth, but we are completely in the dark as to whether our sun runs any
risk of entering the nova stage. If it does, this is probably the greatest of all the risks to which
life on earth is exposed.
Apart from improbable accidents, it seems that if the solar system is left to the natural course
of evolution, the earth is likely to remain a possible abode of life for thousands of millions of
years to come.
(From The Universe Around Us by Sir James Jeans, F.R.S.
The Theory of Continuous Creation
We must move on to consider the explanations that have been offered for this expansion of
the universe. Broadly speaking, the older ideas fall into two groups. One was that the universe
started its life a finite time ago in a single huge explosion, and that the present expansion is a
relic of the violence of this explosion. This big bang idea seemed to me to be unsatisfactory
even before detailed examination showed that it leads to serious difficulties. For when we
look at our own galaxy there is not the smallest sign that such an explosion ever occurred. But
the really serious difficulty arises when we try to reconcile the idea of an explosion with the
requirement that the galaxies have condensed out of diffuse background material. The two
concepts of explosion and condensation are obviously contradictory, and it is easy to show, if
you postulate an explosion of sufficient violence to explain the expansion of the universe, that
condensations looking at all like the galaxies could never have been formed.
We come now to the second group of theories. The ordinary idea that two particles attract
each other is only accepted if their distance apart is not too great. At really large distances, so
the argument goes, the two particles repel each other instead. On this basis it can be shown
that if the density of the background material is sufficiently small, expansion must occur. But
once again there is a difficulty in reconciling this with the requirement that the background
material must condense to form the galaxies.
I should like now to approach more recent ideas by describing what would be the fate of our
observable universe if any of these older theories had turned out to be correct. Every receding
galaxy will eventually increase its distance from us until it passes beyond the limit of the
observable universe-that is to say, they will move to a distance beyond the critical limit of
about two thousand million light years that I have already mentioned. When this happens,
nothing that occurs within them can ever be observed from our galaxy. So if any of the older
theories were right we should end in a seemingly empty universe, or at any rate in a universe
that was empty apart perhaps from one or two very close galaxies that became attached to our
galaxy as satellites.
Although I think there is no doubt that every galaxy we now observe to be receding from us
will, in about ten thousand million years, have passed entirely beyond the limit of vision of an
www.seyfihoca.com
162
observer in our galaxy, yet I think that such an observer would still be able to see about the
same number of galaxies as we do now. By this I mean that new galaxies will have condensed
out of the background material at just about the rate necessary to compensate for those that
are being lost as a consequence of their passing beyond our observable universe. At first sight
it might be thought that this could not go on indefinitely because the material forming the
background would ultimately become exhausted. But again, I do not believe that this is so, for
it seems likely that new material is constantly being created so as to maintain a constant
density in the background material. So we have a situation in which the loss of galaxies,
through the expansion of the universe, is compensated by the condensation of new galaxies,
and this can continue indefinitely.
The idea that matter is created continuously represents our ultimate goal in this series of
lectures. The idea in itself is not new. I know of references to the continuous creation of
matter that go back more than twenty years; and I have no doubt that a close inquiry would
show that the idea, in its vaguest form, goes back very much further than that. What is new is
that it has now been found possible to put a hitherto vague idea in a precise mathematical
form. It is only when this has been done that the consequences of any physical idea can be
worked out and its scientific value assessed.
Now what are the consequences of continuous creation? Perhaps the most surprising result of
the mathematical theory is that the average density of the background material must stay
constant. To achieve this only a very slow creation rate is necessary. The new material does
not appear in a concentrated form in small localized regions but is spread throughout the
whole of space. The average rate of appearance amounts to no more than the creation of one
atom in the course of a year in a volume equal to St. Paul's Cathedral. As you will realize, it
would be quite impossible to detect such a rate of creation by direct experiment.
But although this seems such a slow rate when judged by ordinary ideas, it is not small when
you consider that it is happening everywhere in space. The total rate for the observable
universe alone is about a hundred million, million, million, million, million tons per second.
Do not let this surprise you because, as I have said, the volume of the observable universe is
very large. It is this creation that drives the universe. The new material produces an outward
pressure that leads to the steady expansion. But it does much more than that. With continuous
creation the apparent contradiction between the expansion of the universe and the requirement
that the background material shall be able to condense into galaxies is completely overcome.
For it can be shown that once an irregularity occurs in the background material a galaxy must
eventually be formed. Such irregularities are constantly being produced through the
gravitational action of the galaxies themselves. So the background material must give a steady
supply of new galaxies. Moreover, the created material also supplies unending quantities of
atomic energy. For, by arranging that newly created material is composed of hydrogen, we
explain why, in spite of the fact that hydrogen is being consumed in huge quantities in the
stars, the universe is nevertheless observed to be overwhelmingly composed of it.
So we see that no large-scale changes in the universe can be expected to take place in the
future. Without continuous creation the universe must evolve towards a dead state in which all
the matter is condensed into a vast number of dead stars. With continuous creation, on the
other hand, the universe has an infinite future in which all its present very large-scale features
will be preserved.
(From The Nature of the Universe by Fred Hoyle.)
www.seyfihoca.com
163
The Creation of the Universe
Before we can discuss the basic problem of the origin of our universe, we must ask ourselves
whether such a discussion is necessary. Could it not be true that the universe has existed since
eternity, changing slightly in one way or another in its minor features, but always remaining
essentially the same as we know it today? The best way to answer this question is by
collecting information about the probable age of various basic parts and features that
characterize the present state of our universe.
For example, we may ask a physicist or chemist: 'How old are the atoms that form the
material from which the universe is built?' Only half a century ago such a question would not
have made much sense. However, when the existence of natural radioactive elements was
recognized, the situation. became quite different. It became evident that if the atoms of the
radio-active elements had been formed too far back in time, they would by now have decayed
completely and disappeared. Thus the observed relative abundances of various radio-active
elements may give us some clue as to the time of their origin.
We notice first of all that thorium and the common isotope of uranium (U238) are not
markedly less abundant than the other heavy elements, such as, for example, bismuth,
mercury or gold. Since the half-life periods of thorium and of common uranium are 14 billion
and 4.5 billion years respectively, we must conclude that these atoms were formed not more
than a few billion years ago. On the other hand the fissionable isotope of uranium (U235) is
very rare, constituting only 0.7 % of the main isotope. The half-life of U235 is considerably
shorter than that of U238, being only about 0.9 billion years. Since the amount of fissionable
uranium has been cut in half every 0.9 billion years, it must have taken about seven such
periods, or about 6 billion years, to bring it down to its present rarity, if both isotopes were
originally present in comparable amounts. Similarly, in a few other radio-active elements,
such as radio-active potassium, the unstable isotopes are also always found in very small
relative amounts. This suggests that these isotopes were reduced quite considerably by slow
decay taking place over a period of a few billion years. Of course, there is no a priori reason
for assuming that all the isotopes of a given element were originally produced in exactly equal
amounts. But the coincidence of the results is significant, inasmuch as it indicates the
approximate date of the formation of these nuclei. Furthermore, no radio-active elements with
half-life periods shorter than a substantial portion of a billion years are found in nature,
although they can be produced artificially in atomic piles. This also indicates that the
formation of atomic species must have taken place not much more recently than a few billion
years before the present time. Thus there is a strong argument for assuming that radio-active
atoms, and along with them, all other stable atoms were formed under some unusual
circumstances which must have existed in the universe a few billion years ago.
As the next step in our enquiry, we may ask a geologist: 'How old are the rocks that form the
crust of our globe?' The age of various rocks - that is, the time that has elapsed since their
solidification from the molten state - can be estimated with great precision by the so-called
radio-active clock method. This method, which was originally developed by Lord Rutherford,
is based on the determination of the lead content in various radio-active minerals such as
pitchblende and uraninite. The significant point is that the natural decay of radio-active
materials results in the formation of the so-called radiogenic lead isotopes. The decay of
thorium produces the lead isotope Pb208, whereas the two isotopes of uranium produce Pb207
and Pb206. These radiogenic lead isotopes differ from their companion Pb204, natural lead,
which is not the product of decay of any natural radio-active element.
www.seyfihoca.com
164
As long as the rock material is in the molten state various physical and chemical processes
may separate the newly produced lead from the mother substance. However, after the material
has become solid and ore has been formed, radiogenic lead remains at the place of its origin.
The longer the time period after the solidification of the rock, the larger the amount of lead
deposited by any given amount of radio-active substance. Therefore, if one measures the
relative amounts of deposited radiogenic lead isotopes and the lead-producing radio-active
substances (that is, the ratios: Pb208/Th232, Pb207/U235, Pb206/U238) and if one knows the
corresponding decay rates, one can get three independent estimates of the time when a given
radio-active ore was formed. By applying this method, one gets results of the kind shown in
the following table.
Mineral
Locality
Geological period
Age in years x
106
Pitchblende
Colorado, U.S.A.
Tertiary
58
Pitchblende
Bohemia, Europe
Carboniferous
215
Pitchblende
Belgium Congo, Africa
Pre-Cambrian
580
Uraninite
Wilberforce, Canada
Pre-Cambrian
1,035
Pitchblende
Great Bear Lake, Canada
Pre-Cambrian
1,330
Uranite
Karelia, U.S.S.R.
Pre-Cambrian
1,765
Uranite
Manitoba, Canada
Pre-Cambrian
1,985
The last two minerals are the oldest yet found, and from their age we must conclude that the
crust of the earth is at least 2 billion years old.
A much more elaborate method was proposed recently by the British geologist, Arthur
Holmes. This method goes beyond the formation time of different radio-active deposits; and
claims an accurate figure for the age of the material forming the earth. By applying this
method to the relative amounts of lead isotopes found in rocks of different geological ages,
Holmes found that all curves intersect near the point corresponding to a total age of 3.35
billion years, which must represent the correct age of our earth.
How old is the moon? As was shown by the work of the British astronomer, George Darwin,
the moon is constantly receding from the earth, at the rate of about 5 inches every year.
Dividing the present distance to the moon (239,000 miles) by the estimated rate of recession,
we find that the moon must have been practically in contact with the earth about 4. billion
years ago.
Thus we see that whenever we inquire about the age of some particular part or property of the
universe we always get the same approximate answer - a few billion years old. Thus it seems
that we must reject the idea of a permanent unchangeable universe and must assume that the
basic features of the universe as we know it today are the direct result of some evolutionary
development which must have begun a few billion years ago.
www.seyfihoca.com
165
(From The Creation of the Universe by George Gamow.
Atomic Radiation and Life
The radiation dose given off by an X-ray machine or by isotopes is usually measured by
determining the number of ions produced in a volume of gas. Since these carry an electric
charge there are a number of extremely delicate methods by which they can be detected. The
widely used Geiger counter consists essentially of a wire stretched inside a cylindrical tube, so
arranged that an electric current can pass between the wire and the tube only when there are
ions in the gas. Consequently, when an ionizing particle passes through the tube, an electric
signal is given out. In this way the number of ionizing particles given off by a radio-active
source can be accurately counted. This is called the activity of the material. It is measured in a
unit called the curie after the discoverer of radium. The activity of one gram of radium
together with its decay products is equal to one curie. Every time an atom disintegrates a betaor alpha-ray is given off, together with a certain amount of gamma radiation.
The activity in curies can tell us nothing about the dose of radiation given off by the radioactive material, since the curie measures only the number of ionizing particles emitted,
independent of their range or energy. If, for example, we put next to the skin one curie of
radio-active cobalt, which gives off energetic gamma-rays, the dose received on the surface
will be one five-thousandth part of the dose received from one curie of polonium which gives
off alpha-particles. On the other hand the gamma-rays from the curie of cobalt will penetrate
deeply, while the alpha-rays will not affect anything which lies more than two one-thousandth
of an inch below the surface of the skin.
The best way of defining the dose of radiation which an irradiated material has received is in
terms of energy. We have seen that on exposure to ionizing radiation electrons, or other subatomic particles moving at great speed, lose energy to the surrounding molecules. The amount
of' energy gained by the irradiated substance is clearly the important factor, and will
determine the biological changes produced. The most widely used unit for measuring X-ray
and gamma-ray dosage is the roentgen - named after the discoverer of X-rays. The remarkable
property of ionizing radiation is that the small amount of energy represented by a few hundred
roentgens can kill a man.
The primitive embryonic cell known as the zygote, which is formed after the entry of the
sperm into the ovum, is very sensitive so radiation. For example, 80 per cent of mice, exposed
to 200 rads of X-rays within the first five days after conception, fail to give birth. Smaller
doses give rise to a lower incidence of pre-natal death, but an appreciable reduction in the
average litter size has been observed with 50 rads.
At first the embryo grows by cell division without differentiation and becomes firmly
implanted in the wall of the uterus. This requires about eight days in human beings and five
days in mice. Then differentiation begins, and the individual organs and limbs are formed; the
embryo takes shape. During this period it is in the greatest danger. Now radiation no longer
kills-the damaged embryo is not re-absorbed or aborted, but proceeds to a live birth which is
abnormal. These malformations can be very great, so as to give horrible and distressing
monsters, which are, however, quite capable of living for a time. The incidence is particularly
high in the early stages of the active development of the embryo.
www.seyfihoca.com
166
The period of major organ production is over after about three months in human beings, and
the foetus then develops its finer aspects and generally grows and develops. Exposure to doses
insufficient to produce severe radiation sickness in the mother no longer produces gross
deformities which can be recognized in small experimental animals. But the absence of
striking changes in the newborn does not mean that the irradiation has been without harm.
The general effect is less obvious, but none the less serious, and irradiation at the later stages
of pregnancy results in very marked growth reduction, giving small babies which develop into
smaller adults. Their life span is reduced and their reproductive organs are often affected so
that they grow up sterile. Damage to the brain and eyes was found a few weeks after birth in
all cases which had been irradiated in the foetal stage with zoo rads, and there is a significant
incidence after 200 rads. Since only gross disorders of the brain can be detected in
experimental animals, it seems likely that much smaller doses will give effects which are
serious in man.
The detailed picture of the influence of radiation on prenatal development has been obtained
from studies with animals. Unhappily, sufficient human cases are known to make it certain
that the same pattern also occurs in man; and we can confidently superimpose a human timescale on the mouse data. Some of our information is derived from the survivors of the atom
bombs in Japan. The children of the women who were pregnant and exposed to irradiation at
Nagasaki and Hiroshima are, on average, shorter and lighter and have smaller heads,
indicating an under-developed brain. Some show severe mental deficiencies, while others
were unable to speak normally at five years old.
Most of our knowledge comes from expectant mothers who were irradiated for therapeutic or
diagnostic reasons. Many cases are described in the medical literature of abnormalities
following exposure of the embryo. Most of these arose twenty or thirty years ago at a time
when radiologists did not know of the great radio-sensitivity of the foetus. A detailed survey
showed that, where a mother received several hundred roentgen within the first two months
after the implantation of the embryo, severe mal-development was observed in every child, a
high proportion of whom lived for many years.
(From Atomic Radiation and Life by Peter Alexander.)
Marconi and the invention of Radio
The theory of energy waves - or wireless waves - in :he air was first put forward in 1864, ten
years before Marconi was born by the British physicist James Clerk Maxwell, Professor of
Experimental Physics at Cambridge University. Maxwell produce by brilliant mathematical
reasoning, the theory that 'electro-magnetic disturbances', though invisible to our eves, must
exist in space, and that these waves travel at the same speed as the light waves - that is, at the
rate of 186,000 miles, the equivalent of more than seven times round the world, per second.
This was a very remarkable deduction; but at that time no means of producing and detecting
such waves were known, and Maxwell was, therefore unable to carry out any scientific
experiments to prove the truth of his theory, and had to leave it to others to show how correct
his reasoning had been. Twenty-four years later the German physicist, Heinrich Hertz, was
able to show that, when he produced an electric spark across a gap between two metal balls by
applying an electric current to them, a similar spark would jump across a small gap in a metal
ring a few feet away, though the ring was not connected in any way with the other apparatus.
www.seyfihoca.com
167
The experiment made it clear that the second spark was caused by the first, and the fact that
there was no connexion between the balls and the ring suggested that the spark from the first
apparatus must have been transferred through the intervening space between the two pieces of
apparatus in the form of some kind of wave spreading outwards in all directions like light. In
fact, it was clear that those were the electro-magnetic waves of Maxwell's theory.
Marconi began his work by fixing up a simple contraption, similar to the one used by Hertz in
his discovery, and trying this out on a table. Then, when he, too, had succeeded in making a
spark cross from one apparatus to another, he designed a more elaborate set-up and tried to
reproduce a bigger spark and at a greater distance. He spent weeks making and testing new
pieces of equipment, and then breaking them up again when they proved to be useless. He had
so many failures and disappointments that he sometimes felt quite desperate; but he
persevered, and eventually succeeded in sending a spark the full length of a thirty-foot room.
Though Marconi was naturally delighted with this development, it was clearly only a
beginning; he must now try to make his sparks do something useful. This meant more gadgets
and more experiments. Then, one night after the rest of the family had gone to bed, he
succeeded in making the wireless waves start an electric bell ringing in a room two floors
below his laboratory. He was so excited that he rushed to his mother's bedroom and woke her
up to tell her, and the next day he demonstrated his experiment successfully to his father, too.
He extended the arm attached to one of the metal balls of the transmitting instrument into an
elevated aerial and connected the second arm to earth. On the receiving instrument he also
used an elevated aerial and earth. This led to an important advance, for not only did it greatly
extend the range through which the waves could be transmitted, but it also increased the
volume of sound received. Marconi now found that he could transmit Morse Code signals: 'I
actually transmitted and received intelligible signals,' he said.
Marconi was asked to take his equipment on to the roof of the head office of the General Post
Office at St. Martin's-le-Grand (London) and to send a wireless signal from there to the rooftop of another building some 300 yards away, where he set up a receiver. Both these tests
were successful, and so he was then asked to give a third demonstration - this time on
Salisbury Plain - to a group of high-ranking army and naval officers and government officials.
This third demonstration was naturally a very important one for Marconi. He therefore took
immense pains over his preparations, testing and re-testing his apparatus to make certain that
everything was in order. Then, while the important spectators stood by his receiver nearly two
miles away, he tapped out a Morse signal on his transmitter-and waited anxiously for the
result. The signal came through perfectly.
The following March, a German vessel collided in a fog with the East Goodwin lightship off
the Kentish coast, and, for the first time, a distress signal calling for relief was despatched by
wireless - and was answered. This was a remarkable proof of the importance of radio to
shipping, and not long after most of the larger nations were fitting their ships with wireless
equipment.
'He'll be sending messages across the Atlantic next,' people joked, never for a moment
believing that this might really be possible. But it seemed no impossibility to Marconi, and
shortly after a visit to the United States he seriously set to work to achieve the sending of a
signal from Great Britain to America across the Atlantic Ocean.
www.seyfihoca.com
168
He chose a lonely spot on the south coast of Cornwall called Poldhu, high above the cliffs
near Mullion. Work on building the new station began in October 1900. The following
January the elaborate new plant was installed, and then Marconi, with his usual careful
attention to detail, spent several months testing it and suggesting improvements. Then on 27th
November 1901, Marconi with two of his assistants, Kemp and Paget, sailed for
Newfoundland, 2,000 miles away, where they intended setting up both a receiver for the
reception of the signal from Poldhu and a transmitter for sending a second signal back to
Cornwall. The small party successfully installed their receiver at St. John's in a disused
hospital attached to a naval barracks on a 500-foot hill which, strangely enough, was called
'Signal Hill'.
As the weather was growing rapidly worse, Marconi decided to concentrate upon trying to
receive a signal from Cornwall. So he sent a cable back to Poldhu giving instructions for the
three Morse Code dots representing the letter 'S' to be transmitted at frequent intervals each
day, starting on 11 December. He heard nothing on the first day except for the roar of the gale
outside. On the second day, 12 December, the gale was so strong that it blew away the kite
supporting the important aerial and a second kite had to be hoisted. But that afternoon, just as
he was beginning to think that his experiment had failed, Marconi heard, very faintly, the
sound for which he had been listening, the signal from Poldhu ... ... ... ...
(From Great Inventors by Norman Wymer.)
Particles or Waves?
The most obvious fact about a ray of light, at any rate to superficial observation, is its
tendency to travel in a straight line; everyone is familiar with the straight edges of a sunbeam
in a dusty room. As a rapidly-moving particle of matter also tends to travel in a straight line,
the early scientists, rather naturally, thought of light as a stream of particles thrown out from a
luminous source, like shot from a gun. Newton adopted this view, and added precision to it in
his 'corpuscular theory of light'.
Yet it is a matter of common observation that a ray of light does not always travel in a straight
line. It can be abruptly turned by reflection, such as occurs when it falls on the surface of a
mirror. Or its path may be bent by refraction, such as occurs when it enters water or any liquid
medium; it is refraction that makes our oar look broken at the point where it enters the water,
and makes the river look shallower than it proves to be when we step into it. Even in
Newton's time the laws which governed these phenomena were well known. In the case of
reflection the angle at which the ray of light struck the mirror was exactly the same as that at
which it came off after reflection; in other words, light bounces off a mirror like a tennis ball
bouncing off a perfectly hard tennis-court. In the case of refraction, the sine of the angle of
incidence stood in a constant ratio to the sine of the angle of refraction. We find Newton at
pains to skew that his light-corpuscles would move in accordance with these laws, if they
were subjected to certain definite forces at the surfaces of a mirror or a refracting liquid.
Newton's corpuscular theory met its doom in the fact that when a ray of light falls on the
surface of water, only part of it is refracted. The remainder is reflected, and it is this latter part
that produces the ordinary reflections of objects in a lake, or the ripple of moonlight on the
sea. It was objected that Newton's theory failed to account for this reflection, for if light had
consisted of corpuscles, the forces at the surface of the water ought to have treated all
corpuscles alike; when one corpuscle was refracted all ought to be, and this left water with no
www.seyfihoca.com
169
power to reflect the sun, moon or stars. Newton tried to obviate this objection by attributing
'alternate fits of transmission and reflection' to the surface of the water - the corpuscle which
fell on the surface at one instant was admitted, but the next instant the gates were shut, and its
companion was turned away to form reflected light. This concept was strangely and strikingly
anticipatory of modern quantum theory in its abandonment of the uniformity of nature and its
replacement of determinism by probabilities, but it failed to carry conviction at the time.
And, in any case, the corpuscular theory was confronted by other and graver difficulties.
When studied in sufficiently minute details, light is not found to travel in such absolutely
straight lines as to suggest the motions of particles. A big object, such as a house or a
mountain, throws a definite shadow, and so gives as good protection from the glare of the sun
as it would from a shower of bullets. But a tiny object, such as a very thin wire, hair or fibre,
throws no such shadow. When we hold it in front of a screen, no part of the screen remains
unilluminated. In some way, the light contrives to bend round it, and, instead of a definite
shadow, we see an alternation of light and comparatively dark parallel bands, known as
'interference bands'. To take another instance, a large circular hole in a screen lets through a
circular patch of light. But make the hole as small as the smallest of pinholes, and the pattern
thrown on a screen beyond is not a tiny circular patch of light, but a far larger pattern of
concentric rings, in which light and dark rings alternate - 'diffraction rings'. All the light
which is more than a pinhole's radius from the centre has in some way bent round the edge of
the hole.
Newton regarded these phenomena as evidence that his 'light-corpuscles' were attracted by
solid matter. He wrote:
The rays of light that are in our air, in their passage near the angles of bodies, whether
transparent or opaque (such as the circular and rectangular edges of coins, or of knives, or
broken pieces of stone or glass), are bent or inflected round those bodies, as if they were
attracted to them; and those rays which in their passage came nearest to the bodies are the
most inflected, as if they were most attracted.
Here again Newton was strangely anticipatory of present-day science, his supposed forces
being closely analogous to the 'quantum forces' of the modern wave-mechanics. But they
failed to give any detailed explanation of diffraction-phenomena, and so met with no favour.
In time all these and similar phenomena were adequately explained by supposing that light
consists of waves, somewhat similar to those which the wind blows up on the sea, except that,
instead of each wave being many yards long, many thousands of waves go to a single inch.
Waves of light bend round a small obstacle in exactly the way in which waves of the sea bend
round a small rock. A rocky reef miles long gives almost perfect shelter from the sea, but a
small rock gives no such protection - the waves pass round it on either side, and re-unite
behind it, just as waves of light re-unite behind our thin hair or fibre. In the same way seawaves which fall on the entrance to a harbour do not travel in a straight line across the
harbour but bend round the edges of the breakwater, and make the whole surface of the water
in the harbour rough. The seventeenth century regarded light as a shower of particles; the
eighteenth century, discovering that this was inadequate to account for small-scale
phenomena such as we have just described, replaced the showers of particles by trains of
waves.
(From The Mysterious Universe by Sir James Jeans.)
www.seyfihoca.com
170
Matter, Mass and Energy
At the end of last century, physical science recognized three major conservation laws:
A The conservation of matter
B The conservation of mass
C The conservation of energy
Other minor laws, such as those of the conservation of linear and angular momenta, need not
enter our discussion, since they are mere deductions from the three major laws already
mentioned. Of the three major laws, the conservation of matter was the most venerable. It had
been implied in the atomistic philosophy of Democritus and Lucretius, which supposed all
matter to be made up of untreatable, unalterable and indestructible atoms. It asserted that the
matter content of the universe remained always the same, and the matter content of any bit of
the universe or of any region of space remained the same except in so far as it was altered by
the ingress or egress of atoms. The universe was a stage in which always the same actors-the
atoms-played their parts, differing in disguises and groupings, but without change of identity.
And these actors were endowed with immortality.
The second law, that of the conservation of mass, was of more modern growth. Newton had
supposed every body or piece of substance to have associated with it an unvarying quantity,
its mass, which gave a measure of its 'inertia' or reluctance to change its motion. If one motorcar requires twice the engine power of another to give us equal control over its motion we say
that it has twice the mass of the latter car. The law of gravitation asserts that the gravitational
pulls on two bodies are in exact proportion to their masses, so that if the earth's attraction on
two bodies proves to be the same, their 'masses' must be the same, whence it follows that the
simplest way of measuring the mass of any body is by weighing it.
The third principle, that of conservation of energy, is the most recent of all. Energy can exist
in a vast variety of forms, of which the simplest is pure energy of motion-the motion of a train
along a level track, or of a billiard ball over a table. Newton had shown that this purely
mechanical energy is 'conserved'. For instance, when two billiard balls collide, the energy of
each is changed, but the total energy of the two remains unaltered; one gives energy to the
other, but no energy is lost or gained in the transaction. This, however, is only true if the balls
are 'perfectly elastic', an ideal condition in which the balls spring back from one another with
the same speed with which they approached. Under actual conditions such as occur in nature,
mechanical energy invariably appears to be lost; a bullet loses speed on passing through the
air, and a train comes to rest in time if the engine is shut off In all such cases heat and sound
are produced. Now a long series of investigations has shown that heat and sound are
themselves forms of energy. In a classical series of experiments made in 1840-50, Joule
measured the energy of sound with the rudimentary apparatus of a violoncello string.
Imperfect though his experiments were, they resulted in the recognition of 'conservation of
energy' as a principle which covered all known transformations of energy through its various
modes of mechanical energy, heat, sound and electrical energy. They showed in brief that
energy is transformed rather than lost, an apparent loss of energy of motion being
compensated by the appearance of an exactly equal energy of heat and sound; the energy of
www.seyfihoca.com
171
motion of the rushing train is replaced by the equivalent energy of the noise of the shrieking
brakes, and of the heating of wheels, brake-blocks and rails.
These three conservation laws ought of course to have been treated merely as working
hypotheses, to be tested in every conceivable way and discarded as soon as they showed signs
of failing. Yet so securely did they seem to be established that they were treated as
indisputable universal laws. Nineteenth-century physicists were accustomed to write of them
as though they governed the whole of creation, and on this basis philosophers dogmatized as
to the fundamental nature of the universe.
It was the calm before the hurricane. The first rumble of the approaching storm was a
theoretical investigation by Sir J. J. Thomson, which showed that the mass of an electrified
body could be changed by setting it into motion; the faster such a body moved the greater its
mass became, in opposition to Newton's concept of a fixed unalterable mass. For the moment,
the principle of conservation of mass appeared to have abandoned science.
For a time this conclusion remained of merely academic interest; it could not be tested
observationally because ordinary bodies could neither be charged with sufficient electricity,
nor set into motion with sufficient speed, for the variations of mass predicted by theory to
become appreciable in amount. Then, just as the nineteenth century was drawing to a close,
Sir J. J. Thomson and his followers began to break up the atom, which now proved to be no
more uncuttable, and so no more entitled to the name of 'atom' than the molecule to which the
name had previously been attached. They were only able to detach small fragments, and even
now the complete break-up of the atom into its ultimate constituents has not been fully
achieved. These fragments were found to be all precisely similar, and charged with negative
electricity. They were accordingly named electrons.
These electrons are far more intensely electrified than an ordinary body can ever be. A gram
of gold, beaten as thin as it will go, into a gold leaf a yard square, can with luck be made to
hold a charge of about 60,000 electrostatic units of electricity, but a gram of electrons carries
a permanent charge which is about nine million million times greater. Because of this, and
because electrons can be set into motion by electrical means with speeds of more than a
hundred thousand miles a second, it is easy to verify that an electron's mass varies with its
speed. Exact experiments have shown that the variation is precisely that predicted by theory.
(From The Mysterious Universe by Sir James Jeans.)
Structure of Matter
One topic which we shall certainly discuss is the structure of matter. We shall find that there
is a simple, general pattern in the structure of all the solid materials which principally concern
us, and that electrons are a key part of that pattern.
There are ninety-two separate chemical substances from which the world is made; substances
such as oxygen, carbon, hydrogen, iron, sulphur and silicon. Often, of course, they are
combined together to make more elaborate materials, such as hydrogen and oxygen in water.
But these are the ninety-two elements; elements which the chemists have grouped in a list
called the Periodic Table and which, in various combinations, go to make up all the millions
of compounds that exist.
www.seyfihoca.com
172
If, however, an atom of any one of these elements is examined, it will be found to consist of
an assembly of three different kinds of particle: protons, electrons and neutrons. The atoms of
some elements may contain several hundred particles, while other elements may have less
than ten particles in the atom.
Before we look at the way in which these particles makeup an atom, we need to know
something of their two most important properties: mass and electrical charge. The proton and
the neutron both have approximately the same mass, and the mass of the electron is very
much less - about 1/1840 of the mass of the proton. The electron is negatively charged, and
the proton has a charge of the same size but of positive sign. The neutron carries no charge.
Any object, from an atom upwards in size, will normally contain equal numbers of protons
and electrons, and will thus have no net electrical charge. If electrons are removed in some
way from the object, it will be left with a net positive charge. If electrons are added to the
object, it will become negatively charged.
Two objects which are electrically charged exert a force on each other which is inversely
proportional to the square of their distance apart. If the two charges have the same sign, then
the objects repel each other, and if the charges are of opposite sign, then they attract each
other. In particular, a proton and an electron will attract each other, and the closer they are
together, the greater will be the force.
In general, an atom has a central core called the nucleus, which consists of protons and
neutrons. Surrounding this nucleus is a cloud of electrons. The number of protons in the
nucleus is equal to the number of electrons in the cloud. The total positive charge on the
nucleus due to all the protons is just balanced by the total negative charge of all the electrons
in the cloud, and the atom as a whole is electrically neutral.
To get an idea of the principles common to the structure of all atoms we shall start by
considering the simplest atom-that of the element hydrogen.
Hydrogen is the lightest of all the atoms, with a single electron in the electron `cloud' and a
single proton as its nucleus., All other elements have neutrons as well as protons in the
nucleus; for example, helium, the next simplest atom, has two electrons in the cloud, with two
protons and two neutrons in the nucleus.
In the hydrogen atom the electron rotates round the nucleus-in this case a single proton-rather
like the Earth rotates round the Sun. If we think of the circular electron orbit, then its radius is
such that the electrical attraction between the positively charged proton and the negatively
charged electron is just sufficient to provide the force needed to bend the electron path into a
circle, just as a stone whirled round on a string would fly off at a tangent if it were not
constrained to its circular path by the tension in the string.
This simple system with the electron rotating round the nucleus has a certain amount of
energy associated with it. First, there is the kinetic energy of the moving electron. Any
moving body possesses kinetic energy, and the amount of energy is proportional to the mass
of the body and the square of its velocity. Thus a small, high-speed body like a bullet may
have more kinetic energy than a much larger body moving slowly.
www.seyfihoca.com
173
The second kind of energy associated with the hydrogen atom is potential energy, due to the
fact that a positive and negative charge separated by a certain distance attract each other, and
could be organised to do work in coming together. In the same way, a lake of water at the top
of a mountain has potential energy associated with it, because it could be organised to do
work by running through turbines down into the valley. The higher the lake is above the
valley, the greater will be the potential energy. Similarly, the farther the electron is from the
nucleus in the hydrogen atom, the greater will be the potential energy.
The total energy associated with the hydrogen atom. is the sum of the potential energy and the
kinetic energy. This total depends upon the radius of the orbit in which the electron rotates
and has a minimum value for hydrogen atoms in the normal state. If a normal hydrogen atom
is in some way given a little extra energy, then the electron moves out into an orbit of greater
radius, and the total ; energy associated with the atom is now greater than it was in the normal
state. Atoms possessing more than the normal amount of energy are said to be `excited'.
The way in which atoms receive extra energy to go into an excited state, and the way in which
they give up that energy in returning to the normal state, is of fundamental; importance and
we shall discuss it later as the quantum: theory. In particular, we shall find that energy can
only be given to an atom in packets of certain sizes and that energy is emitted by excited
atoms in similar packets. The wrong size packet will not be accepted by an atom, and an
excited atom will never emit anything other than one of a limited set of packet sizes.
The fact that an atom of any particular element can emit or absorb energy only in packets of
certain sizes is' one consequence of a set of rules which governs the behaviour of electrons in
atoms. These rules also give rise to two other important general properties of the' electrons in
atoms.
The first of these is that, within the atom, an electron may only possess certain energies-there
are certain `permitted energy levels' for the electron. This concept of permitted energy levels
is extended later from the single isolated atom to the electrons in solids, where the electrical
properties are largely determined by the permitted energy levels and which of them are
possessed by electrons.
Another consequence of the rules for electron behaviour within the atom is one which
concerns us immediately, because these rules establish a set of patterns into which the
electrons are arranged in more elaborate atoms than those of hydrogen.
The hydrogen atom is very simple, with its single electron rotating in a circular orbit, which is
of a fixed radius for the normal state of the atom.
Helium is the next simplest atom, with two electrons and, of course, two protons in the
nucleus to balance the negative charge of the electrons. The nucleus also contains two
neutrons, and the two electrons circulate round this nucleus in the same orbit.
When we come to the next element, lithium, which has three orbital electrons, the electrons
are arranged in a new way round the nucleus. The nucleus now contains three protons and
some neutrons.
Two of the electrons are in the same orbit, but the third one is in a different orbit, farther away
from the nucleus. The rules say that the innermost orbit, or shell, is completely filled when it
www.seyfihoca.com
174
has two electrons. Atoms like lithium, with more than two electrons, must start a second orbit
of greater radius to accommodate the third; fourth, etc., electrons. When this second orbit has
eight electrons in it, there is no room for more, and a third orbit of still larger radius has to be
started. Thus sodium, which has eleven orbital electrons, has two in the innermost orbit, eight
in the next orbit and one in the outermost orbit.
In the illustrations of electron orbits for various atoms, the number with the positive sign
indicates the number of protons in the nucleus and thus is equal to the number of orbital
electrons. This number is called the atomic number of the element. There are neutrons in all
the nuclei except hydrogen. The chemical properties of an element are determined by the
electrons, and in this respect it is not surprising that the outermost electrons are most
important because, when two atoms come together in a chemical reaction, it is the electrons in
the outside orbits which will first meet and interact. It is worth noting from the illustrations
that hydrogen, lithium and sodium, all with a single electron in their outside shell, have a
general similarity of chemical behaviour. Similarly, many of the atoms with two electrons-or
with three, etc.- in the outer shell, can be grouped together as being related chemically.
Atoms do not normally have a separate existence, and the simplest form in which we meet
matter in the natural world is the gas. This consists of single molecules of the substance
moving about at random and largely independent of each other except when they collide.
Typical gas molecules are oxygen (O2) and hydrogen (H2). In each of these molecules, two
atoms of the element are joined together to make the stable unit found in natural oxygen or
hydrogen. The atoms in a molecule are bound together by forces due to complex interactions
between the electron systems of the individual atoms. The exact nature of these forces does
not concern us as we shall be more interested in solids than in gases. However, it is important
to notice that molecules, like atoms, emit or absorb energy in packets of certain sizes and that
the electrons in the molecular system also have only certain permitted energies.
Gas molecules may contain atoms of more than one element, e.g. carbon dioxide (CO2), and
in the case of some organic gases may contain many atoms of several different elements and
be very large.
The solid is in many ways similar to a very large molecule. Most of the solids that are
important in electronics are crystalline, and the characteristic feature of a crystal is a regular
arrangement of atoms. The atoms may all be of the same element, as in copper, or they may
be of different elements, as in common salt (NaCl) or copper sulphate (CuSO4).
In the crystal, as in the molecule, the interaction of the orbital electrons binds the individual
atoms into the characteristic pattern or lattice.
www.seyfihoca.com
175
There are certain permitted energy levels for electrons in the solid, just as there are in the
atom and the molecule, and these energy levels will be different in different materials.
One special aspect of electron properties in solids is that certain materials, particularly metals,
contain some electrons which are able to move away from their parent atoms. If a battery is
connected to such a material, electrons will flow through the material, and it is called a
conductor of electricity. If the application of a battery to a material does not cause a flow of
electric current, then the material is called an insulator.
As they flow through a conductor, the electrons which make up the current being driven
round the circuit by the battery give up some of their energy to the main solid structure of the
material, which thus becomes hot. If electrons lose much energy in passing through a
conducting material, then that material is said to have a large resistance to current flow. A
given current flowing: through a high-resistance conductor will generate much more heat than
the same current flowing through a low resistance conductor. Thus the heating element of an
electric fire will be made of high-resistance material, usually a metal alloy. The wires carrying
the current under the floor to the fire will be made of low-resistance material, invariably a
metal. Materials which offer low` resistance to the passage of an electric current are called
good electrical conductors.
This, then, is the way in which the chemical energy stored in a battery is converted into heat
through the action of a flow of electrons in a conductor. If a metal is heated to a sufficiently
high temperature by the current, as in the tungsten lamp filament, then the same mechanism
can provide light - another form of energy.
Sometimes, however, it will be necessary to encourage electrons actually to escape from the
surface of the metal so that, for instance, they can be formed into a beam passing' down a
cathode-ray tube to paint a picture on a screen.
The emission of electrons from a material occurs only if the electrons inside are in some way
given sufficient energy to break through the surface where a sort of energy barrier exists,
called the work function of the material. Materials with low work functions emit electrons
easily because less energy is required for an electron to overcome the energy barrier and
escape.
If a solid is heated, then it receives extra energy. This may be shared between the regular
crystal lattice and some of the electrons which can escape from their parent atoms. If the
electrons thereby acquire sufficient energy to escape from the surface of the solid, then the
process is called thermionic emission because it is brought about by heat. Thermionic
emission provides the electrons which move through the vacuum in a radio valve, but not the
electrons in a transistor, which always remain inside the solid material of which the transistor
is made.
Another way in which electrons may be given sufficient energy to escape is by shining light
onto the surface of the material. If electrons escape as a result of receiving energy from
incident light, then the process is called photoelectric emission. This is the basis of many light
sensitive devices using photoelectric cells.
For a material with a given work function there is a certain critical wavelength for
photoelectric emission to occur. If the wavelength is too long, there will be no emission. In
www.seyfihoca.com
176
general, therefore, ultra-violet or blue light, which has a short wavelength, causes
photoelectric emission from more materials than the longer wavelength red or infra-red
radiation.
Whether an electron stays inside a solid or escapes depends on the amount of energy the
electron possesses. But inside the solid the electrons have a number of permitted energy
levels, and very often important properties of the material depend on which of these energy
levels are occupied. Thus a material may be an insulator if its electrons are in low energy
levels, but it can be made to conduct electricity if sufficient energy is given to it to raise the
electrons to higher permitted levels. Furthermore, if electrons in high energy levels give up
energy and fall to lower permitted levels, then the energy emitted may be useful in special
ways - lasers, for instance, depend upon such emission.
We have seen that electrons are fundamental to the structure of matter. Moreover, that the part
played by an electron in the structure is very much connected with the energy levels permitted
to it, and with the energy it actually possesses. In particular, we have seen that the electrons
flowing as electric current through a solid may give up energy to the crystal structure, so that
heat and perhaps light may be given off by the material.
The Quantum Theory of Radiation
Now, before the quantum theory was put forward, there was no notion of natural units of
radiant energy: it was believed that we could have any amount of energy, as small as we
pleased, radiated by a hot body or a luminous atom. It could, however, be shown
mathematically that, if this were true, we should expect a hot body to radiate nearly all its
energy in the violet and ultraviolet end of the spectrum, which we know to be against the facts
of observation.
The problem was solved in the first year of the present century, when Planck showed that, to
get the right result, it was necessary to make a revolutionary hypothesis: to suppose that
radiant energy was sent out in packets, as it were - in units or atoms of energy, just as matter
existed in atomic units. We cannot have less than an atom of lead, say; any minute piece of
lead must consist of a whole number of atoms. We cannot have an electric charge of less than
an electron. In the same way, we cannot have less than a unit - or quantum, as it is called - of
radiant energy, and any body that sends out or absorbs radiation must deal with one quantum
or a whole number of quanta.
The little parcel of light of one particular frequency in which radiant energy is delivered is
sometimes called a 'light dart', a very expressive term, but is more generally known as a
photon. The photon is simply a quantum of radiant energy, the only object of sometimes using
the new term being that 'quantum' is a more inclusive term, which can be applied to other
things as well as light - for instance, to the vibration of whole atoms and molecules.
The quantum of radiant energy differs from the quantum of electricity, the electron, in a very
important way. The amount of charge is the same on all electrons: there is but one unit. The
magnitude of this unit of radiant energy, however, is different for every different kind-that is,
for every different wave-length - of radiation. It is, in fact, proportional to the frequency, so
that the quantum of energy of extreme visible red radiation is only half that of the extreme
visible violet radiation, which, as we have said before, has double the frequency. The
quantum of an X-radiation is very much greater than the quantum of any visible radiation.
www.seyfihoca.com
177
The quantum of energy corresponding to a given species of radiation is found, then, by
multiplying the frequency by a certain fixed number, which is called Planck's universal
constant, and always indicated by h. Planck's constant enters into every aspect of modern
atomic physics and its numerical value has been found by at least ten different methods,
involving such things as X-ray properties, the distribution of energy in black-body radiation,
the frequencies of spectral lines, and so on. All the methods give values agreeing to within a
few parts in ten thousand.
Light, then, or radiation in general, has a packet property as well as a wave property, and this
is one of the paradoxes of modern physics. Newton's conception of light was a stream of
particles, which he endowed with something in the nature of pulsating properties in an attempt
to account for certain phenomena which we can now easily explain on the wave theory. He
felt the need for the double aspect, the particle and the periodic, and provided for it in his
theory.
If white light could be considered to consist of particles of various sizes, corresponding to all
the different colours which it contains, it would be fairly easy to imagine the amount of
energy in each particle to depend upon its size, and the quantum, or atomic, nature of the
energy would be a natural consequence. We could have one or two or three particles of
sodium light, but not a fraction of a particle. However, to account for the various phenomena
of interference and diffraction, we have to admit that light behaves under some conditions as a
stream of particles, each one of which has a fixed energy belonging to it, and under other
conditions as a wave motion. Both wave aspects and particle aspects are included today in a
comprehensive theory known as wave mechanics.
Thus in geometrical optics we usefully treat light as travelling in straight lines through lenses,
although we know that it has wave properties which must be considered when optical
resolution is in question. In the simple kinetic theory of gases - say, for working out the action
of high-vacuum pumps - we can safely treat atoms as small elastic particles and neglect their
structure. The essential is to know which aspect is the dominating one in the problem
concerned.
Take, first of all, the photo-electric effect. We have seen that, when light or any radiation of
short wave-length falls on a metal, electrons are shot off, the number depending on the
strength of the light; but the speed, or more correctly the energy, on the kind of light. If the
radiation is done up in packets of energy, and this energy can be expended in pushing out
electrons, then clearly we should expect each atom of light either to throw out an electron
with a definite speed, or to do nothing; but not to throw out an electron with some greater
speed. According to the quantum theory, a short wave-length - that is, a high-frequency radiation should therefore throw out electrons with an energy of motion greater than that
which characterizes electrons ejected by radiation of long wave-length.
This is just what is observed; red light does not eject electrons at all from most metals, violet
light drives them out with a small speed, ultra-violet light produces greater speed, and X-rays
throw out very fast electrons. Red light, the quanta of which are very small, on account of its
low frequency, has no effect, because a certain minimum energy is required just to jerk an
electron out of an atom: the energy of the red quanta is, for most metals, less than this
minimum. Allowing for the small energy required just to free the electron, the energy which
the electron acquires can be shown experimentally to be exactly proportional to the frequency
of the radiation. This is now so well established that, in cases where it is difficult to measure
www.seyfihoca.com
178
the frequency of very short ultra-violet waves by ordinary means, the energy of the electrons
which they eject from matter has been determined, and taken as a measure of the wave-length.
The same method has been applied to the exceedingly high-frequency gamma rays of radium.
(From An Approach to Modern Physics by E. N. Da C. Andrade.)
Footprints of the Atom
The Cambridge physicist C. T. R. Wilson was studying the formation of fogs in 1898 when he
started on a train of ideas and discoveries which led ultimately to the perfection of the Wilson
Cloud Chamber as a marvellous aid to nuclear physics. This first fog-making apparatus was,
however, very simple just two glass jars connected by a pipe with a tap in it. One jar
contained moist (saturated) air and the other was pumped empty of air. When the tap was
opened the air expanded quickly into the empty jar. When gases expand very quickly they
cool. You may have noticed that the air rushing out of a bicycle tyre when it is suddenly let
down is quite cold. As a result of such cooling, clouds form in the moist air, since only a
smaller amount of water vapour can be held by the air at the lower temperature.
Fogs cannot form unless the cooling is very marked. This is because small drops tend to
evaporate again more easily than the large drops. It is therefore very difficult for any drops to
begin to form at all unless they form immediately into large drops, as they would if the
cooling were pronounced. The dust particles in ordinary air act as a very convenient
beginning for the drops, because they are already of sufficient size for the drops formed on
them to avoid re-evaporation. This explains why fogs and mists are much more common and
more persistent near large manufacturing towns, where there is a lot of smoke and dust, rather
than in the clear air of the countryside.
One of Wilson's cloud chambers, designed in 1912, is shown in figure 1. The moist air is
contained in a glass-topped cylinder B by a close-fitting piston. The air in flask A is pumped
out by a vacuum pump, while the stopper C is kept closed. When you want to operate the
apparatus this stopper is pulled out so that the air beneath the piston rushes out into A. As
there is nothing to hold up the piston it falls, and thus allows the damp air to expand and cool,
and form a cloud.
www.seyfihoca.com
179
Wilson soon discovered that even if he used very clean air he still occasionally got clouds
with only a moderate amount of cooling when there was some radio-active substance or a
source of the then recently discovered X-rays near his cloud chamber. He soon showed that
this was because the air in his chamber was being 'ionized' by their radio-activity or by the Xrays.
Let us see what happens when air is ionized. The atoms of the various gases that make up air
are all built of a heavy positive nucleus surrounded by very light negative electrons. The
amount of positive charge on the nucleus is exactly balanced by the number of negative
electrons surrounding it, so that the atom as a whole is electrically neutral. Although it takes
quite a hard knock by a particle from one of the powerful 'atom-smashing' machines to break
off a bit of the nucleus, the negative electrons are held to the atom only by the electrostatic
attraction of opposite charges between negative electron and positive nucleus. It is therefore
much easier to remove an electron from the atom, either by strong electro-magnetic radiation
(X-rays) or by collision with another atomic particle like those shot by radio-active
substances.
Before impact the atom was neutral, but after it has lost a negative electron it becomes a
positively charged 'ion'. The electron it has lost is a negative ion; it is free to move until it
joins another positive ion to form a neutral atom. When an atom is split in this way into two
ions, positive and negative, we say it is ionized.
If ions of this type are present in moist air their effect is twofold. First, they attract many more
normal atoms to themselves, with the result that a cluster of atoms is formed. Secondly, the
www.seyfihoca.com
180
charge on the ion reduces the tendency of small drops to evaporate. All this means that the
charged ions in moist cooled air act as suitable nuclei, like dust and smoke, on which fog and
clouds can form.
To make this clear we will follow up what happens when an alpha-particle passes through a
cloud chamber immediately after expansion has taken place, and when the moist air has been
cooled so that it is ready to form clouds. An alpha-particle is a fragment shot out of the
nucleus of a radio-active atom. It is about four times as heavy as a hydrogen atom. It has a
double positive charge, and it will travel a few inches in normal air before it is brought to a
stop by repeated collisions with the atoms in the air. While it is travelling very fast through
the air of a cloud chamber, its positive charge attracts many of the outer negative electrons of
the atoms in the air, which may be drawn i out of their original atoms, leaving them ionized.
The passage of one alpha-particle thus leaves a trail of ionized atoms behind it all along its
track, each of which is capable of acting as a centre about which a tiny drop of water can
form. If the alpha-particle is shot into the moist air of a cloud chamber just after a cooling
expansion has taken place, so that the air is supersaturated, a trail of little water drops appears
in its wake, looking just like the cloud trails left by a high-flying aircraft.
The first time that an atomic nucleus was split artificially was in 1919 when Lord Rutherford
turned a nitrogen atom into an oxygen atom by bombardment with alpha-particles. This
nucleus-splitting reaction has been studied by Professor Blackett with the aid of a cloud
chamber. He obtained an actual photo graph of the famous event. In this photograph a beam
of alpha-particles appears as white trails crossing the chamber. One of them stops half-way as
it hits a nitrogen atom in the air and the lighter trail of a proton (a positively charged hydrogen
atom), knocked out of the nitrogen nucleus, moves off to the left, while the resulting oxygen
nucleus gives a thick short trail to the right. We have come as close as we can to actually
seeing the invisible atom.
(From an article by R. R. Campbell in Adventures in Science edited by B. C. Brookes.)
Splitting the Atom
'The experiments started about four in the afternoon,' recalled a scientist whom Rutherford
had invited one day in 1919 to see what he was doing. 'We went into his laboratory to spend a
preliminary half hour in the dark to get our eyes into the sensitive state necessary for
counting. Sitting there, drinking tea, in the dim light of a minute gas jet at the further end of
the laboratory, we listened to Rutherford talking of all things under the sun. It was curiously
intimate, yet impersonal, and all of it coloured by that characteristic of his of considering
statements independently of the person who put them forward.'
Then Rutherford, in his unassuming white coat, made a last minute inspection tour round his
laboratory, a high and wide room with a cement floor. There was, in one corner, the enormous
column of the condenser, which went right up through the ceiling; and at the other end of the
room a large tube, enthroned on top of a work-bench in the midst of a mass of entangled
electric wires. There was an arc-lamp projector behind the tube, and a screen had been set up
in front of it.
'You know, we might go up through the roof,' warned Rutherford, but the boyish smile under
the big greying moustache belied his words. The blinds were now pulled down over the big
www.seyfihoca.com
181
leaded windows, and bluish-green sparks were seen to jump to and fro in the tube. The screen
lit up. At first there was nothing but a thick grey mist. Then some large objects, like the
shadows of enormous fish, flowed across the screen in a steady stream.
The Professor explained. Alpha particles - helium nuclei - were being hurled through the tube,
in which an artificial mist had been created. It was an adaptation of Wilson's cloud chamber,
filled with nitrogen gas. Suddenly a thick streak appeared on the screen, striking off at right
angles at terrific speed. 'That's it,' said Rutherford. 'The atom has been split!'
The performance was repeated - once, twice, a third time at irregular intervals. Millions of
alpha particles went straight through the nitrogen gas without touching any of its atoms. But
now and then there came a direct hit on a nitrogen nucleus, which split it. 'Where are we
going from here?' mused one of Rutherford's guests. 'Who knows?' he replied, 'We are
entering no-man's land.'
What interested Rutherford in these experiments was the transmutation of one element into
another-which furnished the proof that his theory of what the atom looked like was correct.
When an alpha particle hit a nitrogen nucleus it drove out some of its seven protons. And each
of these loose protons became the nucleus of a hydrogen atom, which has only one proton
with an electron revolving around it. Thus nitrogen changed into hydrogen!
But Rutherford proved yet another theory, which was closely connected with Einstein's hotly
disputed claim-that there is no real difference between mass and energy, and that the
destruction of matter would free its latent energy. Already in 1905, Albert Einstein, then a
young man of 26, had startled the scientific world with his Special Theory of Relativity, in
which he gave the phenomenon of radio-activity an important place within the framework of
his new picture of the universe. He explained that, if matter is converted into energy by the
disintegration of atoms, that process would be represented by a simple little equation: E =
mc².
What does it mean? Basically it says that mass and energy are not different things which have
no relation to one another, but that one can be changed into the other. Einstein's equation
connects the two quantities. E is the energy in ergs released when a mass of m grams is
completely disintegrated. And c is the velocity of light in centimetres per second, and
therefore c² is 900 million million ergs.
This sounded completely fantastic. Even if matter could ever be converted into energy, surely
the energy released in this process would not be of such unimaginable magnitude! There was,
of course, no way of proving or disproving Einstein's equation - until Rutherford showed how
to split the atom. In 1905 no one really believed that Einstein's equation would ever be put to
the test; that man could ever release the incredible forces locked up in the atoms of matter.
Today we know that, if one ounce of matter could be completely destroyed and changed into
energy, it would yield as much power as we derive from burning 100,000 tons of coal in a
conventional power station!
'Atom-splitting' became almost a fashion in the physical laboratories of Europe and America,
and in the 1920s most of the lighter nuclei were being split up by bombarding them with alpha
particles. Only beryllium - the fourth lightest of the elements resisted all attempts to break up
its nucleus. Instead of releasing one of its four protons when hit, it gave off a burst of
radiation more penetrating than even the hard gamma rays. Sir James Chadwick, again at the
www.seyfihoca.com
182
Cavendish Laboratory in Cambridge, proved that this radiation must consist of particles about
as heavy as protons, but without an electric charge. 'If such a neutral particle exists,'
Rutherford had said in 1920, 'it should be able to move freely through matter, and it may be
impossible to contain it in a sealed vessel.' He knew that its discovery would be of great
importance-an electrically neutral particle could be fired into any matter without being
attracted or repelled by protons or electrons.
In 1932 the Joliot-Curies made a radio-active metal bombard beryllium - which is a nonradio-active metal - with its rays. The result was that the beryllium, too, became radio-active even more so than the original source of the rays. Sir James Chadwick's explanation was that
the beryllium nuclei had released their non-electrical particles, which he called neutrons. They
were found to have slightly greater mass than the protons - but neutrons can change to protons
by acquiring a positive electric charge.
The discovery of the neutron not only solved quite a number of problems which had so far
defied the efforts of the scientists, but it also gave an even greater impetus to atom-splitting.
The Americans, true to style, went into the business in a big way. The University of California
built an enormous machine, the cyclotron, for the head of its radiation laboratory, Professor E.
O. Lawrence, who had just come to the conclusion that the heavy hydrogen atom, consisting
of one proton plus one neutron, would make an ideal bullet for shooting up other nuclei.
(From Atomic Energy - A Layman's Guide to the Nuclear Age by E. Larsen.)
THE DEVELOPMENT OF ELECTRICITY
The phenomenon which Thales had observed and recorded five centuries before the birth of
Christ aroused the interest of many scientists through the ages. They made various practical
experiments in their efforts to identify the elusive force which Thales had likened to a 'soul'
and which we now know to have been static electricity.
Of all forms of energy, electricity is the most baffling and difficult to describe. An electric
current cannot be seen. In fact it does not exist outside the wires and other conductors which
carry it. A live wire carrying a current looks exactly the same and weighs exactly the same as
it does when it is not carrying a current. An electric current is simply a movement or flow of
electrons.
Benjamin Franklin, the American statesman and scientist born in Boston in 1706, investigated
the nature of thunder and lightning by flying a child's kite during a thunderstorm. He had
attached a metal spike to the kite, and at the other end of the string to which the kite was tied
he secured a key. As the rain soaked into the string, electricity flowed freely down the string
and Franklin was able to draw large sparks from the key. Of course this could have been very
dangerous, but he had foreseen it and had supported the string through an insulator. He
observed that this electricity had the same properties as the static electricity produced by
friction.
But long before Franklin many other scientists had carried out research into the nature of
electricity.
In England William Gilbert (1544-1603) had noticed that the powers of attraction and
repulsion of two non-metallic rods which he had rubbed briskly were similar to those of
www.seyfihoca.com
183
lodestone and amber - they had acquired the curious quality we call magnetism.
Remembering Thales of old he coined the word 'electricity'.
Otto von Guericke (1602-1686) a Mayor of Magdeburg in Germany, was an amateur scientist
who had constructed all manner of gadgets. One of them was a machine consisting of two
glass discs revolving in opposite directions which produced high voltage charges through
friction. Ramsden and Wimshurst built improved versions of the machine.
A significant breakthrough occurred when Alessandro Volta (1745-1827) in Italy constructed
a simple electric cell (in 1799) which produced a flow of electrons by chemical means. Two
plates, one of copper and the other of zinc, were placed in an acid solution and a current
flowed through an external wire connecting the two plates. Later he connected cells in series
(voltaic pile) which consisted of alternate layers of zinc and copper discs separated by flannel
discs soaked in brine or acid which produced a higher electric pressure (voltage). But Volta
never found the right explanation of why his cell was working. He thought the flow of electric
current was due to the contact between the two metals, whereas in fact it results from the
chemical action of the electrolyte on the zinc plate. However, his discovery proved to be of
incalculable value in research, as it enabled scientists to carry out experiments which led to
the discoveries of the heating, lighting, chemical and magnetic effects of electricity.
One of the many scientists and physicists who took advantage of the 'current electricity' made
possible by Volta's cells was Hans Christian Oersted (1777-1851) of Denmark. Like many
others he was looking for a connection between the age-old study of magnetism and
electricity, but now he was able to pass electric currents through wires and place magnets in
various positions near the wires. His epoch-making discovery which established for the first
time the relationship between magnetism and electricity was in fact an accident.
While lecturing to students he showed them that the current flowing in a wire held over a
magnetic compass needle and at right angles to it (that is east-west) had no effect on the
needle. Oersted suggested to his assistant that he might try holding the wire parallel to the
length of the needle (north- south) and hey presto, the needle was deflected! He had stumbled
upon the electromagnetic effect in the first recorded instance of a wire behaving like a magnet
when a current is passed through it.
A development of Oersted's demonstration with the compass needle was used to construct the
world's first system of signaling by the use of electricity.
In 1837 Charles Wheatstone and William Cooke took out a patent for the world's first Fiveneedle Telegraph, which was installed between Paddington railway station in west London
and West Drayton station a few miles away. The five copper wires required for this system
were embedded in blocks of wood.
Electrolysis, the chemical decomposition of a substance into its constituent elements by the
action of an electric current, was discovered by the English chemists Carlisle and William
Nicholson (1753-1815). If an electric current is passed through water it is broken down into
the two elements of which it is composed -- hydrogen and oxygen. The process is used
extensively in modern industry for electroplating. Michael Faraday (1791-1867) who was
employed as a chemist at the Royal Institution, was responsible for introducing many of the
technical terms connected with electrolysis, like electrolyte for the liquid through which the
electric current is passed, and anode and cathode for the positive and negative electrodes
www.seyfihoca.com
184
respectively. He also established the laws of the process itself. But most people remember his
name in connection with his practical demonstration of electromagnetic induction.
In France Andre-Marie Ampere (1775-1836) carried out a complete mathematical study of the
laws which govern the interaction between wires carrying electric currents.
In Germany in 1826 a Bavarian schoolmaster Georg Ohm (1789- 1854) had defined the
relationship between electric pressure (voltage), current (flow rate) and resistance in a circuit
(Ohm's law) but 16 years had to elapse before he received recognition for his work.
Scientists were now convinced that since the flow of an electric current in a wire or a coil of
wire caused it to acquire magnetic properties, the opposite might also prove to be true: a
magnet could possibly be used to generate a flow of electricity.
Michael Faraday had worked on this problem for ten years when finally, in 1830, he gave his
famous lecture in which he demonstrated, for the first time in history, the principle of
electromagnetic induction. He had constructed powerful electromagnets consisting of coils of
wire. When he caused the magnetic lines of force surrounding one coil to rise and fall by
interrupting or varying the flow of current, a similar current was induced in a neighbouring
coil closely coupled to the first.
The colossal importance of Faraday's discovery was that it paved the way for the generation
of electricity by mechanical means. However, as can be seen from the drawing, the basic
generator produces an alternating flow of current.(A.C.)
Rotating a coil of wire steadily through a complete revolution in the steady magnetic field
between the north and south poles of a magnet results in an electromotive force (E.M.F.) at its
terminals which rises in value, falls back to zero, reverses in a negative direction, reaches a
peak and again returns to zero. This completes one cycle or sine wave. (1Hz in S.I. units).
In recent years other methods have been developed for generating electrical power in
relatively small quantities for special applications. Semiconductors, which combine heat
insulation with good electrical conduction, are used for thermoelectric generators to power
isolated weather stations, artificial satellites, undersea cables and marker buoys. Specially
developed diode valves are used as thermionic generators with an efficiency, at present, of
only 20% but the heat taken away from the anode is used to raise steam for conventional
power generation.
Sir Humphry Davy (1778-1829) one of Britain's leading chemists of the 18th century, is best
remembered for his safety lamp for miners which cut down the risk of methane gas explosions
in mines. It was Davy who first demonstrated that electricity could be used to produce light.
He connected two carbon rods to a heavy duty storage battery. When he touched the tips of
the rods together a very bright white light was produced. As he drew the rods apart, the arc
light persisted until the tips had burnt away to the critical gap which extinguished the light. As
a researcher and lecturer at the Royal Institution Davy worked closely with Michael Faraday
who first joined the institution as his manservant and later became his secretary. Davy's
crowning honour in the scientific world came in 1820, when he was elected President of the
Royal Society.
www.seyfihoca.com
185
In the U.S.A. the prolific inventor Thomas Alva Edison (1847-1831) who had invented the
incandescent carbon filament bulb, built a number of electricity generators in the vicinity of
the Niagara Falls. These used the power of the falling water to drive hydraulic turbines which
were coupled to the dynamos. These generators were fitted with a spinning switch or
commutator (one of the neatest gadgets Edison ever invented) to make the current flow in
unidirectional pulses (D.C.) In 1876 all electrical equipment was powered by direct current.
Today mains electricity plays a vital part in our everyday lives and its applications are
widespread and staggering in their immensity. But we must not forget that popular demand
for this convenient form of power arose only about 100 years ago, mainly for illumination.
Recent experiments in superconductivity, using ceramic instead metal conductors have given
us an exciting glimpse into what might be achieved for improving efficiency in the
distribution of electric power.
Historians of the future may well characterise the 20th century as 'the century of electricity &
electronics'. But Edison's D.C. generators could not in themselves, have achieved the
spectacular progress that has been made. All over the world we depend totally on a system of
transmitting mains electricity over long distances which was originally created by an amazing
inventor whose scientific discoveries changed, and are still changing, the whole world. His
name was scarcely known to the general public, especially in Europe, where he was born.
Who was this unknown pioneer? Some people reckon that it was this astonishing visionary
who invented wireless, remote control, robotics and a form of X-ray photography using high
frequency radio waves. A patent which he took out in the U.S.A. in 1890 ultimately led to the
design of the humble ignition coil which energises billions and billions of spark plugs in all
the motor cars of the world. His American patents fill a book two inches thick. His name was
Nicola Tesla (1856-1943).
Nicola Tesla was born in a small village in Croatia which at that time formed part of the great
Austro-Hungarian Empire. Today it is a northern province of Yugoslavia, a state created after
the 1914-1918 war. Tesla studied at the Graz Technical University and later in Budapest.
Early in his studies he had the idea that a way had to be found to run electric motors directly
from A.C. generators. His professor in Graz had assured him categorically that this was not
possible. But young Tesla was not convinced. When he went to Budapest he got a job in the
Central Telegraph Office, and one evening in 1882, as he was sitting on a bench in the City
Park he had an inspiration which ultimately led to the solution of the problem.
Tesla remembered a poem by the German poet Goethe about the sun which supports life on
the earth and when the day is over moves on to give life to the other side of the globe. He
picked up a twig and began to scratch a drawing on the soil in front of him. He drew four coils
arranged symmetrically round the circumference of a circle. In the centre he drew a rotor or
armature. As each coil in turn was energised it attracted the rotor towards it and the rotary
motion was established. When he constructed the first practical models he used eight, sixteen
and even more coils. The simple drawing on the ground led to the design of the first induction
motor driven directly by A.C.electricity.
Tesla emigrated to the U.S.A. in 1884. During the first year he filed no less than 30 patents
mostly in relation to the generation and distribution of A.C. mains electricity. He designed
and built his 'A.C.Polyphase System' which generated three-phase alternating current at 25
www.seyfihoca.com
186
Hz. One particular unit delivered 422 amperes at 12,000 volts. The beauty of this system was
that the voltage could be stepped down using transformers for local use, or stepped up to
many thousands of volts for transmission over long distances through relatively thin
conductors. Edison's generating stations were incapable of any such thing.
Tesla signed a lucrative contract with the famous railway engineer George Westinghouse, the
inventor of the Westinghouse Air Brake which is used by most railways all over the world to
the present day. Their generating station was put into service in 1895 and was called the
Niagara Falls Electricity Generating Company. It supplied power for the Westinghouse
network of trains and also for an industrial complex in Buffalo, New York.
After ten years Tesla began to experiment with high frequencies. The Tesla Coil which he had
patented in 1890 was capable of raising voltages to unheard of levels such as 300,000 volts.
Edison, who was still generating D.C., claimed A.C. was dangerous and to prove it contracted
with the government to produce the first electric chair using A.C. for the execution of
murderers condemned to death. When it was first used it was a ghastly flop. The condemned
man moaned and groaned and foamed at the mouth. After four minutes of repeated
application of the A.C.voltage smoke began to come out of his back. It was obvious that the
victim had suffered a horribly drawn-out death.
Tesla said he could prove that A.C. was not dangerous. He gave a demonstration of high
voltage electricity flowing harmlessly over his body. But in reality, he cheated, because he
had used a frequency of 10,000 cycles (10 kHz) at extremely low current and because of the
skin effect suffered no harm.
One of Tesla's patents related to a system of lighting using glass tubes filled with fluorine (not
neon) excited by H.F.voltages. His workshop was lit by this method. Several years before
Wilhelm Roentgen demonstrated his system of X-rays Tesla had been taking photographs of
the bones in his hand and his foot from up to 40 feet away using H.F.currents.
More astonishing still is the fact that in 1893, two years before Marconi demonstrated his
system of wireless signaling, Tesla had built a model boat in which he combined power to
drive it with radio control and robotics. He put the small boat in a lake in Madison Square
Gardens in New York. Standing on the shore with a control box, he invited onlookers to
suggest movements. He was able to make the boat go forwards and backwards and round in
circles. We all know how model cars and aircraft are controlled by radio today, but when
Tesla did it a century ago the motor car had not been invented, and the only method by which
man could cover long distances was on horseback!
Many people believe that a modification of Tesla's 'Magnifying Transmitter' was used by the
Soviet Union when suddenly one day in October 1976 they produced an amazing noise which
blotted out all radio transmissions between 6 and 20 MHz. (The Woodpecker) The B.B.C., the
N.B.C. and most broadcasting and telecommunication organisations of the world complained
to Moscow (the noise had persisted continuously for 10 hours on the first day), but all the
Russians would say in reply was that they were carrying out an experiment. At first nobody
seemed to know what they were doing because it was obviously not intended as another form
of jamming of foreign broadcasts, an old Russian custom as we all know.
It is believed that in the pursuit of his life's ambition to send power through the earth without
the use of wires, Tesla had achieved a small measure of success at E.L.F. (extremely low
www.seyfihoca.com
187
frequencies) of the order of 7 to 12 Hz. These frequencies are at present used by the military
for communicating with submarines submerged in the oceans of the world.
Tesla's career and private life have remained something of a mystery. He lived alone and
shunned public life. He never read any of his papers before academic institutions, though he
was friendly with some journalists who wrote sensational stories about him. They said he was
terrified of microbes and that when he ate out at a restaurant he would ask for a number of
clean napkins to wipe the cutlery and the glasses he drank out of. For the last 20 years of his
life until he died during World War II in 1943 he lived the life of a semi-recluse, with a
pigeon as his only companion. A disastrous fire had destroyed his workshops and many of his
experimental models and all his papers were lost for ever.
Tesla had moved to Colorado Springs where he built his largest ever coil which was 52 feet in
diameter. He studied all the different forms of lightning in his unsuccessful quest for the
transmission of power without wires.
In Yugoslavia, Tesla is a national hero and a well-equipped museum in Belgrade contains
abundant proof of the genius of this extraordinary man.
(From: The dawn of amateur radio in the U.K. and Greece: a personal view by Norman F.
Joly.)
The Discovery of X-rays
Except for a brief description of the Compton effect, and a few other remarks, we have
postponed the discussion of X-rays until the present chapter because it is particularly
convenient to treat X-ray spectra after treating optical spectra. Although this ordering may
have given the reader a distorted impression of the historical importance of X-rays, this
impression will be corrected shortly as we describe the crucial role played by X-rays in the
development of modern physics.
X-rays were discovered in 1895 by Roentgen while studying the phenomena of gaseous
discharge. Using a cathode ray tube with a high voltage of several tens of kilovolts, he noticed
that salts of barium would fluoresce when brought near the tube, although nothing visible was
emitted by the tube. This effect persisted when the tube was wrapped with a layer of black
cardboard. Roentgen soon established that the agency responsible for the fluorescence
originated at the point at which the stream of energetic electrons struck the glass wall of the
tube. Because of its unknown nature, he gave this agency the name X-rays. He found that Xrays could manifest themselves by darkening wrapped photographic plates, discharging
charged electroscopes, as well as by causing fluorescence in a number of different substances.
He also found that X-rays can penetrate considerable thicknesses of materials of low atomic
number, whereas substances of high atomic number are relatively opaque. Roentgen took the
first steps in identifying the nature of X-rays by using a system of slits to show that (1) they
travel in straight lines, and that (2) they are uncharged, because they are not deflected by
electric or magnetic fields.
The discovery of X-rays aroused the interest of all physicists, and many joined in the
investigation of their properties. In 1899 Haga and Wind performed a single slit diffraction
experiment with X-rays which showed that (3) X-rays are a wave motion phenomenon, and,
from the size of the diffraction pattern, their wavelength could be estimated to be 10-8 cm. In
www.seyfihoca.com
188
1906 Barkla proved that (4) the waves are transverse by showing that they can be polarized by
scattering from many materials.
There is, of course, no longer anything unknown about the nature of X-rays. They are
electromagnetic radiation of exactly the same nature as visible light, except that their
wavelength is several orders of magnitude shorter. This conclusion follows from comparing
properties 1 through 4 with the similar properties of visible light, but it was actually
postulated by Thomson several years before all these properties were known. Thomson
argued that X-rays are electromagnetic radiation because such radiation would be expected to
be emitted from the point at which the electrons strike the wall of a cathode ray tube. At this
point, the electrons suffer very violent accelerations in coming to a stop and, according to
classical electromagnetic theory, all accelerated charged particles emit electromagnetic
radiations. We shall see later that this explanation of the production of X-rays is at least
partially correct.
In common with other electromagnetic radiations, X-rays exhibit particle-like aspects as well
as wave-like aspects. The reader will recall that the Compton effect, which is one of the most
convincing demonstrations of the existence of quanta, was originally observed with
electromagnetic radiation in the X-ray region of wavelengths.
POLITICS
CROWDS
On October 30, 1938, thousands of Americans in the New York area were terror stricken by a
radio broadcast describing an invasion from Mars. The presentation was merely a
dramatisation of H. G. Wells' fantastic novel The War of the Worlds, but it was presented
with such stark realism - including reports from fictitious astronomers, governmental officials,
and 'eye-witnesses' - that many listeners fled their homes and their communities. In London
on January 15, 1955, a thick belt of darkness, caused by an accumulation of smoke under an
extremely thick layer of cloud, suddenly wrapped itself around the city in the early afternoon.
It lasted only ten minutes; but during this time, women screamed in the streets, others fell to
their knees on the sidewalks and prayed. Some cried out hysterically that they had gone blind.
A man at Croydon groped through the inky blackness shouting, 'The end of the world has
come'. These two episodes are at once identified as cases of panic, an extreme type of crowd
behaviour, which in turn is a variety of collective behaviour. In the interests of coherence of
treatment and because of limitation of space, this chapter will treat only two related aspects of
collective behaviour: crowds and publics.
Ordinarily when we think of the crowd, we picture a group of individuals massed in one
place; but as the opening illustration of the chapter indicates, physical proximity is not
essential to crowd behaviour, especially in a society like ours with instruments of mass
communication like the newspaper and radio. What is crucial to the understanding of the
crowd is the highly emotional responses of individuals when they are released from the
restraints that usually inhibit extreme behaviour. What releases the customary restraints and
leads to crowd behaviour?
Crowd emotionality is perhaps best interpreted in terms of heightened suggestibility, that is,
the tendency of an individual in a crowd to respond uncritically to the stimuli provided by the
other members. The individual learns to make almost automatic responses to the wishes of
www.seyfihoca.com
189
others, particularly those in authority and those he greatly respects. From infancy on, he is so
dependent upon the judgment of others for direction in his own affairs that he comes to lean
heavily on the opinions of others. Moreover, he learns to value highly the esteem in which
other persons hold him, and consequently he courts their favour by conforming to their ways
and wishes. For these reasons, among others, when he finds himself in a congenial crowd of
persons, all of whom are excited, it is natural that he, too, should be affected.
The effect of suggestion is to produce a partial dislocation of consciousness. When we are
critical about a matter, we give it close attention, and our whole interest is centred upon it. But
when a suggestion is made by someone whom we esteem, our attention is divided, partly on
the issue at hand, partly on the person who made the suggestion. The more awesome the
source of the suggestion, the greater the degree of dissociation and the greater the amount of
automatic behaviour.
If the crowd has a leader who is admired, the effect of the suggestion is still further
heightened. The situation is illustrated by hypnotism, where the effectiveness of the
suggestion depends on the attitude of the subject towards the hypnotist. No one can be
hypnotized against his will; and the best results are obtained where close co-operation exists
between subject and experimenter. The effect of suggestion should help us to understand the
frenzy of a camp meeting led by an evangelist like the late Billy Sunday, or the hysteria of a
Nazi mass meeting led by Hitler.
The group factor also influences crowd behaviour through the security which an individual
feels when he is part of the mass. Individuals are less reluctant to join a lynching party than to
commit murder on their own. The explanation would seem to lie partly in the fact that the
action seems more defensible when carried out by the group and partly in the fact that
individual responsibility is blotted out. The participants remain anonymous and there is no
one upon whom the authorities can pin the offence. This condition, in which the group does
not identify its members as individuals and which therefore has been called 'de-individuation',
leads to reduction of inner restraint and to more expressive behaviour.
(From A handbook of sociology by William F. Ogburn and Meyer J. Nimkoff)
DIPLOMACY
It is thus essential, at the outset of this study, to define what the word 'diplomacy' really
means and in what sense, or senses, it will be used in the pages that follow.
In current language this word 'diplomacy' is carelessly taken to denote several quite different
things. At one moment it is employed as a synonym for 'foreign policy', as when we say
'British diplomacy in the Near East has been lacking in vigour'. At another moment it signifies
'negotiation', as when we say 'the problem is one which might well be solved by diplomacy'.
More specifically, the word denotes the processes and machinery by which such negotiation is
carried out. A fourth meaning is that of a branch of the Foreign Service, as when one says 'my
nephew is working for diplomacy'. And a fifth interpretation which this unfortunate word is
made to carry is that of an abstract quality or gift, which, in its best sense, implies skill in the
conduct of international negotiation; and, in its worst sense, implies the more guileful aspects
of tact.
www.seyfihoca.com
190
These five interpretations are, in English-speaking countries, used indiscriminately, with the
result that there are few branches of politics which have been exposed to such confusion of
thought. If, for instance, the word 'army' were used to mean the exercise of power, the art of
strategy, the science of tactics, the profession of a soldier and the combative instincts of man we should expect public discussion on military matters to lead to much misunderstanding.
The purpose of this monograph is to describe, in simple but precise terms, what diplomacy is
and what it is not. In the first two chapters a short sketch will be given of the origins and
evolution of diplomatic practice and theory. The purpose of this historical review will be to
show that diplomacy is neither the invention nor the pastime of some particular political
system, but is an essential element in any reasonable relation
between man and man and between nation and nation. An examination will follow of recent
modifications in diplomatic methods, with special reference to the problems of 'open' and
'secret' diplomacy and to the difficulty of combining efficient diplomacy with democratic
control. Other sections will deal with the actual functioning of modern diplomacy, with the
relation between diplomacy and commerce, with the organization and administration of the
Foreign Service, with diplomacy by Conference and with the League of Nations as an
instrument of negotiation. At the end a reasoned catalogue will be given of current diplomatic
phrases such as may assist the student in understanding the technical language (it is something
more than mere jargon) which diplomacy has evolved.
Yet before embarking upon so wide a field of examination, it is, as has been said, necessary to
define in what sense, or senses, the word 'diplomacy' will be used in this study. I propose to
employ the definition given by the Oxford English Dictionary. It is as follows:
'Diplomacy is the management of international relations by negotiation; the method by
which these relations are adjusted and managed by ambassadors and envoys; the
business or art of the diplomatist'.
By taking this precise, although wide, definition as my terms of reference I hope to avoid
straying, on the one hand into the sands of foreign policy, and on the other into the marshes of
international law. I shall discuss the several policies or systems of the different nations only in
so far as they affect the methods by which, and the standards according to which, such
policies are carried out. I shall mention international law only in so far as it advances
diplomatic theory or affects the privileges, immunities, and actions of diplomatic envoys. And
I shall thus hope to be able to concentrate upon the 'executive' rather than upon the 'legislative'
aspects of the problem.
(From Diplomacy by Harold Nicholson)
WHAT FUTURE FOR AFRICA?
Even if the most ambitious programme of studies of the lives and contributions of individual
African rulers, statesmen, religious leaders, scholars, could be carried out, we should still be a
long way from achieving a reasonable understanding of African history. The principle that the
history of a people cannot be adequately understood as the history of its dynasties, kings,
paramount chiefs, nobility, and top people generally, is as valid for Africa as it is for Europe.
Until we know a great deal more about the conditions of the people-vassal tribes, slaves, with
their differing degrees of servitude, free commoners, the rank-and-file of the army-in the
www.seyfihoca.com
191
various African states at successive periods, we shall remain very much in the dark about the
real character of these systems and of the causes of the changes which they have undergone.
But until very recently such questions have scarcely been raised-let alone answered. Hence
belle-lettrist generalizations about 'the African past', and speculations about 'the African
future' derived from them, are likely in present circumstances to be valueless.
A further reason for our incompetence is that we do not, for the most part, see the
contemporary African situation from at all the same standpoint as the majority of African
intellectuals. We have come to make use of the term 'decolonization'; to take for granted that
Western European colonial power is being withdrawn, more or less rapidly, from the African
continent; and to be worried only about the possible consequences of this withdrawal. Few
Africans, even those whose political thinking approximates most closely to that of
Westerners, regard the phase of history through which we are now passing in this way. They
are far more conscious of the survival of unlimited, naked colonial power in a number of
states - Algeria, the Union of South Africa, Southern Rhodesia, the Portuguese territories; and
the emergence of 'neo-colonialism' (meaning the gradual granting of formal independence,
combined with the continuing effective présence of the metropolitan power) in others the
states of the French Community, or, under very different conditions, the Congo. For Africans
who think continentally, or globally-and there are many of these-'colonialism' remains a
dominant aspect of the contemporary situation. indeed, they think with much greater subtlety
than we about such matters, and are skilled in distinguishing the varying degrees of intensity
with which 'le poids du colonialisme' presses upon the peoples of the various African states.
And 'colonialism', in almost every context, means Western colonialism: naturally, since
Africans do not feel themselves confronted with the fact, or likelihood, of domination from
any other quarter. They are not at all impressed by the thesis, favoured by American publicists
- 'European colonial systems are now virtually liquidated; therefore let us, the Free World,
protect you against the dangers of Soviet or Chinese colonialism'.
There is a deeper sense in which we are ill-equipped to understand the processes of political
and social change taking place in modern Africa.
We are too remote, historically, from our own revolutions three centuries away in the British,
almost two centuries in the American, case. And when African intellectuals, interested in our
histories as well as theirs, draw attention to resemblances between their political theories and
those of the English Levellers, or Thomas Jefferson, or Robespierre, we seldom see the point.
Americans in particular, are inclined to argue that Africans must either be democrats in their
sense of the term-meaning 'Western values', anti-Communism, two-party systems, free
enterprise, church-on-Sundays, glossy magazines, etcetera - or, if they claim to be democrats
in some other sense, they must be crypto-Communists. But, I have argued elsewhere, there is
a strong vein of Rousseauian revolutionary democracy running through the ideologies of most
African radical nationalist parties and popular organizations.
There are, I think, five respects in which these similarities are most marked. There is, first, the
tendency to take certain ethical presuppositions as the point of departure: the purpose of
liberation is to recover the dignity, or assert the moral worth, of African man. Second, there is
the classic conception of 'democracy', as involving, primarily, the transfer of political power
from a European dominant minority-an oligarchy of colonial officials, or some combination
of the two-to an undifferentiated African demos. Third, there is the idea of the radical mass
party, and its associated organizations, as expressing - as adequately as it can be expressed-the
desires and aspirations of the demos, the common people; and as having therefore the right,
www.seyfihoca.com
192
once independence is achieved, to take control of the state, and reconstruct it in the party's
own image and in relation to its own objectives. There is also the constant insistence upon
human equality: 'il n'y a pas de surhommes' as Sekou Touré once put it. Hence it is not only
the privileged status of the European minority that is attacked, but also the special privileges
and claims to authority-deriving from the precolonial or colonial period-of groups within
contemporary African society, chiefs, marabouts, 'bourgeois intellectuals'. They can play a
part within the new system only to the extent that they accept its egalitarian assumptions.
Finally, there is the idea of the artificiality, undesirability, and impermanence of barriersethnic, linguistic, territorial, cultural, 'racial', religious - within humanity, 'Pan-Africanism', in
one of its aspects, being an attempt to apply this idea in the modern African context.
(Thomas Hodgkin from an article in Encounter, June 1961)
NATIONALISM
Language alone is not, of course, enough to explain the rise of modern nationalism. Even
language is a shorthand for the sense of belonging together, of sharing the same memories, the
same historical experience, the same cultural and imaginative heritage. When in the
eighteenth century, nationalism began to take form as a modern movement, its forerunners in
many parts of Europe were not soldiers and statesmen but scholars and poets who sought to
find in ancient legends and half forgotten folksongs the 'soul' of the nation. But it was
language that enshrined the memories, the common experience and the historical record.
Nor could the sense of common tongue and culture have become the political battering ram
that it is in our own times, if it had not been inextricably bound up with the modern political
and economic revolutions of the West-the political drive for democracy and the economic
revolution of science and industry.
Three thousand years ago, in one small area of the Mediterranean world, there came a break
with the earlier traditions of state building which were all despotic. The dominance of a strong
local tribe or conquest by foreign groups had turned the old fragmented tribal societies into
centralized dynastic and imperial states in which the inhabitant was subject absolutely to the
ruler's will. But in the Greek City State, for the first time, the idea was formulated that a man
should govern himself under law, and that he should not be a subject but a free citizen.
After the post-Roman collapse, it re-emerged as a seminal idea in the development of later
European history. Even in the Middle Ages, before there were any fully articulated
democratic systems, two or three of the essential foundations of democracy had appeared. The
rule of law was recognized. The right of the subject to be consulted had called into being the
parliaments and 'estates' of the fourteenth century. And the possibility of a plurality of power through State, through Church, through royal boroughs and free municipalities - mitigated the
centralizing tendencies of government. It was in fact for a restoration of these rights after the
Tudor interregnum that the first modern political revolution, the English Civil War, was
fought.
But if a man had a right to take part in his own government, it followed logically that his
government could not be arbitrarily controlled from elsewhere. It was useless to give him
representation if it did not affect the true centre of power. The American Revolution
symbolized the connexion between the rights of the citizen and the rights of the state. The free
citizen had a right to govern himself, ergo the whole community of free citizens had a right to
www.seyfihoca.com
193
govern itself. This was not yet modern nationalism. The American people did not see
themselves as a national group but as a community of free men 'dedicated to a proposition'.
But within two decades, the identification had been made.
The French Revolution, proclaiming the Rights of Man, formed the new style of nation. The
levee-en-masse which defeated the old dynastic armies of Europe was the first expression of
total national unity as the basis of the sovereign state. Men and nations had equally the right
to self-determination. Men could not be free if their national community was not.
The same revolution quickly proved that the reverse might not be true. The nation could
become completely unfettered in its dealings with other states while enslaving its own
citizens. In fact, over-glorification of the nation might lead inevitably to the extinction of
individual rights. The citizen could become just a tool of the national will, of the so-called
'general will'. But in the first explosion of revolutionary ardour, the idea of the Rights of Man
and of the Rights of the Nation went together. And, formally, that is where they have
remained. At the end of the First World War, it was the world's leading democratic statesman,
President Woodrow Wilson, who wrote the right of self-determination, the right of national
groups to form their own sovereign government, into the Peace Treaties and at no time in
human history have so many independent national states been formed as after the Second
World War.
From all this it will be clear that the development of nationalism is a recognizable, historical
process. It happened in certain countries, it happened in a certain way, and it created a certain
mood which became embodied in the national idea. But with the means of communication
open to the modern world, an idea developed in one place can quickly become the possession
of all mankind. What is certain is that, in the twentieth century, nationalism, the historical
product of certain political institutions, geographical facts, and economic developments in
Western Europe, has swept round the world to become the greatest lever of change in our day.
We see it at the United Nations where, in the course of eleven short years, the number of
sovereign states based upon the principle of nationhood has grown by three score and more.
As I have pointed out, it is unprecedented in human history that such a number of separate,
autonomous, sovereign institutions should come into being in so short a space of time.
(From Five Ideas that Change the World by Barbara Ward)
DEMOCRACY
In the last days of 1945 I received a letter from Gothenburg, from a gathering of progressive
Swedish students, begging to be advised by me as to what was the task of intellectuals in the
present condition of the world. Their letter was dated July 26, the day of my defeat at
Berwick, and I sent them an answer from the heart: 'The task of intellectuals in Sweden, as
elsewhere, is to introduce reason and foresight into practical affairs. Only if it is governed by
reason will democracy be sufficiently successful in practical affairs, to make certain of
survival'.
Democracy is better than despotism, offers the only hope for mankind of freedom, of justice,
and of peace. But is democracy, as we know it, good enough? A general election in any of the
larger democracies today, in the United States or in Britain, in France or in Italy, is not
www.seyfihoca.com
194
conspicuously a feast of reason. If democracy is not all that nineteenth-century fancy used to
paint, how should it be made better? Can it be made to do well enough to be sure of survival?
In an Epilogue to another story I can only ask these questions. I cannot attempt the answers.
But, having regard to Britain's internal revolutions since the beginning of my story, it may be
worthwhile to name some of the problems illustrated by the story, and facing us today. We
have to learn as a democracy to choose our governors wisely, by reason, not greed. We have
in an economically flattened society to find men who will undertake office in a public spirit,
not for personal gain or glory; we must carry on the aristocratic tradition without the
aristocrats. We have to keep open channels for new ideas of unknown men to reach and
influence the temporary holders of power. We seem to have solved for the present the
problems of full employment, but we have not solved two of the problems to which full
employment in a free society gives rise-how to preserve the value of our money against
endless rise of costs, wages and prices, and how without fear of unemployment to secure the
maximum of output.
Democracy must be efficient in practical affairs, as efficient as the nearest despotism.
Democracy must be democratic in substance, not only in form. This means that the process of
choosing and changing holders of power shall be unaffected by privilege of established
organization and wealth, that the holders of political power, when an election comes, shall
compete with their opponents on equal terms. Power must not be used to prolong itself.
Power, the stupid necessary mule, should have neither pride of ancestry nor hope of posterity.
In the leading democracies of today many special measures have been taken to secure this.
But, at any risk of causing offence, a question must be asked about Britain. Is it consistent
with democratic principle that organizations like the trade unions which have received special
privileges for industrial work should become tied to a political party? Ought it to be difficult
for an individual to earn his living by employment without contributing from his wages to the
retention of power by one set of politicians rather than another? A one-party State in any form
is the destruction of freedom.
Democracies need to look within. They must look without as well. They must, in one way or
another, abandon and lead others to abandon any claim to absolute sovereignty-the claim to
kill in one's own cause without selection or limit. The headnote of this Epilogue is not a
paradox but a truism. If with our growing control over nature we could abolish war, we
should be in Utopia. If we cannot abolish war, we shall plunge ever deeper into a hell of evil
imagining and evil doing.
The theme of my story returns at its end. Power as a means of getting things done appeals to
that which men share with brutes; to fear and to greed; power leads those who wield it to
desire it for its own sake, not for the service it may render, and to seek its continuance in their
own hands. Influence as a means of getting things done appeals to that which distinguishes
men from brutes. The way out of the world's troubles today is to treat men as men, to enthrone
influence over power, and to make power revocable.
(From Power and influence: An autobiography by Lord Beveridge)
LOCKE'S POLITICAL THEORY
www.seyfihoca.com
195
Locke's political theory is to be found in his Two Treatises of Civil Government, particularly
in the second of these. The immediate aim of that treatise is apparent: to justify the Revolution
of 1688 and to help 'establish the throne of our great restorer, our present King William'. But
this aim is achieved by securing in turn a great and fundamental political principle, true for
the English nation in 1688 and true, in Locke's opinion, for all well regulated communities
everywhere and at all times, that government must be with the consent of the governed, that a
ruler who has lost the confidence of his people no longer has the right to govern them.
The principle involves a particular view of government and of political community. Locke set
himself to refute two theories which were used to justify privilege, oppression, and political
slavery. The first was the theory of the divine right of kings as put forward by Robert Filmer,
that the king is the divinely ordained father of his people, and that the relation between king
and subjects is precisely the same as that between father and child. Locke ridicules the
comparison. In the modern state, a large, highly complex organization, parental or patriarchal
government is no longer possible, and the claim that it is divinely ordained cannot be
substantiated. The second theory is to be found in its most explicit form in the works of
Hobbes, although Locke does not refer to Hobbes by name, at least in the Treatise.
Government, in this theory, necessarily involves the complete subjection of the governed to
the absolute will of the governor, for without such subjection no civil society is possible.
Locke denies this theory categorically. The facts of human experience are against it and
reason is against it. A political community is possible in which the governor is limited; in
which sovereignty ultimately pertains not to the monarch, as opposed to those whom he
governs, but to the people as a whole. Government becomes an instrument for securing the
lives, property, and well-being of the governed, and this without enslaving the governed in
any way. Government is not their master; it is created by the people voluntarily and
maintained by them to secure their own good. Those who, because of their superior talent,
have been set to rule by the community, rule not as masters over slaves, or even as fathers
over children. They are officers elected by the people to carry out certain tasks. Their powers
are to be used in accordance with 'that trust which is put into their hands by their brethren'.
For Locke government is a 'trust' and a political community is an organization of equals, of
'brothers', into which men enter voluntarily in order to achieve what they cannot achieve
apart.
Such was the view of government which Locke adopted, and the second treatise is an effort to
discover a rational justification of this view. Locke might have appealed to experience and to
history, or again he might have contented himself with showing the public utility of the theory
he advocated. But the late seventeenth century was rationalist and would listen to no
arguments other than rationalist ones, and so Locke analysed the notion of political society in
order to prove rationally that it was from the first a community of free individuals and that it
remained so throughout. He spoke in the language of his day and he made use of the theories
of his day. In particular, he borrowed two concepts from earlier political theories, the law of
nature and the social contract.
(From John Locke by Richard I. Aaron)
THE SEARCH FOR WORLD ORDER
Before I leave the subject of disarmament there is one further point of importance. Some
Western writers, and some people in Russia too, argue that the best way to minimize the
explosive quality of the present arms race is somehow to develop a stable balance of terror or
www.seyfihoca.com
196
deterrence. This means developing nuclear weapons and delivery systems so strong and so
varied that no surprise attack could knock out the power to retaliate.
I can see some force in this argument. Effective deterrence depends to some extent on mutual
conviction that the other man can and will do what he threatens if he is attacked. And it may
be that this is, for the time being, the only practical way of curbing hasty action. But, in fact,
attempting to produce stability in this way also means continuing the arms race. Because, as
the power to retaliate increases, there is bound to be a corresponding search for improved
weapons which will increase the element of surprise. In any case, inaction through fear, which
is the basis of deterrence, is not a positive way to secure peace-at any rate in the long run. I
feel bound to doubt whether safety, as Winston Churchill once claimed, can really become the
'sturdy child of terror'.
It is important to remember that neither the League of Nations nor, so far, the United Nations
has contemplated the abolition of all armaments. The Covenant of the League spoke of: 'the
reduction of national armaments to the lowest point consistent with national safety and the
enforcement by common action of international obligations'. The first article of the Charter of
the United Nations charges the Organization with the duty of suppressing 'acts of aggression
and other breaches of the peace', and Article 51 allows the organization to use force for this
purpose. Indeed, right at the beginning a Military Staffs Committee was set up at United
Nations headquarters and charged with the strategic direction of whatever military forces
were to be made available to the Security Council.
In practice, however, the United Nations does not have any military force permanently at its
disposal or any staff to plan operations in advance and direct them when they become
necessary. Whatever operations the Organization has undertaken have been done on an
entirely ad hoc and improvised basis. In fact, in 1958 Mr. Hammarskjold himself argued
against the creation of a permanent United Nations military force. One of the main reasons for
this failure to develop a United Nations peace-keeping capacity in terms of military forces has
undoubtedly been the opposition of some of the Great Powers. And it must be admitted that
there is no prospect of the United Nations coercing the Great Powers into keeping the peace at
present. But perhaps we can make a virtue of necessity here.
I have tried to suggest that international agreements, like any system of municipal law,
demand a sanction of force if observance is normally to be guaranteed and non-observance
controlled before it explodes into general disorder. In other words, legislative decision
demands as its corollary some form of executive action. It was surely this which Mr.
Hammarskjold had in mind in presenting his last annual report as Secretary General. Some
people, he said, wanted the United Nations to work simply as a conference system producing
reconciliation by discussion. Others - and clearly himself among them-looked upon the
Organization primarily as a dynamic instrument of government through which they, jointly
and for the same purpose, should seek such reconciliation but through which they should also
try to develop forms of executive action undertaken on behalf of all members, aiming at
forestalling conflicts and resolving them, once they have arisen, by appropriate diplomatic or
political means. The word 'military' was not used. But at that very moment, the United
Nations had in the Congo, and largely through Mr. Hammarskjold's efforts, a military force
expressly designed to re-establish order and to prevent civil strife from exploding into general
war.
www.seyfihoca.com
197
It seems to me that any international organization designed to keep the peace must have the
power not merely to talk but also to act. Indeed I see this as the central theme of any progress
towards an international community in which war is avoided not by chance but by design. Nor
need our present limitations daunt us. This is a slow process in which experience grows into
habit, and habit into trust. Many people have already suggested how this development could
be encouraged. The United Nations could have a bigger permanent staff to act as observers
and intelligence officers in potential trouble spots. Here would be part of the political basis of
control. It could develop much more detailed methods in advance for drawing on national
armed forces when police action becomes inevitable, even without possessing a big military
establishment of its own. It could prepare training manuals for the police action its forces are
likely to undertake, and for which the ordinary soldier is not normally trained. And it could
begin to hold under its own control small specialist staffs, for example, multilingual
signallers, and some small stocks of equipment such as transport aircraft, which its operations
almost inevitably demand.
The fact that coercion of the Great Powers is impossible does not invalidate any of these
suggestions. If these Powers can, for the time being, avoid major war among themselves by
nuclear deterrence, then the likeliest explosive situations will occur in areas not of vital
interest to them. It is there that the United Nations can experiment and develop. Nor can a
firm beginning be made otherwise. At present the United Nations Organization, in the words
of a recent writer, 'is not ... the parliament and government of mankind but an institution of
international diplomacy'. It can only hope to grow from the one into the other by admitting its
present limitations and, more than that, by beginning to practise its own terms and conditions.
If a start could be made now-and even if only in miniature--international government might
finally emerge.
(Norman Gibbs from an article in The Listener, December 28th, 1961)
THE DECLARATION OF INDEPENDENCE
In Congress, July 4, 1776
THE UNANIMOUS DECLARATION OF THE THIRTEEN UNITED STATES OF
AMERICA
When in the Course of human Events, it becomes necessary for one People to dissolve the
Political Bands which have connected them with another, and to assume among the Powers of
the Earth, the separate and equal Station to which the Laws of Nature and of Nature's God
entitle them, a decent Respect to the Opinions of Mankind requires that they should declare
the causes which impel them to the Separation.
We hold these Truths to be self-evident, that all Men are created equal, that they are endowed
by their Creator with certain unalienable Rights, that among these are Life, Liberty and the
Pursuit of Happiness -- That to secure these Rights, Governments are instituted among Men,
deriving their just Powers from the Consent of the Governed, that Whenever any Form of
Government becomes destructive of these Ends, it is the Right of the People to alter or to
abolish it, and to institute new Government, laying its Foundation on such Principles, and
organizing its Powers in such Form, as to them shall seem most likely to effect their Safety
and Happiness. Prudence, indeed, will dictate that Governments long established should not
be changed for light and transient Causes; and accordingly all Experience hath shewn, that
www.seyfihoca.com
198
Mankind are more disposed to suffer, while Evils are sufferable, than to right themselves by
abolishing the Forms to which they are accustomed. But When a long Train of Abuses and
Usurpations, pursuing invariably the same Object, evinces a Design to reduce them under
absolute Despotism, it is their Right, it is their Duty, to throw off such Government, and to
provide new Guards for their future Security. Such has been the patient Sufferance of these
Colonies; and such is now the Necessity which constrains them to alter their former Systems
of Government. The History of the present King of Great- Britain is a History of repeated
Injuries and Usurpations, all having in direct Object the Establishment of an absolute Tyranny
over these States. To prove this, let Facts be submitted to a candid World.
He has refused his Assent to Laws, the most wholesome and necessary for the public Good.
He has forbidden his Governors to pass Laws of immediate and pressing Importance, unless
suspended in their Operation till his Assent should be obtained; and When so suspended, he
has utterly neglected to attend to them.
He has refused to pass other Laws for the Accommodation of large Districts of People, unless
those People would relinquish the Right of Representation in the Legislature, a Right
inestimable to them, and formidable to Tyrants only.
He has called together Legislative Bodies at Places unusual, uncomfortable, and distant from
the Depository of their public Records, for the sole Purpose of fatiguing them into
Compliance with his Measures.
He has dissolved Representative Houses repeatedly, for opposing with manly Firmness his
Invasions on the Rights of the People.
He has refused for a long Time, after such Dissolutions, to cause others to be elected;
whereby the Legislative Powers, incapable of the Annihilation, have returned to the People at
large for their exercise; the State remaining in the mean time exposed to all the Dangers of
Invasion from without, and the Convulsions within.
He has endeavoured to prevent the Population of these States; for that Purpose obstructing the
Laws for Naturalization of Foreigners; refusing to pass others to encourage their Migrations
hither, and raising the Conditions of new Appropriations of Lands.
He has obstructed the Administration of Justice, by refusing his Assent to Laws for
establishing Judiciary Powers.
He has made Judges dependent on his Will alone, for the Tenure of their Offices, and the
Amount and Payment of their Salaries.
He has erected a Multitude of new Offices, and sent hither Swarms of Officers to harrass our
People, and eat out their Substance.
He has kept among us, in Times of Peace, Standing Armies, without the consent of our
Legislatures.
He has affected to render the Military independent of and superior to the Civil Power.
www.seyfihoca.com
199
He has combined with others to subject us to a Jurisdiction foreign to our Constitution, and
unacknowledged by our Laws; giving his Assent to their Acts of pretended Legislation:
For quartering large Bodies of Armed Troops among us;
For protecting them, by a mock Trial, from Punishment for any Murders which they should
commit on the Inhabitants of these States:
For cutting off our Trade with all Parts of the World:
For imposing Taxes on us without our Consent:
For depriving us, in many Cases, of the Benefits of Trial by Jury:
For transporting us beyond Seas to be tried for pretended Offences:
For abolishing the free System of English Laws in a neighbouring Province, establishing
therein an arbitrary Government, and enlarging its Boundaries, so as to render it at once an
Example and fit Instrument for introducing the same absolute Rules into these Colonies:
For taking away our Charters, abolishing our most valuable Laws, and altering fundamentally
the Forms of our Governments:
For suspending our own Legislatures, and declaring themselves invested with Power to
legislate for us in all Cases whatsoever.
He has abdicated Government here, by declaring us out of his Protection and waging War
against us.
He has plundered our Seas, ravaged our Coasts, burnt our Towns, and destroyed the Lives of
our People.
He is, at this Time, transporting large Armies of foreign Mercenaries to compleat the Works
of Death, Desolation, and Tyranny, already begun with circumstances of Cruelty and Perfidy,
scarcely paralleled in the most barbarous Ages, and totally unworthy the Head of a civilized
Nation.
He has constrained our fellow Citizens taken Captive on the high Seas to bear Arms against
their Country, to become the Executioners of their Friends and Brethren, or to fall themselves
by their Hands.
He has excited domestic Insurrections amongst us, and has endeavoured to bring on the
Inhabitants of our Frontiers, the merciless Indian Savages, whose known Rule of Warfare, is
an undistinguished Destruction, of all Ages, Sexes and Conditions.
In every stage of these Oppressions we have Petitioned for Redress in the most humble
Terms: Our repeated Petitions have been answered only by repeated Injury. A Prince, whose
Character is thus marked by every act which may define a Tyrant, is unfit to be the Ruler of a
free People.
www.seyfihoca.com
200
Nor have we been wanting in Attentions to our British Brethren. We have warned them from
Time to Time of Attempts by their Legislature to extend an unwarrantable Jurisdiction over
us. We have reminded them of the Circumstances of our Emigration and Settlement here. We
have appealed to their native Justice and Magnanimity, and we have conjured them by the
Ties of our common Kindred to disavow these Usurpations, which, would inevitably interrupt
our Connections and Correspondence. They too have been deaf to the Voice of Justice and of
Consanguinity. We must, therefore, acquiesce in the Necessity, which denounces our
Separation, and hold them, as we hold the rest of Mankind, Enemies in War, in Peace,
Friends.
We, therefore, the Representatives of the UNITED STATES OF AMERICA, in GENERAL
CONGRESS, Assembled, appealing to the Supreme Judge of the World for the Rectitude of
our Intentions, do, in the Name, and by Authority of the good People of these Colonies,
solemnly Publish and Declare, That these United Colonies are, and of Right ought to be,
FREE AND INDEPENDENT STATES; that they are absolved from all Allegiance to the
British Crown, and that all political Connection between them and the State of Great-Britain,
is and ought to be totally dissolved; and that as FREE AND INDEPENDENT STATES, they
have full Power to levy War, conclude Peace, contract Alliances, establish Commerce, and to
do all other Acts and Things which INDEPENDENT STATES may of right do. And for the
support of this Declaration, with a firm Reliance on the Protection of divine Providence, we
mutually pledge to each other our Lives, our Fortunes, and our sacred Honor.
DECLARATION OF THE RIGHTS OF MAN AND OF THE CITIZEN
Approved by the National Assembly of France, August 26, 1789
The representatives of the French people, organized as a National Assembly, believing that
the ignorance, neglect, or contempt of the rights of man are the sole cause of public calamities
and of the corruption of governments, have determined to set forth in a solemn declaration the
natural, unalienable, and sacred rights of man, in order that this declaration, being constantly
before all the members of the Social body, shall remind them continually of their rights and
duties; in order that the acts of the legislative power, as well as those of the executive power,
may be compared at any moment with the objects and purposes of all political institutions and
may thus be more respected, and, lastly, in order that the grievances of the citizens, based
hereafter upon simple and incontestable principles, shall tend to the maintenance of the
constitution and redound to the happiness of all. Therefore the National Assembly recognizes
and proclaims, in the presence and under the auspices of the Supreme Being, the following
rights of man and of the citizen:
Articles:
1. Men are born and remain free and equal in rights. Social distinctions may be
founded only upon the general good.
2. The aim of all political association is the preservation of the natural and
imprescriptible rights of man. These rights are liberty, property, security, and
resistance to oppression.
3. The principle of all sovereignty resides essentially in the nation. No body nor
individual may exercise any authority which does not proceed directly from the
nation.
www.seyfihoca.com
201
4. Liberty consists in the freedom to do everything which injures no one else; hence
the exercise of the natural rights of each man has no limits except those which
assure to the other members of the society the enjoyment of the same rights. These
limits can only be determined by law.
5. Law can only prohibit such actions as are hurtful to society. Nothing may be
prevented which is not forbidden by law, and no one may be forced to do anything
not provided for by law.
6. Law is the expression of the general will. Every citizen has a right to participate
personally, or through his representative, in its foundation. It must be the same for
all, whether it protects or punishes. All citizens, being equal in the eyes of the law,
are equally eligible to all dignities and to all public positions and occupations,
according to their abilities, and without distinction except that of their virtues and
talents.
7. No person shall be accused, arrested, or imprisoned except in the cases and
according to the forms prescribed by law. Any one soliciting, transmitting,
executing, or causing to be executed, any arbitrary order, shall be punished. But any
citizen summoned or arrested in virtue of the law shall submit without delay, as
resistance constitutes an offense.
8. The law shall provide for such punishments only as are strictly and obviously
necessary, and no one shall suffer punishment except it be legally inflicted in virtue
of a law passed and promulgated before the commission of the offense.
9. As all persons are held innocent until they shall have been declared guilty, if arrest
shall be deemed indispensable, all harshness not essential to the securing of the
prisoner's person shall be severely repressed by law.
10. No one shall be disquieted on account of his opinions, including his religious
views, provided their manifestation does not disturb the public order established by
law.
11. The free communication of ideas and opinions is one of the most precious of the
rights of man. Every citizen may, accordingly, speak, write, and print with freedom,
but shall be responsible for such abuses of this freedom as shall be defined by law.
12. The security of the rights of man and of the citizen requires public military forces.
These forces are, therefore, established for the good of all and not for the personal
advantage of those to whom they shall be intrusted.
13. A common contribution is essential for the maintenance of the public forces and for
the cost of administration. This should be equitably distributed among all the
citizens in proportion to their means.
14. All the citizens have a right to decide, either personally or by their representatives,
as to the necessity of the public contribution; to grant this freely; to know to what
uses it is put; and to fix the proportion, the mode of assessment and of collection
and the duration of the taxes.
15. Society has the right to require of every public agent an account of his
administration.
16. A society in which the observance of the law is not assured, nor the separation of
powers defined, has no constitution at all.
17. Since property is an inviolable and sacred right, no one shall be deprived thereof
except where public necessity, legally determined, shall clearly demand it, and then
only on condition that the owner shall have been previously and equitably
indemnified.
www.seyfihoca.com
202
Written by The Marquis de Lafayette, with help from his friend and neighbor, American
envoy to France, Thomas Jefferson.
French King Louis XVI signed this document, under duress, but never intended to support it.
PSYCHOLOGY
Society and Intelligence
Intelligence tests have been constructed of three kinds. Verbal paper-and-pencil tests, nonverbal paper-and-pencil tests, where the tasks are presented by means of pictures and
diagrams, and performance tests which require the manipulation of objects. Some, such as the
Binet test and the performance tests, are given to subjects separately; most verbal and nonverbal tests can be done by a group of subjects writing at the same time.
The subjects are told to do their tasks within a certain time, their results are marked, and the
result of each is compared with a scale indicating what may be expected of children of the
same age, i.e. what marks are expected of the relatively few bright ones, what marks are
expected of the few dull ones, and what marks are expected of the bulk of the population with
whom the comparison is being made. This 'calibration' of the test has been made beforehand
and we are not concerned with the methods employed. One thing, however, we have to notice,
and that is that the assessment of the intelligence of any subject is essentially a comparative
affair.
The results of assessment are expressed in various ways, the most familiar being in terms of
what is called the Intelligence Quotient. For our purposes we need not consider how this has
been devised, it is enough to say that an I.Q. round about 100 is 'average', while more than
105 or less than 95 are above or below the average respectively.
Now since the assessment of intelligence is a comparative matter we must be sure that the
scale with which we are comparing our subjects provides a 'valid' or 'fair' comparison. It is
here that some of the difficulties, which interest us, begin. Any test performed involves at
least three factors: the intention to do one's best, the knowledge required for understanding
what you have to do, and the intellectual ability to do it. The first two must be held equal for
all who are being compared, if any comparison in terms of intelligence is to be made. In
school populations in our culture these assumptions can be made with fair plausibility, and the
value of intelligence testing has been proved up to the hilt. Its value lies, of course, in its
providing a satisfactory basis for prediction. No one is in the least interested in the marks little
Basil gets on his test, what we are interested in is whether we can infer from his mark on the
test that Basil will do better or worse than other children of his age at other tasks which we
think require 'general intelligence'. On the whole such inference can be made with a certain
degree of confidence, but only if Basil can be assumed to have had the same attitude towards
the test as the others with whom he is being compared, and only if he was not penalized by
lack of relevant information which they possessed.
It is precisely here that the trouble begins when we use our tests for people from different
cultures. If, as happens among the Dakota Indians, it is indelicate to ask a question if you
think there is someone present who does not know the answer already, this means that a
Dakota child's test result is not comparable with the results of children brought up in a less
sensitive environment. Porteous found difficulty among the Australian aborigines. They were
www.seyfihoca.com
203
brought up to believe that all problems had to be discussed in the group, and they thought it
very eccentric to be expected to solve one by oneself.
Supposing, however, a satisfactory attitude towards the test can be assumed, what about
equality in relevant knowledge? In a society where children play with bricks, performance
tests involving the manipulation of little cubes present an easier problem than they would in a
society where such toys were unknown. Bartlett reports that a group of East African natives
were unable to arrange coloured pegs in an alternating series, but they planted trees according
to the same plan in everyday life.
Then there is the story of the little boy in Kentucky who was asked a test question: 'If you
went to a store and bought six cents worth of candy and gave the clerk ten cents what change
would you receive?' The boy replied: 'I never had ten cents and if I had I wouldn't spend it on
candy and anyway candy is what mother makes.' The tester reformulated the question: 'If you
had taken ten cows to pasture for your father and six of them strayed away, how many would
you have left to drive home?' The boy replied: 'We don't have ten cows, but if we did and I
lost six I wouldn't dare go home.' Undeterred the tester pressed his question: 'If there were ten
children in your school and six of them were sick with the measles how many would there be
in school?' The answer came: 'None, because the rest would be afraid of catching it too.'
Thus all intercultural comparisons of intelligence are vitiated by the lack of true
comparability, and any generalizations about 'racial' differences in intellectual competence
which do not take account of this are worthless. So are many comparisons which have been
made between children of different social classes.
(From Social Psychology, by W. J. H. Sprott.)
The Pressure to Conform
Suppose that you saw somebody being shown a pair of cards. On one of them there is a line,
and on the other three lines. Of these three, one is obviously longer than the line on the other
card, one is shorter, and one the same length. The person to whom these cards are being
shown is asked to point to the line on the second card which is the same length as the one on
the first. To your surprise, he makes one of the obviously wrong choices. You might suppose
that he, or she, perhaps suffers from distorted vision, or is insane, or perhaps merely cussed.
But you might be wrong in all these suggestions; you might be observing a sane, ordinary
citizen, just like yourself. Because, by fairly simple processes, sane and ordinary citizens can
be induced to deny the plain evidence of their senses-not always, but often. In recent years
psychologists have carried out some exceedingly interesting experiments in which this sort of
thing is done.
The general procedure was first devised by Dr Asch in the United States. What happens is
this: Someone is asked to join a group who are helping to study the discrimination of length.
The victim, having agreed to this seemingly innocent request, goes to a room where a number
of people-about half a dozen-and the experimenter are seated. Unbeknown to our victim, none
of the other people in the room is a volunteer like himself; they are all in league with the
experimenter. A pair of cards, like those I have described, is produced; and everyone in turn is
asked which of the three lines on the second card is equal to the line on the first. They all,
without hesitation, pick-as they have been told to pick-the same wrong line. Last of all comes
www.seyfihoca.com
204
the turn of our volunteer. In many cases the volunteer, faced with this unanimity, denies the
plain evidence of his senses, and agrees.
An ingenious variation of this experiment was devised by Stanley Milgram of Harvard. He
used sounds instead of lines, and the subjects were merely asked to state which of two
successive sounds lasted longer. The volunteer would come into a room where there was a
row of five cubicles with their doors shut, and coats hanging outside, and one open cubicle for
him. He would sit in it and don earphones provided. He would hear the occupants of the other
cubicles tested in turn, and each would give a wrong answer. But the other cubicles were, in
fact, empty, and what he heard were tape-recordings manipulated by the experimenter.
Milgram conducted a whole series of experiments in this way, in which he varied
considerably the pressure put upon the subjects. As expected, their conformity varied with the
pressure, but, over all, he clearly showed that, faced with the unanimous opinion of the group
they were in, people could be made to deny the obvious facts of the case in anything up to 75
per cent of the trials.
The victim of brainwashing can be induced to assert falsehoods, as we well know. But he is
subjected to terrible and continuous stress: to hunger, sleeplessness, cold, and fear. The people
we have been discussing were free of all these things, and subjected to nothing more than the
complete agreement of the group in which they found themselves. Nevertheless, they too
could be made to assert manifest falsehoods. I find this more than a trifle alarming -and very
thought-provoking.
You may reply that there is no cause for alarm, because in real situations the total unanimity
of a group is rare. The more usual case concerns the effects of what we might call a 'pressure
group'. This has been examined, at least partially, by W. M. and H. H. Kassarjian, in
California. They used the 'group in a room' and 'lines on cards' situation; and I imagine that
they must be kindly people, because they made things much easier for their volunteers. In the
first place, the genuine volunteers were in a majority: twenty out of thirty. Secondly, the
volunteers never had to make their selections out loud, but always enjoyed the anonymity of
paper and pencil. The experimenter explained that some people would be asked to declare
their choices publicly, and asked only his primed collaborators. Thus each volunteer heard the
views of only a third of the group he was in. This third was unanimous, and the volunteers
probably concluded that they expressed a majority view, but they were not put in a glaring
minority of one, and their choice was secret. Nevertheless, a substantial distortion was still
produced: almost, though not quite, as large as in the harsher situations we looked at first. So
there is only small comfort here.
I am aware that there is grave danger in taking results obtained in the special and carefully
simplified situation of the laboratory, or of the clinic, and applying them directly to the
immensely complicated affairs of normal life. But these results seem to me so interesting and
so suggestive that, in spite of the obvious risks, it may be worth while to see where they may
shed a little light. In speculating thus, I am stepping outside the proper bounds of scientific
rigour; so, if I only make myself a laughing-stock, it is my own fault.
Whether one line is or is not the same length as another is a matter fairly easy to judge as a
rule. But many things-and many more important things-are by no means so clear cut. If we are
asked which of two cars or two schools is the better, or which of two 'pop' songs or two
politicians is the worse, we may be genuinely perplexed to answer. We may guess that in such
doubtful cases the 'majority effect' or the 'pressure group effect' would be even more
www.seyfihoca.com
205
pronounced. Recent experiments by A. E. M. Seaborne suggest that they would not always be,
but it seems generally likely. Can we observe such effects taking place around us now, or
having taken place in the past? I think we can, and that they help us a little to understand the
massive inertia of commonly held ideas, and the fantastic standing of some more ephemeral
ones.
(From an article by Max Hammerton in The Listener, October 18th, 1962.)
Learning to Live With the Computer
A rapid pace of technological advance has been accepted by many manufacturing industries
for some time now, but for the office worker, who has led a sheltered existence in
comparison, radical changes are a new experience. With the advent of electronic data
processing techniques and, especially, computers, this situation has altered very swiftly.
Office staff are suddenly finding themselves exposed to the traumatic consequences of
scientific progress.
Most offices, by the very nature of their structure and function, are geared to stability or slow
change. Accelerated change of the kind that a computer brings is likely to prove disrupting
and disturbing. This is because people in stable organizations tend to expect a steady
continuation of existing arrangements, and because departments unaccustomed to change
frequently find they have become too inflexible to assimilate it without stress. Social as well
as technical factors are therefore highly relevant to a successful adaptation to new techniques.
Research into the social and organizational problems of introducing computers into offices
has been in progress in the social science department in Liverpool University for the past four
years. My colleagues and I have shown that many firms get into difficulties with their new
computers partly because of lack of technical knowledge and experience, but also because
they have not been sufficiently aware of the need to understand and plan for the social as well
as the technical implications of change. In the firms we have been studying, change has
usually been seen simply as a technical problem to be handled by technologists. The fact that
the staff might regard the introduction of a computer as a threat to their security and status has
not been anticipated. Company directors have been surprised when, instead of cooperation,
they encountered anxiety and hostility.
Once the firm has signed the contract to purchase a computer, its next step, one might expect,
would be to 'sell' the idea to its staff, by giving reassurances about redundancy, and
investigating how individual jobs will be affected so that displaced staff can be prepared for a
move elsewhere. In fact, this may not happen. It is more usual for the firm to spend much time
and energy investigating the technical aspects of the computer, yet largely to ignore the
possibility of personnel difficulties. This neglect is due to the absence from most firms of
anyone knowledgeable about human relations. The personnel manager, who might be
expected to have some understanding of employee motivation, is in many cases not even
involved in the changeover.
Again, because the changeover is seen only as a technical problem, little thought is given to
communication and consultation with staff. Some firms go so far as to adopt a policy of
complete secrecy, telling their staff nothing. One director told us: 'If we are too frank, we may
create difficulties for ourselves.' This policy was applied to managers as well as clerks
because, it was explained, 'our managers will worry if they find out they will lose workers and
www.seyfihoca.com
206
so have their empires reduced'. Several months after the arrival of the computer, the sales
manager in this firm had still not been given full information on the consequences of this
change.
One computer manufacturer with extensive American experience has tried to give advice to
firms purchasing its computers on how and when to communicate. It suggests to customers
that as soon as the contract is signed management should hold a meeting with all office staff
and explain what a computer is, what the firm's plans are, how it proposes to use the
computer, and how staff will be affected. Management should also explain at this meeting that
there will be some redundancy, but that it will be absorbed by the normal processes of people
changing jobs and retiring. This manufacturer tells us that he frequently encounters resistance
to this approach. Directors often take the line: 'No, don't tell anyone.'
The real bogey of the computer is that it is likely-or even intended-to displace staff. So it
constitutes a major threat to staff security, and for this reason alone is likely to be resisted. An
important part of the preparations for a machine must be, therefore, the estimating of the
number of redundancies, and identifying jobs which will be eliminated or reduced in scope by
the machine.
Briefly, I would offer the following advice to firms embarking on computers. First, do not
think your problems will be predominantly technical, because this is unlikely. Secondly, get
as much information as you can on both social and technical problems before you sign the
contract to purchase; other firms' experience can be a useful guide here. And thirdly, plan well
in advance, communicate and consult with all your staff and do not be surprised if they take a
sour view of the proposed change. No one likes to think that he can be replaced by a machine.
(From an article by Enid Mumford in The New Scientist, 18th June, 1964.)
Forgetting
In 1914, Freud published an English edition of his The Psychopathology of Everyday Life. In
this book he endeavours to show that many 'lapses of memory' and 'slips of the tongue' are not
inexplicable accidents but can be readily understood if fitted into the personality picture of the
individual. The reader is recommended to look at this well-written book for himself and
discover the wealth of intriguing anecdotal evidence with which Freud supports and develops
his thesis.
Freud is at his best when discussing those seemingly accidental mistakes of speech and
writing where one word is substituted for another and, especially, where the substitute word
means the opposite of the word intended. A physician is writing out a prescription for an
impecunious patient who asks him not to give her big bills because she cannot swallow themand then says that, of course, she meant pills. An arrogant lecturer says that he could count the
number of real authorities on his subject on one finger - he means the fingers of one hand. A
President of the Austrian House of Deputies is opening a session from which he fears little
good will come and announces that, since such and such a number of gentlemen are present,
he declares the session as closed; amid laughter, he corrects his mistake and declares the
session as opened. All of these examples clearly derive from the person saying what he
actually thinks without checking himself to make his insincere but diplomatic statement. No
doubt we have all encountered similar examples in our everyday life. Certainly writers of
fiction have long been aware of this phenomenon, and have exploited it to good dramatic
www.seyfihoca.com
207
effect by putting such lapsus linguae into the mouths of characters. In Shakespeare's Merchant
of Venice, for example, Portia has lost her affections to Bassanio but is under a vow not to
reveal it. She directs a speech to this welcome suitor in which, throughout, her love for him is
thinly disguised and finishes with the words: 'One half of me is yours, the other half yours Mine own, I would say.' The same expression of our thoughts and wishes is seen in some
erroneously carried-out actions. Thus, one physician reports that he is quite often disturbed in
the midst of engrossing work at home by having to go to hospital to carry out some routine
duty. When this happens he is apt to find himself trying to open the door of his laboratory
with the key of his desk at home. The two keys are quite unlike each other and the mistake
does not occur under normal circumstances but only under conditions where he would rather
be at home. His error seems to express his wish.
When Freud begins to discuss 'lapses of memory' in terms of repression, he seems to move on
less firm ground. He does not, of course, claim that all lapses are due to repression. His
concern is to show that at least some are and, to this end, he gives examples in which a name
or a word is unexpectedly forgotten and proceeds to demonstrate that the forgotten item is
associated either directly or indirectly with unpleasant circumstances. Here we may cite two
of his most convincing examples. The first concerns a man (X) who repeatedly forgot the
name of an old acquaintance and business associate (Y). When he required to correspond with
Y, he had to ask other people for his name. It transpired that Y had recently married a young
woman X himself had hoped to marry. Thus X had good reason to dislike his happy rival and
want to forget all about him. The second example concerns a man who set out to recite a
poem, got so far, and then could recall no more although he knew the poem well. The line on
which he blocked was descriptive of a pine-tree which is covered 'with the white sheet'. Why
should this phrase have been forgotten? Asked to relate what came to his mind when he
thought of this phrase, it was found that it immediately reminded him of the white sheet which
covers a dead body, and of the recent death of his brother from a heart condition which was
common in his family and from which he feared he too might die. The phrase referring to the
white sheet appears to have been forgotten because it was associated with circumstances
which the man did not wish to recall. In Freud's other examples, the link between the
forgotten item and some unpleasant circumstance is not so easily demonstrated.
(From Memory: Facts and Fallacies, by Ian Hunter.)
Adolescence
The period of adolescence has fascinated people of all ages. Even Aristotle turned aside from
his philosophical and ethical speculations to make a study of the adolescent. He realistically
described a boy's voice as 'the bleating of a billy goat'. He also characterized the adolescent as
being 'high-minded', but somewhat cynically put this down to lack of experience! Plato
devoted much time and thought to discovering how best to bring up youth to true citizenship.
Growing-up. Adolescence means 'growing-up' and strictly speaking should apply to a child
from birth to maturity. Why then do we use it for this teenage period alone? Because when we
speak of the adolescent as 'growing-up', we mean that the youth is leaving behind the phase of
protective childhood and is becoming independent, capable of going out to fend for himself.
Girls of this age used to be called 'flappers', a very descriptive term, for they are figuratively
trying out their wings. Very often, like fledglings, both boys and girls require a gentle push
off! Sometimes they push off too soon and hurt themselves.
www.seyfihoca.com
208
Venturesomeness. A characteristic of 'growing-up' is a desire to be venturesome - so unlike
the dependence of the child and the set ways of the adult. The adolescent seeks for new
experience in life, and likes roughing it. In their camps and hiking, for example, boys and
girls seek uncomfortable and difficult conditions-and then set about making themselves
comfortable in them. They deliberately seek difficulties in order to overcome them.
Responsibility. The adolescent also loves responsibility. The boy likes to be given the job of
packing the luggage in the car; the girl, the responsibility of getting the younger children
ready for the trip. This is a natural urge and requires expression.
Relation to life. The healthy adolescent boy or girl likes to do the real things in life, to do the
things that matter. He would rather be a plumber's mate and do a real job that requires doing
than learn about hydrostatics sitting at a desk, without understanding what practical use they
are going to be. A girl would rather look after the baby than learn about child care.
Logically we should learn about things before doing them and that is presumably why the
pundits enforce this in our educational system. But it is not the natural way-nor, I venture to
think, the best way. The adolescent wants to do things first for only then does he appreciate
the problems involved and want to learn more about them.
They do these things better in primitive life, for there at puberty the boy joins his father in
making canoes, patching huts, going out fishing or hunting, and preparing weapons of war.
He is serving his apprenticeship in the actual accomplishments of life. It is not surprising that
anthropologists find that the adolescents of primitive communities do not suffer from the
same neurotic 'difficulties' as those of civilized life. This is not, as some assume, because they
are permitted more sexual freedom, but because they are given more natural outlets for their
native interests and powers and are allowed to grow up freely into a full life of responsibility
in the community.
In the last century this was recognized in the apprenticeship system, which allowed the boy to
go out with the master carpenter, thatcher, or ploughman, to engage in the actual work of
carpentry, roof-mending, or ploughing, and so to learn his trade. It was the same in medicine,
in which a budding young doctor of sixteen learnt his job by going round with the general
practitioner and helping with the blood-letting and physic. In our agricultural colleges at the
present time young men have to do a year's work on a farm before their theoretical training at
college. The great advantage of this system is that it lets the apprentice see the practical
problems before he sets to work learning how to solve them, and he can therefore take a more
intelligent interest in his theoretical work. That is also why a girl should be allowed to give
expression to her natural desire to look after children, and then, when she comes up against
difficulties, to learn the principles of child care.
Since more knowledge of more things is now required in order to cope with the adult world,
the period of growing-up to independence takes much longer than it did in a more primitive
community, and the responsibility for such education, which formerly was in the hands of the
parents, is now necessarily undertaken by experts at school. But that should not make us lose
sight of the basic principle, namely the need and the desire of the adolescent to engage
responsibly in the 'real' pursuits of life and then to learn how-to learn through responsibility,
not to learn before responsibility.
(From Childhood and Adolescence, by J. A. Hadfield)
www.seyfihoca.com
209
Body Language.
What does scientific literature tell us about the idea that body language reflects our real
feelings? One experiment carried out about 10 years ago by Ross Buck from Carnegie-Mellon
University in Pennsylvania suggests that spontaneous facial expression is not a very good
index of real emotional state. Buck and his colleagues tested the accuracy with which people
could identify the emotions felt by another person. They presented one set of subjects with
colour slides involving a variety of emotionally-loaded visual stimuli - such as “scenic” slides
(landscapes, etc), “maternal” slides (mothers and young children), disgusting slides (severe
facial injuries and burns) and unusual slides (art objects). Unknown to these subjects, they
were being televised and viewed by another matched set of subjects, who were asked to
decide, on the basis of the televised facial expressions, which of the four sets of slides had just
been viewed. This experiment involved both male and female pairs, but no pairs comprising
both men and women; that is men observed only men, and women observed women. Buck
found that the female pairs correctly identified almost 40 per cent of the slides used - this was
above the level which would be predicted by chance alone. (Chance level is 25 per cent here,
as there were four classes of slide). But male pairs correctly identified only 28 per cent of
slides - not significantly above chance level. In other words, this study suggests that facial
expression is not a very good index of “real” feeling - and in the case of men watching and
interpreting other men, is almost useless.
Paul Ekman from the University of California has conducted a long series of experiments on
nonverbal leakage (or how nonverbal behaviour may reveal real inner states) which has
yielded some more positive and counter-intuitive results. Ekman has suggested that nonverbal
behaviour may indeed provide a clue to real feelings and has explored in some detail people
actively involved in deception, where their verbal language is not a true indication of how
they really feel. Ekman here agrees with Sigmund Freud, who was also convinced of the
importance of nonverbal behaviour in spotting deception when he wrote: “He that has eyes to
see and ears to hear may convince himself that no mortal can keep a secret. If his lips are
silent, he chatters with his finger-tips; betrayal oozes out of him at every pore.”
Ekman predicted that the feet and legs would probably hold the best clue to deception because
although the face sends out very quick instantaneous messages, people attend to and receive
most feedback from the face and therefore try to control it most. In the case of the feet and
legs the “transmission time” is much longer but we have little feedback from this part of the
body. In other words, we are often unaware of what we are doing with our feet and legs.
Ekman suggested that the face is equipped to lie the most (because we are often aware of our
facial expression) and to “leak” the most (because it sends out many fast momentary
messages) and is therefore going to be a very confusing source of information during
deception. The legs and feet would be the primary source of nonverbal leakage and hold the
main clue to deception. The form the leakage in the legs and feet would take would include
“aggressive foot kicks, flirtatious leg displays, abortive restless flight movements”. Clues to
deception could be seen in “tense leg positions, frequent shifts of leg posture, and in restless
or repetitive leg and foot movements.”
Ekman conducted a series of experiments to test his speculations, some involving psychiatric
patients who were engaging in deception, usually to obtain release from hospital. He made
films of interviews involving the patients and showed these, without sound, to one of two
groups of observers. One group viewed only the face and head, the other group, the body from
the neck down. Each observer was given a list of 300 adjectives describing attitudes,
www.seyfihoca.com
210
emotional state, and so on, and had to say which adjectives best described the patients. The
results indicated quite dramatically that individuals who utilised the face tended to be misled
by the patients, whereas those who concentrated on the lower body were much more likely to
detect the real state of the patients and not be misled by the attempted deception.
These studies thus suggest that some body language may indeed reflect our real feelings, even
when we are trying to disguise them. Most people can, however, manage to control facial
expression quite well and the face often seems to provide little information about real feeling.
Paul Ekman has more recently demonstrated that people can be trained to interpret facial
expression more accurately but this, not surprisingly, is a slow laborious process. Ekman’s
research, suggests that the feet and legs betray a great deal about real feelings and attitudes
but the research is nowhere near identifying the meanings of particular foot movements. Ray
Birdwhistell of the Eastern Pennsylvania Psychiatric Institute has gone some way towards
identifying some of the basic nonverbal elements of the legs and feet, and as a first
approximation has identified 58 separate elements. But the meanings of these particular
elements is far from clear and neither are the rules for combining the elements into larger
meaningful units. Perhaps in years to come we will have a “language” of the feet provided
that we can successfully surmount the problems described earlier in identifying the basic
forms of movement following Birdwhistell’s pioneering efforts, of how they may combine
into larger units, and in teaching people how they might make sense of apparently
contradictory movements.
In the meantime, if you go to a party and find someone peering intently at your feet - beware.
DISTANCE REGULATION IN ANIMALS
Comparative studies of animals help to show how man's space requirements are influenced by
his environment. In animals we can observe the direction, the rate, and the extent of changes
in behaviour that follow changes in space available to them as we can never hope to do in
men. For one thing, by using animals it is possible to accelerate time, since animal
generations are relatively short. A scientist can, in forty years, observe four hundred and forty
generations of mice, while he has in the same span of time seen only two generations of his
own kind. And, of course, he can be more detached about the fate of animals.
In addition, animals don’t rationalise their behaviour and thus obscure issues. In their natural
state, they respond in an amazingly consistent manner so that it is possible to observe repeated
and virtually identical performances. By restricting our observations to the way animals
handle space, it is possible to learn an amazing amount that is translatable to human terms.
Territoriality, a basic concept in the study of animal behaviour, is usually defined as
behaviour by which an organism characteristically lays claim to an area and defends it against
members of its own species. It is a recent concept, first described by the English ornithologist
H. B. Howard in his Territory in Bird Life, written in 1920. Howard stated the concept in
some detail, though naturalists as far back as the seventeenth century had taken note of
various events which Howard recognised as manifestations of territoriality.
Territoriality studies are already revising many of our basic ideas of animal life and human
life as well. The expression “free as a bird” is an encapsulated form of man’s conception of
his relation to nature. He sees animals as free to roam the world, while he himself is
imprisoned by society. Studies of territoriality show that the reverse is closer to the truth and
www.seyfihoca.com
211
that animals are often imprisoned in their own territories. It is doubtful if Freud, had he
known what is known today about the relation of animals to space, could have attributed
man’s advances to trapped energy redirected by culturally imposed inhibitions.
Many important functions are expressed in territoriality, and new ones are constantly being
discovered. H. Hediger, Zurich’s famous animal psychologist, described the most important
aspects of territoriality and explained succinctly the mechanisms by which it operates.
Territoriality, he says, insures the propagation of the species by regulating density. It provides
a frame in which things are done - places to learn, places to play, safe places to hide. Thus it
coordinates the activities of the group and holds the group together. It keeps animals within
communicating distance of each other, so that the presence of food or an enemy can be
signalled. An animal with a territory of its own can develop an inventory of reflex responses
to terrain features. When danger strikes, the animal on its home ground can take advantage of
automatic responses rather than having to take time to think about where to hide.
The psychologist C. R. Carpenter, who pioneered in the observation of monkeys in a native
setting, listed thirty-two functions of territoriality, including important ones relating to the
protection and evolution of the species. The list that follows is not complete, nor is it
representative of all species, but it indicates the crucial nature of territoriality as a behavioural
system, a system that evolved in very much the same way as anatomical systems evolved. In
fact, differences in territoriality have become so widely recognised that they are used as a
basis for distinguishing between species, much as anatomical features are used.
Territoriality offers protection from predators, and also exposes to predation the unfit who are
too weak to establish and defend a territory. Thus, it reinforces dominance in selective
breeding because the less dominant animals are less likely to establish territories. On the other
hand territoriality facilitates breeding by providing a home base that is safe. It aids in
protecting the nests and the young in them. In some species it localises waste disposal and
inhibits or prevents parasites. Yet one of the most important functions of territoriality is
proper spacing, which protects against over-exploitation of that part of the environment on
which a species depends for its living.
In addition to preservation of the species and the environment, personal and social functions
are associated with territoriality. C. R. Carpenter tested the relative roles of sexual vigour and
dominance in a territorial context and found that even a desexed pigeon will in its own
territory regularly win a test encounter with a normal male, even though desexing usually
results in loss of position in a social hierarchy. Thus, while dominant animals determine the
general direction in which the species develops, the fact that the subordinate can win (and so
breed) on his home grounds helps to preserve plasticity in the species by increasing variety
and thus preventing the dominant animals from freezing the direction which evolution takes.
Territoriality is also associated with status. A series of experiments by the British
ornithologist A. D. Bain on the great tit altered and even reversed dominance relationships by
shifting the position of feeding stations in relation to birds living in adjacent areas. As the
feeding station was placed closer and closer to a bird’s home range, the bird would accrue
advantages it lacked when away from its own home ground.
Man, too, has territoriality and he has invented many ways of defending what he considers his
own land, turf, or spread. The removal of boundary markers and trespass upon the property of
another man are punishable acts in much of the Western world. A man’s home has been his
www.seyfihoca.com
212
castle in English common law for centuries, and it is protected by prohibitions on unlawful
search and seizure even by officials of his government. The distinction is carefully made
between private property, which is the territory of an individual, and public property, which is
the territory of the group.
This cursory review of the functions of territoriality should suffice to establish the fact that it
is a basic behavioural system characteristic of living organisms including man.
An observation and an explanation
It is worth looking at one or two aspects of the way a mother behaves towards her baby. The
usual fondling, cuddling and cleaning require little comment, but the position in which she
holds the baby against her body when resting is rather revealing. Careful American studies
have disclosed the fact that 80 per cent of mothers cradle their infants in their left arms,
holding them against the left side of their bodies. If asked to explain the significance of this
preference most people reply that it is obviously the result of the predominance of righthandedness in the population. By holding the babies in their left arms, the mothers keep their
dominant arm free for manipulations. But a detailed analysis shows that this is not the case.
True, there is a slight difference between right-handed and left-handed females, but not
enough to provide an adequate explanation. It emerges that 83 per cent of right-handed
mothers hold the baby on the left side, but then so do 78 per cent of left-handed mothers. In
other words, only 22 per cent of the left-handed mothers have their dominant hands free for
actions. Clearly there must be some other, less obvious explanation.
The only other clue comes from the fact that the heart is on the left side of the mother’s body.
Could it be that the sound of her heartbeat is the vital factor? And in what way? Thinking
along these lines it was argued that perhaps during its existence inside the body of the mother,
the growing embryo becomes fixated (‘imprinted’) on the sound of the heart beat. If this is so,
then the re-discovery of this familiar sound after birth might have a calming effect on the
infant, especially as it has just been thrust into a strange and frighteningly new world outside.
If this is so then the mother, either instinctively or by an unconscious series of trials and
errors, would soon arrive at the discovery that her baby is more at peace if held on the left
against her heart, than on the right.
This may sound far-fetched, but tests have now been carried out which reveal that it is
nevertheless the true explanation. Groups of new-born babies in a hospital nursery were
exposed for a considerable time to the recorded sound of a heartbeat at a standard rate of 72
beats per minute. There were nine babies in each group and it was found that one or more of
them was crying for 60 per cent of the time when the sound was not switched on, but that this
figure fell to only 38 per cent when the heartbeat recording was thumping away. The
heartbeat groups also showed a greater weight-gain than the others, although the amount of
food taken was the same in both cases. Clearly the beatless groups were burning up a lot more
energy as a result of the vigorous actions of their crying.
Another test was done with slightly older infants at bedtime. In some groups the room was
silent, in others recorded lullabies were played. In others a ticking metronome was operating
at the heartbeat speed of 72 beats per minute. In still others the heartbeat recording itself was
played. It was then checked to see which groups fell asleep more quickly. The heartbeat group
dropped off in half the time it took for any of the other groups. This not only clinches the idea
that the sound of the heart beating is a powerfully calming stimulus, but it also shows that the
www.seyfihoca.com
213
response is a highly specific one. The metronome imitation will not do - at least, not for
young infants.
So it seems fairly certain that this is the explanation of the mother’s left-side approach to
baby-holding. It is interesting that when 466 Madonna and child paintings (dating back over
several hundred years) were analysed for this feature, 373 of them showed the baby on the left
breast. Here again the figure was at the 80 per cent level. This contrasts with observations of
females carrying parcels, where it was found that 50 per cent carried them on the left and 50
per cent on the right.
What other possible results could this heartbeat imprinting have? It may, for example, explain
why we insist on locating feelings of love in the heart rather than the head. As the song says:
‘You gotta have a heart!’ It may also explain why mothers rock their babies to lull them to
sleep. The rocking motion is carried on at about the same speed as the heartbeat, and once
again it probably ‘reminds’ the infants of the rhythmic sensations they became so familiar
with inside the womb, as the great heart of the mother pumped and thumped away above
them.
Nor does it stop there. Right into adult life the phenomenon seems to stay with us. We rock
with anguish. We rock back and forth on our feet when we are in a state of conflict. The next
time you see a lecturer or an after-dinner speaker swaying rhythmically from side to side,
check his speed for heartbeat time. His discomfort at having to face an audience leads him to
perform the most comforting movements his body can offer in the somewhat limited
circumstances; and so he switches on the old familiar beat of the womb.
Wherever you find insecurity, you are liable to find the comforting heartbeat rhythm in one
kind of disguise or another. It is no accident that most folk music and dancing has a
syncopated rhythm. Here again the sounds and movements take the performers back to the
safe world of the womb.
From The Naked Ape by Desmond Morris. (Jonathan Cape and McGraw Hill, 1967)
GESTURES
A gesture is any action that sends a visual signal to an onlooker. To become a gesture, an act
has to be seen by someone else and has to communicate some piece of information to them. It
can do this either because the gesturer deliberately sets out to send a signal - as when he
waves his hand - or it can do it only incidentally - as when he sneezes. The hand-wave is a
Primary Gesture, because it has no other existence or function. It is a piece of communication
from start to finish. The sneeze, by contrast, is a secondary, or Incidental Gesture. Its primary
function is mechanical and is concerned with the sneezer’s personal breathing problem. In its
secondary role, however, it cannot help but transmit a message to his companions, warning
them that he may have caught a cold.
Most people tend to limit their use of the term ‘gesture’ to the primary form - the hand-wave
type - but this misses an important point. What matters with gesturing is not what signals we
think we are sending out, but what signals are being received. The observers of our acts will
make no distinction between our intentional Primary Gestures and our unintentional,
incidental ones. In some ways, our Incidental Gestures are the more illuminating of the two, if
only for the very fact that we do not think of them as gestures, and therefore do not censor and
www.seyfihoca.com
214
manipulate them so strictly. This is why it is preferable to use the term ‘gesture’ in its wider
meaning as an ‘observed action’.
A convenient way to distinguish between Incidental and Primary Gestures is to ask the
question: Would I do it if I were completely alone? If the answer is No, then it is a Primary
Gesture. We do not wave, wink, or point when we are by ourselves; not, that is, unless we
have reached the unusual condition of talking animatedly to ourselves.
INCIDENTAL GESTURES
Mechanical actions with secondary messages
Many of our actions are basically non-social, having to do with problems of personal body
care, body comfort and body transportation; we clean and groom ourselves with a variety of
scratchings, rubbings and wipings; we cough, yawn and stretch our limbs; we eat and drink;
we prop ourselves up in restful postures, folding our arms and crossing our legs; we sit, stand,
squat and recline, in a whole range of different positions; we crawl, walk and run in varying
gaits and styles. But although we do these things for our own benefit, we are not always
unaccompanied when we do them. Our companions learn a great deal about us from these
‘personal’ actions - not merely that we are scratching because we itch or that we are running
because we are late, but also, from the way we do them, what kind of personalities we possess
and what mood we are in at the time.
Sometimes the mood-signal transmitted unwittingly in this way is one that we would rather
conceal, if we stopped to think about it. Occasionally we do become self-consciously aware of
the ‘mood broadcasts’ and ‘personality displays’ we are making and we may then try to check
ourselves. But often we do not, and the message goes out loud and clear.
For instance, if a student props his head on his hands while listening to a boring lecture, his
head-on-hands action operates both mechanically and gesturally. As a mechanical act, it is
simply a case of supporting a tired head - a physical act that concerns no one but the student
himself. At the same time, though, it cannot help operating as a gestural act, beaming out a
visual signal to his companions, and perhaps to the lecturer himself, telling them that he is
bored.
In such a case his gesture was not deliberate and he may not even have been aware that he
was transmitting it. If challenged, he would claim that he was not bored at all, but merely
tired. If he were honest - or impolite - he would have to admit that excited attention easily
banishes tiredness, and that a really fascinating speaker need never fear to see a slumped,
head-propped figure like his in the audience.
In the schoolroom, the teacher who barks at his pupils to ‘sit up straight’ is demanding, by
right, the attention-posture that he should have gained by generating interest in his lesson. It
says a great deal for the power of gesture-signals that he feels more ‘attended-to’ when he
sees his pupils sitting up straight, even though he is consciously well aware of the fact that
they have just been forcibly un-slumped, rather than genuinely excited by his teaching.
Many of our Incidental Gestures provide mood information of a kind that neither we nor our
companions become consciously alerted to. It is as if there is an underground communication
system operating just below the surface of our social encounters. We perform an act and it is
www.seyfihoca.com
215
observed. Its meaning is read, but not out loud. We ‘feel’ the mood, rather than analyse it.
Occasionally an action of this type becomes so characteristic of a particular situation that we
do eventually identify it - as when we say of a difficult problem: ‘That will make him scratch
his head’, indicating that we do understand the link that exists between puzzlement and the
Incidental Gesture of head-scratching. But frequently this type of link operates below the
conscious level, or is missed altogether.
Where the links are clearer, we can, of course, manipulate the situation and use our Incidental
Gestures in a contrived way. If a student listening to a lecture is not tired, but wishes to insult
the speaker, he can deliberately adopt a bored, slumped posture, knowing that its message will
get across. This is a Stylized Incidental Gesture - a mechanical action that is being artificially
employed as a pure signal. Many of the common ‘courtesies’ also fall into this category - as
when we greedily eat up a plate of food that we do not want and which we do not like, merely
to transmit a suitably grateful signal to our hosts. Controlling our Incidental Gestures in this
way is one of the processes that every child must learn as it grows up and learns to adapt to
the rules of conduct of the society in which it lives.
EXPRESSIVE GESTURES
Biological gestures of the kind we share with other animals
Primary Gestures fall into six main categories. Five of these are unique to man, and depend on
his complex, highly evolved brain. The exception is the category I called Expressive Gestures.
These are gestures of the type which all men, everywhere, share with one another, and which
other animals also perform. They include the important signals of Facial Expression, so
crucial to daily human interaction.
All primates are facially expressive and among the higher species the facial muscles become
increasingly elaborate, making possible the performance of a whole range of subtly varying
facial signals. In man this trend reaches its peak, and it is true to say that the bulk of nonverbal signalling is transmitted by the human face.
The human hands are also important, having been freed from their ancient locomotion duties,
and are capable, with their Manual Gesticulations, of transmitting many small mood changes
by shifts in their postures and movements, especially during conversational encounters. I am
defining the word ‘gesticulation’, as distinct from ‘gesture’, as a manual action performed
unconsciously during social interactions, when the gesticulator is emphasizing a verbal point
he is making.
These natural gestures are usually spontaneous and very much taken for granted. Yes, we say,
he made a funny face. But which way did his eyebrows move? We cannot recall. Yes, we say,
he was waving his arms about as he spoke. But what shape did his fingers make? We cannot
remember. Yet we were not inattentive. We saw it all and our brains registered what we saw.
We simply did not need to analyse the actions, any more than we had to spell out the words
we heard, in order to understand them. In this respect they are similar to the Incidental
Gestures of the previous category, but they differ, because here there is no mechanical
function - only signalling. This is the world of smiles and sneers, shrugs and pouts, laughs and
winces, blushes and blanches, waves and beckons, nods and glares, frowns and snarls. These
are the gestures that nearly everyone performs nearly everywhere in the world. They may
differ in detail and in context from place to place, but basically they are actions we all share.
www.seyfihoca.com
216
We all have complex facial muscles whose sole job it is to make expressions, and we all stand
on two feet rather than four, freeing our hands and letting them dance in the air evocatively as
we explain, argue and joke our way through our social encounters. We may have lost our
twitching tails and our bristling fur, but we more than make up for it with our marvellously
mobile faces and our twisting, spreading, fluttering hands.
In origin, our Expressive Gestures are closely related to our Incidental Gestures, because their
roots also lie in primarily non-communicative actions. The clenched fist of the gesticulator
owes its origin to an intention movement of hitting an opponent, just as the frown on the face
of a worried man can be traced back to an ancient eye-protection movement of an animal
anticipating physical attack. But the difference is that in these cases the link between the
primary physical action and its ultimate descendant, the Expressive Gesture, has been broken.
Smiles, pouts, winces, gapes, smirks, and the rest, are now, for all practical purposes, pure
gestures and exclusively communicative in function.
Despite their worldwide distribution, Expressive Gestures are nevertheless subject to
considerable cultural influences. Even though we all have an evolved set of smiling muscles,
we do not all smile in precisely the same way, to the same extent, or on the same occasions.
For example, all children may start out as easy-smilers and easy-laughers, but a local tradition
may insist that, as the youngsters mature, they must hide their feelings, and their adult
laughter may become severely muted as a result. These local Display Rules, varying from
place to place, often give the false impression that Expressive Gestures are local inventions
rather than modified, but universal, behaviour patterns.
MIMIC GESTURES
Gestures which transmit signals by imitation
Mimic Gestures are those in which the performer attempts to imitate, as accurately as
possible, a person, an object or an action. Here we leave our animal heritage behind and enter
an exclusively human sphere. The essential quality of a Mimic Gesture is that it attempts to
copy the thing it is trying to portray. No stylized conventions are applied. A successful Mimic
Gesture is therefore understandable to someone who has never seen it performed before. No
prior knowledge should be required and there need be no set tradition concerning the way in
which a particular item is represented. There are four kinds of Mimic Gesture:
First, there is Social Mimicry, or ‘putting on a good face’. We have all done this. We have all
smiled at a party when really we feel sad, and perhaps looked sadder at a funeral than we feel,
simply because it is expected of us. We lie with simulated gestures to please others. This
should not be confused with what psychologists call ‘role-playing’. When indulging in Social
Mimicry we deceive only others, but when role-playing we deceive ourselves as well.
Second, there is Theatrical Mimicry - the world of actors and actresses, who simulate
everything for our amusement. Essentially it embraces two distinct techniques. One is the
calculated attempt to imitate specifically observed actions. The actor who is to play a general,
say, will spend long hours watching films of military scenes in which he can analyse every
tiny movement and then consciously copy them and incorporate them into his final portrayal.
The other technique is to concentrate instead on the imagined mood of the character to be
portrayed, to attempt to take on that mood, and to rely upon it to produce, unconsciously, the
necessary style of body actions.
www.seyfihoca.com
217
In reality, all actors use a combination of both these techniques, although in explaining their
craft they may stress one or other of the two methods. In the past, acting performances were
usually highly stylized, but today, except in pantomime, opera and farce, extraordinary
degrees of realism are reached and the formal, obtrusive audience has become instead a
shadowy group of eavesdroppers. Gone are the actor’s asides, gone are the audience
participations. We must all believe that it is really happening. In other words, Theatrical
Mimicry has at last become as realistic as day-to-day Social Mimicry. In this respect, these
first two types of mimic activity contrast sharply with the third, which can be called Partial
Mimicry.
In Partial Mimicry the performer attempts to imitate something which he is not and never can
be, such as a bird, or raindrops. Usually only the hands are involved, but these make the most
realistic approach to the subject they can manage. If a bird, they flap their ‘wings’ as best they
can; if raindrops, they describe a sprinkling descent as graphically as possible. Widely used
mimic gestures of this kind are those which convert the hand into a ‘gun’, an animal of some
sort, or the foot of an animal; or those which use the movements of the hand to indicate the
outline shape of an object of some kind.
The fourth kind of Mimic Gesture can best be called Vacuum Mimicry, because the action
takes place in the absence of the object to which it is related. If I am hungry, for example, I
can go through the motions of putting imaginary food into my mouth. If I am thirsty, I can
raise my hand as if holding an invisible glass, and gulp invisible liquid from it.
The important feature of Partial Mimicry and Vacuum Mimicry is that, like Social and
Theatrical Mimicry, they strive for reality. Even though they are doomed to failure, they make
an attempt. This means that they can be understood internationally. In this respect they
contrast strongly with the next two types of gesture, which show marked cultural restrictions.
SCHEMATIC GESTURES
Imitations that become abbreviated or abridged
Schematic Gestures are abbreviated or abridged versions of Mimic Gestures. They attempt to
portray something by taking just one of its prominent features and then performing that alone.
There is no longer any attempt at realism.
Schematic Gestures usually arise as a sort of gestural shorthand because of the need to
perform an imitation quickly and on many occasions. Just as, in ordinary speech, we reduce
the word ‘cannot’ to ‘can’t’, so an elaborate miming of a charging bull becomes reduced
simply to a pair of horns jabbed in the air as a pair of fingers.
When one element of a mime is selected and retained in this way, and the other elements are
reduced or omitted, the gesture may still be easy to understand, when seen for the first time,
but the stylization may go so far that it becomes meaningless to those not ‘in the know’. The
Schematic Gesture then becomes a local tradition with a limited geographical range. If the
original mime was complex and involved several distinctive features, different localities may
select different key features for their abridged versions. Once these different forms of
shorthand have become fully established in each region, then the people who use them will
become less and less likely to recognize the foreign forms. The local gesture becomes ‘the’
gesture, and there quickly develops, in gesture communication, a situation similar to that
www.seyfihoca.com
218
found in linguistics. Just as each region has its own verbal language, so it also has its own set
of Schematic Gestures.
To give an example: the American Indian sign for a horse consists of a gesture in which two
fingers of one hand ‘sit astride’ the fingers of the other hand. A Cistercian monk would
instead signal ‘horse’ by lowering his head slightly and pulling at an imaginary tuft of hair on
his forehead. An Englishman would probably crouch down like a jockey and pull at imaginary
reins. The Englishman’s version, being closer to a Vacuum Mimic Gesture, might be
understood by the other two, but their gestures, being highly schematic, might well prove
incomprehensible to anyone outside their groups.
Some objects, however, have one special feature that is so strongly characteristic of them that,
even with Schematic Gestures, there is little doubt about what is being portrayed. The bull,
mentioned above, is a good example of this. Cattle are nearly always indicated by their horns
alone, and the two horns are always represented by two digits. In fact, if an American Indian,
a Hindu dancer, and an Australian Aborigine met, they would all understand one another’s
cattle signs, and we would understand all three of them. This does not mean that the signs are
all identical. The American Indian’s cattle sign would represent the bison, and the horns of
bison do not curve forward like those of domestic cattle, but inward, towards each other. The
American Indian’s sign reflects this, his hands being held to his temples and his forefingers
being pointed inward. The Australian Aborigine instead points his forefingers forward. The
Hindu dancer also points forward, but rather than using two forefingers up at the temples,
employs the forefinger and little finger of one hand, held at waist height. So each culture has
its own variant, but the fact that horns are such an obvious distinguishing feature of cattle
means that, despite local variations, the bovine Schematic Gesture is reasonably
understandable in most cultures.
SYMBOLIC GESTURES
Gestures which represent moods and ideas
A Symbolic Gesture indicates an abstract quality that has no simple equivalent in the world of
objects and movements. Here we are one stage further away from the obviousness of the
enacted Mimic Gesture.
How, for instance, would you make a silent sign for stupidity? You might launch into a fullblooded Theatrical Mime of a drooling village idiot. But total idiocy is not a precise way of
indicating the momentary stupidity of a healthy adult. Instead, you might tap your forefinger
against your temple, but this also lacks accuracy, since you might do precisely the same thing
when indicating that someone is brainy. All the tap does is to point to the brain. To make the
meaning more clear, you might instead twist your forefinger against your temple, indicating ‘a
screw loose’. Alternatively, you might rotate your forefinger close to your temple, signalling
that the brain is going round and round and is not stable.
Many people would understand these temple-forefinger actions, but others would not. They
would have their own local, stupidity gestures, which we in our turn would find confusing,
such as tapping the elbow of the raised forearm, flapping the hand up and down in front of
half-closed eyes, rotating a raised hand, or laying one forefinger flat across the forehead.
www.seyfihoca.com
219
The situation is further complicated by the fact that some stupidity signals mean totally
different things in different countries. To take one example, in Saudi Arabia stupidity can be
signalled by touching the lower eyelid with the tip of the forefinger. But this same action, in
various other countries, can mean disbelief, approval, agreement, mistrust, scepticism,
alertness, secrecy, craftiness, danger, or criminality. The reason for this apparent chaos of
meanings is simple enough. By pointing to the eye, the gesturer is doing no more than stress
the symbolic importance of the eye as a seeing organ. Beyond that, the action says nothing, so
that the message can become either: ‘Yes, I see’, or ‘I can’t believe my eyes’, or ‘Keep a
sharp look-out’, or ‘I like what I see’, or almost any other seeing signal you care to imagine.
In such a case it is essential to know the precise ‘seeing’ property being represented by the
symbolism of the gesture in any particular culture.
So we are faced with two basic problems where Symbolic Gestures are concerned: either one
meaning may be signalled by different actions, or several meanings may be signalled by the
same action, as we move from culture to culture. The only solution is to approach each culture
with an open mind and learn their Symbolic Gestures as one would their vocabulary.
As part of this process, it helps if a link can be found between the action and the meaning, but
this is not always possible. In some cases we simply do not know how certain Symbolic
Gestures arose. It is clear that they are symbolic because they now represent some abstract
quality, but how they first acquired the link between action and meaning has been lost
somewhere in their long history. A good instance of this is the ‘cuckold’ sign from Italy. This
consists of making a pair of horns, either with two forefingers held at the temples, or with a
forefinger and little finger of one hand held in front of the body. There is little doubt about
what the fingers are meant to be: they are the horns of a bull. As such, they would rate as part
of a Schematic Gesture. But they do not send out the simple message ‘bull’. Instead they now
indicate ‘sexual betrayal’. The action is therefore a Symbolic gesture and, in order to explain
it, it becomes necessary to find the link between bulls and sexual betrayal.
Historically, the link appears to be lost, with the result that some rather wild speculations have
been made. A complication arises in the form of the ‘horned hand’, also common in Italy,
which has a totally different significance, even though it employs the same motif of bull’s
horns. The Y horned hand is essentially a protective gesture, made to ward off imagined
dangers. Here it is clear enough that it is the bull’s great power, ferocity and masculinity that
is being invoked as a symbolic aid to protect the gesturer. But this only makes it even more
difficult to explain the other use of the bull’s-horns gesture as a sign of a ‘pathetic’ cuckold.
A suggested explanation of this contradiction is that it is due to one gesture using as its
starting point the bull’s power, while the other - the cuckold sign - selects the bull’s frequent
castration. Since the domestication of cattle began, there have always been too many bulls in
relation to cows. A good, uncastrated bull can serve between 50 and 100 cows a year, so that
it is only necessary to retain a small proportion of intact bulls for breeding purposes. The rest
are rendered much more docile and easy to handle for beef production, by castration. In folklore, then, these impotent males must stand helplessly by, while the few sexually active bulls
‘steal their rightful females’; hence the symbolism of: bull = cuckold.
A completely different explanation once offered was that, when the cuckold discovers that his
wife has betrayed him, he becomes so enraged and jealous that he bellows and rushes
violently about like a ‘mad bull’.
www.seyfihoca.com
220
A more classical interpretation involves Diana the Huntress, who made horns into a symbol of
male downfall. Actaeon, another hunter, is said to have sneaked a look at her naked body
when she was bathing. This so angered her that she turned him into a horned beast and set his
own hounds upon him, who promptly killed and ate him.
Alternatively, there is the version dealing with ancient religious prostitutes. These ladies
worshipped gods who wore ‘horns of honour’ - that is, horns in their other role as symbols of
power and masculinity - and the gods were so pleased with the wives who became sacred
whores that they transferred their godly horns on to the heads of the husbands who had
ordered their women to act in this role. In this way, the horns of honour became the horns of
ridicule.
As if this were not enough, it is also claimed elsewhere, and with equal conviction, that
because stags have horns (antlers were often called horns in earlier periods) and because most
stags in the rutting season lose their females to a few dominant males who round up large
harems, the majority of ‘horned’ deer are unhappy ‘cuckolds’.
Finally, there is the bizarre interpretation that bulls and deer have nothing to do with it.
Instead, it is thought that the ancient practice of grafting the spurs of a castrated cockrel on to
the root of its excised comb, where they apparently grew and became ‘horns’, is the origin of
the symbolic link between horns and cuckolds. This claim is backed up by the fact that the
German equivalent word for ‘cuckold’ (hahnrei) originally meant ‘capon’.
If, after reading these rival claims, you feel that all you have really learned is the meaning of
the phrase ‘cock-and-bull story’, you can be forgiven. Clearly, we are in the realm of fertile
imagination rather than historical record. But this example has been dealt with at length to
show how, in so many cases, the true story of the origin of a Symbolic Gesture is no longer
available to us. Many other similarly conflicting examples are known, but this one will suffice
to demonstrate the general principle.
There are exceptions, of course, and certain of the Symbolic Gestures we make today, and
take for granted, can easily be traced to their origins. ‘Keeping your fingers crossed’ is a good
example of this. Although used by many non-Christians, this action of making the cross,
using only the first and second fingers, is an ancient protective device of the Christian church.
In earlier times it was commonplace to make a more conspicuous sign of the cross (to cross
oneself) by moving the whole arm, first downwards and then sideways, in front of the body,
tracing the shape of the cross in the air. This can still be seen in some countries today in a
non-religious context, acting as a ‘good luck’ protective device. In more trivial situations it
has been widely replaced, however, by the act of holding up one hand to show that the second
finger is tightly crossed over the first, with the crossing movement of the arm omitted.
Originally this was the secret version of ‘crossing oneself’ and was done with the hand in
question carefully hidden from view. It may still be done in this secret way, as when trying to
protect oneself from the consequences of lying, but as a ‘good luck’ sign it has now come out
into the open. This development is easily explained by the fact that crossing the fingers lacks
an obvious religious character. Symbolically, the finger-crossing may be calling on the
protection of the Christian God, but the small finger action performed is so far removed from
the priestly arm crossing action, that it can without difficulty slide into everyday life as a
casual wish for good fortune. Proof of this is that many people do not even realize that they
are demanding an act of Christian worship - historically speaking - when they shout out:
‘Keep your fingers crossed!’
www.seyfihoca.com
221
TECHNICAL GESTURES
Gestures used by specialist minorities
Technical Gestures are invented by a specialist minority for use strictly within the limits of
their particular activity. They are meaningless to anyone outside the specialization and operate
in such a narrow field that they cannot be considered as playing a part in the mainstream of
visual communication of any culture.
Television-studio signals are a good example of Technical Gestures in use today. The studio
commentator we see on our screens at home is face to face with a ‘studio manager’. The
manager is linked to the programme director in the control room by means of headphones and
conveys the director’s instructions to the commentator by simple visual gestures. To warn the
commentator that he will have to start speaking at any moment, the manager raises a forearm
and holds it stiffly erect. To start him speaking, he brings the forearm swiftly down to point at
the commentator. To warn him that he must stop speaking in a few seconds, the manager
rotates his forearm, as if it were the hand of a clock going very fast - ‘Time is running out
fast.’ To ask him to lengthen the speaking time and say more, he holds his hands together in
front of his chest and pulls them slowly apart, as if stretching something – ‘stretch it out.’ To
tell the speaker to stop dead this instant, the manager makes a slashing action with his hand
across his throat - ‘Cut!’ There are no set rules laid down for these signals. They grew up in
the early days of television and, although the main ones listed here are fairly widespread
today, each studio may well have its own special variants, worked out to suit a particular
performer.
Other Technical Gestures are found wherever an activity prohibits verbal contact. Skindivers,
for instance, cannot speak to one another and need simple signals to deal with potentially
dangerous situations. In particular they need gestures for danger, cold, cramp and fatigue.
Other messages, such as yes, no, good, bad, up and down, are easily enough understood by
the use of everyday actions and require no Technical Gestures to make sense. But how could
you signal to a companion that you had cramp? The answer is that you would open and close
one hand rhythmically - a simple gesture, but one that might nevertheless save a life.
Disaster can sometimes occur because a Technical Gesture is required from someone who is
not a specialist in a technical field. Suppose some holiday-makers take out a boat, and it sinks,
and they swim to the safety of a small, rocky island. Wet and frightened, they crouch there
wondering what to do next, when to their immense relief a small fishing-boat comes chugging
towards them. As it draws level with the island, they wave frantically at it. The people on
board wave back, and the boat chugs on and disappears. If the stranded holiday-makers had
been marine ‘specialists’, they would have known that, at sea, waving is only used as a
greeting. To signal distress, they should have raised and lowered their arms stiffly from their
sides. This is the accepted marine gesture for ‘Help!’
Ironically, if the shipwrecked signallers had been marine experts and had given the correct
distress signal, the potential rescue boat might well have been manned by holiday-makers,
who would have been completely nonplussed by the strange actions and would probably have
ignored them. When a technical sphere is invaded by the non-technical, gesture problems
always arise.
www.seyfihoca.com
222
Firemen, crane-drivers, airport-tarmac signalmen, gambling-casino croupiers, dealers at
auctions, and restaurant staff, all have their own special Technical Gestures. Either because
they must keep quiet, must be discreet, or cannot be heard, they develop their own sets of
signals. The rest of us can ignore them, unless we, too, wish to enter their specialized spheres.
CODED GESTURES
Sign-language based on a formal system
Coded Gestures, unlike all others, are part of a formal system of signals. They interrelate with
one another in a complex and systematic way, so that they constitute a true language. The
special feature of this category is that the individual units are valueless without reference to
the other units in the code. Technical Gestures may be systematically planned, but, with them,
each signal can operate quite independently of the others. With Coded Gestures, by contrast,
all the units interlock with one another on rigidly formulated principles, like the letters and
words in a verbal language.
The most important example is the Deaf-and-dumb Sign Language of hand signals, of which
there is both a one-handed and a two-handed version. Also, there is the Semaphore Language
of arm signals, and the Tic-tac Language of the race course. These all require considerable
skill and training and belong in a totally different world from the familiar gestures we employ
in everyday life. They serve as a valuable reminder, though, of the incredibly sensitive
potential we all possess for visual communication. It makes it all the more plausible to argue
that we are all of us responding, with greater sensitivity than we may realize, to the ordinary
gestures we witness each day of our lives.
(From Manwatching by Desmond Morris.)
Adaptive control of reading rate
One important factor in reading is the voluntary, adaptive control of reading rate, i.e. the
ability to adjust the reading rate to the particular type of material being read.
Adaptive reading means changing reading speed throughout a text in response to both the
difficulty of material and one’s purpose in reading it. Learning how to monitor and adjust
reading style is a skill that requires a great deal of practice.
Many people, even college students are unaware that they can learn to control their reading
speed. However, this factor can be greatly improved with a couple of hundred hours of work,
as opposed to the thousands of hours needed to significantly alter language comprehension.
Many college reading skills programmes include a training procedure aimed at improving
students’ control of reading speed. However, a number of problems are involved in successfully implementing such a programme. The first problem is to convince the students that they
should adjust their reading rates. Many students regard skimming as a sin and read everything
in a slow methodical manner. On the other hand some students believe that everything,
including difficult mathematical texts, can be read at the rate appropriate for a light novel.
There seems to be evidence that people read more slowly than necessary. A number of studies
on college students have found that when the students are forced to read faster than their selfimposed rate, there is no loss in retention of information typically regarded as important.
www.seyfihoca.com
223
The second problem involved in teaching adaptive reading lies in convincing the students of
the need to be aware of their purposes in reading. The point of adjusting reading rates is to
serve particular purposes. Students who are unaware of what they want to get out of a reading
assignment will find it difficult to adjust their rates appropriately. They should know in
advance what they want.
Once these problems of attitude are overcome, a reading skills course can concentrate on
teaching the students the techniques for reading at different rates. Since most students have
had little practice at rapid reading, most of the instruction focuses on how to read rapidly.
Scanning is a rapid reading technique appropriate for searching out a piece of information
embedded in a much larger text - for example a student might scan this passage for an
evaluation of adaptive reading. A skilled scanner can process 10,000 or more words per
minute. Obviously, at this rate scanners only pick up bits and pieces of information and skip
whole paragraphs. It is easy for scanners to miss the target entirely, and they often have to
rescan the text. Making quick decisions as to what should be ignored and what should be
looked at takes practice. However, the benefits are enormous. I would not be able to function
as an academic without this skill because I would not be able to keep up with all the
information that is generated in my field.
Skimming is the processing of about 800-1500 words a minute - a rate at which identifying
every word is probably impossible. Skimming is used for extracting the gist of the text. The
skill is useful when the skimmer is deciding whether to read a text, or is previewing a text he
wants to read, or is going over material that is already known.
Both scanning and skimming are aided by a knowledge of where the main points tend to be
found in the text. A reader who knows where an author tends to put the main points can read
selectively. Authors vary in their construction style, and one has to adjust to author
differences, but some general rules usually apply. Section headings, first and last paragraphs
in a section, first and last sentences in a paragraph, and highlighted material all tend to convey
the main points.
Students in reading skills programmes often complain that rapid reading techniques require
hard work and that they tend to regress towards less efficient reading habits after the end of
the programme. Therefore, it should be emphasised that the adaptive control of the reading
rate is hard work because it is a novel skill. Older reading habits seem easy because they have
been practised for longer. As students become more practised in adjusting reading rate, they
find it easier. I can report that after practising variable reading rates for more than ten years, I
find it easier to read a text using an adjustable rate than to read at a slow methodical word by
word rate. This is something of a problem for me because part of my professional duties is to
edit papers that I would not normally process word by word. I find it very painful to have to
read at this rate.
SOCIOLOGY
RATIONAL AND IRRATIONAL ELEMENTS IN CONTEMPORARY SOCIETY
Imagine yourselves standing at a street corner of a large and busy city. Everything in front of
you is bustling, moving. Here, to your left, a man laboriously pushes a wheelbarrow. There in
measured trot, a horse is pulling a carriage; on all sides you see a constant stream of cars and
www.seyfihoca.com
224
buses. Above you, somewhere in the distance, can be heard the buzzing noise of an aeroplane.
In all this there is nothing unusual; nothing that would to-day call for surprise or
astonishment; it is only when concentrated analysis has revealed the problematic aspect of
even the most obvious things in life that we discover sociological problems underlying these
everyday phenomena. Wheelbarrow, carriage, automobile, and aeroplane are each typical of
the means of conveyance in different phases of historical development. They originate in
different times, thus they represent different phases of technical development; and yet they are
all used simultaneously. This particular phenomenon has been called the law of the
'contemporaneousness of the non-contemporaneous'. However well these different phases of
history seem to exist side by side in the picture before us, in certain situations and under
particular circumstances they can lead to the most convulsive disturbances in our social life.
No sooner does this thought occur to us than we can see a different picture unfolding itself.
The pilot who only a minute ago seemed to be flying quietly above us hurls a hurricane of
bombs and in the twinkle of an eye lays waste everything and annihilates everybody
underneath him. You know that this idea is far from being a figment of the imagination, and
the uneasiness which its horror awakens in you leads involuntarily to a modification of your
previous admiration of human progress. In his scientific and technical knowledge man has,
indeed, made miraculous strides forward in the span of time that separates us from the days
when the carriage came into use; but is human reason and rationality, in other than the
technical field, today so very different from what it was in the distant past of which the
wheelbarrow is a symbol? Do our motives and impulses really operate on a higher plane than
those of our ancestors? What, in essence, does the action of the pilot who drops bombs
signify?
Surely this: that man is availing himself of the most up-to-date results of technical ingenuity
in order to satisfy ancient impulses and primitive motives. If, therefore, the city is destroyed
by the deadly means of modern warfare this must be attributed solely to the fact that the
development of man's technical powers over nature is far ahead of the development of his
moral faculties and his knowledge of the guidance and government of society. The
phenomenon suggested by this whole analogy can now be given a sociological designation; it
is the phenomenon of a disproportionate development of human faculties. This phenomenon
of a disproportionate development can be observed not only in the life of groups but also in
that of individuals. We know from child-psychology that a child may be intellectually
extremely precocious, whilst the development of his moral or temperamental qualities has
been arrested at an infantile stage. Such an unevenly balanced development of his various
faculties may be a source of acute danger to an individual; in the case of society, it is nothing
short of catastrophic.
(From Hobhouse Memorial Lecture, 1940, by Karl Mannheim)
SOCIAL LIFE IN A PROVINCIAL UNIVERSITY
A number of assumptions about the way of life in a pro-vincial university are current today. I
used myself to hold a number of them but in the course of an inquiry in King's College,
Newcastle, on which this paper is based, I modified my views considerably.
It was with a view to seeing how much truth there was in these assumptions that an inquiry
was undertaken at King's College, Newcastle, during the academic years 1952-54. One-tenth
of the undergraduate population was interviewed; in order to check whether conditions in
www.seyfihoca.com
225
Newcastle were peculiar, or common to other northern university cities, a small sample of
Liverpool students living at home were also seen, which made a total of 373.
First the question of the students' homes: university teachers may find it deplorable that many
of the students live at home with parents or other guardians, but this is by no means the
unanimous view of the students themselves. Nearly half of them, both at Liverpool and
Newcastle, consciously prefer it, and think it has many advantages over going away to
another university. About an equal number regret it, and a small number are undecided either
way. Those who prefer living at home ascribe their preference to a number of factors,
principally home cooking, good facilities for working, without the distractions of resident life,
and the companionship of those for whom they feel affection.
Not a few of those who wished themselves away said 'It's hopeless to try and work at home';
and indeed the first question we must ask, if we wish to get a true picture of the home
student's life, is: how good a place is the home for a young man or woman who needs to
spend many hours each week in quiet, uninterrupted reading? It might be supposed that
household chores would affect the picture, and students were asked to say how much time
they spent on them daily. The results were classified as 'Heavy', 'Negligible' and 'Neither'.
Counted under negligible were those who replied 'None at all' or 'Hardly any' or 'I wash up
occasionally'. These accounted for just 44 per cent of the Newcastle and the Liverpool
students. The criterion by which to judge of 'Heavy' chores was whether the student himself
felt his share to be a heavy one. It is possible that some of the others, not so classified, spent
as many hours on house-work; but what is undertaken from choice, rather as a hobby and
relaxation, is bound to seem different from a regular responsibility, undertaken from
necessity. In this category were found 12 per cent Newcastle students and only 5 out of the 67
in Liverpool. As one would expect, more women than men are found in this group, and with
both, those whose responsibilities are heavy are usually the victims of some kind of family
emergency, such as the illness of the mother or (occasionally) the father, or the death of the
mother. The remainder, though spending an amount of time on domestic chores which is by
no means negligible, are not oppressed by it, and do not feel it a hindrance to their work. It
may be taken that their own instinct is sound in this, and that if they had not been doing
house-work they would have been doing something other than reading.
Much more fundamental to the matter we are considering is the question of where they do
their work when they are at home. The students interviewed, then, were asked the question
'Where do you usually work when at home?' In Newcastle 30 Out of 152 give the family
living-room as their usual place, and another 18 say they sometimes work there. But 95, or 62
per cent, have special provision made for them, either by some heating arrangement in their
bedroom, or by the putting of a fire in another living-room. Much depends on how one looks
at it. In comparison with the Oxford college or the hall of residence, where every student has
his own room, it seems regrettable that as many as a third of these students worked in the
family room, either sometimes or always. But if one reflects that in all probability most of
these people, when they were grammar school pupils, worked in the family room, then the
figure of 62 per cent for whom special provision is made indicates a real effort on the part of
these families to meet the felt needs of a university student, for heating is a heavy item in a
family budget. The question also arises in connection with students living in lodgings, who
will be considered later. The culprit seems to be chiefly that unique social institution the
English bedroom, on account of which about half the available living space in a house is
unheated and unfurnished so as to render it unusable for anything but sleeping.
www.seyfihoca.com
226
(Alice Eden From an article in The British Journal of Sociology, December 1959)
THE MENACE OF OVER-POPULATION
The essential fact about the population problem is well known. It is simply that world
population is increasing at a rate with which food and other production may not be able to
keep pace and will certainly not be able to overtake sufficiently to raise the standard of living
in the underdeveloped countries. World population is thought to have increased by 1,000
million since 1800 - that is, by more than 60 per cent (Figure 1) - to raise the estimated total in
1956 to about 2,700 million, and the present rate of increase, which far exceeds the estimates
of 25 years ago, is such as to conjure up visions of staggering numbers in the foreseeable
future.
The total, now thought to be approaching 3,000 million, is not, of course, equally distributed
in proportion to the land areas. Europe, including the U.S.S.R., is about averagely populated,
Africa, North and South America and Oceania are under-populated and Asia is greatly overpopulated (Figure 2 a, b). The rate of increase is also very uneven. Figure 3 shows birth-rates
and death-rates since 1947 for four contrasting areas, the difference in the two rates
representing natural increase. In Singapore a very high and almost static birth-rate, coupled
with a rapid decline in the death-rate, is giving rise to serious alarm. Many other regions, such
as Malaya, Ceylon and Mexico, are in the same position to a greater or lesser extent. In Japan
a decline in the death-rate has been offset by a dramatic drop of nearly 50 per cent in the
birth-rate brought about mainly by quasi-legalized abortion. In the United States a low and
slightly declining death-rate has combined with a substantial and recently static birth-rate to
give a formidable rate of natural increase. In Great Britain a death-rate which is not
particularly low combines with a low birth-rate to give a very small rate of increase. Those
Asian countries in which the rate of increase is still slow are being held back by a high deathrate rather than a low birth-rate, and the same applies to Africa. Unless steps are taken now, it
is only a matter of time before the population explosion extends also to these areas.
The differing patterns of birth-rates and death-rates cause differences in the age structure of
populations (Figure 4). A high death-rate and a high, static birth-rate mean that each age
group comprises fewer people than the one before, giving the pyramidal 'profile' typical of
India and many other countries today and of Great Britain a century ago. The same applies as
a whole to South America, which demographically is essentially a country of the young. The
profile for the United Kingdom, in which five years ago there were about as many people
aged 40 to 45 as there were infants 0 to 5 years old, shows the effect of an erratic but overall
decrease in the birth-rate over many years.
The explosive growth of world-population has not been caused by a sudden increase in human
fertility, and probably owes little in any part of the world to an increase in birth-rate. It has
been caused almost entirely by advances in the medical and ancillary sciences, and the
consequent decrease of the death-rate in areas where the birth-rate remains high. This is of
some biological interest. Nature takes as her motto that nothing succeeds like excess, and any
living thing, including Man, if able to reproduce without restraint to the limit of its capacity,
would soon inundate all parts of the world where it could exist. As it is, biological populations
are kept severely in check by limiting factors, of which the most important are limitations of
food supply, disease and enemies, and fluctuations in natural populations are determined by
fluctuations in these limiting factors. Generally speaking relaxation of one factor, after a
period of expansion, brings into operation one of the other two.
www.seyfihoca.com
227
In the l00 years before the second World War, the expectation of life at birth in England and
Wales rose from about 40 years to over 60 years - that is, one year every five years. In India
the expectation of life at birth is about 40 years, and is said to be increasing by 2½ years every
five years. Even if the birth-rate were at no more than replacement level, the increasing
expectation of life would add enormously to the population of India. True, the expectation of
life is not likely to increase indefinitely at the present velocity, but it has a very long way to
go in many countries of the world. Even in the developed countries it has some way to go
before everyone dies essentially of senility, and in the meantime increasing longevity will
reinforce natural reproductivity. With present birth-rate and death-rate trends, the world is
threatened with astronomical numbers of people. To quote from the preface of a United
Nations report, The Future Growth of World Population (1958) : '. . . it took 200,000 years for
the world's human population to reach 2,500 million; it will now take a mere 30 years to add
another 2,000 million. With the present rate of increase, it can be calculated that in 600 years
the number of human beings on Earth will be such that there will be only one square metre for
each to live on. It goes without saying that this can never take place, something will happen to
prevent it.' The human race will have to decide whether that 'something' is to be pleasant or
unpleasant.
(A. S . Parkes from an article in The New Scientist )
www.seyfihoca.com
228
CHANGES IN ENGLISH SOCIAL LIFE AFTER 1918
The most remarkable outward change of the Twenties was in the looks of women in the
towns. The prematurely aged wife was coming to be the exception rather than the rule.
Children were fewer and healthier and gave less trouble; labour-saving devices were
introduced, especially for washing, cleaning, and cooking-the introduction of stainless plate
and cutlery saved an appreciable amount of time daily and this was only one of a hundred
such innovations. Provisioning also had become very much easier. The advertising of branded
goods was simplifying shopping problems. Housewives came to count on certain brands of
goods, which advertisers never allowed them to forget. The manufacturers' motto was: 'Swear
not by the moon, the inconstant moon, but swear by constant advertising'. They made things
very easy for the housewives by selling their foods in the nearest possible stage to tablereadiness: the complicated processes of making custard, caramel, blanc-mange, jelly, and
other puddings and sweets, were reduced to a single, short operation by the use of prepared
powders. Porridge had once been the almost universal middle class breakfast food. It now no
longer took twenty minutes to cook, Quick Quaker Oats reducing the time to two; but even so,
cereals in the American style, eaten with milk, began to challenge porridge and bacon and
eggs in prosperous homes, and the bread and margarine eaten by the poor. At first the only
choice was Force and Grape-Nuts; but soon there was a bewildering variety of different
www.seyfihoca.com
229
'flakes'; and grains of rice, wheat and barley 'puffed' by being fired at high velocity from a sort
of air gun. Bottled and tinned goods grew more and more various and plentiful. When the war
ended the only choice was soup, salmon, corned beef, Californian fruits, and potted meat; but
by the Thirties almost every kind of domestic and foreign fruit, meat, game, fish, vegetable
could be bought, even in country groceries. Foodstuffs that needed no tin-opener were also
gradually standardized: eggs, milk, and butter were graded and guaranteed and greengrocers
began selling branded oranges and bananas. Housewives could send or ring up for goods
without inspecting them; more and more shops called daily or weekly for orders and delivered
free of charge as light commercial vans displaced the horse and cart. The fish-van brought
fresh fish to the door even in inland towns and villages. The cleanest and neatest shops
secured the best custom; flies and wasps disappeared from grocers' counters, finding no open
pots of treacle or boxes of sugar to attract them, and the butchers began keeping their carcases
in refrigerators out of sight, not suspended bleeding from hooks in the full glare of the sun. By
the Thirties cellophane, a cheap wood-pulp product, was coming into general use for keeping
dry groceries and cigarettes fresh and clean, and soon also covered baskets of strawberries,
lumps of dates, and even kippers and other cured fish.
Woolworth's stores were the great cheap providers of household utensils and materials. There
had been a few '6½d. Bazaars' before the war, but the Woolworth system was altogether new.
It worked by small profits and quick returns in a huge variety of classified and displayed cutprice goods; some, such as excellent glass and hardware, were even sold below cost price to
attract custom. The Daily Herald reported in 1924 that the T.U.C. was reviewing complaints
about working conditions in Woolworth's-'the well-known bazaar-owners'-and that this was
the more serious because the stores were patronized chiefly by the working-class. But the firm
never had any difficulty in engaging unskilled sales-girls at a low wage; for 'the local
Woolworth's' was increasingly the focus of popular life in most small towns. And the name of
Woolworth was a blessed one to the general public; wherever a new branch was opened the
prices of ironmongers, drapers, and household furnishers in the neighbourhood would drop
twopence in the shilling. The middle class at first affected to despise Woolworth's goods, but
they soon caught the working-class habit and would exclaim brightly among themselves: 'My
dear - guess where I got this amazing object - threepence at Maison Woolworth! I don't know
how they do it!'
Woolworth's, the Building Societies, and the Instalment System made it financially possible
for people of small means to take over new houses. The instalment or 'never-never' system
was being applied to all major household purchases, such as furniture, sewing-machines,
vacuum-cleaners, gas-ovens, wireless sets. A Punch illustration showed a young mother,
watching her husband writing out the monthly cheque to pay off the maternity-home debt:
'Darling, only one more instalment and Baby will be ours'.
(From The Long Weekend by Robert Graves and Alan Hodge)
SCIENTIFIC METHOD IN THE SOCIAL SCIENCES
Even the social scientist who is occupied with the study of what are called institutions must
draw his ultimate data (with one important exception mentioned below) from the experience
of the senses. Suppose, for instance, that he is engaged on a study of the role of trade unions
in contemporary England. The abstract conception 'trade union' is simply a shorthand for
certain types of behaviour by certain people, of which we can only be aware by sensory
perception. It means men sitting in a room and making certain sounds in the conduct of a
www.seyfihoca.com
230
'trade union meeting', or handing over to other persons tangible objects (money) as their
subscriptions to the union. Anyone who wishes to make a study of trade unions, or even of the
more abstract conception 'trade unionism', can only do so by personally observing such
behaviour, or by using his eyes and ears on books and speeches made by other people who
have themselves made such observations (or who have in their turn heard or seen records of
such observations made by others). Even such comments on a union meeting as that it was
'orderly' or 'peaceful' are fundamentally statements about its physical properties: an orderly
meeting is presumably one in which people do not make noises by banging upon the table or
speaking very loudly.
This dependence of social studies upon sense perception is certainly a wholesome reminder of
the fundamental homogeneity of the original data of science. For knowledge of the external
world, whether of things or of people, we continually come back to our five senses in the end.
Nevertheless, if a great mass of data relevant to social science is sensory, we have, I think,
also to admit an important collection that is not namely the whole body of primary mental or
psychological experience. Perception of mental pleasure and pain appears to have the same
universality as sensory experience. At all levels of culture, sensations of simple happiness and
unhappiness are as general as are the experiences of seeing and hearing. It is of course true
that no person can experience the feelings of anyone other than himself; but equally no one
can see with another's eyes or hear with another's ears. The grounds for belief in the sense
experiences of other people and the grounds for belief in their primitive psychological
experiences are thus both equally shaky, or equally firm. We derive our conviction that other
people experience emotion from the fact that they say so, and from analogies between their
behaviour and our own: we derive our conviction that they see and hear from exactly the same
evidence.
The irresistibility of psychological experience is perhaps slightly more disputable. If one's
eyes are open and one looks in a certain quarter one cannot help seeing. Is it equally true that
one cannot help a feeling of pleasure or pain or shock or excitement? Essentially, I should say
that it is. But it is clear that primitive emotional reactions can be inhibited: one can, for
example contrive not to be depressed by an event. Nevertheless, if we stand back from all
philosophical niceties and ask ourselves whether psychological sensation ought to be omitted
from the data of the social sciences on the ground that it is doubtfully 'primitive', there cannot,
I think, be much doubt about the answer. We must conclude with Bertrand Russell 'that there
is knowledge of private data, and there is no reason why there should not be a science of
them'. Equally, if we consider whether the similarities or the differences, in this matter of
universality-plus-irresistibility, between psychological and sensory experience are the more
impressive, we are surely bound to come down on the side of the similarities. Certainly, social
studies which consistently ignored human feelings would be worse than laughable.
(From Testament for Social Science by Barbara Wootton)
The Troubles of shopping in Russia
A large crowd gathered outside a photographic studio in Arbat Street, one of the busiest.
shopping streets in Moscow, recently. There was no policeman within sight and the crowd
was blocking the pavement. The centre of attraction - and amusement - was a fairly welldressed man, perhaps some official, who was waving his arm out of the ventilation window of
the studio and begging to be allowed out. The woman in charge of the studio was standing
outside and arguing with him. The man had apparently arrived just when the studio was about
www.seyfihoca.com
231
to close for lunch and insisted upon taking delivery of some prints which had been promised
to him. He refused to wait so the staff had locked the shop and gone away for lunch. The
incident was an extreme example of the common attitude in service industries in the Soviet
Union generally, and especially in Moscow. Shop assistants do not consider the customer as a
valuable client but as a nuisance of some kind who has to he treated with little ceremony and
without concern for his requirements.
For nearly a decade, the Soviet authorities have been trying to improve the service facilities.
More shops are being opened, more restaurants are being established and the press frequently
runs campaigns urging better service in shops and places of entertainment. It is all to no avail.
The main reason for this is shortage of staff. Young people are more reluctant to make a
career in shops, restaurants and other such establishments. Older staff are gradually retiring
and this leaves a big gap. It is not at all unusual to see part of a restaurant or a shop roped off
because there is nobody available to serve. Sometimes, establishments have been known to be
closed for several days because of this.
One reason for the unpopularity of jobs in the service industries is their low prestige. Soviet
papers and journals have reported that people generally consider most shop assistants to be
dishonest and this conviction remains unshakeable. Several directors of business
establishments, for instance, who are loudest in complaining about shortage of labour, are also
equally vehement that they will not let their children have anything to do with trade.
The greatest irritant for the people is not the shortage of goods but the time consumed in
hunting for them and queuing up to buy them. This naturally causes ill-feeling between the
shoppers and the assistants behind the counters, though often it may not be the fault of the
assistants at all. This too, damages hopes of attracting new recruits. Many educated
youngsters would be ashamed to have to behave in such a negative way.
Rules and regulations laid down by the shop managers often have little regard for logic or
convenience. An irate Soviet journalist recently told of his experiences when trying to have an
electric shaver repaired. Outside a repair shop he saw a notice: ‘Repairs done within 45
minutes.’ After queuing for 45 minutes he was asked what brand of shaver he owned. He
identified it and was told that the shop only mended shavers made in a particular factory and
he would have to go to another shop, four miles away. When he complained, the red-faced
girl behind the counter could only tell him miserably that those were her instructions.
All organisations connected with youth, particularly the Young Communist League
(Komsomo1), have been instructed to help in the campaign for better recruitment to service
industries. The Komsomol provides a nicely-printed application form which is given to
anyone asking for a job. But one district head of a distribution organisation claimed that in the
last in years only one person had come to him with this form. ‘We do not need fancy paper.
We do need people!’ he said. More and more people are arguing that the only way to solve the
problem is to introduce mechanisation. In grocery stores, for instance, the work load could be
made easier with mechanical devices to move sacks and heavy packages.
The shortages of workers are bringing unfortunate consequences in other areas. Minor rackets
flourish. Only a few days ago, Pravda, the Communist Party newspaper, carried a long
humorous feature about a plumber who earns a lot of extra money on the side and gets
gloriously drunk every night. He is nominally in charge of looking after 300 flats and is paid
for it. But whenever he has a repair job to do, he manages to screw some more money from
www.seyfihoca.com
232
the flat dwellers, pretending that spare parts are required. Complaints against him have no
effect because the housing board responsible is afraid that they will be unable to get a
replacement. In a few years’ time, things could be even worse if the supply of recruits to these
jobs dries up altogether.
TECHNOLOGY
SEDUCED BY TECHNOLOGY
My neighbour, Nick, is a soft-spoken, easy-going fellow who owns a big, ungainly dog
named Duffy and has a passion for music. He has been a freelance musician all his adult life
and plays the double bass for a living. But things have changed in the music business: it's not
easy to earn a living as a freelance musician today as it was a few years ago.
A lot of the well-paid 'session work' has disappeared and been replaced by pre-programmed,
computerised synthesizers. Nick still plays in the 'pit' when he can, in touring musicals like
Miss Saigon or Phantom of the Opera. But today he also works part-time in a music shop,
helping to ship out trumpets and French horns to school bands and re-stocking inventory
when new shipments arrive.
Nick is not untypical these days. In fact his story is just one of millions that unveil the other
side of the computer revolution - the human costs and consequences of the new 'wired world'
which receive little attention from government bureaucrats or industry boosters.
Fantastic, science-fiction tinged claims about the benefits of the coming 'information age' are
hard to escape. The press is full of hacks extolling the liberating virtues of electronic mail and
tub-thumping about how the Internet will unite the masses in a sort of electronic, Jeffersonian
democracy (at least those with a personal computer, modem and enough spare cash to pay the
monthly hook-up fee).
'If you snooze, you lose' is the underlying message. Jump on board now or be brushed aside as
the new high-tech era reshapes the contours of modern life. This is not the first time that
technology has been packaged as a panacea for social progress. I can still recall a youthful
Ronald Reagan touting for General Electric on American television back in the 1950s:
'Progress is our most important product,' the future President intoned.
That ideology of progress is welded as firmly now as it was to the power-loom in the early
nineteenth century, the automobile in the 1920s or nuclear power in the 1960s. Yet the
introduction of all these technologies had disastrous side effects. The power-loom promised
cheap clothing and a wealthier Britain but produced catastrophic social dislocation and job
loss. The car promised independence and freedom and delivered expressways choked with
traffic, suburbanization, air pollution and destructive wars fought over oil supplies. Nuclear
power promised energy 'too cheap to meter' and produced Chernobyl and Three Mile Island.
There is a lesson here that can and should be applied to all new technologies - and none more
so than computers. One of the last century's more astute analysts of communications
technologies, Marshall McLuhan, said it best: 'We shape our tools and thereafter our tools
shape us.' In his cryptic way McLuhan was simply summing up what a small band of dogged
critics have been saying for decades. Technology is not just hardware - whether it's a hammer,
www.seyfihoca.com
233
an axe or a desk-top PC with muscular RAM and a pentium chip. Limiting it in this way
wrenches technology from its social roots. The conclusion? It's not 'things' that are the
problem, it's people.
This has the simple attraction of common sense. Yet the more complex truth is that
technologies carry the imprint of the cultures from which they issue. They arise out of a
system, a social structure: They are grafted on to it,' argues Canadian scientist Ursula
Franklin, 'and they may reinforce or destroy it, often in ways that are neither foreseen or
foreseeable.' What this means is that technology is never neutral. Even seemingly benign
technologies can have earth-shaking, unintended, social consequences.
The American writer Richard Sclove outlines what happened when water was piped into the
homes of villagers in Ibieca in north eastern Spain in the early 1970s. The village fountain
soon disappeared as the centre of community life when families gradually bought washing
machines and women were released from scrubbing laundry by hand. But at the same time the
village's social bonds began to fray. Women no longer shared grievances and gossip together;
when men stopped using their donkeys to haul water the animals were seen as uneconomical.
Tractors began to replace donkeys for field work, thus increasing the villager's dependence on
outside jobs to earn cash needed to pay for their new machines. The villagers opted for
convenience and productivity. But, concludes Sciove: 'They didn't reckon on the hidden costs
of deepening inequality, social alienation and community dissolution.'
When it comes to introducing new technologies we need to look less at how they influence
our lives as individuals and more at how they impact on society as a whole.
Let's consider computers for a moment. Over the last decade new technologies based on
micro-electronics and digitized data have completely changed the way information is
transmitted and stored, And word processors and electronic mail have made writing and
sending messages around the globe both cheap and quick. 'Surfing the net' (clicking around
the Internet in a random fashion for fun and entertainment) has become the fashionable way to
spend your leisure time. But these are benefits filtered through the narrow prism of personal
gain.
What happens when we step back and examine the broad social impact? How else are
computers used? Let's look at just four examples:
The money maze: The computer that allows us to withdraw cash from an automatic teller, day
or night, is the same technology that makes possible the international capital market. Freed
from the shackles of government regulation corporate money managers now shift billions of
dollars a day around the globe. 'Surfing the yield curve,' big money speculators can move
funds at lightning speed, day and night - destabilizing national economies and sucking
millions out of productive long-term investment. The global foreign exchange trade alone is
now estimated at more than $1.3 billion a day.
Computer games: They come in all shapes and sizes and you can find them as easily in Kuala
Lumpur as in Santiago or Harare. They vary from the jolt-a-second, shoot -'em-up games
(often offensively sexist) to the mesmerizing hand-held and usually more innocuous variety.
Now think of 'Desert Storm', the world's first (and certainly not the last) electronic war. Lethal
firepower as colourful blips on our TV screens, charred bodies reduced to the arching trail of
an explosive starburst.
www.seyfihoca.com
234
The ability to kill and maim large numbers of our fellow human beings is not a new skill.
We've been able to destroy human life many times over for more than half a century and
computers have not changed that reality. What they have done is sideline human decisionmaking in favour of computer programs - making catastrophe ever more likely. As one
software engineer has pointed out, complex computer programs require maintenance just like
mechanical systems. The problem is that 'every feature that is added and every bug that is
fixed adds the possibility of some new interaction between parts of the program' - thus
making the software less rather than more reliable.
Information as power: This is the mantra of those who suggest that both the Internet and the
World Wide Web will establish a new on-line paradigm of decentralized power, placing real
tools for liberation into the hands of the marginalized and the poor. That's a tall order but it is
nonetheless true that the new communications technologies can be used positively by political
dissidents and human-rights activists. Examples abound. At the 1995 UN Conference on
Women in Beijing the proceedings were posted instantaneously over the Net thus bringing
thousands of women, who would have otherwise been left out, into the discussions.
This 'high-tech jujitsu', as critic Jerry Mander calls it is both valiant and necessary. But it
doesn't change the key fact that computers contribute more to centralization than to
decentralization. They help activists, but they help the centralizing forces of corporate
globalization even more. This is what the communications theorist Harold Tunis described as
the 'bias' of technology in the modern era. Computers, as the most powerful of modern
communications tools, reflect their commercial and military origins.
Efficiency and employment:
Technology has always destroyed jobs. In the economy of industrial society that is its main
purpose - to replace labour with machines, thereby reducing the unit cost of production while
increasing both productivity and efficiency. In theory this spurs growth: producing more and
better jobs, higher wages and an increased standard of living. This is the credo of orthodox
economics and there are still many true believers.
But evidence to support this view in the real world of technology is fading fast. More
widespread is the pattern detailed in a recent Statistics Canada report which underlined the
growth of a 'two-tiered' labour market in that country. On the top tier: long hours of overtime
by educated, experienced and relatively well-paid workers. And on the bottom: a large group
of low- paid, unskilled and part-time workers 'who can be treated as roughly interchangeable' .
And then there are those who miss out altogether - the chronic jobless, the socially
marginalized who form a permanent and troubling underclass.
This same trend is repeated throughout the industrialized world. In the US author Jeremy
Rifkin says less than 12 per cent of Americans will work in factories within a decade and less
than two per cent of the global work force will be engaged in factory work by the year 2020.
'Near-workerless factories and virtual companies' are already looming on the horizon, Rifkin
claims. The result? 'Every nation will have to grapple with the question of what to do with the
millions of people whose labour is needed less or not at all in an ever-more-automated
economy.'
Computerization is at the core of the slimmed down, re-engineered workplace that freemarket boosters claim is necessary to survive the new, lean-and-mean global competition.
www.seyfihoca.com
235
Even factory jobs that have relocated to the Third World are being automated quickly. In the
long run machines will do little to absorb the millions of young people in Asia, Africa and
Latin America who will be searching for work in the coming decades. Slowly, that sobering
fact is beginning to strike home. A Chinese government official recently warned that
unemployment in the world's most populous nation could soon leap to 268 million as Chinese
industries modernize and automate.
In the long run computers don't eliminate work, they eliminate workers. But in a social system
based on the buying and selling of commodities this may have an even more pernicious effect.
With fewer jobs there is less money in circulation; market demand slackens, reinforcing
recession and sending the economy into a tailspin. The impact of automation on jobs is a
dilemma which can no longer be ignored.
Though thinkers in the green movement have been grappling with this issue for over a decade,
most governments and even fewer business people are prepared to grasp the nettle. Both cling
to the increasingly flimsy belief that economic growth spurred by an increasing consumption
of the earth's finite resources will solve the problem. It won't. And serious questions need now
to be raised about alternatives.
First, we need to think about democratizing the process of introducing new technologies into
society and into the workplace. At the moment these decisions are left typically in the hands
of bureaucrats and corporations who base their decisions on the narrow criteria of profit and
loss. This bunkered mindset that equates technological innovation with social progress needs
to be challenged.
But there is also the critical issue of the distribution of work and income in a world where
waged labour is in a steady, inexorable decline. We can't continue to punish and stigmatize
those who are unable to find jobs just because there aren't enough to go around. Instead, we
need to think creatively about how to redefine work so that people can find self-esteem and
social acceptance outside of wage labour. This may mean redesigning jobs so that workers
have more control and input into decisions about which technologies to adopt and what
products to make. Up to now this has been exclusively a management prerogative. But it also
means developing strategies to cut the average work week - without cutting pay. This would
be one way of sharing the wealth created by new technology and of creating jobs at the same
time, Hard work also needs to go into designing a plan for a guaranteed annual social wage.
This is a radical (some would say outrageous) idea for societies like ours that have anchored
their value systems on the bedrock of wage labour.
But how can we deny people the basic rights of citizenship and physical well-being simply
because the economic system is no longer capable of providing for them?
(By: Wayne. Elwood, New Internationalist, 12,1996)
Blowing hot and cold on British windmills
Last year the Department of Energy and the Science Research Council together spent less
than £1 million in research into wind energy, although £100 million each year goes into
nuclear research and development. In sharp contrast the USA has been spending some 60
million dollars each year on wind energy and now plans a 1,000 million dollar demonstration
and "commercialisation" programme.
www.seyfihoca.com
236
In Germany, Denmark, and Sweden large programmes are under way and a number of
megawatt size windmills have been or are being built. In the USA, competitive electricity
prices are already envisaged when the present models of machines can be produced in large
numbers (bringing down costs). The British Wind Energy Association brings together
scientists, engineers, and entrepreneurs from industry, universities, and various Government
bodies. With some 120 members, it now presents a respected view on the subject of wind
energy.
But the activity has remained small, and in my view seriously under-funded compared with
many of the other developed countries of the world. The UK is still without a single really
large windmill (in the megawatt range) and this means that we are failing to build up the
practical experience which is essential if any serious progress is to be made.
We in the UK are doing some very nice work on many of the associated problems of wind
energy. But none of these "generic studies" can replace real life experience with one or two
very big windmills.
It is in the light of this background that we must examine the CEGB decision to press ahead
with a rather complete and ambitious programme.
1. An island site is to be sought and a megawatt size windmill is to be purchased and
erected by 1985.
2. A smaller (100kW) windmill is to be bought and set up, as soon as possible, so that
an early experience may be built up in the CEGB of windmill operation and
integration into the grid.
3. International collaboration will be sought for research on offshore windmills.
It is my guess that the CEGB might well wish to build its first windmill group or "cluster" in
the period 1985-1990. This first cluster might have 10 machines in it. In the long term we are
thinking, of course, of clusters of 400-1,000 machines (each of about 4MW). Such a cluster
would provide an output similar in magnitude to a modern coal fired or nuclear powered
station.
The importance of this CEGB decision is that the United Kingdom is at last moving forward
towards the building of its first multi-megawatt windmill station. This is a milestone for those
who believe that the so called "renewable" energy sources (wind, wave and sun) have an
important part to play in our energy future. Even more important from the point of view of
UK industry is the fact that it is the potentially largest UK customer (the CEGB) who is taking
the initiative.
Meanwhile, why has the CEGB opted for a lowland windmill? Good lowland sites offer
average wind speeds of about six metres per second, whereas hilltop and offshore sites can
offer average wind speeds in excess of eight metres/second. This ratio of 1.33 in windspeed
actually represents a factor of 1.33 cubed (i.e. about 2.5) in available energy, for a given size
of windmill.
Three or four years ago UK interest centred on hilltop and coastal sites. Developments since
1977 have greatly changed the picture. Dr Peter Musgrove pioneered in the UK the idea of
putting windmills in the shallow waters of the North Sea. He pointed out that there are vast
areas of shallow waters (less than 30 metres deep) off the east coast of the UK. He showed
www.seyfihoca.com
237
that a number of windmill clusters in these shallow waters could meet up to 30 per cent of UK
annual electricity needs.
To their credit the Department of Energy took up these ideas and funded a detailed study. No
insuperable technical difficulties were found and the estimated building and running costs
would indicate a price for electricity which could become competitive in the near future with
nuclear or coal-generated electricity.
The CEGB has given the production costs for electricity for stations currently being built.
Nuclear power from Dungeness B is put at 2.62p/kWh: coal at Drax B is 3.59p/kWh: oil at
Littlebrook D is 6.63p/kWh.
The Taylor Woodrow led study came up with a number of different figures for the cost of
North Sea electricity depending on the assumptions made. Let me pick out the figure of
4.20p/kWh which was based on the windmills having a diameter of 100 metres (which is the
largest size presently being built in the world). The figure taken for average wind speed is
9.5m/s.
Thus North Sea-generated electricity looks like being close to competitive on present fuel
costs. This figure of 4.2p/kWh could come down dramatically if we find that we are able to
build much larger diameter windmills than the present 100 metres. This is because foundation
and lower costs were dominating the picture. Larger windmills would mean fewer windmills
and hence lower overall foundation and tower expenditure.
The CEGB has chosen, in the meantime, to go for the on-land option for their initial
programme. This is an eminently sensible decision. A number of windmill designs have been
developed in the USA and elsewhere for machines to operate in moderate wind regions. The 6
metres/second average wind speeds that we get in many lowland areas of the UK are
considered to be satisfactory speeds for wind turbine operation in the USA. Given a lower
wind speed you simply design a larger diameter wind turbine.
What now remains to be seen is not so much whether you can build large windmills or
whether they will be economically viable, but whether they will be environmentally and
socially acceptable, placed for example in the windy lowlands of Lincolnshire and East
Anglia. The CEGB search for its first site should bring out some interesting attitudes.
I have seen the 200 ft. Mod 1 Windmill which is on a hilltop near the small town of Boone in
North Carolina. Even from a distance of only two miles it is far from obtrusive. In fact I found
it a most attractive sight - but perhaps I am biased. The locals nonetheless are very proud of it!
Professor N. H. Lipman of the Department of Engineering at Reading University, is a member
of the Reading Energy Group.
(From an article by Norman Lipman in The Guardian)
Direct use of solar radiation
Simply because a very large energy flux falls on Britain through the year, it is wrongly
assumed by many that the life to which they have become accustomed can be supported with
solar energy. Even with the reduced level of energy use we managed to achieve with proper
www.seyfihoca.com
238
insulation, the average level of radiation occurring during the two months when we need
warmth most - December and January - is down to an average over the period of 15 watts per
square metre. This is vanishingly small. To make any use of such small radiation requires that
correspondingly very large areas be set aside to low temperature collection, which will be
redundant during the summer months when average radiation, over the month of June for
example, is 225 watts per square metre. Thus, unless you can set aside at least 40 square
metres of cheap solar collecting surface, facing due south at an angle of around 70° to the
horizontal, it is best to forget energy self-sufficiency from the sun alone.
It is not so silly, however, to use the sun when it does shine hot to provide between 50 and
90% of your hot water, usually during the months of April to September when it is warm
enough. 4 square metres of collecting surface plumbed into the hot water system of your
house can provide up to 100kw-h/m² per year. At 1p per kw-h this could be worth £24 a year,
and at 2p per kwh, which is the price we are likely to be paying soon, it would be worth £48 a
year.
For anyone wishing to embark on this course it should be remembered that while solar
collectors are very efficient in the lower temperature ranges up to around 10-17°C, this
'efficiency' falls off very quickly when the temperature of the heat transfer fluid inside the
collector rises above 35-40°C. In tests carried out on an unglazed solar collector we found that
we could achieve an average of about 50% efficiency at 40°C. By putting a clear plastic
glazing on top we increased the average efficiency from about 50 to 60% in the higher
temperature ranges. It is really not worth spending a great deal on glazing, since the extra
energy you collect will take years to pay back the additional cost of the framing and glazing,
in terms of hot water.
I doubt whether it is worth spending money on making your own solar collector. The cheapest
do-it-yourself collector is a scrap steel radiator stripped down and painted black. Insulation
and aluminium foil underneath stops radiation and conduction losses. So far as possible,
collectors should face south. The cheapest place to site them is on the south wall of the house.
The most expensive place is on the south roof, since you will need the services, or skills, of a
roofer and carpenter, in addition to those of a plumber and electrician, unless you have these
skills yourself.
Industrial Robots
There are few micro-electronic applications more likely to raise fears regarding future
employment opportunities than robots for the very obvious reason that such machines directly
replace human labour. The emotive nature of the subject inevitably gives rise to
misapprehensions.
It is necessary first to define an industrial robot. Alternative definitions and classifications
abound but basically a robot is a machine which moves, manipulates, joins or processes
components in the same way as human hand or arm. It consists basically of three elements:
the mechanical structure (including the artificial wrist and gripper), the power unit (hydraulic,
pneumatic or, increasingly, electrical) and the control system (increasingly mini-computers
and microprocessors). However, the essential characteristic of a robot is that it can be
programmed. Thus many devices (often called robots) would be better termed 'numerically
controlled arms', since they are mechanical arms controlled by rudimentary (non-computer)
software and as such are not radically different to much existing automation equipment. There
www.seyfihoca.com
239
are reportedly about 20,000 of the latter in use in Japan, and perhaps several thousand in the
United Kingdom. A robot, however, is here defined as a hybrid of mechanical, electrical and
computing engineering.
Most robots in current use handle fairly straightforward tasks such as welding and spraying
where the software programmes controlling the machines are not very complex. However, the
newer machines, usually referred to as 'universal' but which are still under development, will
be able to perform more complex assembly tasks (for example, carburettor assembly).
Table 1 gives some world-wide estimates of robot diffusion. The table is based on a number
of different studies and must be treated with caution since there are problems of definition:
some companies producing medium technology robots do not classify them as robots; and
many large companies are known to have developed robots in-house, but there is little
statistical information on them.
Table l Estimates of the international diffusion of robots in 1978
Japan
3500
USA
2500
Europe
2000
of which
West Germany
600
Sweden
600
Italy
300-400
France
200
United Kingdom
100
World-wide
8000
The distribution of robots according to applications is again very difficult to estimate, but
figures released by the Unimate Company of America indicate that by June 1977 they had
installed around 1600 machines world-wide. Of these, virtually 50 per cent were used for spot
welding, about 11 per cent for die-casting and another 5 per cent for machine loading. Other
surveys have suggested that a significant proportion of robots are used for coating, mainly for
paint spraying.
What is clear from surveys to date is that the automobile and metal working and forming
industries have been the biggest users of robots. Indeed, the relative decline of these industries
in Britain is probably a significant cause of the relatively small use of robots in this country.
With so little happening in Britain, and indeed only marginal penetration world-wide, it is
difficult to generalize either upon current motives for investment in them or indeed the likely
pattern of future applications; but obviously these are the matters that will determine the
employment impact.
The reasons most commonly cited for the introduction of robots into the work place are:
1. improvement in productivity both in work-rate and quality of output;
2. improvement of working conditions, most usually where health hazards are
involved;
3. improved flexibility of production systems;
www.seyfihoca.com
240
4. greater effective management for production control.
Balanced against these, however, are factors limiting the spread of robots such as:
1.
2.
3.
4.
high price;
hardware and software problems;
lack of knowledge, and perhaps even fear of robots;
competition from alternative automation systems.
It might be thought that union resistance should be added to the list; but it does not appear to
be a restraining influence overseas, and in our discussions on the subject with British
manufacturers it has not been placed high on their list of constraints. More often it has been
stressed that the relatively low cost of labour in this country greatly reduces the economic
justification for investment in robots.
(From The Manpower Implications of Micro-Electronic Technology, MSC.)
Tomorrow's Phone Calls
One day we all may find it useful to have a facility for sending documents, writing and
pictures across the telephone lines. A detector at the sending end would quickly transmit the
signals representing the document to a printer at the receiving end where the documents
would be accurately and quickly reproduced.
View-phone would become an easy facility to provide; not that the findings in America
indicate an overwhelming demand for us all to be seen as well as heard on the telephone.
Conference telephone facilities could become widely available and multi-channel cable
television could also consume some of that capacity, but perhaps the greatest use, initially at
least, will be in what our Post Office now calls Prestel.
The system will link the subscriber's television set to a public computer via his telephone. By
dialling the appropriate number on the telephone and connecting the television receiver to the
circuit, the user will have at his disposal sixty thousand or so pages of information, and all
sorts of services which the telephone and the television on their own could never provide.
The television receiver is fitted with a small memory capable of storing the digital
information necessary to generate one page of display. The conversion to Prestel also requires
fitting the television with a key pad, rather like the finger panel of a pocket calculator.
On dialling the central computer the massive memory there transmits the signals to the small
memory on the television set. These signals generate the first page of the instruction
sequence. The picture welcomes you to Prestel and gives you any message it has stored for
you in its memory. You, of course, can leave a message for someone else to pick up. Prestel
talks to you in that strange computer fashion of asking you questions to which you can answer
'Yes', 'No', or give a number.
By asking questions the Prestel computer finds out what services you require, whether you
want to find out the closing prices on the Stock Exchange, what is on at the theatre or
something more complicated. Your answers on your key pad dictate when the memory in
www.seyfihoca.com
241
your set will receive a new set of instructions from central control and what those instructions
will be.
Leaving a complicated message may prove difficult but as long as you are prepared to accept
the alternatives Prestel offers, your wife can, for instance, learn whilst you are on your way,
that you are arriving at the station at the time you have keyed in from your number pad, or
that you are not coming home at all.
Then there are more subtle applications. On file in the main memory could be a whole host of
valuable data ranging from, for instance, the Highway Code, to when you are likely to be
approved for a mortgage. Because the computer can respond to your 'Yes', 'No' or number
answers, it can actually give you advice.
One programme I tried was designed to help those anxious to adopt a child. The first question
was, 'Are you applying on behalf of yourself only? If so key zero, if not key one.' I keyed
zero, for the sake of argument, and up came question two. 'Are you the parent of the child?'
Again zero for no. 'Are you over twenty-five?' I lied a little and said 'no' again and this proved
too much for the computer. 'You are not eligible to adopt,' came back the answer.
If I had told the truth about being over twenty-five that would have made a difference and
another question might have emerged before a final answer was given. Exactly the same
technique is being used in some hospitals now for routine diagnosis of patients' complaints.
The computers in this case are not public ones on an open network, but maybe the time will
come when the Prestel computer will also tell you what is the matter with you, even if it takes
a while longer for it actually to prescribe a treatment.
Prestel will also play games with you. Computerised noughts and crosses, mazes, and
problems like balancing fuel consumption in the retro-rockets of your Mars lander against the
gravitational pull of the planet so that you land at a comfortable speed without running out of
fuel, are all in the compendium at computer headquarters.
In fact it is easy to forget that what Prestel is really trying to do is get you to use the phone,
The Post Office does not want to engage in games or even take over completely from the
Citizen's Advice Bureau, but it is in business to sell telephone calls, and the advent of Prestel
gives it a mighty potent marketing weapon. But Prestel and all other new services which will
emerge in its wake can only work if the capacity for these extra services is built into the
system. It is optical fibre communication which promises to make that possible.
(From Tomorrow's World by M. Blackstad, BBC Publications)
Coal
The rapid growth of steam power relied directly on large supplies of its only fuel: coal. There
was a great increase in the amount of coal mined in Britain.
Coal was still cut by hand and the pits introduced few really new techniques. The increased
demand was met by employing more miners and by making them dig deeper. This was made
possible by more efficient steam pumps and steam-driven winding engines which used wire
ropes to raise the coal to the surface, by better ventilation and by the miner's safety lamp
which detected some dangerous gases. By the 1830s scams well below 1000 feet were being
www.seyfihoca.com
242
worked in south Durham and the inland coalfields of Lancashire and Staffordshire. Central
Scotland and south Wales were mined more intensively later. By the end of the nineteenth
century the best and most accessible seams were worked out. As the miners followed the
eastern scams of the south Yorkshire-north Midlands area, which dipped further below the
surface, shafts of 3,000 feet were not uncommon.
Some of the work in mines was done by women and children. Boys and girls were often put in
charge of the winding engines or of opening and shutting the trap doors which controlled the
ventilation of the mines. Then they had to crouch all day in the same spot by themselves in the
dark. When these evils were at last publicized in 1842 by a Royal Commission, many mines
no longer employed women, but Parliament made it illegal for them all. It also forbade them
to employ boys under the age of ten. The limit, which was very difficult to enforce, was
increased to twelve in the 1870s. Subsequently it rose with the school leaving age.
Mining was very dangerous. Loose rocks were easily dislodged and the risk of being killed or
injured by one was always greater in the tall scams where they had further to fall. In the north
of England fatal accidents were not even followed by inquests to discover why they had
happened until after 1815. Few safety precautions were taken before the mid-nineteenth
century. The mine owners insisted that they were not responsible. The men were most
reluctant to put up enough props to prevent the roof from falling in and to inspect the winding
gem: and other machinery on which their lives depended. If they did, they spent less time
mining and so earned less money because the miners' pay was based not on how long they
worked but on how much coal they extracted. They preferred to take risks.
The deeper seams contained a dangerous gas called 'fire-damp' which could be exploded by
the miners' candles. The safety lamp, which was invented in the early nineteenth century, did
not really solve this problem, but it was often used to detect gas and so made the mining of
deeper seams possible. There the air was more foul, the temperature higher (one pit paid the
men an extra 6d a day for working in 130°F) and the risk of fire-damp even greater. In the
1840s a series of terrible explosions in the deeper mines led to stricter regulations, which
inspectors helped enforce. The inspectors were particularly keen on proper ventilating
machines and, although deeper shafts were sunk, they did not become more dangerous.
However, many serious accidents still occurred.
(From Britain Transformed, Penguin Books)
The Medium Is the Message
In a culture like ours, long accustomed to splitting and dividing all things as a means of
control, it is sometimes a bit of a shock to be reminded that, in operational and practical fact,
the medium is the message. This is merely to say that the personal and social consequences of
any medium - that is, of any extension of ourselves - result from the new scale that is
introduced into our affairs by each extension of ourselves, or by any new technology. Thus,
with automation, for example, the new patterns of human association tend to eliminate jobs, it
is true. That is the negative result. Positively, automation creates roles for people, which is to
say depth of involvement in their work and human association that our preceding mechanical
technology had. destroyed. Many people would be disposed to say that it was not the
machine, but what one did with the machine, that was its meaning or message. In terms of the
ways in which the machine altered our relations to one another and to ourselves, it mattered
not in the least whether it turned out cornflakes or Cadillacs. The restructuring of human work
www.seyfihoca.com
243
and association was shaped by the technique of fragmentation that is the essence of machine
technology. The essence of automation technology is the opposite. It is integral and
decentralist in depth, just as the machine was fragmentary, centralist, and superficial in its
patterning of human relationships.
The instance of the electric light may prove illuminating in this connection. The electric light
is pure information. It is a medium without a message, as it were, unless it is used to spell out
some verbal ad or name. This fact, characteristic of all media, means that the "content" of any
medium is always another medium. The content of writing is speech, just as the written word
is the content of print, and print is the content of the telegraph. If it is asked, "What is the
content of speech?," it is necessary to say, "It is an actual process of thought, which is in itself
nonverbal." An abstract painting represents direct manifestation of creative thought processes
as they might appear in computer designs. What we are considering here, however, are the
psychic and social consequences of the designs or patterns as they amplify or accelerate
existing processes. For the "message" of any medium or technology is the change of scale or
pace or pattern that it introduces into human ; affairs. The railway did not introduce
movement or transportation or wheel or road into human society, but it accelerated and
enlarged the scale of previous human functions, creating totally new kinds of cities and new
kinds of work and leisure. This happened whether the railway functioned in a tropical or a
northern environment, and is quite independent of the freight or content of the railway
medium. The airplane, on the other hand, by accelerating the rate of transportation, tends to
dissolve the railway form of city, politics, and association, quite independently of what the
airplane is used for.
Let us return to the electric light. Whether the light is being used for brain surgery or night
baseball is a matter of indifference. It could be argued that these activities are in some way the
"content" of the electric fight, since they could not exist without the electric light. This fact
merely underlines the point that "the medium is the message" because it is the . medium that
shapes and controls the scale and form of human association and action. The content or uses
of such media are as diverse as they are ineffectual in shaping the form of human association.
Indeed, it is only too typical that the "content" of any medium blinds us to the character of the
medium. It is only today that industries have become aware of the various kinds of business in
which they are engaged. When IBM discovered that it was not in the business of making
office equipment or business machines, but that it was in the business of processing
information, then it began to navigate with clear vision. The General Electric Company makes
a considerable portion of its profits from electric light bulbs and lighting systems. It has not
yet discovered that, quite as much as A.T.&T., it is in the business of moving information.
The electric light escapes attention as a communication medium just because it has no
"content." And this makes it an invaluable instance of how people fail to study media at all.
For it is not till the electric light is used to spell out some brand name that it is noticed as a
medium. Then it is not the light but the "content" (or what is really another medium) that is
noticed. The message of the electric light is like the message of electric power in industry,
totally radical, pervasive, and decentralized. For electric light and power are separate from
their uses, yet they eliminate time and space factors in human association exactly as do radio,
telegraph, telephone, and TV, creating involvement in depth.
(From Understanding media by Marshall McLuhan)
THE DEVELOPMENT OF ELECTRICITY
www.seyfihoca.com
244
The phenomenon which Thales had observed and recorded five centuries before the birth of
Christ aroused the interest of many scientists through the ages. They made various practical
experiments in their efforts to identify the elusive force which Thales had likened to a 'soul'
and which we now know to have been static electricity.
Of all forms of energy, electricity is the most baffling and difficult to describe. An electric
current cannot be seen. In fact it does not exist outside the wires and other conductors which
carry it. A live wire carrying a current looks exactly the same and weighs exactly the same as
it does when it is not carrying a current. An electric current is simply a movement or flow of
electrons.
Benjamin Franklin, the American statesman and scientist born in Boston in 1706, investigated
the nature of thunder and lightning by flying a child's kite during a thunderstorm. He had
attached a metal spike to the kite, and at the other end of the string to which the kite was tied
he secured a key. As the rain soaked into the string, electricity flowed freely down the string
and Franklin was able to draw large sparks from the key. Of course this could have been very
dangerous, but he had foreseen it and had supported the string through an insulator. He
observed that this electricity had the same properties as the static electricity produced by
friction.
But long before Franklin many other scientists had carried out research into the nature of
electricity.
In England William Gilbert (1544-1603) had noticed that the powers of attraction and
repulsion of two non-metallic rods which he had rubbed briskly were similar to those of
lodestone and amber - they had acquired the curious quality we call magnetism.
Remembering Thales of old he coined the word 'electricity'.
Otto von Guericke (1602-1686) a Mayor of Magdeburg in Germany, was an amateur scientist
who had constructed all manner of gadgets. One of them was a machine consisting of two
glass discs revolving in opposite directions which produced high voltage charges through
friction. Ramsden and Wimshurst built improved versions of the machine.
A significant breakthrough occurred when Alessandro Volta (1745-1827) in Italy constructed
a simple electric cell (in 1799) which produced a flow of electrons by chemical means. Two
plates, one of copper and the other of zinc, were placed in an acid solution and a current
flowed through an external wire connecting the two plates. Later he connected cells in series
(voltaic pile) which consisted of alternate layers of zinc and copper discs separated by flannel
discs soaked in brine or acid which produced a higher electric pressure (voltage). But Volta
never found the right explanation of why his cell was working. He thought the flow of electric
current was due to the contact between the two metals, whereas in fact it results from the
chemical action of the electrolyte on the zinc plate. However, his discovery proved to be of
incalculable value in research, as it enabled scientists to carry out experiments which led to
the discoveries of the heating, lighting, chemical and magnetic effects of electricity.
One of the many scientists and physicists who took advantage of the 'current electricity' made
possible by Volta's cells was Hans Christian Oersted (1777-1851) of Denmark. Like many
others he was looking for a connection between the age-old study of magnetism and
electricity, but now he was able to pass electric currents through wires and place magnets in
www.seyfihoca.com
245
various positions near the wires. His epoch-making discovery which established for the first
time the relationship between magnetism and electricity was in fact an accident.
While lecturing to students he showed them that the current flowing in a wire held over a
magnetic compass needle and at right angles to it (that is east-west) had no effect on the
needle. Oersted suggested to his assistant that he might try holding the wire parallel to the
length of the needle (north- south) and hey presto, the needle was deflected! He had stumbled
upon the electromagnetic effect in the first recorded instance of a wire behaving like a magnet
when a current is passed through it.
A development of Oersted's demonstration with the compass needle was used to construct the
world's first system of signaling by the use of electricity.
In 1837 Charles Wheatstone and William Cooke took out a patent for the world's first Fiveneedle Telegraph, which was installed between Paddington railway station in west London
and West Drayton station a few miles away. The five copper wires required for this system
were embedded in blocks of wood.
Electrolysis, the chemical decomposition of a substance into its constituent elements by the
action of an electric current, was discovered by the English chemists Carlisle and William
Nicholson (1753-1815). If an electric current is passed through water it is broken down into
the two elements of which it is composed -- hydrogen and oxygen. The process is used
extensively in modern industry for electroplating. Michael Faraday (1791-1867) who was
employed as a chemist at the Royal Institution, was responsible for introducing many of the
technical terms connected with electrolysis, like electrolyte for the liquid through which the
electric current is passed, and anode and cathode for the positive and negative electrodes
respectively. He also established the laws of the process itself. But most people remember his
name in connection with his practical demonstration of electromagnetic induction.
In France Andre-Marie Ampere (1775-1836) carried out a complete mathematical study of the
laws which govern the interaction between wires carrying electric currents.
In Germany in 1826 a Bavarian schoolmaster Georg Ohm (1789- 1854) had defined the
relationship between electric pressure (voltage), current (flow rate) and resistance in a circuit
(Ohm's law) but 16 years had to elapse before he received recognition for his work.
Scientists were now convinced that since the flow of an electric current in a wire or a coil of
wire caused it to acquire magnetic properties, the opposite might also prove to be true: a
magnet could possibly be used to generate a flow of electricity.
Michael Faraday had worked on this problem for ten years when finally, in 1830, he gave his
famous lecture in which he demonstrated, for the first time in history, the principle of
electromagnetic induction. He had constructed powerful electromagnets consisting of coils of
wire. When he caused the magnetic lines of force surrounding one coil to rise and fall by
interrupting or varying the flow of current, a similar current was induced in a neighbouring
coil closely coupled to the first.
The colossal importance of Faraday's discovery was that it paved the way for the generation
of electricity by mechanical means. However, as can be seen from the drawing, the basic
generator produces an alternating flow of current.(A.C.)
www.seyfihoca.com
246
Rotating a coil of wire steadily through a complete revolution in the steady magnetic field
between the north and south poles of a magnet results in an electromotive force (E.M.F.) at its
terminals which rises in value, falls back to zero, reverses in a negative direction, reaches a
peak and again returns to zero. This completes one cycle or sine wave. (1Hz in S.I. units).
In recent years other methods have been developed for generating electrical power in
relatively small quantities for special applications. Semiconductors, which combine heat
insulation with good electrical conduction, are used for thermoelectric generators to power
isolated weather stations, artificial satellites, undersea cables and marker buoys. Specially
developed diode valves are used as thermionic generators with an efficiency, at present, of
only 20% but the heat taken away from the anode is used to raise steam for conventional
power generation.
Sir Humphry Davy (1778-1829) one of Britain's leading chemists of the 18th century, is best
remembered for his safety lamp for miners which cut down the risk of methane gas explosions
in mines. It was Davy who first demonstrated that electricity could be used to produce light.
He connected two carbon rods to a heavy duty storage battery. When he touched the tips of
the rods together a very bright white light was produced. As he drew the rods apart, the arc
light persisted until the tips had burnt away to the critical gap which extinguished the light. As
a researcher and lecturer at the Royal Institution Davy worked closely with Michael Faraday
who first joined the institution as his manservant and later became his secretary. Davy's
crowning honour in the scientific world came in 1820, when he was elected President of the
Royal Society.
In the U.S.A. the prolific inventor Thomas Alva Edison (1847-1831) who had invented the
incandescent carbon filament bulb, built a number of electricity generators in the vicinity of
the Niagara Falls. These used the power of the falling water to drive hydraulic turbines which
were coupled to the dynamos. These generators were fitted with a spinning switch or
commutator (one of the neatest gadgets Edison ever invented) to make the current flow in
unidirectional pulses (D.C.) In 1876 all electrical equipment was powered by direct current.
Today mains electricity plays a vital part in our everyday lives and its applications are
widespread and staggering in their immensity. But we must not forget that popular demand
for this convenient form of power arose only about 100 years ago, mainly for illumination.
Recent experiments in superconductivity, using ceramic instead metal conductors have given
us an exciting glimpse into what might be achieved for improving efficiency in the
distribution of electric power.
Historians of the future may well characterise the 20th century as 'the century of electricity &
electronics'. But Edison's D.C. generators could not in themselves, have achieved the
spectacular progress that has been made. All over the world we depend totally on a system of
transmitting mains electricity over long distances which was originally created by an amazing
inventor whose scientific discoveries changed, and are still changing, the whole world. His
name was scarcely known to the general public, especially in Europe, where he was born.
Who was this unknown pioneer? Some people reckon that it was this astonishing visionary
who invented wireless, remote control, robotics and a form of X-ray photography using high
frequency radio waves. A patent which he took out in the U.S.A. in 1890 ultimately led to the
design of the humble ignition coil which energises billions and billions of spark plugs in all
www.seyfihoca.com
247
the motor cars of the world. His American patents fill a book two inches thick. His name was
Nicola Tesla (1856-1943).
Nicola Tesla was born in a small village in Croatia which at that time formed part of the great
Austro-Hungarian Empire. Today it is a northern province of Yugoslavia, a state created after
the 1914-1918 war. Tesla studied at the Graz Technical University and later in Budapest.
Early in his studies he had the idea that a way had to be found to run electric motors directly
from A.C. generators. His professor in Graz had assured him categorically that this was not
possible. But young Tesla was not convinced. When he went to Budapest he got a job in the
Central Telegraph Office, and one evening in 1882, as he was sitting on a bench in the City
Park he had an inspiration which ultimately led to the solution of the problem.
Tesla remembered a poem by the German poet Goethe about the sun which supports life on
the earth and when the day is over moves on to give life to the other side of the globe. He
picked up a twig and began to scratch a drawing on the soil in front of him. He drew four coils
arranged symmetrically round the circumference of a circle. In the centre he drew a rotor or
armature. As each coil in turn was energised it attracted the rotor towards it and the rotary
motion was established. When he constructed the first practical models he used eight, sixteen
and even more coils. The simple drawing on the ground led to the design of the first induction
motor driven directly by A.C.electricity.
Tesla emigrated to the U.S.A. in 1884. During the first year he filed no less than 30 patents
mostly in relation to the generation and distribution of A.C. mains electricity. He designed
and built his 'A.C.Polyphase System' which generated three-phase alternating current at 25
Hz. One particular unit delivered 422 amperes at 12,000 volts. The beauty of this system was
that the voltage could be stepped down using transformers for local use, or stepped up to
many thousands of volts for transmission over long distances through relatively thin
conductors. Edison's generating stations were incapable of any such thing.
Tesla signed a lucrative contract with the famous railway engineer George Westinghouse, the
inventor of the Westinghouse Air Brake which is used by most railways all over the world to
the present day. Their generating station was put into service in 1895 and was called the
Niagara Falls Electricity Generating Company. It supplied power for the Westinghouse
network of trains and also for an industrial complex in Buffalo, New York.
After ten years Tesla began to experiment with high frequencies. The Tesla Coil which he had
patented in 1890 was capable of raising voltages to unheard of levels such as 300,000 volts.
Edison, who was still generating D.C., claimed A.C. was dangerous and to prove it contracted
with the government to produce the first electric chair using A.C. for the execution of
murderers condemned to death. When it was first used it was a ghastly flop. The condemned
man moaned and groaned and foamed at the mouth. After four minutes of repeated
application of the A.C.voltage smoke began to come out of his back. It was obvious that the
victim had suffered a horribly drawn-out death.
Tesla said he could prove that A.C. was not dangerous. He gave a demonstration of high
voltage electricity flowing harmlessly over his body. But in reality, he cheated, because he
had used a frequency of 10,000 cycles (10 kHz) at extremely low current and because of the
skin effect suffered no harm.
www.seyfihoca.com
248
One of Tesla's patents related to a system of lighting using glass tubes filled with fluorine (not
neon) excited by H.F.voltages. His workshop was lit by this method. Several years before
Wilhelm Roentgen demonstrated his system of X-rays Tesla had been taking photographs of
the bones in his hand and his foot from up to 40 feet away using H.F.currents.
More astonishing still is the fact that in 1893, two years before Marconi demonstrated his
system of wireless signaling, Tesla had built a model boat in which he combined power to
drive it with radio control and robotics. He put the small boat in a lake in Madison Square
Gardens in New York. Standing on the shore with a control box, he invited onlookers to
suggest movements. He was able to make the boat go forwards and backwards and round in
circles. We all know how model cars and aircraft are controlled by radio today, but when
Tesla did it a century ago the motor car had not been invented, and the only method by which
man could cover long distances was on horseback!
Many people believe that a modification of Tesla's 'Magnifying Transmitter' was used by the
Soviet Union when suddenly one day in October 1976 they produced an amazing noise which
blotted out all radio transmissions between 6 and 20 MHz. (The Woodpecker) The B.B.C., the
N.B.C. and most broadcasting and telecommunication organisations of the world complained
to Moscow (the noise had persisted continuously for 10 hours on the first day), but all the
Russians would say in reply was that they were carrying out an experiment. At first nobody
seemed to know what they were doing because it was obviously not intended as another form
of jamming of foreign broadcasts, an old Russian custom as we all know.
It is believed that in the pursuit of his life's ambition to send power through the earth without
the use of wires, Tesla had achieved a small measure of success at E.L.F. (extremely low
frequencies) of the order of 7 to 12 Hz. These frequencies are at present used by the military
for communicating with submarines submerged in the oceans of the world.
Tesla's career and private life have remained something of a mystery. He lived alone and
shunned public life. He never read any of his papers before academic institutions, though he
was friendly with some journalists who wrote sensational stories about him. They said he was
terrified of microbes and that when he ate out at a restaurant he would ask for a number of
clean napkins to wipe the cutlery and the glasses he drank out of. For the last 20 years of his
life until he died during World War II in 1943 he lived the life of a semi-recluse, with a
pigeon as his only companion. A disastrous fire had destroyed his workshops and many of his
experimental models and all his papers were lost for ever.
Tesla had moved to Colorado Springs where he built his largest ever coil which was 52 feet in
diameter. He studied all the different forms of lightning in his unsuccessful quest for the
transmission of power without wires.
In Yugoslavia, Tesla is a national hero and a well-equipped museum in Belgrade contains
abundant proof of the genius of this extraordinary man.
(From: The dawn of amateur radio in the U.K. and Greece: a personal view by Norman F.
Joly.)
The Autonomous House
www.seyfihoca.com
249
The autonomous house on its site is defined as a house operating independently of any inputs
except those of its immediate environment. The house is not linked to the mains services of
gas, water, electricity or drainage, but instead uses the income-energy sources of sun, wind
and rain to service itself and process its own wastes. In some ways it resembles a land-based
space station which is designed to provide an environment suitable for life but unconnected
with the existing life-support structure of Earth. The autonomous house uses the life-giving
properties of the Earth but in so doing provides an environment for the occupants without
interfering with or altering these properties.
Although the self-serviced house provides a useful starting-point for experiments in
autonomy, as it forms a small unit that can be designed, built and tested within a relatively
short time, the idea can be expanded to include self-sufficiency in food, the use of on-site
materials for building and the reduction of the building and servicing technology to a level
where the techniques can be understood and equipment repaired by a person without recourse
to specialized training. Although it is possible to survive with pre-industrial technology, this
is not what is proposed by autonomous living. At present, however, technology appears to be
exploited for its own sake, without thought to its benefits, uses or effects on people or the
external environment. We are persuaded to expect a higher material standard of living when,
for the majority, the standard that we already have in the West is perfectly adequate. A
marginal increase in this standard can only be made with the use of yet greater quantities of
the existing resources of the Earth. What are essentials for the American way of life (full
central heating, air conditioning, a car per person) are considered, albeit less so now, as
luxuries for Europeans, and what are considered necessary for a satisfactory European life
(enough to eat, a home and fuel to heat it, access to transport) would be luxuries for the ‘third
world’. If we cannot find a way of levelling standards rationally while there is time left to
consider the problem, then the levelling may be forced on us as the lack of fossil fuels on
which western economy so critically depends precipitates a collapse which must change our
way of living if we are to survive at all.
The autonomous house is not seen as a regressive step. It is not simply a romantic vision of
‘back to the land’, with life again assuming a rural pace and every man dependent upon
himself and his immediate environment for survival. Rather, it is a different direction for
society to take. Instead of growth, stability is the aim; instead of working to earn money to
pay other people to keep him alive, the individual is presented with the choice of selfautonomy or working to pay for survival. No such choice exists at present. ‘Dropping out’
now is a game for those with private means.
Stability would be an obvious goal were it not for the fact that society is so geared to growth
in every sense. A stable population, making only what it actually needs, with each article
being considered with regard to the material it is made of and what is to be done with it once
its useful life is over, and finding all its power from what can be grown or from the sun,
would give man back a true place in the world’s system. However, a consumer society can
exist only by living off the capital resources of the Earth, whether the stored fuels or the
reserves of oxygen for running the machinery of the growth economy; and, as has frequently
been shown, these reserves are not infinite. The oil shortage in 1974 gave a taste of enforced
‘no growth’ economy, and our survival at whatever price or hardship will be a first lesson in
stability. Whether this lesson will provide the impetus for yet more growth from a nuclearbased economy, or whether it could form the basis of a more rational society, remains to be
seen. The autonomous house would only form a very small part of this total picture, but it is
an object that can be grasped and realized in material terms at present.
www.seyfihoca.com
250
However, the attractive idea of a house generating its own power and recycling its own wastes
is almost as difficult to realize as the idea of a stable economy. Apart from the physical
limitations of income-energy sources, the system can be made only marginally competitive
with existing methods of servicing houses. This difficulty could be removed if autonomy did
not have to fit within the present system. At the moment, however, with houses already more
expensive than most people can afford, the idea of an increased capital cost for houses, even
though future running costs would be reduced, could never be accepted.
The idea of autonomy probably arose from two quests. The first was to gain free power for
house heating, etc., so that conventional fuels need not be bought, and the second was to free
the planning of communities. At present any new building must link to an existing or purposebuilt service network. Cities, therefore, expand around their edges in order to keep houses on
the mains, although expansion is limited by the size of the existing servicing plants. Removal
of this restraint would enable houses to be built virtually anywhere, and communities would
be formed for a more logical reason than the need to be fed and watered at a central point.
Existing cities can be likened to babies in that they are serviced completely from the outside
and the control of their functions is at the will of a very few people. If any one person declares
a state of emergency, half a million people may sit in the dark unable to help themselves.
Autonomy could provide for every community to become adult. Each person or community
would be in control of his own heating, lighting, food production, etc. A real decentralization
of control would be achieved and every person would become self-governing.
How desirable such decentralization is in political terms, with removal of choice from the few
to the many, is open to discussion. An autonomous country would mean one where there
would be no growth in the economy, where population size was strictly controlled, where a
higher standard of living could not be expected, where resources were shared equally between
every man, where freedom to act was curtailed by the need to survive. The society would be
unlike any that we know at the moment. It would encompass something of many previous
political doctrines but it would be aimed at providing for the survival of mankind, given that
our present method of living off capital cannot go on for all time.
Any acceptance of the desirability of autonomy can only be based on faith. If you believe that
it is important for man to be part of his natural ecology, to know how survival is
accomplished, to be in control of his own life, then autonomy is a logical outcome. If,
however, you believe that mankind has always solved every problem that arises, that
eventually some way will be found for dealing with nuclear waste after a given number of
years of research and that the benefits of cheap nuclear power outweigh the possible dangers,
then there is no case for autonomy and the status quo will be maintained.
(From The autonomous house - design and planning for self-sufficiency by Brenda and
Robert Vale)
TWENTIETH CENTURY DISCOVERY
1 - War Against The Six-Legs: The Discovery of insecticides and pesticides
Just about the greatest problem we all face now is our own numbers. We crowd the earth
more thickly now than we ever have before and this is creating strains.
www.seyfihoca.com
251
Before the invention of agriculture about 8500 B.C., man lived on the animals he could catch
and kill and on the plants he could find that were good to eat. At that time, there weren't many
human beings on Earth. One careful guess is that there were only eight million people on the
whole planet. (That's about the population of New York City today. Imagine New Yorkers
being the only people alive and that they were spread over the entire planet.)
The reason there were so few then was that there are only so many animals to be caught and
only so many plants to be found. If, for some reason, there were suddenly more people, some
of them would be sure to starve to death. The population would shrink again.
Once agriculture was developed, people deliberately grew large quantities of plants that could
be eaten. There was more food to be found in one spot and more people could eat well.
Population increased.
By the time of Julius Caesar, in 50 B.C., there were fifty million people living on agriculture
around the shores of the Mediterranean Sea. Another fifty million were living in China and
another fifty million in the rest of the world. The total for the world was 150 million but that
was still less than the population of the United States alone today.
Population continued to increase and by 1600 A.D., it had reached 500 million.
After that, the increase became so rapid that we can speak of a "population explosion." New
continents had been discovered with large tracts of land into which people could push and
where they could begin to farm. The Industrial Revolution came and made it possible to farm
more efficiently and ship food greater distances.
By 1800, the world population was 900 million; by 1900, it was 1,600,000,000. Now, it is
about 3,500,000,000. Three and a half billion people are alive today.
In recent years, medical advances have placed many diseases under control. The death rate
has dropped and with fewer people dying, population is increasing faster than ever. The world
population doubled between 1900 and 1969, a period of sixty-nine years. It will double again,
in all likelihood, between 1969 and 2009, a period of only forty years.
When the twenty-first century opens, and the youngsters of today are in middle life and
raising a family, the world population will be something like 6,500,000,000. The United
States alone will have a population of 330 million.
Naturally, this can't continue forever. There comes a point when the number of men, women,
and children is too great to feed and take care of. If the numbers become too great, there will
be famine and disease. Desperate, hungry men will fight and there will be wars and revolts.
With this in mind, many people are trying to discover ways of limiting the population by
controlling the number of births. It seems to make sense that no more children should be born
than we can feed and take care of. It is no act of kindness to bring a child into the world who
must starve, or live a miserable, stunted life.
It is possible that kind and intelligent ways of controlling birth will be accepted and that
human population will reach some reasonable level and stay there. It will take time for this to
come to pass, however, and no matter what we do the figure of 6,500,000,000 will probably
www.seyfihoca.com
252
be reached. Even if it goes no higher, we will have to count on feeding and taking care of this
number.
This will be difficult. At this very time, when the world population is only 3,500,000,000, we
are having difficulty. Large sections of the world are poorly fed. There are perhaps 300
million children in the world who are so badly underfed that they may have suffered
permanent brain damage and will therefore never be quite as intelligent and useful as they
might have been if they had only received proper food. Nations such as India face famine and
would have seen millions die already if it were not that the United States shipped them huge
quantities of grain out of its own plentiful supplies. But American supplies are dwindling fast,
and when they are gone, what will happen to nations like India?
There are no longer large empty spaces of good land which farmers can utilize. The fertile
areas of the world are all in use. We have to try to find less easy solutions. We can bring
water to dry areas. We can use chemicals to restore the fertility of soil which has been fading
out after centuries of farming. We can use more fish from the ocean; and perhaps we can even
grow plants in the sea.
Actually, mankind has been steadily increasing food production since World War II. The
trouble is that this food increase has barely matched the population increase. Despite all the
extra food, each individual today gets no more than he used to get twenty years ago. The
percentage of hungry people in the world stays the same.
And as the population rises ever faster, it is important that the food supply increase ever faster
also. It is important to feed the ever-increasing numbers of human beings until such time as
the population can come under control.
One way of doing so, without having to increase the size of our farmlands one bit, would be
to prevent any of our precious food from being eaten by creatures other than humans. Farmers
are always on the watch for hawks that eat their chickens, coyotes that eat their lambs, crows
that eat their corn.
These are creatures we can see and do something about. We can lay traps, or shoot, or set up
scarecrows.
But hawks, and coyotes, and crows are nothing at all compared to an enemy that is much
smaller, much more dangerous, and until very recently, almost impossible to fight.
These are the insects; the little buzzing, flying six-legged creatures that we find everywhere.
Insects are the most successful form of animal life on earth. There are nearly a million
different kinds (or "species") of insects known, and perhaps another two million species exist
that have not yet been discovered and described. This is far more than the total number of
different species of all other animals put together.
The number of individual insects is incredible. In and above a single acre of moist soil there
may be as many as four million insects of hundreds of different species. There may be as
www.seyfihoca.com
253
many as a billion billion (1,000,000,000,000,000,000) insects living in the world right nowover 300 million insects for each man, woman, and child alive.
Almost all the different species of insects are harmless to man. They are, indeed, useful in the
scheme of life. Many insects serve as the food supply for the pleasant songbirds we all enjoy.
Other insects help pollinate plants, and without that the plants would die.
Some insects are directly useful to man. The bee produces honey and wax, the silkworm
produces silk, and certain scale insects produce a brilliant red dye. Some insects, such as
locusts, are even eaten by men in some areas of the world. To be sure, there are some species
of insects that are troublesome. Perhaps 3,000 species at most (out of a possible three million)
are nuisances. These include the mosquitoes, flies, fleas, lice, wasps, hornets, weevils,
cockroaches, carpet beetles, and so on.
As a result, people come to dislike "bugs" and get the urge to swat or crush anything with six
legs that flies or crawls. This is wrong, though. We don't want really to wipe out all insects
because a few are bothersome. Insects, as I said, are necessary to the scheme of life.
In fact, all the different species of creatures are useful to each other. Even killer animals are
useful to the creatures they kill.
As an example, mountain lions kill deer. Now deer are pretty animals while mountain lions
seem to be dangerous killers that deserve to be wiped out. It has happened that men have
killed the mountain lions in some areas and freed the deer from the danger.
That does not do the deer a favour!
While the mountain lions were active they killed some deer but never very many. What's
more, they usually killed old or sick deer, for the strong young ones had a better chance to get
away. The mountain lions kept the numbers of deer down and there was that much more food
for those that were left.
Once the mountain lions were gone, the deer population increased quickly. Even the old and
sick had a chance to live. All the deer searched the countryside for food and in no time the
area was stripped bare. Starvation gripped the herd and all became weak and sick. They began
to die and in the end there were far fewer deer than there had been in the days when the
mountain lions were active.
So you see, the deer depend for their life and health on the very animals that seem to be
killing them.
The way in which different species of animals depend upon one another results in a "balance
of nature." The numbers of any particular species stay about the same for long periods of time
because of this balance. Even if the balance is temporarily upset, when one species grows
unusually numerous or unusually rare, the food supplies drop, or increase, in. such a way that
the proper number is restored.
The study of this balance of nature is called "ecology" and it has grown to be one of the
branches of science that is of greatest interest to mankind, for we have badly upset the balance
of nature and are upsetting it worse each year.
www.seyfihoca.com
254
In the end, we might suffer as the deer suffer when the mountain lions are gone, and scientists
are anxious to prevent this if possible. By studying the principles of ecology, they hope to
learn how best to prevent it.
Actually, insects wouldn't have grown to be such nuisances, if mankind hadn't upset the
balance of nature many thousands of years ago when he first developed agriculture. Once he
began to plant barley, for instance, he saw to it that many acres of land produced hardly
anything but barley, barley, barley. All the other plants that might have been growing on those
acres he wiped out as much as possible. They were "weeds."
Animals that lived on those weeds were starved out. On the other hand, animals that lived on
barley multiplied, for suddenly they had a huge food supply.
In this way, agriculture encouraged certain insects to multiply and what had been just a
nuisance became a great danger. As an example, locusts may suddenly multiply and swarm
down on fields in gigantic armies of billions. This happened frequently in ancient times and
even the Bible describes such a locust plague in the book of Joel. Locusts would sweep across
the fields, eating everything green. When they left, a barren waste would remain.
This would be a disaster, for large numbers of people would be depending upon those
vanished crops. Widespread famine would be the result.
Nor could anything be done about it. People were completely helpless as they watched their
food disappear. They might go out and try to kill locusts, but no matter how hard they worked
at it, there would be ten thousand left alive for every one they killed.
Even today, although scientists have discovered ways of fighting insects, there is serious
trouble in some places and at some times. This is especially true in the less-developed
countries where scientific methods of fighting insects are least available-and where the
population can least afford the loss.
In India, for instance, there is an insect called the "red cotton bug" which lives on the cotton
plant. If cotton plants were growing wild, some of them might be affected by the bug, but the
plants would be few in number and would be spread widely apart. The bugs would not have
much to eat and would find it difficult to get from one plant to the other. The number of red
cotton bugs would therefore remain small and the cotton plants themselves would be only
slightly damaged. They would continue to grow quite well.
In large cotton fields, however, the bugs have a tremendous food supply, with one plant right
on top of the other. The bugs increase in numbers, therefore, and become a huge horde. Each
year, half of all the cotton grown in India is destroyed by them.
Even in the United States, we have trouble. An insect called the "boll weevil" feeds on the
cotton plant in this country. We can fight the boll weevil better than the Indians can fight the
cotton bug. Still, as a result of the boll weevil damage, each pound of cotton produced in the
United States costs ten cents (about 10d.) more than it would if the boll weevil didn't exist.
www.seyfihoca.com
255
The losses resulting from insect damage in the United States alone run to something like eight
billion dollars each year. Man himself has also vastly increased in numbers since agriculture
was developed. Before that, small groups of men hunted through wide stretches of forests.
They offered only a small target for fleas and lice.
After the appearance of agriculture, farming communities were established. These were much
larger than hunting bands, and in such communities, men lived huddled together. Fleas and
lice multiplied and men had to do a great deal more scratching. Mosquitoes, too, gained a
much larger food supply and increased in numbers.
You might think that insects like termites and boll weevils did real damage and that fleas and
lice were just nuisances, but that is wrong. The insects that bite and sting human beings can be
terrible dangers; and this was something that wasn't discovered until the opening of the
twentieth century.
The discovery came in connection with yellow fever. This is a rapidly spreading disease that
can kill vast numbers of people. Nowadays it is rarely heard of in the United States but in
previous centuries, it would suddenly flare up in huge epidemics that would lay whole cities
low. Twenty times in the history of the city of Philadelphia, yellow fever epidemics raged
across it. New York had fifteen epidemics.
There seemed no way of preventing the epidemics. They struck out of nowhere and suddenly
people were dying on every side. The United States military forces grew particularly
interested in the problem in 1898.
That year they fought a short war with Spain. Most of the fighting took place in Cuba where
few Americans were killed by Spanish guns, but many died of yellow fever. What people
didn't understand was how the yellow fever passed from one person to another. Was it by
infected clothing, by polluted water, or how?
In 1899, the American government sent to Cuba a team of doctors headed by Walter Reed.
Their mission was to find out how yellow fever was spread. Yellow fever does not attack
animals so the mission had to work with human beings, and that meant using themselves as
guinea pigs.
They handled the clothing and bedding of people sick with yellow fever yet didn't come down
with it themselves. Walter Reed remembered that a few people had advanced the notion some
years before that mosquitoes might carry the disease. They would bite sick men and suck in
infected blood, then pass the infection to the next person they bit.
Reed's group checked this. They introduced mosquito netting to keep mosquitoes away from
certain houses. Sure enough, they found that people protected by mosquito netting usually
didn't get the disease even when it was striking all around.
They went on to something more daring. They captured mosquitoes in rooms where there
were men sick with yellow fever and then allowed those mosquitoes to bite them. Some of the
group soon came down with yellow fever and one of them, Jesse William Lazear, died.
www.seyfihoca.com
256
A mosquito bite is more than a nuisance, then. Mosquitoes of a certain species can pass on a
deadly disease with their bite.
Yellow fever struck the United States again, for the last time, in 1904, with New Orleans the
victim. But Reed had shown how to fight the disease. The mosquitoes were kept away with
netting. The places where they bred were wiped out. As a result, yellow fever is no longer a
serious danger in the United States. There hasn't been an epidemic in this country in over
sixty years.
Another species of mosquito was found to spread the disease called malaria. Malaria isn't as
dramatic as yellow fever. It isn't as rapid a killer. Besides, there is a drug, quinine (obtained
from the bark of a South American tree), that, for centuries now, has been known to control
the disease.
Even so, malaria is the most widespread disease in the world-or it was. As late as 1955, there
were estimated to be no less than 250 million people in the world who were ill with malaria.
Each year 2,500,000 people died of it. Those who didn't die were greatly weakened and
couldn't work as healthy people could. Entire nations were greatly reduced in vigour and in
the ability to help themselves because so many individuals among them were malarial. And
all the result of mosquito bites.
Certain species of insects in Africa, called the "tsetse fly," spread sleeping sickness, a brain
infection that usually ends in death. This disease spread into eastern Africa at the beginning of
the twentieth century and between 1901 and 1906 it killed 200,000 people in Uganda. About
two out of every three people in the affected areas died.
The disease also affects horses and cattle. It is the tsetse fly more than anything else-more
than the heat, the jungle, or the large wild animals-that keeps sections of Africa from
advancing.
Naturally, men were anxious to kill insects. Insects were starving mankind, eating his grain
and fruits and fibres, too Insects were killing men with their infected bites. Men had to strike
back.
One way was to poison insects. Suppose, for instance, you sprayed your crops with a solution
of "Paris green," a deadly poison compound containing copper and arsenic.
Paris green did not affect the plants. The plants lived on carbon dioxide in the air and on
certain minerals which they absorbed from the soil. If there was some poison on their leaves,
that made no difference.
Any insect trying to feed on the leaves that were coated with Paris green would, however, die
at once. Insects simply could not live on sprayed plants and the plants grew large and ripe
without being bothered. Paris green was an "insecticide," a word meaning "insect-killer."
(Nowadays, the word is used less often because insects are not the only kind of creature we
want to kill. There are also worms and snails, mice and rats, even rabbits-all of which become
serious problems if they grow too numerous. They are all lumped together as "pests" and any
www.seyfihoca.com
257
chemical used to kill any of them is a "pesticide." In this chapter, though, I will be talking
chiefly about insects and I will continue to use the word insecticide.)
Paris green and other mineral insecticides have their drawbacks. For one thing, they are just as
poisonous to human beings as they are to insects. Foods which have been sprayed with these
solutions must be carefully washed, or they could be deadly.
And, of course, plants are washed, naturally, by rain. The rain tends to remove some of the
mineral poison and drip it down to the soil. Little by little, the soil accumulates copper,
arsenic, and other elements which will reach the roots of the plants eventually. There they do
affect plants and the soil will after a while become poisonous to them.
What's more, such mineral insecticides can't be used on human beings themselves. Sometimes
it would be most useful if we could use them so, to destroy insects that live directly on people.
Mosquitoes and flies may bite people and annoy them (or sometimes transmit diseases that
kill them) but at least they don't actually live on people. If we want to attack them, we can
keep them off by netting, spray the places where they land with poison, or find the stagnant
pools or garbage where they breed and either remove or spray them.
But what about the fleas and lice that live in human clothing or hair? In many parts of the
world even today there are no automatic washers in which clothes can be washed every
couple of days. There isn't even a supply of soap or of clean running water. The poorer people
have very little in the way of clothing and- if there is a cold season they must simply wear the
same clothes all winter long.
Naturally, the fleas and lice in that clothing have a happy hunting ground all winter long. This
was all the more true if people were forced to crowd into small dirty hovels or tenements. If
anyone happened not to have fleas and lice, he quickly caught them from others.
This could be extremely serious because typhus, a disease always present among the poor,
every once in a while became epidemic and spread everywhere. It was most likely to be found
among poor, dirty people huddled together on ships, for instance, or in jails. It was
particularly dangerous during wars when many thousands of soldiers might be penned up in a
besieged town or in lines of trenches or in prisoners' camps.
When thousands of Irish emigrated to America after the potato blight brought famine to
Ireland in the 1840s, half of them sickened with typhus on the way here. In World War I,
typhus did more damage among the miserable soldiers in eastern and south-eastern Europe
than the guns did.
The little country of Serbia drove back the armies of much larger Austria-Hungary several
times in 1914 and 1915, but then typhus struck and crippled the small nation. The Austrians
dared not invade while the epidemic was raging but afterwards they marched in and what was
left of the Serbian army could not stop them.
By the time of World War I, however, doctors knew very well what was causing the spread of
typhus. They had learned that from a French physician, Charles Nicolle, who, in 1903, had
been appointed director of a medical institute in Tunis in North Africa. (Tunis belonged to
France at the time.)
www.seyfihoca.com
258
Tunis was riddled with typhus but Nicolle noticed a very curious thing. The disease was
infectious only outside the hospital, not inside. Doctors visiting patients in their homes caught
typhus. Medical attendants who admitted patients into the hospital caught it. But once the
patients were in the hospital, they stopped being infectious, even though they might
be sicker than ever. Doctors and nurses who tended typhus patients inside the hospital never
caught typhus themselves. Nicolle decided that something happened at the moment that
patients entered the hospital that changed everything. For one thing, the patient had removed
the clothes he was wearing and took a bath. The clothes were got rid of and the infectiousness
disappeared.
By that time the word was about that mosquitoes spread yellow fever and malaria, so it didn't
seem hard to believe that maybe typhus fever was spread by the lice in the dirty clothes.
Nicolle worked with animals, first with chimpanzees, and then with guinea pigs, and he
proved his case completely. Typhus would spread by a louse bite, not otherwise.
Nor is typhus the only disease to be spread by such body insects. There is a dreaded disease
called "plague." In the fourteenth century, it spread all across Europe and killed one out of
every three human beings on the continent. It was called "the Black Death" then.
This disease is spread by fleas. The fleas that are most dangerous live on rats and wherever
the rats spread, so do the fleas. When a flea bites a sick rat, then jumps on a human being and
bites him, it is usually all up with the human.
These are hard diseases to conquer. Rats are difficult creatures to get rid of. Even today they
infest American slums and are a downright danger to sleeping babies. Body lice or fleas are
even harder to eliminate.
After all you can't avoid lice and fleas by something as simple as mosquito netting. You must
wash clothes and body regularly, but how can you ask people to do that who have no soap and
no clean water?
It would be helpful if you could spray the bodies and clothes with insecticide, but you would
have to find one that would kill the insects without killing the person. Certainly Paris green
wouldn't do.
Instead of minerals, then, the search was on for some suitable organic substance. An organic
substance is one that has a structure similar to the compounds contained in living tissue. There
are many millions of different organic substances, and no two species of creatures act exactly
alike in response to particular organic substances.
Might it not be possible to find an organic substance which would interfere with some of the
chemical reactions that go on in insects, but not in other kinds of animals.
In 1935, a Swiss chemist, Paul Muller, began to search for such a compound. He wanted one
that could be easily made and would therefore be cheap. It had to be without an unpleasant
odour. It had to kill insects but be reasonably harmless to other kinds of life.
www.seyfihoca.com
259
He narrowed down the search by studying different classes of organic compounds and then
following up those classes that showed at least a little promise. He would study the chemical
structure of those compounds that showed a little promise and would then try a slightly
different compound to see if that had more promise. If it did, he would study the difference in
structure and see how to make a still further difference that would be better still.
It took four years but in September of 1939 (the very month in which World War II started),
Muffler came across a compound called "dichlorodiphenyltrichloroethane." That is a long
name even for chemists and it is usually referred to by its initials, as DDT. This compound
had first been prepared and described in 1874 but at that time there seemed nothing unusual
about it. Now, however, Muller discovered that DDT was the very thing he was looking for. It
was cheap, stable, and odourless, fairly harmless to most forms of life, but meant death to
insects.
By 1942, preparations containing DDT were beginning to be manufactured for sale to the
public, and in 1943, it had its first dramatic use. The city of Naples, in Italy, had been
captured by Allied forces and, as winter came on, typhus began to spread.
It wasn't possible to make the population strip off their clothes, burn them, and put on new
clothes, so something else was done. Soldiers and civilians were lined up and sprayed with a
DDT solution. The lice died and typhus died with them. For the first time in human history, a
winter epidemic of typhus had been stopped in its tracks.
To show that this was no accident the same thing was done in Japan in late 1945, after the
American occupation. Since World War II, DDT and other organic insecticides have been
used in large quantities. Tens of thousands of tons are produced each year. The United States
alone spent over a billion dollars for such insecticides in the single year of 1966. Not only are
our crops saved but the various insect-spread diseases are all but wiped out. Since DDT wipes
out mosquitoes and flies, as well as lice, malaria is now almost unknown in the United States.
Less than a hundred cases a year are reported and almost all are brought in from abroad.
Yet this does not represent a happy ending. The use of organic insecticides has brought
troubles in its train. Sometimes such insecticides don't work because they upset the balance of
nature.
For instance, DDT might be fairly deadly to an insect we want to kill, but even more deadly to
another insect that lives on the first one. Only a few harmful insects survive but their insect
enemies are now all dead. In a short time, the insects we don't want are more numerous than
they were before the use of DDT.
Then, too, organic insecticides don't kill all species of insects. Some insects have a chemical
machinery that isn't affected by these poisons; they are "resistant." It may happen that a
resistant insect could do damage to our crops but usually doesn't because some other insect is
more numerous and gets the lion's share of the food.
If DDT kills the damaging insect, but leaves the resistant insect behind, then that resistant
insect can multiply enormously. It then becomes a great danger and DDT can't touch it.
www.seyfihoca.com
260
In fact, even among those species of insects that are killed by DDT there are always a few
individuals that differ chemically from the rest and are resistant. They survive when all other
individuals are killed. They multiply and then a whole species of resistant insects comes into
existence.
Thus, as the years pass, DDT has become less effective on the house fly, for instance. Some
resistance was reported as early as 1947, and this has been growing more serious. By now
almost every species of insect has developed resistance, including the body louse that spreads
typhus.
Finally, even though organic insecticides are not very poisonous to creatures other than
insects, they are not entirely harmless either. If too much insecticide is used, some birds can
be poisoned. Fish are particularly easy to kill, and if insecticides are used on water to kill
young insects, young fish may also go in great numbers.
Organic insecticides are also washed into the soil. Eventually, they are broken down by
bacteria but not very quickly. Some accumulates in the soil, then in the plants that grow in the
soil, then in the animals that eat the plants. All animals, including man, have a little bit of
DDT inside ourselves. Not enough to hurt us so far, but it is there.
For that reason, attempts have been made to control insects by means that don't involve
chemicals.
For one thing, there are certain strains of plants which are naturally resistant to particular
insects. These strains might be cultivated.
Then, too, crops might be rotated; one crop might be grown one year, another crop the next.
In this way, an insect which flourished one year might drop to very low levels the next when
the wrong plants were grown, plants it could not eat. It would have to start from scratch again
and its numbers would stay low. Or else one might break up the fields so that not too large an
area would be devoted to a single crop. That would make it harder for an insect to spread.
Here's something else-insects have their enemies. The enemy might exist in one part of the
world but not in another. It might be another insect or some kind of parasite. If it could be
introduced in places where the insect we were after was flourishing, the numbers of that insect
might be brought under control.
Modern science has worked up a number of additional devices for killing insects. Bats eat
insects and locate them by emitting very shrill squeaks, squeaks too shrill for us to hear. The
sound waves of these squeaks bounce off the insect, and the bat, by listening for the echo,
knows where the insect is.
Naturally, insects have developed an instinctive avoidance of such a sound. If a device is set
up to send out these shrill "ultrasonic" squeaks, insects stay away from a wide area near it.
Another device is just the opposite-to attract rather than to repel. Insects can find each other
over large distances because they can smell each other. Female moths give off an odour that a
www.seyfihoca.com
261
male moth of the same species can detect many hundreds of yards away. Female moths can
tell by smell a good spot on which to lay eggs.
Chemists have worked to isolate the chemicals that give off this attractive odour. Once they
isolate it, they can place it on certain spots to attract insects. If those spots are sprayed with
insecticide, too, insects could die in great numbers. Only a little insecticide would have to be
used; it wouldn't have to be spread everywhere; and it would be less likely to affect other
forms of life.
Or else a female could be induced to lay eggs in an unsuitable place by means of a sprayed
odour, so that the eggs would not develop.
Then, too, male insects can be subjected to radioactivity that destroys some of their internal
organs so they cannot fertilize the female's eggs. If such sterilized males are released, the
females end up laying eggs that cannot develop. An insect called the "screwworm," which
infests cattle in south-eastern United States, was almost wiped out by this method.
But all that mankind is doing today is not yet enough. The insecticides are too poisonous and
the other methods are a little too fancy for undeveloped countries where the insect menace is
greatest. Is there something better we can do to help feed the doubled population of 2000?
Actually, the 1960s are seeing the development of an exciting new way of battling insects, a
way that makes the insects fight themselves, so to speak. To understand how this should be,
let's consider how insects grow.
An insect has two chief stages to its life. In its young form, it is a "larva"; later on, it is an
"adult." The two forms are very often completely different in appearance.
Consider the caterpillar, for instance. It is a larva, a wingless, wormlike creature with stumpy
little leg-like structures. Eventually, though, it becomes a moth or butterfly, with the usual six
legs of the insect, and often with gorgeous wings. Similarly, the housefly develops out of its
egg as a tiny, wormlike "maggot."
The reason for two such different forms is that the two have widely different specialities. The
larva spends almost all its time eating and growing. It is almost what we might call an eating
machine with all its makeup concentrated on that. The adult, on the other hand, is an egglaying machine. Sometimes adult insects do nothing but lay eggs. Mayflies live less than a
day after they reach the adult stage and don't even have proper eating apparatus. In their short
adult life they just lay eggs; they don't have to eat.
The change from larva to adult is called "metamorphosis." Sometimes the metamorphosis is
not a very startling one. A young grasshopper looks quite grasshopperish, for instance.
Where the metamorphosis is a thoroughgoing one, as in the case of the caterpillar, the insect
must pause in its life cycle to make the enormous change within its body. It is almost as
though it must go back into an original egg stage and start again. It becomes a motionless,
apparently dead object, slowly changing within and making itself over until it is ready to step
forth as an adult. In this motionless intermediate stage it is called a "pupa."
www.seyfihoca.com
262
There are insect species which act in such a way as to protect this defenceless pupa stage. In
its final period as a larva, it will produce thin jets of liquid from special openings in its
abdomen. These jets harden into a tough fibre which the larva weaves round about itself until
it is completely enclosed. This is the "cocoon" within which the pupa remains hidden till
metamorphosis is done. It is the fibre from the cocoon of the silkworm moth that provides
mankind with silk.
All this requires careful organization. For instance, it is a problem for a larva just to grow.
The larva is surrounded by a thin, but tough, cuticle made of a substance called "chitin." This
protects it and gives it an anchor for its muscles, but chitin doesn't expand with the body.
As a larva grows, its cuticle becomes tighter and tighter about it. Finally, the cuticle splits and
is shed. The larva is said to "moult." From the split cuticle, the larva wriggles. It is expanded
now and is distinctly bigger now that the cuticle which had been holding it in like a tight
girdle is gone. A new, but roomier, cuticle quickly forms and within it the larva grows again.
But what makes the cuticle split at just the right time? The answer is that there is an automatic
chemical control involved. Any living creature is a complex setup of automatic self-regulating
chemical machinery. This is true even of the human being and it was only at the very opening
of-the twentieth century that biologists began to have an inkling as to how some of this
machinery worked.
In the human being there is a large gland called the pancreas. It manufactures a digestive juice
which enters the small intestine and mixes with food emerging from the stomach. The
interesting thing is that the pancreas doesn't bother wasting its juice when the small intestine
is empty. Nothing happens until food enters the small intestine and then, instantly, the
pancreatic juice starts flowing.
Something automatic must be involved and in 1902, two English biologists, William
Maddock Bayliss and Ernest Henry Starling, discovered what it was.
The food in the stomach is mixed with a strongly acid juice. When the food emerges from the
stomach and enters the small intestine, the touch of its acidity has a chemical effect on the
intestinal lining and causes it to produce a substance which Bayliss and Starling called
"Secretin."
Secretin is discharged into the bloodstream and is carried to all the body. When it reached the
pancreas, it brings about a chemical effect that causes the pancreas to begin to manufacture
and discharge its juice.
Secretin is a substance which rouses the pancreas to activity. In 1905, Bayliss suggested that
secretin, and all other substances like it, be called "hormones," from a Greek word meaning
"to arouse."
The process of moulting seems to be an automatic process controlled by a hormone. As the
larva grows, there is growing pressure from the cuticle. When the pressure reaches a certain
point, a hormone is triggered. It pours into the larva's bloodstream and when it reaches the
cuticle that cuticle is made to split.
www.seyfihoca.com
263
The hormone that does this has been given the name "ecdysone," from a Greek word meaning
"to moult."
But moulting doesn't go on forever. After a certain number of moults, there is a sudden
change. Instead of continuing to grow in order to prepare the way for still another moult, the
larva begins to undergo metamorphosis instead.
Can this be because a second hormone is involved? Is there a second hormone that suddenly
appears after a certain number of moults and brings about the metamorphosis?
Not quite. In 1936, an English biologist, Vincent Brian Wigglesworth, was working with a
certain disease-spreading, blood-sucking bug called Rhodnius. In the course of his
experiments, he thought it would be useful to see what would happen if he cut off the head of
the larva of these bugs.
Naturally, if you cut off the head of a mammal or a bird, the creature would die and that
would be all. An insect, however, is far less dependent on its head, and life could continue in
some ways.
But different parts of the body produce different hormones and some can be produced in the
head. By cutting off the head of a larva, Wigglesworth could tell what hormones the insect
head might be producing. After all, the headless larva would grow differently than one with a
head would and the differences might be at least partly due to the missing head-hormones.
Wigglesworth did indeed notice a change. As soon as he had cut off the head, the larva went
into a moult and emerged as an adult. (Rhodnius was not one of the bugs that went through a
pupa stage.)
It did this even when it was nowhere near ready for such a change. It hadn't moulted enough
times; it was far too small. Yet it did change and a miniature adult would appear.
But if metamorphosis was caused by the production of a hormone, how could cutting off the
head produce it? Cutting off the head should cause the loss of a hormone, not its production.
Wigglesworth argued that the head produced a hormone that prevented metamorphosis. As
long as it was produced ecdysone, the moulting hormone, did its work; the larva moulted and
grew, moulted and grew. At a certain point, though, in the course of the life of the normal
insect, something happened which cut off the supply of this head hormone. Without that
hormone, ecdysone couldn't work even though it was present, and metamorphosis began.
If the head were cut off, the supply of the hormone was destroyed at once and metamorphosis
began even though the insect body wasn't really ready for it.
Wigglesworth called this hormone from the insect head "juvenile hormone" because it kept
the insect in its juvenile, or youthful, form. He also located tiny glands, barely visible without
a microscope, behind the brain of the larva and these, Wigglesworth was certain, produced the
hormone.
www.seyfihoca.com
264
What Wigglesworth found to be true of Rhodnius was true of other insects, too; of the
silkworm caterpillar, for instance. It seems that all insects that undergo metamorphosis do so
because the supply of juvenile hormone stops at a certain time.
Wigglesworth's suggestion about the glands in the head was quickly shown to be correct. In
1938, a French biologist, Jean Bounhiel, worked out a delicate technique for removing the
tiny hormone-producing glands from a small silkworm caterpillar and placing them in a large
one.
The large silkworm caterpillar was about ready to enter its pupal stage, which meant that its
glands had stopped producing juvenile hormone. The glands from the small caterpillar,
however, were still capable of producing the hormone. When the glands from the small
caterpillar were grafted into
the large one, the large caterpillar suddenly found itself with a new supply of juvenile
hormone. Instead of entering the pupal stage, it continued to moult one or two extra times.
Naturally, it continued to grow, too, and when it finally did switch to the pupa, it was a
considerably larger-than-normal one, and out of it emerged a considerably larger-than-normal
adult moth.
At this point, Carroll Williams of Harvard University stepped onto the scene. He transferred
hormone-producing glands, not to another larva, but to the pupa of a silkworm. The pupa was
well along in metamorphosis. It wasn't supposed to be exposed to any juvenile hormone at all;
it was past that stage. But what if juvenile hormone were forced upon it?
Williams had his answer at once. The presence of juvenile hormone seemed to stop the
metamorphosis, or at least slow it down. When the adult moth finally appeared it was
incomplete. Some of it had not changed over.
Williams found that the more gland material he inserted into the pupa, the more incomplete
the metamorphosis. He could use the amount of incompleteness of metamorphosis to judge
how much juvenile hormone were present in the glands at different stages in the life of the
larva.
He could also determine if there were juvenile hormone anywhere else in an insect body, and
here he stumbled over something that was a complete surprise.
In 1956, Williams found that an insect called the "Cecropia moth" produced a large quantity
of juvenile hormone just before entering the adult stage, after having passed through the pupa
stage entirely without it. Why they do this nobody knows.
This juvenile hormone is stored in the abdomen of the moth for some reason. Only the male
moth does it, not the female. Only one other kind of moth, as far as is known, stores juvenile
hormone in this fashion. All other insects do not.
Even if biologists don't know the reason for any of this, it still turned out to be a useful fact.
The tiny glands that produce juvenile hormone in larvas contain so little that it is just about
www.seyfihoca.com
265
impossible to extract a useful amount. The reserve supply in the abdomen of the male
Cecropia moth is so large, on the other hand, that the hormone can be produced in visible
quantities.
Williams produced an extract from the abdomens of many such moths; a few drops of golden
oil that contained huge quantities of juvenile hormone. Now he had plenty of material with
which he could experiment further.
One Cecropia abdomen supplied enough hormone to block completely the metamorphosis of
ten pupas of almost any kind of moth or butterfly. The extract did not even have to be injected
into the pupa. If some were just applied to the skin of the pupa, enough hormone leaked into
the inner tissues to upset the metamorphosis.
The metamorphosis could be so badly upset, if enough juvenile hormone were used, that the
pupa could not develop at all. It simply died.
The thought at once occurred to Williams that here might be a potential insecticide that would
have great advantages over every other kind known. After all, it turned the insect's own
chemistry against itself.
An insect couldn't grow resistant to juvenile hormone, as it could to any other sort of
insecticide. It had to respond to its own hormones. If it didn't, it would die.
In other words, an insect had to respond to juvenile hormone at the right time or it would die.
And if it did respond at the right time, then it would also respond at the wrong time and still
die. Either way, the insect was dead.
Even more important, the juvenile hormone would be no danger to forms of life other than
insects. It affected only insects and has no effect whatever (as far as has been found so far) on
any form of life other than insects.
Of course, it is one thing to kill a few pupas in a laboratory and quite another to kill vast
quantities out in the fields. Thousands of tons of insecticides are needed for the work that
must be done and it would be impossible to get thousands of tons out of Cecropia moths.
If only the chemical structure of the juvenile hormone were known. It would then be possible
to manufacture it from other chemicals; or else manufacture something that was close enough
to do the job. Unfortunately, the structure was not known.
Williams and a colleague, John Law, sat in their Harvard Laboratories one summer day in
1962, wondering if they could reason out what the structure might be. A lab assistant,
listening to them, suggested a particular type of compound as a joke.
John Law thought he would go along with the gag. It wouldn't be too difficult to make the
compound, or something with a name very like the lab assistant's joke. With scarcely any
trouble at all, he produced an oily solution of a mixture of substances which he intended to
show the young assistant and say, "Well, here is the compound you joked about."
Still as long as he had it, he tried it first on insect pupas. To John Law's everlasting surprise, it
worked! It worked unbelievably well. It was over a thousand times as powerful as the extract
www.seyfihoca.com
266
from Cecropia abdomens. An ounce of Law's solution would kill all the insects over an area
of two and one-half acres-at least all the insects that were metamorphosing.
This substance is "synthetic juvenile hormone." It contains at least four different chemicals,
and none of them seems to have a structure like that of the natural hormone.
Synthetic juvenile hormone works on all insects tested, including the mosquito that spreads
yellow fever and the louse that spreads typhus. Yet it doesn't affect any creature other than
insects. It would be no danger to birds, fish, mammals, or man.
Still, killing all insects is a little too much. That would upset the balance of nature.
We want to kill only certain insects, only one species out of a thousand. This could be done
perhaps with the natural juvenile hormone. Each different group of species of insects
manufactures its own kind of juvenile hormone which works for itself but not for others.
Perhaps then, you can use a particular juvenile hormone and get just the insect you're after
and no other kind.
For instance, a biologist in Prague, Czechoslovakia, named Karel Sláma, was trying to make
natural juvenile hormone work on a harmless insect called the "red linden bug." He used the
technique developed by Carroll Williams, but the extract from Cecropia moths didn't affect
the red linden bugs. It might kill moths and butterflies but it had no effect at all on the red
linden bugs. The red linden bugs must have a juvenile hormone so different from those of
moths and butterflies that the effects didn't cross.
Williams heard of these experiments and was most curious. In the summer of 1965, Williams
asked Sláma to bring his red linden bugs to Harvard and to come with them. Sláma came, and
together the two men began to grow the bugs. In Prague, Sláma had grown them by the tens
of thousands and their way of growing was always the same. The larvas went through exactly
five moults and then moved into the adult stage. (The red linden bug does not go through a
pupa stage.)
Yet at Harvard this did not happen. Bug after bug went through the fifth moult. Then, instead
of changing into an adult, they stayed larvas and tried to moult a sixth time. Usually, they
didn't make it, but died. In the few cases where a bug survived the sixth moult, they died
when they attempted a seventh moult. About 1,500 insects died in the Harvard laboratories,
where none had died in Prague.
Why? It was as though the bugs had received a dose of juvenile hormone and couldn't stop
being larvas-but no juvenile hormone had been given them.
Williams and Sláma tried to think of all possible differences between the work at Harvard and
the work in Prague. In Harvard, the red linden bugs were surrounded by all sorts of other
insects which were involved in juvenile hormone experiments. Perhaps some of the hormone
got across somehow. The other insects were therefore removed but the red linden bugs still
died.
Could the glassware have been contaminated during cleaning? Maybe. So Williams ordered
new glassware that had never been used. The bugs still died.
www.seyfihoca.com
267
Could there be something wrong with the city water? Williams got spring water, but the bugs
still died.
Altogether fourteen different possibilities were thought of and thirteen were cancelled out.
One thing remained, and one only Strips of paper were placed into the jars in which the red linden bugs were grown. They were
slanted against the sides, as a kind of path for the bugs to scurry along. (That seemed to keep
them more active and in better health.) Of course, the paper used at Harvard was not the same
as the paper used in Prague. Williams was, in fact, using strips of ordinary paper towels
produced by a local manufacturer.
Williams proceeded to check that. He used strips of chemically pure filter paper instead. At
once, the bugs stopped dying.
There was something in the paper towels that acted like juvenile hormone and upset the
chemical machinery of the larvas. It kept them moulting after they should have stopped doing
so and that killed them. Williams called the mysterious substance that did this the "paper
factor." Later, it received the more chemical sounding name of "juvabione."
Williams and Sláma went on to try all kinds of paper. They found that almost any American
newspaper and magazine contained the factor. Red linden bug larvas that crawled over them
never made it to the adult stage. On the other hand, paper from Great Britain, the European
continent, or Japan, did not have it and the bugs lived well on such paper. (That's why they
lived in Prague.)
Could it be that American manufacturers put something in paper that other manufacturers did
not? A check with the manufacturers showed they didn't. Well, then, what about the trees
from which the paper was made.
They began to test extracts from the trees and found one called the "balsam fir" which was
much used for American paper but which did not grow in Europe. It was particularly rich in
paper factor, and this paper factor could be obtained from the tree in large quantities.
Here is an interesting point. The paper factor works on only one group of insect species, the
one to which the red linden bug happens to belong. If Sláma had brought with him some
insect from another group of species, the paper factor might have gone undiscovered.
The paper factor is an example of an insecticide that will kill only one small group of insects
and won't touch anything else. Not only are fish, birds, and mammals safe, but so are all
insects outside that one group of species.
To be sure, the red linden bug is harmless and there is no purpose in killing it, but the red
cotton bug, which eats up half of India's cotton crop, is closely related to it. The red cotton
bug can also be hit by the paper factor and experiments are underway to see how well it will
work in India's cotton fields.
Paper factor catches bugs at the time of their metamorphosis. This is better than nothing but it
still isn't quite as good as it might be. By the time the metamorphosis is reached, the insect has
spent a lot of time as a larva-eating, eating, eating.
www.seyfihoca.com
268
Then any insects that happen to survive the paper factor for some reason can lay a great many
eggs. They will develop into larvas that will eat and eat and eat and will only be caught at the
metamorphosis.
It would be better if insects were caught at the beginning of the larval stage rather than at the
end.
And they can be! It turns out that the eggs, like the period of metamorphosis, must be free of
juvenile hormone. In 1966, Sláma placed eggs of the red linden bug on paper containing the
factor and-if the eggs were fresh enough and weren't already on the point of hatching-they
didn't hatch.
Then he tried it on adult females that were ready to lay eggs but hadn't laid them yet. He
placed a drop of the factor on the adult's body and found that it worked its way inside and,
apparently, even into the eggs. At least such a female laid eggs that didn't hatch.
The paper factor was more valuable than ever now, for it could be used to catch the insects at
the very beginning of their life.
But why should the balsam fir possess a compound that acts like juvenile hormone? The
answer seems clear. Insects eat plants and plants must have developed methods of selfprotection over the millions of years of evolution.
A good method of self-protection is for the plants to develop substances that taste bad to
insects or that kill them. Plants which happen to develop such substances live and flourish
better than those that don't.
Naturally, a plant would develop a substance that would affect the particular insects that are
dangerous to it. It seemed that if biologists were to make extracts from a large variety of
plants, they might find a variety of substances that would kill this type of insect or that. In the
end, they would have, perhaps, a whole collection of insecticides to use on particular insect
pests. We would be able to attack only the insects we want to attack and leave the rest of
nature alone. By 1968, indeed, some fifteen such plant self-defence chemicals were located.
Then, too, in 1967, Williams took another step in this direction, while with an expedition
exploring the interior of the South American country Brazil. There the large river Rio Negro
flows into the Amazon. The name means "Black River" because its waters are so dark.
Williams -noticed there were surprisingly few insects about the banks of the river and
wondered if the trees might not contain considerable paper factor of different kinds. Then he
wondered if the darkness of the river water might not come from its soaking substances out of
the trees that lined its bank. If so, the river water might contain all kinds of paper factors.
Tests have shown that the Rio Negro does have insecticide properties. Perhaps many different
paper factors will be extracted from it in endless supply. Perhaps other rivers may be found to
be as useful.
www.seyfihoca.com
269
In 1968, Sláma synthesized a hormone like compound which was the most powerful yet. An
insect treated with such a hormone would pass a bit of it on to any other insect with which it
mated. One treated insect could sterilize hundreds of others.
So things look quite hopeful. Between the supplies found in nature and between the chemicals
that can be formed in the laboratory, we may get our insect enemies at last.
This will mean that man's supply of food and fibre will increase. It will mean that a number of
diseases will no longer threaten him, and he will be able to work harder to produce goods.
In that case, we may well be able to feed, clothe, and take care of all the billions who will
swell Earth's population in the next forty years or so.... And by that time we may have learned
to control our own numbers and we will then be safe.
From: Twentieth Century Discovery by Isaac Azimov
2 - In The Beginning: The Origin Of Life
The first chapter dealt with a scientific search that had a very practical goal-ways of killing
dangerous insects. When you solve the problem, there is no mistake about it; the insects die.
But there are also problems that are much more difficult to tackle; problems that are so
complicated it is even hard to tell whether we are on the road to solving them, or just in a
blind alley. Yet they are problems so important that man's curiosity forces him to tackle them
anyway.
Consider the question: What is life?
There is no plain answer yet and some scientists wonder if there ever can be. Even the
simplest form of life is composed of very complex substances that are involved in so many
numerous complicated chemical changes that it is almost hopeless to try to follow them. What
parts of those changes make up the real basis of life? Do any of them?
The problem is so enormous that it is like a continent that must be explored at different points.
One group of explorers follows a huge river inland; another group may follow jungle trails
elsewhere; while a third sets up a camel caravan into the desert.
In the same way, some biologists analyze the behaviour of living animals under various
conditions; others study the structure of portions of tissue under microscopes while still others
separate certain chemicals from tissue and work with them. All such work contributes in its
own way to increasing knowledge concerning life and living things.
Enormous advances have indeed been made. The two greatest biological discoveries of the
nineteenth century were 1) that all living creatures are constructed of cells, and 2) that life has
slowly developed from simple creatures to more complex ones.
The first discovery is referred to as the "cell theory," the second as the "theory of evolution."
Both theories made the problem of life seem a little simpler. Cells are tiny droplets of living
substance marked off from the rest of the world by a thin membrane. They are surprisingly
www.seyfihoca.com
270
alike no matter what creature they are found in. A liver cell from a fish and one from a dog
are much more similar than the whole fish and dog are.
Perhaps if one could work out all the details of what makes individual cells alive, it would not
be so difficult to go on and get an understanding about whole creatures.
Then, too, consider that there was a gradual development of complex organisms from simpler
ones. In that case, it might well be that all creatures that exist today developed from the same
very simple one that existed long ages ago. There would then be only one form of life,
existing in many different varieties. If you understood what made a housefly alive, or even a
germ, you ought then understand what makes a man alive.
But these nineteenth century theories also raised a new problem. The more people
investigated cells and evolution, the more clear it became that all living creatures came from
other living creatures; all cells came from other cells. New life, in other words, is only formed
from old life. You, for example, are born of your parents.
Yet there must have been a time in the early history of the Earth when there was no life upon
it. How, then, did life come to be? This is a crucial question, for if scientists knew how the
first life was formed on a world that had no life on it, they might find they had taken a big
step forward in understanding what life is.
Some nineteenth century scientists were aware of this question and understood its importance.
Charles Darwin, the English biologist who first presented the theory of evolution to the world
in its modern form, speculated on the subject. In a letter written to a friend in 1871, he
wondered if the kind of complex chemicals that make up living creatures might not have been
formed somewhere in a "warm little pond" where all the ingredients might be present.
If such complex compounds were formed nowadays, tiny living creatures existing in that
pond would eat them up at once. In a world where there was no life, however, such
compounds would remain and accumulate. In the end, they might perhaps come together in
the proper way to form a very simple kind of life.
But how can one ever find out? No one can go back billions of years into the past to look at
the Earth as it was before life was on it. Can one even be sure what conditions were like on
such an Earth, what chemicals existed, how they would act?
So fascinating was the question of life's origin, however, that even if there was no real
information, some scientists were willing to guess.
The twentieth century opened with a very dramatic guess that won lots of attention. The
person making the guess was a well-known Swedish chemist, Svante August Arrhenius. In
1908, he published a book, Worlds in the Making, in which he considered some new
discoveries that had recently been made.
It had just been shown that light actually exerted a push against anything it shone upon. This
push was very small, but if the light were strong and an object were tiny, the light-push would
be stronger than gravity and would drive the object away from the Sun.
www.seyfihoca.com
271
The size of particles that could most easily be pushed by sunlight was just about the size of
small cells. Suppose cells were blown, by air currents, into the thin atmosphere high above the
Earth's surface. Could they then be caught by the push of sunlight and driven away from the
Earth altogether? Could they then go wandering through space?
That might be so but wouldn't the cells then die after having been exposed to the vacuum of
outer space?
Not necessarily. It had also been discovered that certain bacterial cells could go into a kind of
suspended animation. If there was a shortage of food or water, they could form a thick wall
about themselves. Within the wall, the bit of life contained in the cell could wait for years, if
necessary, without food or water from the outside. They could withstand freezing cold or
boiling heat. Then, when conditions had improved, the wall would break away and the
bacterial cell could start to live actively once more.
Such walled cells in suspended animation are called "spores." Arrhenius argued that such
spores, driven by the push of light, could wander through space for many years, perhaps for
millions of years, without dying.
Eventually, such spores might strike some object. It might be some tiny asteroid or some other
cold world without air or water. The spore would have to remain a spore forever, until even its
patient spark of life winked out. Or it might strike a world so hot as to cause it to scorch to
death.
But what if the spore struck a world with a warm, pleasant atmosphere and with oceans of
water? Then it would unfold and begin to live actively. It would divide and redivide and form
many cells like itself. Over long periods of time, these cells would grow more complicated.
They would evolve and form many-celled creatures. In the end, the whole planet would
become a home for millions of species of life.
Is that how life originated on Earth itself, perhaps? Once long ago; billions of years ago; did a
spore from a far distant planet make its way into Earth's atmosphere? Did it fall into Earth's
ocean and begin to grow? Is all the life on Earth, including you and I, the descendant of that
little spore that found its way here?
It was a very attractive theory and many people were pleased with it, but alas, there were two
things wrong with it.
In the first place, it wouldn't work. It was true that bacterial spores would survive many of the
conditions of outer space, but not all. After Arrhenius' book had been published, astronomers
began to learn more about what it was like in outer space. They learned more about the sun's
radiation for instance.
The sun gives out not visible light alone, but all kinds of similar radiation that might be less
energetic or more energetic than light itself.
It radiates infrared waves and radio waves, which are less energetic than ordinary light. It also
radiates ultraviolet waves and x rays, which are more energetic than ordinary light. The more
energetic radiation is dangerous to life.
www.seyfihoca.com
272
Much of the energetic radiation is absorbed by Earth's atmosphere. None of the x rays and
very little ultraviolet manage to make their way down to Earth's surface, under a blanket of air
miles and miles thick. Even so, if we stand on the beach on a sunny summer day, enough
ultraviolet light reaches us to penetrate the outer layers of the skin and to give us sunburn (if
we are fair-skinned).
In outer space, the ultraviolet light and x rays are present in full force. They easily penetrate a
spore wall and kill the spark of life inside.
If spores were drifting towards our solar system from other stars, they might strike the
outermost planets without harm, but on Pluto or on Neptune they would find conditions too
cold for development. As they drifted inward towards Earth, they would be coming into
regions where sunlight was stronger and stronger. Long before they could actually reach our
planet, the energetic radiation in sunlight would have killed them.
It would seem then that spores, giving rise to the kind of life we now have on Earth, couldn't
possibly have reached Earth alive.
Then, too, another flaw in Arrhenius's theory is that it doesn't really answer the question of
how life began. It just pushes the whole problem back in time. It says that life didn't begin on
Earth but on some other planet far away and long ago and that it reached our world from that
other planet. In that case, how did life start on that other planet? Did it reach that other planet
from still another planet?
We can go back and back that way but we must admit that originally life must have started on
some planet from nonliving materials. Now that is the question. How did life do that? And if
life started somewhere from non-living materials, it might just as well have done so on Earth.
So don't let's worry about the possibility of life starting elsewhere and reaching Earth. Let us
concentrate on asking how life might have started on earth itself from non-living materials.
Naturally, we ought to try to make the problem as simple as possible. We wouldn't expect
non-living substances to come together and suddenly form a man, or even a mouse, or even a
mosquito. It would seem reasonable that before any creature even as complicated as a
mosquito was formed, single cells would have come into existence; little bits of life too small
to be seen except under a microscope.
Creatures exist, even today, that are made up of just one cell. The amoeba is such a creature.
Thousands of different species of one-celled plants and animals exist everywhere. There are
also the bacteria, which are composed of single cells even smaller than those of the one-celled
plants and animals.
But these cells are complicated, too; very complicated. They are surrounded by membranes
made up of many thousands of complex molecules arranged in very intricate fashion. Inside
that membrane are numerous small particles that have a delicately organized structure.
www.seyfihoca.com
273
It seems hopeless to expect the chemicals in a non-living world to come together and
suddenly form even as much as a modern bacterial cell. We must get down to things that are
even simpler.
Every cell contains chemicals that don't seem to exist in the non-living world. When such
chemicals are found among non-living surroundings, we can be sure that those surroundings
were once alive, or that the substances were originally taken from living cells.
This seems to be so clear that early in the nineteenth century chemists began to speak of two
kinds of substances. Chemicals that were associated with living creatures, or organisms, were
called "organic." Those that were not were "inorganic."
Thus, wood and sugar are two very common organic substances. They are certainly not alive
in themselves. You may be sitting in a wooden chair, and you can be sure that it is no more
alive than if it were made of stone. However, that wood, as you know very well, was once
part of a living tree.
Again, the sugar you put on your morning cereal is certainly not alive. Still, it was once part
of a living sugar cane or sugar beet plant.
Salt and water, on the other hand, are inorganic substances. They are found in all living
organisms, to be sure; your own tears, for instance, are nothing but a solution of salt in water.
However, they are not found only in organisms and did not originate only in organisms. There
is a whole ocean of salt water that we feel pretty sure existed in some form or other before life
appeared on this planet.
(Beginning in the middle of the nineteenth century, chemists began to form new compounds
that were not to be found in nature. They were very similar in many ways to organic
compounds, though they were never found in living organisms or anywhere else outside the
chemists' test tubes. These "synthetic" compounds were, nevertheless, lumped together with
the organic group because of the similarity in properties.)
It would seem then we could simplify our problem. Instead of asking how life began out of
non-living substances, we could begin by asking how organic substances came to be formed
out of inorganic substances in the absence of life.
To answer that question, we ought to know in what way organic substances differ from
inorganic ones.
Both organic and inorganic substances are made up of "molecules"; that is, of groups of atoms
that cling together for long periods of time. Organic molecules are generally larger and more
complicated than inorganic ones. Most inorganic molecules are composed of a couple of
dozen atoms at most; sometimes only two or three atoms. Organic molecules, however,
usually contain well over a dozen atoms and may, indeed, be made up of hundreds, thousands,
or even millions of atoms.
www.seyfihoca.com
274
When we ask how organic compounds may be formed from inorganic compounds, then, we
are really asking how large and complicated molecules might be formed from small and
simple ones.
Chemists know that to force small and simple molecules to join together to form large and
complicated ones, energy must be added. This is no problem, really, for a very common
source of a great deal of energy is sunlight, and in the early lifeless Earth, sunlight was
certainly blazing down upon the ocean. We will come back to that later.
It is also true that the different kinds of atoms within molecules cannot change their nature
under ordinary circumstances. The large organic molecules in living matter must be formed
from small and simple molecules that contain the same kinds of atoms.
We must ask ourselves what kinds of atoms organic molecules contain.
There are over a hundred different kinds of atoms known today (each kind making up a
separate "element"). Over eighty are found in reasonable quantities in the inorganic
substances making up the Earth's crust. Only half a dozen of these elements, however, make
up the bulk of the atoms in organic molecules.
The six types of atoms occurring most frequently in organic molecules are carbon, hydrogen,
oxygen, nitrogen, phosphorus, and sulphur. We can let each one be represented by its initial
letter: C, H, O, N, P,. and S. The initial letters could also stand for a single atom of each
element. C could be a carbon atom, H a hydrogen atom, and so on.
Of these elements, carbon is, in a way, the crucial one. Carbon atoms can combine with each
other to form long chains, which can branch in complicated ways. They can also form single
rings or groups of rings; or, for that matter, rings with chains attached. To the carbon atoms
arranged in any of these ways, other atoms can be attached in different manners.
These complicated chains and rings of carbon atoms are found only in organic compounds,
never in inorganic compounds. It is this which makes organic molecules larger and more
complicated than inorganic ones.
Carbon atoms can be hooked together in so many ways, and can attach other atoms to
themselves in so many ways that there is almost no end to the different variations. And
each different variation is a different substance with different properties.
Hundreds of thousands of different organic compounds are known today. Every year many
more organic compounds are discovered and there is no danger of ever running out of new
ones. Uncounted trillions upon trillions of such compounds can exist.
This seems to make the problem of the origin of life more difficult again. If we are trying to
find out how organic substances are formed from inorganic ones, and if there are uncounted
trillions upon trillions of organic substances possible, how can we decide which organic
substance ought to be formed and which were formed in the past.
www.seyfihoca.com
275
Suppose, though, we can narrow down the choice. Not all organic compounds are equally
vital to life. Some of them seem to be more central to the basic properties of life than others
are.
All cells without exception, whether plant, animal, or bacterial, seem to be built about two
kinds of substances that are more important than any others. These are "proteins" and "nucleic
acids."
Even viruses can be included here. They are tiny objects, far smaller than even the smallest
cells, yet they seem to be alive since they can invade cells and multiply there. They, too,
contain proteins and nucleic acids. Some viruses, in fact, contain practically nothing else but
proteins and nucleic acids.
Now we have narrowed the problem. We must not ask how organic compounds were built up
out of inorganic ones, but how proteins and nucleic acids were built up out of them.
That still leaves matters complicated enough. Both proteins and nucleic acids are made up of
very large molecules, often containing millions of atoms. It is too much to expect that small
inorganic molecules would come together suddenly to form a complete molecule of protein or
nucleic acid.
Let's look more closely at such giant molecules. Both proteins and nucleic acids are composed
of simpler structures strung together like beads on a necklace. Both protein and nucleic acid
molecules can be treated chemically in such a way that the string breaks and the individual
"building blocks" separate. They can then be studied separately.
In the case of the protein molecule, the building blocks are called "amino acids." The
molecule of each amino acid is built around a chain of three atoms, two of which are carbon
and one nitrogen. We can write this chain as -C-C-N-.
There would be different atoms attached to each of these. The atoms attached to the carbon
and nitrogen atoms at the end are always the same in all the amino acids obtained from
proteins (with a minor exception we needn't worry about). The carbon atom in the middle,
however, can have any of a number of different atom-groupings attached to it. If we call this
atom-grouping R, then the amino acid would look like this:
Each different structure for R results in a slightly different amino acid. Altogether there are
nineteen different amino acids that are found in almost every protein molecule. The simplest
R consists of just a hydrogen atom. The rest all contain different numbers of carbon and
hydrogen atoms, while some contain one or two oxygen atoms in addition, or one or two
nitrogen atoms, or even one or two sulphur atoms. Individual amino acids are made up of
from eleven to twenty-six atoms.
Although there are only nineteen different amino acids in most proteins, they can be put
together in many different ways, each way making up a slightly different molecule. Even a
www.seyfihoca.com
276
middle-sized protein molecule is made up of several hundred of these amino acids and the
number of different combinations is enormous.
Imagine yourself to be given several hundred beads of nineteen different colours and that you
set to work to string them. You could make necklaces of many trillions of different colour
combinations. In the same way, you could imagine protein molecules of many trillions of
different amino acid combinations.
In thinking of the origin of life, then, you don't have to worry, just at first, about forming
complicated protein molecules. That would come later. To begin with, it would be satisfying
to know whether the amino acid building blocks could be formed and, if so, how.
The nucleic acids are both simpler and more complicated than the protein. Nucleic acid
molecules are made up of fewer different kinds of building blocks but the individual building
block is more complicated.
The huge nucleic acid molecule is made up of long chains of smaller compounds known as
"nucleotides," each of which is made up of about three dozen atoms. These include carbon,
hydrogen, oxygen, nitrogen, and phosphorus.
An individual nucleotide molecule is made up of three parts. First there is a one-ring or tworing combination made up of carbon and nitrogen atoms. If there is only one ring, this portion
is called a "pyrimidine"; two rings is a "purine."
The second portion is made up of a ring of carbon and oxygen atoms. This comes in two
varieties. One is called "ribose"; the other, with one fewer oxygen atom, is "deoxyribose."
Both these compounds belong to the class called sugars.
Finally, the third part is a small atom group containing a phosphorus atom. It is the
"phosphate group."
We might picture a nucleotide as follows:
purine or
pyrimidine
-
ribose or
phosphate
deoxyribose
group
There are two kinds of nucleic acid molecules. One of them is built up of nucleotides that all
contain ribose. This is, there fore, "ribosenucleic acid" or RNA. The other is built up of
nucleotides that all contain deoxyribose; "deoxyribosenucleic acid" or DNA.
In both cases, individual nucleotides vary in the particular kind of purine or pyrimidine they
contain. Both RNA and DNA are made up of chains of four different nucleotides. Even
though there are only four different nucleotides, so many of them are present in each
enormous nucleic acid molecule that they can be arranged in trillions upon trillions of
different ways.
Now that we have decided we want to form amino acids and nucleotides out of inorganic
compounds, we must ask out of what inorganic compounds we can expect them to be formed.
www.seyfihoca.com
277
We must have inorganic compounds, to start with, that contain the right atoms: carbon,
hydrogen, oxygen, and the rest.
To begin with, there is the water molecule in the oceans. That is made up of two hydrogen
atoms and an oxygen atom and it can therefore be written H2O. Then there is the carbon
dioxide of the air, which dissolves in the ocean water and which is made up of a carbon atom
and two oxygen atoms, C02. Water and carbon dioxide can supply carbon, hydrogen, and
oxygen, three of the necessary elements.
Also dissolved in ocean water are substances that are called nitrates, sulphates, and
phosphates. They contain nitrogen atoms, sulphur atoms, and phosphorus atoms respectively.
These substances all have certain properties in common with ordinary table salt and can be
lumped together as "salts."
What we have to ask ourselves now is this: Is it possible that once long ago, when the world
was young, water, carbon dioxide, and salts combined to form amino acids and nucleotides. If
so, how was it done?
There are certain difficulties in this thought.
To begin with, in order for water, carbon dioxide, and salts to form amino acids and
nucleotides, oxygen atoms must be discarded. There is much more oxygen in water, carbon
dioxide, and salts, than there is in amino acids and nucleotides.
But Earth's atmosphere contains a great deal of oxygen. To discard oxygen, when oxygen is
already all about, is very difficult. It is like trying to bail the water out of a boat that is resting
on the lake bottom.
Secondly, it takes energy to build up amino acids and nucleotides out of simple inorganic
molecules and the most likely source is sunlight. Just sunlight isn't enough, however. To get
enough energy, you must use the very energetic portion of the sunlight; you must use
ultraviolet waves.
But very little of the ultraviolet waves gets down to the surface of the Earth. The air absorbs
most of it. When scientists studied the situation more closely it turned out that it was the
oxygen in the air that produced the substance that absorbed the ultraviolet.
So oxygen was a double villain. It kept the ultraviolet away from the surface of the Earth and
its presence made it very difficult to discard excess oxygen.
To be sure, the plant life that covers the land and fills the sea is carrying through just the sort
of thing we are talking about and doing it right now. Plants absorb water, carbon dioxide, and
salts and use the energy of sunlight to manufacture all sorts of complicated organic
compounds out of them. In doing so, they discard oxygen and pour it into the atmosphere.
However, to do this, plants make use of visible light, not ultraviolet waves. Visible light
(unlike ultraviolet waves) can penetrate the atmosphere easily, so that it is available for the
plants to use. Visible light has considerably less energy than ultraviolet waves but the plants
make use of it anyway.
www.seyfihoca.com
278
You might wonder if this could not have happened on the early Earth. Suppose the energy of
visible light had been used to build up the amino acids and nucleotides.
It doesn't seem likely, though, that it could have happened that way. The reason it happens
now is that plants make use of a complicated chemical system that includes a substance
known as "chlorophyll." Chlorophyll is an organic compound with a most complicated
molecule that is formed only by living organisms.
In thinking of the early Earth, a planet without life on it, we must suppose that chlorophyll
was absent. Without chlorophyll, the energy of visible light is not enough to form amino acids
and nucleotides. The more energetic ultraviolet waves are necessary and that can't pass
through our atmosphere.
We seem to be stuck.
But then, in the 1920s, an English biochemist, John Burdon Sanderson Haldane, suggested
that oxygen had not always existed in Earth's atmosphere.
After all, plant life is always using up carbon dioxide and producing oxygen, as it forms
organic substances from inorganic substances. Might it not be that all the oxygen that is now
in the Earth's atmosphere is the result of plant action? Before there was life, and therefore
before there were plants, might not the atmosphere have been made up of nitrogen and carbon
dioxide, instead of nitrogen and oxygen, as today?
If that were the case, ultraviolet waves could get right down to the Earth's surface without
being much absorbed. And, of course, oxygen could be discarded with much greater ease.
The suggestion turned the whole question in a new direction. It wasn't proper to ask how
amino acids and nucleotides might be formed from small compounds that are now available
under conditions as they exist now. Instead we must ask how amino acids and nucleotides
might be formed from small compounds that would be available when the Earth was a young
and lifeless planet under conditions as they existed then.
It became necessary to ask, then, what kind of an atmosphere and ocean the Earth had before
life developed upon it.
That depends on what the universe is made up of, generally. In the nineteenth century, ways
were worked out whereby the light from the stars could be analyzed to tell us what elements
were to be found in those stars (and even in the space between the stars).
Gradually, during the early decades of the twentieth century, astronomers came more and
more to the conclusion that by far the most common atoms in the universe were the two
simplest: hydrogen and helium. In general, you can say that 90 percent of all the atoms in the
universe are hydrogen and 9 percent are helium. All the other elements together make up only
1 percent or less. Of these other elements, the bulk was made up of carbon, nitrogen, oxygen,
sulphur, phosphorus, neon, argon, silicon, and iron.
www.seyfihoca.com
279
If that is so, then you might expect that when a planet forms out of the dust and gas that fills
certain sections of space, it ought to be mostly hydrogen and helium. These are the gases that
would make up most of the original atmosphere.
Helium atoms do not combine with any other atoms, but hydrogen atoms do. Because
hydrogen atoms are present in such quantities, any type of atom that can combine with
hydrogen will do so.
Each carbon atom combines with four hydrogen atoms to form "methane" (CH4). Each
nitrogen atom combines with three hydrogen atoms to form "ammonia" (NH3). Each sulphur
atom combines with two hydrogen atoms to form "hydrogen sulphide" (H2S). And, of course,
oxygen atoms combine with hydrogen to form water.
These hydrogen-containing compounds are all gases, or liquids that can easily be turned into
gases, so they would all be found in the primitive atmosphere and ocean.
The silicon and iron atoms, together with those of various other fairly common elements such
as sodium, potassium, calcium, and magnesium, don't form gases. They make up the solid
core of the planet.
This sort of logic seems reasonable, for a large, cold planet like Jupiter was found, in 1932, to
have just this sort of atmosphere. Its atmosphere is chiefly hydrogen and helium, and it
contains large quantities of ammonia and methane.
Jupiter is a huge planet, however, with strong gravitation. Smaller planets like Earth, Venus,
or Mars, have gravitation that is too weak to hold the very small and very nimble helium
atoms or hydrogen molecules. (Each hydrogen molecule is made up of two hydrogen atoms,
H2.)
On Earth, therefore, we would expect the very early atmosphere to contain mostly ammonia,
methane, hydrogen sulphide, and water vapour. Most of the water would go to make up the
ocean and in that ocean would be dissolved ammonia and hydrogen sulphide. Methane is not
very soluble but small quantities would be present in the ocean also.
If we began with such an atmosphere, would it stay like that forever? Perhaps not. Earth is
fairly close to the sun and a great deal of ultraviolet waves strike the Earth's atmosphere.
These ultraviolet waves are energetic enough to tear apart molecules of water vapour in the
upper atmosphere and produce hydrogen and oxygen.
The hydrogen can't be held by Earth's gravity and drifts off into space, leaving the oxygen
behind. (Oxygen forms molecules made up of two oxygen atoms each, 02, and these are heavy
enough to be held by Earth's gravity.)
The oxygen does not remain free, however. It combines with the carbon and hydrogen atoms
in methane to form carbon dioxide and water. It wouldn't combine with the nitrogen atoms of
ammonia, but it would combine with the hydrogen to form water, leaving the nitrogen over to
form molecules made up of two atoms each (N2).
www.seyfihoca.com
280
Little by little, as more and more water is broken apart by ultraviolet light, all the ammonia
and methane in the atmosphere is converted to nitrogen and carbon dioxide. In fact, the
planets Mars and Venus seem to have a nitrogen plus carbon dioxide atmosphere right now.
You might wonder, though, what could happen if all the ammonia and methane were
converted to nitrogen and carbon dioxide and if water molecules continued to break up into
hydrogen and oxygen. The oxygen would not have anything more to combine with. Perhaps it
would gradually accumulate in the air.
This, however, would not happen. As free oxygen accumulates, the energy of sunlight turns
some of it into a three-atom combination called "ozone" (O3). This ozone absorbs the
ultraviolet light of the sun and because the ozone layer forms about fifteen miles high in the
atmosphere, the ultraviolet light is shielded from the regions of the atmosphere where water
vapour exists.
No further water molecules can be broken up and the whole process comes to an end before
oxygen can really fill the atmosphere. It is only later on when plants develop and make use of
chlorophyll to tap the energy of visible light which can get through the ozone layer that the
process begins again. After plants come on the scene, the atmosphere fills with oxygen.
So we have three atmospheres for Earth. The first, "Atmosphere I" was chiefly ammonia,
methane, and water vapour, with an ocean containing much ammonia in solution.
"Atmosphere II" was chiefly nitrogen, carbon dioxide, and water vapour, with an ocean
containing much carbon dioxide in solution. Our present atmosphere "Atmosphere III," is
chiefly nitrogen, oxygen, and water vapour, with an ocean in which only small quantities of
gas are dissolved.
Atmosphere III formed only after life had developed, so life must have originated in the first
place in either Atmosphere I or Atmosphere II (or possibly while Atmosphere I was changing
into Atmosphere II) .
Haldane had speculated that life had originated in Atmosphere II, but a Russian biochemist,
Alexander Ivanovich Oparin, thought otherwise.
In 1936, he published a book called The Origin of Life, which was translated into English in
1938. Oparin was the first to go into the problem of the origin of life in great detail, and he
felt that life must have originated in Atmosphere I.
How was one to decide which was the correct answer? How about experiment? Suppose you
were actually to start with a particular mixture of gases that represents an early atmosphere
and add energy in the way it might have been added on the early Earth. Will more
complicated compounds be formed out of simple ones? And if they are, will they be the kind
of compounds that are found in living creatures?
The first scientist who actually tried the experiment was Melvin Calvin at the University of
California.
www.seyfihoca.com
281
In 1950, he began to work with a portion of Atmosphere 11- carbon dioxide and water
vapour. The fact that he left out nitrogen meant that he couldn't possibly form nitrogen
containing molecules, like amino acids and nucleotides. However, he was curious to see what
he would get.
What he needed, to get anything at all, was a source of energy. He might have used ultraviolet
waves, the most likely source on the early Earth, but he preferred not to.
Instead, he made use of the energy of certain kinds of atoms that were always exploding.
They were "radioactive" atoms. The radioactive elements on Earth are very slowly breaking
down so that every year there are very slightly less than the year before. Several billion years
ago there must have been twice as much radioactivity in the Earth's crust as there is now. The
energy of radioactivity could therefore have been important in forming life.
Since Melvin Calvin was engaged in experimental work that made use of radioactive
substances, he had a good supply
of them to work with. He bombarded his gas mixture with flying particles released by
radioactive atomic explosions. After a while, he tested the gas mixture and found that in
addition to carbon dioxide and water, he had some very simple organic molecules in solution.
He had, for instance, a molecule containing one carbon atom, two hydrogen atoms, and one
oxygen atom (CH2O), which was well known to chemists under the name of "formaldehyde."
He also had formic acid, which has a second oxygen atom, and has a formula written HCOOH
by chemists.
This was just a beginning but it showed a few important things. It showed that molecules
could be made more complicated under early Earth conditions. For another the complicated
molecules contained less oxygen than the original molecules, so that oxygen was being
discarded.
In 1953 came an important turning point, something that was the key discovery in the search
for the origin of life. It came in the laboratories of Harold Clayton Urey at the University of
Chicago.
Urey was one of those who had tried to reason out the atmosphere of the early Earth, and, like
Oparin, he felt it was in Atmosphere I that life might have got its start. He suggested to one of
his students, Stanley Lloyd Miller, that he set up an experiment in which energy would be
added to a sample of Atmosphere I. (At the time Miller was in his early twenties, working for
his Ph.D. degree.)
Miller set up a mixture of ammonia, methane, and hydrogen in a large glass vessel. In another
glass vessel, he boiled water. The steam that was formed passed up a tube and into the gas
mixture. The gas mixture was pushed by the steam through another tube back into the boiling
water. The second tube was kept cool so that the steam turned back into water before dripping
back into the hot water.
The result was that a mixture of ammonia, methane, hydrogen, and water vapour was kept
circulating through the system of vessels and tubes, driven by the boiling water. Miller made
www.seyfihoca.com
282
very certain that everything he used was completely sterile; that there were no bacteria or
other cells in the water or in the gases. (If he formed complicated compounds he wanted to
make sure they weren't formed by living cells.)
Next, energy had to be supplied. Urey and Miller reasoned that two likely sources of energy
were ultraviolet light from the sun and electric sparks from lightning. (There may have been
numerous thunderstorms in Earth's early days.)
Of the two, ultraviolet light is easily absorbed by glass and there was a problem as to how to
get enough energy through the glass into the chemicals within. Miller therefore thought that as
a first try he would use an electric spark like a small bolt of lightning. Through the gas in one
portion of the system he therefore set up a continuing electric spark.
Now it was only necessary to wait.
Something was happening. The water and gases were colourless to begin with, but by the end
of one day, the water had turned pink. As the days continued to pass, the colour grew darker
and darker, till it was a deep red.
After a week, Miller was ready to see what he had formed in his water reservoir. Fortunately,
he had at his disposal a new technique for separating and identifying tiny quantities of
chemical substances. This is called "paper chromatography" and it had been first developed in
1944 by a group of English chemists.
Like Calvin, Miller found that the simple gas molecules had combined with each other to
form more complicated molecules, discarding oxygen atoms.
Again like Calvin, Miller found that formic acid was an important product. He also found,
however, that compounds had been formed which were similar to formic acid but were more
complicated. These included acetic acid, glycolic acid,
and lactic acid, all substances that were intimately associated with life.
Miller had begun with a nitrogen-containing gas, ammonia, which Calvin had lacked. It is not
surprising, therefore, that Miller ended up with some molecules that contained nitrogen as
well as carbon, hydrogen, and oxygen. He found some hydrogen cyanide, for instance, which
is made up of a carbon atom, a hydrogen atom, and a nitrogen atom in its molecule (HCN).
He also found urea, which has molecules made up of two nitrogen atoms, four hydrogen
atoms, a carbon atom, and an oxygen atom (NH2CONH2).
Most important of all, though, Miller discovered among his products two of the nineteen
amino acid building blocks that go to make up the various protein molecules. These were
"glycine" and "alanine," the two simplest of all the amino acids, but also the two that appear
most frequently in proteins.
With a single experiment, Miller seemed to have accomplished a great deal. In the first place,
these compounds had formed quickly and in surprisingly large quantities. One-sixth of the
methane with which he had started had gone into the formation of more complex organic
compounds.
www.seyfihoca.com
283
He had only worked for a week, and with just a small quantity of gas. How must it have been
on the early Earth, with its warm ocean, full of ammonia, and with winds of methane blowing
over it, all baking under the sun's ultraviolet radiation or being lashed by colossal lightning
bolts for a billion years?
Millions of tons of these complex compounds must have been formed, so that the ocean
became a kind of "warm soup." Secondly, the kind of organic molecules formed in Miller's
experiment proved particularly interesting. Among the first compounds formed were simple
amino acids, the building blocks of proteins. In fact, the path taken by the simple molecules as
they grew more complex seemed pointed directly towards life. No molecules were formed
that seemed to point in an unfamiliar direction.
Suppose that, as time went on, more and more complicated molecules were built up, always in
the direction of compounds now involved with life and not in other directions. Gradually,
bigger and bigger molecules would form as building blocks would join together. Finally,
something like a real protein molecule and nucleic acid molecule would form and these would
eventually associate with each other in a very simple kind of cell.
All this would take a lot of time, to be sure. But then, there was a whole ocean of chemicals to
work with, and there was lots of time - a billion years, at least.
Miller's experiment was only a beginning, but it was an extremely hopeful beginning. When
its results were announced, a number of biochemists (some of whom were already thinking
and working in similar directions) began to experiment in this fashion.
In no time at all, Miller's work was confirmed; that is, other scientists tried the same
experiment and got the same results. Indeed, Philip Hauge Abelson, working at the Carnegie
Institution of Washington, tried a variety of experiments with different gases in different
combinations.
It turned out that as long as he began with molecules that included atoms of carbon, hydrogen,
oxygen, and nitrogen somewhere in their structure, he always found amino acids included
among the substances formed. And they were always amino acids of the kind that served as
protein building blocks.
Nor were electric discharges the only source of energy that would work. In 1959, two German
scientists, Wilhelm Groth and H. von Weyssenhoff, tried ultraviolet waves and they also got
amino acids.
It could be no accident. There was a great tendency for atoms to click together in such a way
as to produce amino acids. Under the conditions that seemed to have prevailed on the early
Earth, it appeared impossible not to form amino acids.
By 1968, every single amino acid important to protein structure had been formed in such
experiments. The last to be formed were certain important sulphur-containing amino acids,
according to a report from Pennsylvania State University and from George Williams
University in Montreal.
www.seyfihoca.com
284
Perhaps other important compounds also couldn't help but form. Perhaps they would just
naturally come together to form the important large molecules of living tissue.
If that is so, life may be no "miracle." It couldn't help forming, any more than you can help
dropping downward if you jump off a building. Any planet that is something like the Earth,
with a nearby sun and a supply of water and an atmosphere full of hydrogen compounds,
would then have to form life. The kinds of creatures that eventually evolved on other worlds
would be widely different and might not resemble us any more than an octopus resembles a
gorilla. But, the chances are, they would be built up of the same chemical building blocks as
we.
More and more, scientists are beginning to think in this way, and they are beginning to
speculate that life may be very common in the universe.
Of course, on planets that are quite different from Earth; much bigger and colder, like Jupiter,
or much smaller and hotter, like Mercury, our kind of life could not form. On the other hand,
other kinds of life, based on other types of chemistry, might be formed. We have no way of
telling.
But we are getting ahead of ourselves. Miller's experiments were enough to start speculation
of this sort, but it was still important to check matters. A couple of amino acids weren't
enough. What about the nucleotides, which served as building blocks for nucleic acids? (Since
the 1940s, biochemists have come to believe that nucleic acids are even more important than
proteins.)
One could repeat Miller's experiment for longer and longer periods, hoping that more and
more complicated molecules would be formed. However, as more and more kinds of
compounds were formed, there would be less and less of each separate kind, and it would
become more difficult to spot each one.
Possibly, one could start with bigger and bigger quantities of gases in the first place. Even so,
the large number of complicated molecules that would be formed would confuse matters.
It occurred to some experimenters to begin not at the beginning of Miller's experiment, but at
its end. For instance, one of the most simple products of Miller's experiment was hydrogen
cyanide, HCN.
Suppose you assumed that this gas was formed in quantity in Earth's early ocean and then
started with it. In that way you would begin partway along the road of development of life and
carry it on further.
At the University of Houston, a Spanish-born biochemist, Juan Oro, tried just this in 1961. He
found that not only amino acids were formed once he added HCN to the starting mixture, but
individual amino acids were hooked together in short chains, in just the way in which they are
hooked together in proteins.
Even more interesting was the fact that purines were formed, the double rings of carbon and
nitrogen atoms that are found in nucleotides. A particular purine called "adenine" was
www.seyfihoca.com
285
obtained. This is found not only in nucleic acids but in other important compounds associated
with life.
As the 1960s opened, then, imitations of the chemical environment of the early Earth were
being made to produce not only the building blocks of the proteins, but the beginnings of the
nucleotide building blocks of the nucleic acids.
It was just the beginnings in the latter case. The nucleotides contained not only purines but
also the somewhat similar, but simpler, one-ringed compounds, the pyrimidines. Then there
were the sugars, ribose and deoxyribose. And, of course, there was the phosphate group.
The experimenters bore on. All the necessary purines and pyrimidines were formed. The
sugars proved particularly easy. Sugar molecules are made up of carbon, hydrogen, and
oxygen atoms only. No nitrogen atoms are needed. That reminded one of Calvin's original
experiment. Calvin had obtained formaldehyde (CH20) from carbon dioxide and water. What
if one went a step farther and began with formaldehyde and water.
In 1962, Oro found that if he began with formaldehyde in water and let ultraviolet waves fall
upon it, a variety of sugar molecules were formed, and among them were ribose and
deoxyribose.
What next?
Purines and pyrimidines were formed. Ribose and deoxyribose were formed. Phosphate
groups didn't have to be formed. They existed in solution in the ocean now, and very likely
did then, in just the form they existed in inorganic molecules.
One researcher who drove onward was a Ceylon-born biochemist, Cyril Ponnamperuma, at
Ames Research Center at Moffett Field, California. He had conducted experiments in which
he had, as a beginning, formed various purines with and without hydrogen cyanide. He had
formed them through the energy of beams of electrons (very light particles) as well as
ultraviolet waves.
In 1963, he, along with Ruth Mariner and Carl Sagan, began a series of experiments in which
he exposed a solution of adenine and ribose to ultraviolet waves. They hooked together in just
the fashion they were hooked together in nucleotides. If the experimenters began with
phosphate also present in the mixture, then the complete nucleotide was
formed. Indeed, by 1965, Ponnamperuma was able to announce that he had formed a double
nucleotide, a molecule consisting of two nucleotides combined in just the fashion found in
nucleic acids.
By the middle 1960s, then, it seemed clear to biochemists that the conditions on the early
Earth were capable of leading to the formation of a wide variety of substances associated with
life. These would certainly include the amino acids and nucleotides, those building blocks that
go to make up the allimportant proteins and nucleic acids. Furthermore, these building blocks
hook together under early conditions to make up the very chains out of which proteins and
nucleic acids are formed.
www.seyfihoca.com
286
All the raw materials for life were there on the early Earth, all the necessary chemicals. But
life is more than just chemicals. There are all sorts of chemical changes going on in living
organisms, and they must be taken into account. Atoms and groups of atoms are shifting here,
shifting there, coming apart and reuniting in different ways.
Many of these changes won't take place unless energy is supplied. If we're dealing with the
ocean, the energy is supplied by the sun's ultraviolet radiation, or in other ways. But what
happens inside the tiny living creatures once they come into existence?
Actually, there are certain chemicals in living creatures which break up easily, releasing
energy. Such chemicals make it possible for important chemical changes to take place that
would not take place without them. Without such chemicals life as we know it would be
impossible no matter how many proteins and nucleic acids built up in the early ocean.
Could it be that some of the energy of sunlight went into the production of these energy-rich
compounds? In that case, everything necessary for life might really be supplied.
The best-known of the energy-rich compounds is one called "adenosine triphosphate," a name
that is usually abbreviated as ATP. It resembles a nucleotide to which two additional
phosphate groups (making three altogether) have been added.
If, then, adenine, ribose, and phosphate groups are exposed to ultraviolet waves and if they
hook together to form a nucleotide containing one phosphate group, perhaps we can go
farther. Perhaps longer irradiation or the use of more phosphate to begin with will cause them
to hook together to form ATP, with three phosphate groups. Ponnamperuma tried, and it
worked. ATP was formed.
In 1967 a type of molecule belonging to a class called "porphyrins" was synthesized from
simpler substances by Ponnamperuma. Belonging to this class is the important chlorophyll
molecule in green plants.
No one doubts now that all the necessary chemicals of life could have been produced in the
oceans of the early Earth by chemical reactions under ultraviolet.
To be sure, the life that was formed at first was probably so simple that we might hesitate to
call it life. Perhaps it consisted of a collection of just a few chemicals that could bring about
certain changes that would keep the collection from breaking apart. Perhaps it would manage
to bring about the formation of another collection like itself.
It may be that life isn't so clear-cut a thing that we can point a finger and say: Right here is
something that was dead before and is now alive.
There may be a whole set of more and more complex systems developing over hundreds of
millions of years. To begin with, the systems would be so simple that we couldn't admit they
were alive. To end with, they would be so complex that we would have to admit they were
indeed alive. But where, in between, would be the changeover point?
We couldn't tell. Maybe there is no definite changeover point. Chemical systems might just
slowly become more and more "alive" and where they passed the key point, no one could say.
www.seyfihoca.com
287
With all the successful production of compounds that followed the work of Calvin and Miller,
there still remained the question of how cells were formed. The experimenters who formed
compounds recognized that that question would have to be answered somehow.
No one type of compound is living, all by itself. Everything that seems living to us is a
mixture of all sorts of substances which are kept close together by a membrane and which
react with each other in a very complicated way.
There are viruses, to be sure, which are considered alive and which sometimes consist of a
single nucleic acid molecule wrapped in a protein shell. Such viruses, however, don't really
get to work in a truly living way till they can get inside some cell. In there, they make use of
cell machinery.
Haldane, who had started the modern attack on the problem, wondered how cells might have
formed. He pointed out that when oil is added to water, thin films of oil sometimes form
bubbles in which tiny droplets of water are enclosed.
Some of the compounds formed by the energy of ultraviolet light are oily and won't mix with
water. What if they were to form a little bubble and just happen to enclose a proper mixture of
protein, nucleic acid, and other things? Today's cell membrane may be the development of
that early oily film.
Oparin, the Russian biochemist, went into further detail. He showed that proteins in solution
might sometimes gather together into droplets and form a kind of skin on the outside of those
droplets.
The most eager experimenter in this direction, once Miller's work had opened up the problem,
was Sidney W. Fox at the University of Miami. It seemed to him that the early Earth must
have been a hot planet indeed. Volcanoes may have kept the dry land steaming and brought
the ocean nearly to a boil.
Perhaps the energy of heat alone was sufficient to form complex compounds out of simple
ones.
To test this, Fox began with a mixture of gases like that in Atmosphere I (the type that Oparin
suggested and Miller used) and ran them through a hot tube. Sure enough, a variety of amino
acids, at least a dozen, were formed. All the amino acids that were formed happened to be
among those making up proteins. No amino acids were formed that were not found in
proteins.
Fox went a step farther. In 1958, he took a bit of each of the various amino acids that are
found in protein, mixed them together, and heated the mixture. He found that he had driven
the amino acids together, higgledy-piggledy, into long chains which resembled the chains in
protein molecules. Fox called these chains "proteinoids" (meaning "protein-like"). The
likeness was a good one. Stomach juices, which digest ordinary protein, would also digest
proteinoids. Bacteria, which would feed and grow on ordinary protein, would also feed and
grow on proteinoids.
www.seyfihoca.com
288
Most startling of all, when Fox dissolved the proteinoids in hot water and let the solution cool,
he found that the proteinoids clumped together in little spheres about the size of small
bacteria. Fox called these "microspheres."
These microspheres are not alive, but in some ways they behaved as cells do. They are
surrounded by a kind of membrane. Then, by adding certain chemicals to the solution, Fox
could make the microspheres swell or shrink, much as ordinary cells do. The microspheres
can produce buds, which sometimes seem to grow larger and break off. Microspheres can
divide in two or cling together in chains.
Not all scientists accept Fox's arguments, but what if, on the early Earth, more and more
complicated substances were built up, turning the ocean into the "warm soup" we spoke of.
What if these substances formed microspheres? Might it not be that, little by little, as the
substances grew more complicated and the microspheres grew more elaborate, that eventually
an almost-living cell would be formed? And after that, a fully living one?
Before life began, then, and before evolutionary changes in cells led to living creatures that
were more and more complicated, there must first have been a period of "chemical evolution."
In this period, the very simplest gases of the atmosphere and ocean gradually become more
and more complicated until life and cells formed.
All these guesses about the origin of life, from Haldane on, are backed up by small
experiments in the laboratory and by careful reasoning. Is it possible that we might find traces
of what actually happened on the early Earth if we look deep into the Earth's crust.
We find out about ordinary evolution by studying fossils in the crust. These are the remains of
ancient creatures, with their bones or shells turned to stone. From these stony remains we can
tell what they looked like and how they must have lived.
Fossils have been found deep in layers of rock that must be 600 million years old. Before that
we find hardly anything. Perhaps some great catastrophe wiped out the earlier record. Perhaps
forms of life existed before then that were too simple to leave clear records.
Actually, in the 1960s discoveries were reported of traces left behind by microscopic one-cell
creatures in rocks that are more than two billion years old. Prominent in such research is Elso
Sterrenberg Barghoorn of Harvard. It is a good guess that there were simple forms of life on
Earth at least as long as three billion years ago.
If we are interested in discovering traces of the period of chemical evolution, then, we must
search for still older rocks. In them, we might hope to find chemicals that seem to be on the
road to life.
But will chemicals remain unchanged in the Earth for billions of years? Can we actually find
such traces if we look for them?
Certainly the important chemicals of life, the proteins and nucleic acids, are too complex to
remain unchanged for long after the creature they were in dies and decomposes. In a very
short time, it would seem, they must decompose and fall apart.
www.seyfihoca.com
289
And yet, it turns out, sometimes they linger on, especially when they are in a particularly
well-protected spot. Abelson, one of the people who experimented with early atmospheres,
also worked with fossils. He reasoned that living bones and shell contain protein. Bones may
be 50 percent protein. Clam shells have much less, but there is some. Once such bones and
shells are buried deep in the Earth's crust, remaining there for millions of years while they
turned to stone, it might be that some of the protein trapped between thin layers of mineral
might survive.... Or at least they might break down to amino acids or short chains of amino
acids that might survive.
Painstakingly, Abelson dissolved these ancient relics and analyzed the organic material he
extracted. There were amino acids present all right, exactly the same amino acids that are
present in proteins of living creatures. He found some even in a fossil fish which might have
been 300 million years old.
Apparently, then, organic compounds last longer than one might think and Melvin Calvin
began the search for "chemical fossils" in 1961.
In really old rocks, it is unlikely that the organic chemicals would remain entirely untouched.
The less hardy portions would be chipped away. What would linger longest would be the
chains and rings of carbon atoms, with hydrogen atoms attached. These compounds of carbon
and hydrogen only are called "hydrocarbons."
Calvin has isolated hydrocarbons from ancient rocks as much as three billion years old. The
hydrocarbons have molecules of a complicated structure that looked very much as though
they could have originated from chemicals found in living plants.
J. William Schopf of Harvard, a student of Barghoorn, has gone even further. He has detected
traces of 22 different amino acids in rocks more than three billion years old.
They are probably the remnants of primitive life. It is necessary now to probe farther back and
find chemical remnants that precede life and show the route actually taken.
Very likely it will be the route worked out by chemists in their experiments, but possibly it
won't be.
We must wait and see. And perhaps increasing knowledge of what went on in the days of
Earth's youth will help us understand more about life now.
From: Twentieth Century Discovery by Isaac Azimov
3 - Littler And Littler: The Structure Of Matter
One of the words that fascinates scientists in the 1960s is "quark."
No one has ever seen a quark or come across one in any way. It is far too small to see and no
one is even sure it exists. Yet scientists are anxious to build enormous machines costing
hundreds of millions of dollars to try to find quarks, if they exist.
www.seyfihoca.com
290
This is not the first time scientists have looked for objects they weren't sure existed, and were
too small to see even if they did exist. They were doing it as early as the very beginning of the
nineteenth century.
In 1803, an English chemist, John Dalton, suggested that a great many chemical facts could
be explained if one would only suppose that everything was made up of tiny particles, too
small to be seen under any microscope. These particles would be so small that there couldn't
be anything smaller. Dalton called these particles "atoms" from Greek words meaning "not
capable of being divided further." Dalton's suggestion came to be called the "atomic theory."
No one was sure that atoms really existed, to begin with, but they did turn out to be very
convenient. Judging by what went on in test tubes, chemists decided that there were a number
of different kinds of atoms.
When a particular substance is made up of one kind of atom only, it is an "element." Iron is an
element, for instance, and is made up only of iron atoms. Gold is an element; so is the oxygen
in the air we breathe.
Atoms can join together into groups and these groups are called "molecules." Oxygen atoms
get together in groups of two and these two-atom oxygen groups are called oxygen molecules.
The oxygen in the air is made up of oxygen molecules, not of separate oxygen atoms.
Atoms of different elements can come together to form molecules of "compounds." Water is a
compound with molecules made up of two hydrogen atoms and one oxygen atom.
Dalton and the nineteenth century chemists who followed him felt that every atom was just a
round little ball. There was no reason to think there was anything more to it than that. They
imagined that if an atom could be seen under a very powerful microscope, it would turn out to
be absolutely smooth and shiny, without a mark.
They were also able to tell that the atom was extremely small. They weren't quite certain
exactly how small it was but nowadays we know that it would take about 250 million atoms
laid side by side to stretch across a distance of only one inch.
The chief difference between one kind of atom and another kind, in the nineteenth century
view, lay in their mass, or weight. Each atom had its own particular mass, or "atomic weight."
The hydrogen atom was the lightest of all, and was considered to have an atomic weight of l.
An oxygen atom was about sixteen times as massive as a hydrogen atom, so it had an atomic
weight of 16. A mercury atom had an atomic weight of 200, and so on.
As the nineteenth century wore on, the atomic theory was found to explain more and more
things. Chemists learned how atoms were arranged inside molecules and how to design new
molecules so as to form substances that didn't exist in nature.
By the end of the century, the atomic theory seemed firmly established. There seemed no
room for surprises.
And then, in 1896, there came a huge surprise that blew the old picture into smithereens. The
chemists of the new twentieth century were forced into a new series of investigations that led
them deep into the tiny atom.
www.seyfihoca.com
291
What happened in 1896 was that a French physicist, Antoine Henri Becquerel, discovered
quite by accident that a certain substance had properties no one had ever dreamed of before.
Becquerel had been interested in x rays, which had only been discovered the year before. He
had samples of a substance containing atoms of the heavy metal uranium in its molecules.
This substance gave off light of its own after being exposed to sunlight and Becquerel
wondered if this light might include x rays.
It didn't, but Becquerel found it gave off mysterious radiations of some kind; radiations that
went right through black paper and fogged a photographic film. It eventually turned out that it
was the uranium atoms that were doing it. The uranium atoms were exploding and hurling
small fragments of themselves in every direction.
Scientists had never expected atoms could explode, but here some of them were doing it. A
new word was invented. Uranium was "radioactive."
Other examples of radioactivity were found and physicists began to study the new
phenomenon with great interest as the twentieth century opened.
One thing was clear at once. The atom was not just a hard, shiny ball that could not be divided
into smaller objects. Small as it was, it had a complicated structure and was made up of many
objects still smaller than atoms. This had to be, for the uranium atom, in exploding, hurled
outward some of these still smaller "subatomic particles."
One of the most skillful of the new experimenters was a New Zealander, Ernest Rutherford.
He used the subatomic particles that came flying out of radioactive elements and made them
serve as bullets. He aimed them at thin films of metal and found they passed right through the
metal without trouble. Atoms weren't hard shiny balls at all. Indeed, they seemed to be mostly
empty space.
But then, every once in a while, one of the subatomic bullets would bounce off sharply. It had
hit something hard and heavy somewhere in the atom.
By 1911, Rutherford was able to announce that the atom was not entirely empty space. In the
very centre of the atom was a tiny "atomic nucleus" that contained almost all the mass of the
atom. This nucleus was so small that it would take about 100,000 of them, placed side by side,
to stretch across the width of a single atom.
Outside the nucleus, filling up the rest of the atom, were a number of very light particles
called "electrons." Each different kind of atom had its own particular number of electrons.
The hydrogen atom had only a single electron; the oxygen atom had eight; the iron atom had
twenty-six; the uranium atom had ninety-two, and so on.
All electrons, no matter what atom they are found in, are alike in every way. All of them, for
instance, carry an electric charge. There are two kinds of electric charges-positive and
negative. All electrons carry a negative electric charge and the charge is always of exactly the
same size. We can say that every electron has a charge of just -1.
www.seyfihoca.com
292
The atomic nucleus has an electric charge, too, but a positive one. The charge on the nucleus
just balances the charge on the electrons. A hydrogen atom has a single electron with a charge
of -l. Therefore, the charge on the hydrogen nucleus is +1.
An oxygen atom has eight electrons with a total charge of -8. The oxygen nucleus has a
charge of +8, therefore. You can see, then, that the iron nucleus would have to have a charge
of +26, the uranium nucleus one of +92, and so on.
Both parts of the atom-the tiny nucleus at the centre and the whirling electrons outside-have
been involved in unusual discoveries since Rutherford made his announcement in 1911. In
this chapter, however, we are going to be concerned only with the nucleus.
Naturally, physicists were interested in knowing whether the atomic nucleus was a single
particle. It was so much smaller than the atom that it would seem reasonable to suppose that
here at last was something as small as it could be. The atom had proved a surprise, however,
and scientists were not going to be too sure of the nucleus either.
Rutherford bombarded atoms with subatomic particles, hoping to discover something about
the nucleus if he hit them enough times.
He did. Every once in a while, when one of his subatomic bullets hit a nucleus squarely, that
nucleus changed its nature. It became the nucleus of a different variety of atom. Rutherford
first discovered this in 1919.
This change of one nucleus into another made it seem as though the nucleus had to be a
collection of still smaller particles. Changes would come about because the collection of still
smaller particles was broken apart and rearranged.
The smallest nucleus was that of the hydrogen atom. That had a charge of +1 and it did indeed
seem to be composed of a single particle. Nothing Rutherford did could break it up (nor have
we found a way to do so even today). Rutherford therefore considered it to be composed of a
single particle which he called a "proton."
The proton's charge, +1, was exactly the size of the electron's, but it was of the opposite kind.
It was a positive electric charge, rather than a negative one.
The big difference between the proton and electron, however, was in mass. The proton is
1,836 times as massive as the electron though to this day physicists don't know why that
should be so.
It soon seemed clear that the nuclei of different atoms had different electric charges because
they were made up of different numbers of protons. Since an oxygen nucleus had a charge of
+8, it must contain eight protons. In the same way, an iron nucleus contained twenty-six
protons and a uranium nucleus ninety-two protons.
This is why the nucleus contains just about all the mass of the atom, by the way. The nucleus
is made up of protons which are so much heavier than the electrons that circle about outside
the nucleus.
www.seyfihoca.com
293
But a problem arose at this point that plagued physicists all through the 1920s. The protons
could account for the electric charge of the nucleus, but not for all its mass. Because the
oxygen nucleus had a charge of +8, it therefore had to contain eight protons, but it also had a
mass that was sixteen times as great as a single proton and therefore twice as great as all eight
protons put together. Where did the extra mass come from?
The uranium nucleus had a charge of +92 and therefore had to contain ninety-two protons.
However, the mass of the uranium nucleus was two and a half times as great as all those
ninety-two protons put together. Where did that come from?
Physicists tried to explain this in several ways that proved to be unsatisfactory. A few,
however, speculated that there might be particles in the nucleus that were as heavy as protons
but that didn't carry an electric charge.
Such uncharged particles, if they existed, would add to the mass of the nuclei without adding
to the electric charge. They would solve a great many puzzles concerning the nucleus, but
there was one catch.
There seemed no way of detecting such uncharged particles, if they existed. To see why this is
so, let's see how physicists were detecting ordinary charged particles in the 1920s.
Physicists used a number of techniques for the purpose, actually, but the most convenient had
been invented in 1911 by a Scottish physicist, Charles Thomson Rees Wilson.
He had begun his career studying weather and he grew interested in how clouds came to form.
Clouds consist of very tiny droplets of water (or particles of ice) but these don't form easily in
pure air. Instead, each one forms about a tiny piece of dust or grit that happens to be floating
about in the upper air. In the absence of such dust, clouds would not form even though the air
was filled with water vapour to the very limit it would hold, and more.
It turned out also that a water droplet formed with particular ease, if it formed about a piece of
dust that carried an electric charge.
With this in mind, Wilson went about constructing a small chamber into which moist air
could be introduced. If the chamber were expanded, the air inside would expand and cool.
Cold air cannot hold much water vapour, so as the air cooled the vapour would come out as a
tiny fog.
But suppose the moist air introduced into the chamber were completely without dust. Then
even if the chamber were expanded and the air cooled, a fog would not form.
Next suppose that a subatomic particle comes smashing through the glass and streaks into the
moist air in the chamber. Suppose also that the particle is electrically charged.
Electric charges have an effect on one another. Similar charges (two negatives or two
positives) repel each other; push each other away. Opposite charges (a negative and a
positive) attract each other.
If a negatively charged particle, like an electron, rushes through the air, it repels other
electrons it comes near. It pushes electrons out of the atoms with which it collides. A
www.seyfihoca.com
294
positively charged particle, like a proton, attracts electrons and pulls them out of the atom. In
either case, atoms in the path of electrically charged particles lose electrons.
What is left of the atom then has a positive electrical charge, because the positive charge on
the nucleus is now greater than the negative charge on the remaining electrons. Such an
electrically charged atom is called an "ion."
Water droplets, which form with particular ease about electrically charged dust particles, also
form with particular ease about ions. If a subatomic particle passes through the moist air in the
cloud chamber just as that air is cooled, droplets of water will form about the ions that the
subatomic particle leaves in its track. The path of the subatomic particle can be photographed
and the particle can be detected by the trail it leaves.
Suppose a cloud chamber is placed near a magnet. The magnet causes the moving subatomic
particle to curve in its path. It therefore leaves a curved trail of dewdrops.
The curve tells volumes. If the particle carries a positive electric charge, it curves in one
direction and if it carries a negative electric charge it curves in the other. The more massive it
is, the more gently it curves. The larger its charge, the more sharply it curves.
Physicists took many thousands of photographs of cloud chambers and studied the trails of
dewdrops. They grew familiar with the kind of tracks each particular kind of particle left.
They learned to tell from those tracks what was happening when a particle struck an atom, or
when two particles struck each other.
Yet all of this worked well only for charged particles. Suppose a particle didn't carry an
electric charge. It would have no tendency to pull or push electrons out of an atom. The atoms
would remain intact and uncharged. No ions would be formed and no water droplets would
appear. In other words, an uncharged particle would pass through a cloud chamber without
leaving any sign.
Still, might it not be possible to detect an uncharged particle indirectly? Suppose you faced
three men, one of whom was invisible. You would only see two men and if none of them
moved you would have no reason to suspect that the third man existed. If, however, the
invisible man were suddenly to push one of his neighbours, you would see one of the men
stagger. You might then decide that a third man was present but invisible.
Something of the sort happened to physicists in 1930. When a certain metal called beryllium
was exposed to a spray of subatomic particles, a radiation was produced by it which could not
be detected by cloud chamber.
How did anyone know there was that radiation present then? Well, if paraffin were placed
some distance away from the beryllium, protons were knocked out of it. Something had to be
knocking out those protons.
In 1932, an English physicist, James Chadwick, argued that the radiation from beryllium
consisted of uncharged particles. These particles were electrically neutral and they were
therefore called "neutrons."
www.seyfihoca.com
295
Neutrons were quickly studied, not by cloud chamber, but by the manner in which they
knocked atoms about, and much was learned. It was found that the neutron was a massive
particle, just a trifle more massive than the proton. Where the proton was 1,836 times as
massive as the electron, the neutron was 1,839 times as massive as the electron.
Physicists now found that they had a description of the structure of the nucleus that was better
than anything that had gone before. The nucleus consisted of both protons and neutrons. It
was the neutrons that accounted for the extra mass of the nucleus.
Thus, the oxygen nucleus had a charge of +8 but a mass of 16. That was because it was made
up of 8 protons and 8 neutrons. The uranium nucleus had a charge of +92 and a mass of 238;
it was made up of 92 protons and 146 neutrons. The atomic nucleus, small as it was, did
indeed consist of still smaller particles (except in the case of hydrogen). Indeed, the nuclei of
the more complicated atoms were made up of a couple of hundred smaller particles.
This does not mean that there weren't some serious questions raised by this proton-neutron
theory of nucleus structure. For instance, protons are all positively charged and positively
charged particles repel each other. The closer they are, the more strongly they repel each
other. Inside the atomic nucleus, dozens of protons are pushed together so closely they are
practically touching. The strength of the repulsion must be enormous and yet the nucleus
doesn't fly apart.
Physicists began to wonder if there was a special pull, or force, that held the protons together.
This force had to be extremely strong to overcome the "electromagnetic force" that pushed
protons apart. Furthermore, the new force had to operate only at very small distances, for
when protons were outside nuclei, they repelled each other with no sign of any attraction.
Such a strong attraction that could be felt only within nuclei is called a "nuclear force."
Could such a nuclear force exist? A Japanese physicist, Hideki Yukawa, tackled the problem
shortly after the neutron was discovered. He carefully worked out the sort of thing that would
account for such an extremely strong and extremely short-range force.
In 1935, he announced that if such a force existed, then it might be built up by the constant
exchange of particles by the protons and neutrons in the nucleus. It would be as though the
protons and neutrons were tossing particles back and forth and held firmly together as long as
they were close enough to toss and catch. As soon as the neutrons and protons were far
enough apart so that the particles couldn't reach, then the nuclear force would be no longer
effective.
According to Yukawa, the exchange particle should have a mass intermediate between that of
the proton and the electron. It was therefore eventually named a "meson" from a Greek work
meaning "intermediate."
But did the meson really exist?
The best way of deciding whether it existed and if Yukawa's theory was actually correct was
to detect the meson inside the nucleus, while it was being tossed back and forth between
www.seyfihoca.com
296
protons and neutrons. Unfortunately, that seemed impossible. The exchange took place so
quickly and it was so difficult to find out what was going on deep inside the nucleus, that
there seemed no hope.
But perhaps the meson could be somehow knocked out of the nucleus and detected in the
open. To do that the nucleus would really have to be made to undergo a hard collision.
According to a theory worked out by the German-Swiss physicist, Albert Einstein, in 1905,
matter and energy are two different forms of the same thing. Matter is, however, a very
concentrated form of energy. It would take the energy produced by burning twenty million
gallons of petrol to make one ounce of matter.
To knock a meson out of the nucleus of an atom would be very much like creating the amount
of matter in a meson. To produce that quantity of matter doesn't really take much energy, but
that energy has to be concentrated into a single tiny atomic nucleus and doing that turns out to
be very difficult.
All through the 1930s and 1940s, physicists devised machines for pushing subatomic particles
by electromagnetic forces and making them go faster and faster, piling up more and more
energy, until finally, crash-they were sent barreling into a nucleus.
Gradually, more and more energy was concentrated into these speeding particles. Such energy
was measured in "electron volts" and by the 1940s particles with energies of ten million
electron volts (10 Mev) were produced. This sounds like a great deal, and it is, but it still
wasn't enough to form mesons.
Fortunately, physicists weren't entirely stopped. There is a natural radiation ("cosmic rays")
striking the Earth all the time. This is made up of subatomic particles of a wide range of
energies; some of them are enormously energetic.
They originate somewhere deep in outer space. Even today, physicists are not entirely certain
as to the origin of cosmic rays or what makes them possess so much energy. Still, the energy
is there to be used.
Cosmic rays aren't the perfect answer. When physicists produce energetic particles, they can
aim them at the desired spot. When cosmic rays bombard Earth, they do so without aiming.
Physicists must wait for a lucky hit; when a cosmic ray particle with sufficient energy just
happens to hit a nucleus in the right way. And then he must hope that someone with a
detecting device happens to be at the right place and at the right moment.
For a while, though, it seemed that the lucky break had taken place almost at once. Even
while Yukawa was announcing his theory, an American physicist, Carl David Anderson, was
high on Pike's Peak in Colorado, studying cosmic rays.
The cosmic ray particles hit atoms in the air and sent other particles smashing out of the atoms
and into cloud chambers. When there was finally a chance to study the thousands of
photographs that had been taken, tracks were found which curved in such a way as to show
that the particle that caused them was heavier than an electron but lighter than the proton. In
1936, then, it was announced that the meson had been discovered.
www.seyfihoca.com
297
Unfortunately, it quickly turned out that this meson was a little too light to be the particle
called for by Yukawa's theory. It was wrong in several other ways, too.
Nothing further happened till 1947. In that year, an English physicist, Cecil Frank Powell,
was studying cosmic rays far up in the Bolivian Andes. He wasn't using cloud chambers, but
special photographic chemicals which darkened when a subatomic particle struck them.
When he studied the tracks in these chemicals, he found that he, too, had a meson, but a
heavier one than had earlier been found. Once there was a chance to study the new meson it
turned out to have just the properties predicted by Yukawa.
The first meson that had been discovered, the lighter one, was named the "mu-meson." The
heavier one that Powell had discovered was the "pi-meson." ("Mu" and "pi" are letters of the
Greek alphabet. Scientists often use Greek letters and Greek words in making up scientific
names.)
It is becoming more and more common to abbreviate the names of these mesons. The light
one is called the "muon" and the heavy one the "pion."
The new mesons are very unstable particles. They don't last long once they are formed. The
pion only lasts about twenty-five billionths of a second and then it breaks down into the
lighter muon. The only reason the pion can be detected at all is that when it is formed it is
usually travelling at enormous speed, many thousands of miles a second. Even in a billionth
of a second it has a chance to travel a few inches, leaving a trail as it does so. The change in
the kind of trail it leaves towards the end shows that the pion has disappeared and a muon has
taken its place.
The muon lasts much longer, a couple of millionths of a second, and then it breaks down,
forming an electron. The electron is stable and, if left to itself, will remain unchanged forever.
By the end of the 1940s, then, the atomic nucleus seemed to be in pretty good shape. It
contained protons and neutrons and these were held together by pions flashing back and forth.
Chemists worked out the number of protons and neutrons in every different kind of atom and
all seemed well.
But it did seem that there ought to be two kinds of nucleithe kind that exists all about us and a
sort of mirror image that in the late 1940s, no one had yet seen.
That possibility had first been suggested in 1930 by an English physicist, Paul Adrien
Maurice Dirac. He calculated what atomic structure ought to be like according to the latest
theories and it seemed to him that every particle ought to have an opposite number. This
opposite could be called an "antiparticle."
In addition to an electron, for instance, there ought also to be an "antielectron" that would
have a mass just like that of an electron but would be opposite in electric charge. Instead of
having a charge of -1, it would have one of +1.
In 1932, C. D. Anderson (who was later to discover the muon) was studying cosmic rays. He
noticed on one of his photographs a cloud-chamber track which he easily identified as that of
www.seyfihoca.com
298
an electron. There was only one thing wrong with it; it curved the wrong way. That meant it
had a positive charge instead of a negative one.
Anderson had discovered the antielectron. Because of its positive charge, it is usually called a
"positron." The existence of the antielectron was strong evidence in favor of Dirac's theory,
and as time went on more and more antiparticles were uncovered.
The ordinary muon, for instance, has a negative charge of -l, like the electron, and it is usually
called the "negative muon." There is an antimuon, exactly like the muon except that it has a
positive charge of +1 like the positron. It is the "positive muon."
The ordinary pion is a "positive pion" with a charge of +1. The antipion is the "negative pion"
with a charge of -1.
By the close of the 1940s, it seemed quite reasonable to suppose that there were ordinary
nuclei made up of protons and neutrons with positive pions shifting back and forth among
them; and that there were also "antinuclei" made up of "antiprotons" and "antineutrons" with
antipions shifting back and forth.
Physicists didn't really feel they actually had to detect antiprotons and antineutrons to be sure
of this but, of course, they would have liked to.
To detect antiprotons is even more difficult than to detect pions. An antiproton is as massive
as a proton, which means it is seven times as massive as a pion. To form an antiproton
requires a concentration of seven times as much energy as to form a pion.
To form a pion required several hundred million electron volts, but to form an antiproton
would require several billion electron volts. (A billion electron volts is abbreviated "Bev.")
To be sure, there are cosmic ray particles that contain several Bev of energy, even several
million Bev. The higher the energy level required, however, the smaller the percentage of
cosmic ray particles possessing that energy. The chances that one would come along energetic
enough to knock antiprotons out of atoms just when a physicist was waiting to take a picture
of the results were very small indeed.
However, the machines for producing man-made energetic particles were becoming ever
huger and more powerful. By the early 1950s, devices for producing subatomic particles with
energies of several Bev were built. One of these was completed at the University of California
in March 1954. Because of the energy of the particles it produced, it was called the
"Bevatron."
Almost at once, the Bevatron was set to work in the hope that it might produce antiprotons. It
was used to speed up protons until they possessed 6 Bev of energy and then those protons
were smashed against a piece of copper. The men in charge of this project were an Italianborn physicist, Emilio Segré, and a young American, Owen Chamberlain.
In the process, mesons were formed; thousands of mesons for every possible antiproton. The
mesons, however, were much lighter than antiprotons and moved more quickly. Segré's group
set up detecting devices that would react in just the proper manner to pick up heavy, slow-
www.seyfihoca.com
299
moving, negatively charged particles. When the detecting devices reacted properly, only
something with exactly the properties expected of an antiproton could have turned the trip.
By October 1955, the detection devices had been tripped sixty times. It could be no accident.
The antiproton was there and its discovery was announced.
The antiproton was the twin of the proton. The great difference was that the proton had a
charge of +l and the antiproton had a charge of -1.
Once enough antiprotons were produced for study, it was found that occasionally one would
pass close by a proton and the opposite charges would cancel. The proton would become a
neutron and the antiproton would become an antineutron.
You might wonder how you could tell an antineutron from a neutron since both are
uncharged. The answer is that although the neutron and antineutron have no electric charge,
they spin rapidly in a way that causes them to behave like tiny magnets. The neutron is like a
magnet that points in one direction while the antineutron is like a magnet that points in the
opposite direction.
By the mid-1950s, it was clear that antiprotons and antineutrons existed. But could they
combine to form an antinucleus?
Physicists were sure they could but the final answer did not come till 1965. In that year, at
Brookhaven National Laboratories in Long Island, New York, protons with energies of 7 Bev
were smashed against a beryllium target. Several cases of an antiproton and antineutron in
contact were produced and detected.
In the case of ordinary particles, there is an atomic nucleus that consists of one proton and one
neutron. This is the nucleus of a rare variety of hydrogen atom that is called "deuterium." The
proton-neutron combination is therefore called a "deuteron."
What had been formed at Brookhaven was an "antideuteron." It is the very simplest
antinucleus that could be formed of more than one particle, but that is enough. It proved that it
could be done. It was proof enough that matter could be built up out of antiparticles just as it
could be built of ordinary particles. Matter built up of antiparticles is "antimatter."
When the existence of antiparticles was first proposed, it was natural to wonder why if they
could exist, they weren't anywhere around us. When they were detected at last, they were
found only in tiny quantities and even those quantities didn't last long.
Consider the positron, or antielectron. All around us, in every atom of all the matter we can
see and touch on Earth, are ordinary electrons. Nowhere are there any antielectrons to speak
of. Occasionally, cosmic ray particles produce a few or physicists form a few in the
laboratory. When they do, those antielectrons disappear quickly.
As an antielectron speeds along, it is bound to come up against one of the trillions of ordinary
electrons in its immediate neighbourhood. It will do that in perhaps a millionth of a second.
www.seyfihoca.com
300
When an electron meets an antielectron, both particles vanish. They are opposites and cancel
out. It is like a peg falling into a hole which it fits exactly. Peg and hole both disappear and
nothing is left but a flat surface.
In the case of the electron and antielectron, however, not everything disappears. Both electron
and antielectron have mass, exactly the same amount of the same kind of mass. (We only
know of one kind of mass so far.) When the electron and antielectron cancel out, the mass is
left over and that turns into energy.
This happens with all other particles and antiparticles. A positive muon will cancel a negative
muon; a negative pion will cancel a positive pion; an antiproton will cancel a proton, and so
on. In each case both particles disappear and energy takes their place. Naturally, the more
massive the particles, the greater the amount of energy that appears.
It is possible to reverse the process, too. When enough energy is concentrated into a small
space, particles may be formed out of it. A particle is never formed out of energy by itself,
however. If an electron is formed, an antielectron must be formed at the same time. If a proton
is formed, an antiproton must be formed at the same time.
When Segré and Chamberlain set about forming antiprotons, they had to allow for twice as
much energy as would be sufficient just for an antiproton. After all, they had to form a proton
at the same time.
Since this is so, astronomers are faced with a pretty problem. They have worked up many
theories of how the universe came to be, but in all the theories, it would seem that
antiparticles ought to be formed along with the particles. There should be just as much
antimatter as there is matter.
Where is all this antimatter? It doesn't seem to be around. Perhaps it has combined with
matter and turned into energy. In that case, why is there all the ordinary matter about us left
over. There should be equal amounts of each, and each set should cancel out the other
completely.
Some astronomers suggest that there are two separate universes, one made out of matter (our
own) and another made out of antimatter. Other astronomers think there is only one universe
but that some parts of it (like the parts near ourselves) are matter, while other parts are
antimatter.
What made the matter and antimatter separate into different parts of the universe, or even into
different universes, no one can yet say. It may even be possible that for some reason we don't
understand, only matter, and no antimatter, was formed to begin with.
The problem of the universe was something for astronomers to worry about, however.
Physicists in 1947 were quite satisfied to concentrate on particles and antiparticles and leave
the universe alone.
And physicists in that year seemed to have much ground for satisfaction. If they ignored the
problem of how the universe began and just concentrated on how it was now, they felt they
www.seyfihoca.com
301
could explain the whole thing in terms of a little over a dozen particles altogether. Some of
these particles they had actually detected. Some they had not, but were sure of anyway.
Of course, not everything was absolutely clear, but what mysteries existed ought to be cleared
up, they hoped, without too much trouble.
The particles they knew, or strongly suspected they were soon going to know, fell into three
groups, depending on their mass. There were the light particles, the middle-sized particles,
and the heavy particles. These were eventually given Greek names from words meaning light,
middle-sized, and heavy: "leptons," "mesons," and "baryons."
The leptons, or light particles, include the electron and the antielectron, of course. In order to
explain some of the observed facts about electrons, the Austrian physicist Wolfgang Pauli
suggested, in 1931, that another kind of particle also existed. This was a very small one,
possibly with no mass at all, and certainly with no charge. It was called a "neutrino." This tiny
particle was finally detected in 1956. There was not only a neutrino but also an "antineutrino."
Although the muon was considered a meson, to begin with, it was soon recognized as a kind
of heavy electron. All its properties but mass were identical with those of the electron. Along
with the muon, a neutrino or antineutrino is also formed as in the case of the electron. In 1962,
this muonneutrino was found to be different from the electron-neutrino.
Two other particles might be mentioned. Light, together with other radiation similar to it (like
x rays, for instance) behaves in some ways as though it were composed of particles. These
particles are called "photons."
There is no antiparticle for a photon; no antiphoton. The photon acts as its own opposite. If
you were to fold a sheet of paper down the middle and put the particles on one side and the
antiparticles on the other, you would have to put the photon right on the crease.
Then, too, physicists speculate that the reason different objects pull at each other
gravitationally is because there are tiny particles called "gravitons" flying between them.
Some of the properties of the graviton have been worked out in theory; for instance, it is its
own antiparticle. The graviton is so tiny, however, and so hard to pin down, that it has not yet
been detected.
This is the total list of leptons so far, then:
1.
2.
3.
4.
5.
6.
the graviton
the photon
the electron and the antielectron
the electron-neutrino and the electron-antineutrino
the negative muon and the positive muon
the muon-neutrino and the muon-antineutrino
The leptons pose physicists some problems. Does the graviton really exist? Why does the
muon exist; what is the purpose of something that is just a heavy electron? Why and how are
the muon-neutrinos different from the electron-neutrinos? These puzzles are intriguing but
they don't drive physicists to despair.
www.seyfihoca.com
302
In 1947, only three particles were coming to be known which would now be considered
mesons. Two of them were the positive pion and the negative antipion. The third was a
neutral pion which, like the photon and the graviton, was its own antiparticle.
Only four particles were known in 1947 that would now be classified as baryons. These are
the proton, antiproton, neutron, and antineutron. Both antiproton and antineutron had not yet
actually been detected, but physicists were quite sure they existed.
The situation with regard to the nucleus seemed particularly well settled. There was the
nucleus made up of protons and neutrons held together by pions, and the antinucleus made up
of antiprotons and antineutrons held together by antipions. All seemed well.
But in 1947, the very year which saw the discovery of the pion and the apparent solution of
the problem of the nucleus, there began a new series of discoveries that upset the applecart
again.
Two English physicists, George Dixon Rochester and Clifford Charles Butler, studying
cosmic rays with cloud chambers in 1947, came across an odd V-shaped track. It was as
though some neutral particle, which left no track, had suddenly broken into two particles,
which each had a charge and left a track, and which hastened away in different directions.
The particle that moved off in one direction and formed one branch of the V seemed to be a
pion, but the other was something new. From the nature of the track it left, it seemed to be as
massive as a thousand electrons, or as three and a half pions. It was half as massive as a
proton.
Nothing like such a particle had ever been suspected of existing. It caught the world of
physicists by surprise, and at first all that could be done was to give it a name. It was called a
"V-particle," and the collision that produced it was a "V-event."
Once physicists became aware of V-events, they began to watch for them and, of course, soon
discovered additional ones. By 1950, V-particles were found which seemed to be actually
more massive than protons or neutrons. This was another shock. Somehow physicists had
taken it for granted that protons and neutrons were the most massive particles there were.
The astonished physicists began to study the new particles carefully. The first V-particle to be
discovered, the one that was only half as massive as a proton, was found to have certain
properties much like those of the pion. The new particle was therefore classified as a meson.
It was called a "K-meson" and the name was quickly abbreviated to "kaon." There were four
of these: a positive kaon, a negative antikaon, a neutral kaon, and a neutral antikaon.
The other V-particles discovered in the early 1950s were all more massive than the proton and
were grouped together as "hyperons." There were three kinds of these and each kind was
given the name of a Greek letter. The lightest were the "lambda particles," which were about
20 percent heavier than protons. These came in two varieties, a lambda and an antilambda,
both of them uncharged.
www.seyfihoca.com
303
Next lightest were the "sigma particles," which were nearly 30 percent heavier than the
proton. There was a positive sigma, a negative, and a neutral, and each had its antiparticle.
That meant six sigma particles altogether.
Finally, there were the "xi particles," which were 40 percent heavier than the proton. There
was a negative xi particle and a neutral one (no positive variety) and each had its antiparticle,
making four altogether.
All these hyperons, an even dozen of them, had many properties that resembled those of the
proton and neutron. They were therefore lumped with them as baryons. Whereas there had
been four baryons known, or suspected, in 1947, there were sixteen in 1957.
But then things grew rapidly more complicated still. Partly, it was because physicists were
building machines capable of producing particles with more and more energy. This meant that
nuclei were being smashed into with greater and greater force and it was possible to turn the
energy into all sorts of particles.
What's more, physicists were developing new and better means of detecting particles. In 1952,
a young American physicist, Donald Arthur Glaser, got an idea for something that turned out
to be better than the cloud chamber. It was, in fact, rather the reverse of the cloud chamber.
A cloud chamber contains gas that is on the point of turning partly liquid. Charged particles,
racing through, help the liquid to form and leave trails of water droplets.
But suppose it were the reverse. Suppose there was a chamber which contained liquid that
was on the point of boiling and turning into gas. Charged particles passing through the liquid
would form ions. The liquid immediately around the ion would boil with particular ease and
form small bubbles of gas. The tracks would be gas bubbles in liquid, instead of liquid drops
in gas.
This new kind of detecting device was called a "bubble chamber."
The advantage of a bubble chamber is that the liquid it contains is much denser than the air in
a cloud chamber. There are more atoms and molecules in the liquid for a flying particle to
collide with. More ions are formed and a clearer trail is left behind. Particles that could
scarcely be seen in a cloud chamber are seen very clearly in a bubble chamber.
By using bubble chambers and finding many more kinds of tracks, physicists began to
suspect, by 1960, that there were certain particles that came into existence very briefly. They
were never detected but unless they existed there was no way of explaining the tracks that
were detected.
These new particles were very short-lived indeed. Until now the most unstable particles that
had been detected lasted for a billionth of a second or so. That was a long enough time for
them to make visible tracks in a bubble chamber.
The new particles, however, broke down in something like a hundred thousandth of a billionth
of a billionth of a second. In that time, the particle has only a chance to travel about the width
of a nucleus before breaking down.
www.seyfihoca.com
304
These new particles were called "resonance particles" and different varieties have been
deduced in great numbers since 1960. By now there are over a hundred baryons known that
are heavier than protons. The heaviest are over twice as massive as protons.
Some of the new particles are mesons, all of them heavier than the pion. There are about sixty
of these.
In the 1960s then, physicists were faced with the problem of finding some way of accounting
for a large number of massive particles for which they could think of no uses and whose
existence they couldn't predict.
At first all that physicists could do was to study the way in which one particle broke down
into another; or the way in which one particle was built up into another when energy was
added. Some changes could take place, while some changes could not. Particle A might
change into particles B and C, but never into particles D and E.
Physicists tried to work out rules which would explain why some changes could take place
and some could not. For instance, a neutron couldn't change into only a proton, because the
proton has a positive electric charge and that can't be made out of nothing.
A neutron might, however, change into a proton plus an electron. In that case, a positive and a
negative charge would be formed simultaneously. Together, they might be considered as
balancing each other, so it would be as though no charge at all were formed.
But then to balance certain other qualities, such as the particle spin, more was required. In the
end, it turned out that a neutron had to break down to three particles: a proton, an electron,
and an antineutrino.
Matters such as electric charge and particle spin were enough to explain the events that were
known in the old days when only a dozen or so different particles were known. In order to
explain all the events that took place among nearly 200 particles, more rules had to be worked
out. Quantities such as "isotopic spin," "hypercharge," "parity," and so on, had to be taken
into account.
There is even something called "strangeness." Every particle is given a "strangeness number"
and if this is done correctly, it turns out that whenever one group of particles changes into
another group, the total strangeness number isn't altered.
The notion of strangeness made it plainer that there were actually two kinds of nuclear forces.
The one that had first been proposed by Yukawa and that involved pions was an extremely
strong one. In the course of the 1950s, however, it became clear that there was also a much
weaker nuclear force, only about a hundred trillionths as strong as the strong one.
Changes that took place under the influence of the strong nuclear force took place extremely
rapidly-just long enough to allow a resonance particle to break down. Changes that took place
under the influence of the weak nuclear force took much longer-at least a billionth of a second
or so.
www.seyfihoca.com
305
Only the baryons and the mesons could take part in strong force changes. The leptons took
part only in weak-force changes. The baryons and mesons are therefore lumped together
sometimes as "hadrons."
Even when physicists gradually worked out the rules that showed what particle changes could
take place and what couldn't take place, they were very unsatisfied. They didn't understand
why there should be so many particles.
More and more physicists began to wonder if the actual number of particles was unimportant.
Perhaps particles existed in families and they ought to concentrate on families of particles.
For instance, the first two baryons discovered were the proton and the neutron. They seemed
two completely different particles because there was an important unlikeness about them. The
proton had a positive electric charge and the neutron had no electric charge at all.
This seemed to be an enormous difference. Because of it, a proton could be detected easily in
a cloud chamber and a neutron couldn't. Because of it a proton followed a curved path when
brought near a magnet but a neutron didn't.
And yet when the strong nuclear force was discovered, it was found that it affected protons
and neutrons exactly the same, as though there were no difference between them. If the proton
and neutron are considered from the standpoint of the strong nuclear force only, they are
twins.
Could it be, then, that we ought to consider the proton and neutron as two forms of a single
particle which we might call the "nucleon" (because it is found in the nucleus)? Certainly, that
might simplify matters.
You can see what this means if you consider people. Certainly, a husband and a wife are two
different people, very different in important ways. To the income tax people, however, they
are just one tax-paying combination when they file a joint return. It doesn't matter whether the
husband makes the money, or the wife, or both make half; in the return it is all lumped
together. For tax purposes we simply have a taxpayer in two different forms, husband and
wife.
After 1960, when the resonance particles began to turn up, physicists began to think more and
more seriously of particle families. In 1961, two physicists, Murray Gell-Mann in the United
States and Yuval Ne'eman in Israel, working separately, came up with very much the same
scheme for forming particle families.
To do this, one had to take all the various particle properties that physicists had worked out
and arrange them in a very regular way. There were eight different kinds of properties that
Gell-Mann worked with in order to set up a family pattern. Jokingly, he called his system the
"Eightfold Way," after a phrase in the teachings of the Indian religious leader Buddha. The
more formal name of his scheme is "SU (3) symmetry."
In what turned out to be the most famous example of SU (3) symmetry, Gell-Mann prepared a
family of ten particles. This family of ten can be pictured as follows. Imagine a triangle made
www.seyfihoca.com
306
up of four objects at the bottom, three objects above them, two objects above them, and one
object all by itself at the apex.
The four objects at the bottom are four related "delta particles" each about 30 percent heavier
than a proton. The chief difference among them is the electric charge. The four delta particles
have charges of -1, 0, +1, and +2.
Above these are three "sigma particles" more massive than the deltas and with charges of -1,
0, and +1. Above that are two "xi particles," which are still more massive and which have
charges of -1, and 0. Finally, at the apex of the triangle is a single particle that is most massive
of all and that has a charge of -1. Gell-Mann called this last particle the "omegaminus"
particle, because "omega" is the last letter in the Greek alphabet and because the particle has a
negative electric charge.
Notice that there is a regular way in which mass goes up and the number of separate particles
goes down. Notice also that there is a regular pattern to the electric charges: -1, 0, +1, +2 for
the first set; then -1, 0, +l; then -1, 0; finally -1.
Other properties also change in a regular way from place to place in the pattern. The whole
thing is very neat indeed. There was just one problem. Of the ten particles in this family, only
nine were known. The tenth particle, the omegaminus at the apex, had never been observed. If
it did not exist the whole pattern was ruined. Gell-Mann suggested that it did exist; that if
people looked for it and knew exactly what they were looking for, they would find it.
If Gell-Mann's pattern was correct, one ought to be able to work out all the properties of the
omega-minus by taking those values that would fit into the pattern. When this was done, it
was found that the omega-minus would have to be a most unusual particle for some of its
properties were like nothing yet seen.
For one thing, if it were to fit into its position at the top of the triangle it would have to have
an unusual strangeness number. The deltas at the bottom of the triangle had a strangeness
number of 0, the sigmas above them a strangeness number of -1, and the xis above them one
of -2. The omega-minus particle at the top would therefore have to have a strangeness number
of -3. No strangeness number that large had ever been encountered and physicists could
scarcely bring themselves to believe that one would be.
Nevertheless, they began to search for it.
The instrument for the purpose was at Brookhaven, where, as the 1960s opened, an enormous
new device for speeding particles was put into operation. It could speed up particles to the
point where they would possess energies as high as 33
Bev. This was more than five times the quantity of energy that was enough to produce
antiprotons some years before.
In November 1963, this instrument was put to work in the search for the omega-minus
particle. Along with it was a new bubble chamber that contained liquid hydrogen. Hydrogen
was liquid only at very frigid temperatures, hundreds of degrees below the ordinary zero.
www.seyfihoca.com
307
The advantage to the use of liquid hydrogen was that hydrogen nuclei were made up of single
protons (except for the very rare deuterium form of the element). Nothing else could supply
so many protons squeezed into so small a space without any neutrons present to confuse
matters.
The liquid hydrogen bubble chamber was nearly seven feet across and contained over 900
quarts of liquid hydrogen. There would be very little that would escape it.
Physicists had to calculate what kind of particle collisions might possess sufficient energy
plus all the necessary properties to form an omega-minus particle, if one could be formed at
all. You would have to have a collision that would supply the necessary strangeness number
of -3, for instance. It would also have to be a collision that would supply no quantity of
something called "isotopic spin," for the isotopic spin of omega-minus would have to be 0 if it
were to fit Gell-Mann's pattern.
It was finally decided that what was needed was to smash high-energy negative kaons into
protons. If everything went right, an occasional collision should produce a proton, a positive
kaon, a neutral kaon, and an omega-minus particle.
A beam of 5 Bev negative kaons was therefore shot into the liquid hydrogen bubble chamber
and by January 30, 1964, fifty thousand photographs had been taken. Nothing unusual was
found in any of them.
On January 31, however, a photograph appeared in which a series of tracks were produced
which seemed to indicate that an omega-minus particle had been formed and had broken
down to form other particles. If certain known and easily recognized particles were followed
backward, and it were calculated what kind of particles they must have come from, and then
those were followed backward, one reached the very brief existence of an omega-minus
particle.
A few weeks later, another photograph showed a different combination of tracks which could
be worked backward to an omega-minus particle.
In other words, a particle had been detected which had broken down in two different ways.
Both breakdown routes were possible for the omega-minus particle if it had exactly the
properties predicted by Gell-Mann. Since then, a number of other omega-minus particles have
been detected, all with exactly the predicted properties.
There seemed no question about it. The omega-minus particle did exist. It had never been
detected because it was formed so rarely and existed so briefly. Now that physicists had been
told exactly what to look for and where to look for it, however, they had found it.
Physicists are now satisfied that they must deal with particle families. There are arguments as
to exactly how to arrange these families, of course, but that will probably be straightened out.
But can matters become simpler still? It has often happened in the history of science that
when matters seemed to grow very complicated, it could all be made simpler by some basic
discovery.
www.seyfihoca.com
308
For instance, there are uncounted millions of different kinds of materials on Earth, but
chemists eventually found they were all formed out of a hundred or so different kinds of
elements, and that all the elements were made up, in the main, of three kinds of particles:
protons, neutrons, and electrons.
In the twentieth century, as physicists looked more and more closely at these subatomic
particles and found that nearly two hundred of them existed altogether, naturally they began to
think of going deeper still. What lies beyond the protons and neutrons?
It is a case of digging downward into the littler and littler and littler. First to atoms, then
beyond that to the nucleus, then beyond that to the proton and neutron, and now beyond that
to-what?
Gell-Mann, in working out his family patterns, found that he could arrange them by letting
each particle consist of three different symbols in different combinations. He began to wonder
if these different symbols were just mathematical conveniences or if they were real objects.
For instance, you can write one dollar as $1.00, which is the same as writing 100¢. This
would make it seem that there are one hundred cents in a dollar, and there certainly are. But
does this mean that if you were to take a paper dollar bill and tear it carefully apart you would
find a hundred one-cent pieces in it? Of course not!
The question was, then, if you tore a proton apart, would you find the three smaller objects
that represented the three symbols used by Gell-Mann.
Gell-Mann decided to give the particles a name at least. He happened to think of a passage in
Finnegan's Wake by James Joyce. This is a very difficult book in which words are
deliberately twisted so as to give them more than one meaning. The passage he thought of was
a sentence that went "Three quarks for Muster Mark."
Since three of these simple particles were needed for each of the different baryons, Gell-Mann
decided, in 1963, to call them "quarks."
If the quarks were to fit the picture, they would have to have some very amazing properties.
The most amazing was that they would have to have fractional electric charges.
When the electron was first discovered, its electric charge was set at -1 for simplicity's sake.
Since then, all new particles discovered have either no electric charge at all or have one that is
exactly equal to that of the electron or to an exact multiple of that charge. The same held for
positive charges.
In other words, particles can have charges of 0, -1, +1, -2, +2, and so on. What has never been
observed has been any fractional charge. No particle has ever yet been found to have a charge
of +11/2 or -21/3.
Yet a fractional charge was exactly what the quarks would have to have. Charges of -1/3 and
+2/3s would have to be found among them.
An immense search is now on for the quarks, for if they are found, they will simplify the
physicist's picture of the structure of matter a great deal.
www.seyfihoca.com
309
There is one important difficulty. Gell-Mann's theory makes it quite plain that when quarks
come together to form ordinary subatomic particles, the process gives off a great deal of
energy. In fact, almost all the mass of the quarks is given off as energy and only about onethirtieth is left to form the particle. This means that quarks are thirty times as massive as the
particles they produce.
(This sounds strange, but think about it. Suppose you see three balloons blown up almost to
bursting. Would you suppose it were possible to squeeze them into a small box just an inch
long in each direction? All you would have to do would be to let the air out of the balloons
and what is left can easily be packed away in a small box. Similarly, when three quarks
combine, you "let the mass out" and what is left can easily fit into a proton.)
If you want to form a quark by breaking apart a proton or some other particle, then you have
to supply all the energy that the quarks gave up in the first place. You have to supply enough
energy to form a group of particles thirty times as massive as a proton. You would need at
least fifteen times as much energy as was enough to form a proton and antiproton in the
1950s, and probably even more.
There is no instrument on Earth, not even Brookhaven's 33-Bev colossus, that can supply the
necessary energy. Physicists have two things they can do. First, they can turn to the
astronomers and ask them to watch for any sign of quarks in outer space. There are cosmic
ray particles with sufficient energy to form quarks. Most cosmic ray particles are protons and
if two of them smash together hard enough they may chip themselves into quarks.
However, this would happen very rarely and so far astronomers have detected nothing they
could identify as quarks. The second possibility is to build a device that will produce particles
with sufficient energy to form quarks. In January 1967, the government of the United States
announced plans to build such an instrument in Weston, Illinois.
This will be a huge device, nearly a mile across. It will take six or seven years to build and
will cost 375 million dollars. Once it is completed, it will cost 60 million dollars a year to run.
But when it is done, physicists hope it will produce streams of particles with energies up to
200 Bev. This may be enough to produce quarks-or to show that they probably don't exist.
Physicists are awaiting the completion of the new instrument with considerable excitement
and the rest of us should be excited, also. So far, every new advance in the study of the atom
has meant important discoveries for the good of mankind.
By studying atoms in the first place, chemists learned to put together a variety of dyes and
medicines, fertilizers and explosives, alloys and plastics that had never existed in nature.
By digging inside the atom and studying the electron, physicists made possible the production
of such devices as radio and television.
The study of the atomic nucleus gave us the various nuclear bombs. These are not very
pleasant things, to be sure, but the same knowledge also gave us nuclear power stations. It
may make possible the production of so much cheap energy that our old planet may possibly
reach towards a new era of comfort and ease.
www.seyfihoca.com
310
Now physicists are trying to find the quarks that lie beyond the subatomic particle. We can't
predict what this will result in, but it seems certain there will be results that may change the
world even more than plastics, and television, and atomic power.
We will have to wait and see. Once. the new device is put into action at Weston, it is just
possible we may not have to wait long.
From: Twentieth Century Discovery by Isaac Azimov
4 - A New Look At The Planets: Distance In Our Solar System
The study of the planets reached a peak in the nineteenth century and then, towards its end,
seemed to die down. Other subjects began to interest astronomers much more. There was
nothing left, it would appear, for twentieth century astronomers to do about planets.
If, indeed, the planets seemed worked out by 1900, that is not surprising. After all,
astronomers had been dealing with them for over 2,000 years, and what more could be left?
To be sure, the ancients had got off on the wrong foot. The Greeks had worked out careful
and interesting theories concerning the motions of the pl
Download