ENTROPY & INFORMATION a physicist point of view Jean V. Bellissard Georgia Institute of Technology & Institut Universitaire de France ENTROPY: Some history Carnot’s Principle: • Sadi CARNOT • 1825: • Reflexions sur la Puissance Motrice du Feu Carnot’s Principle: • Sadi CARNOT • 1825: • Reflexions sur la Puissance Motrice du Feu A steam machine needs 2 sources of heat: - a hot one: temperature Th - a cold one: temperature Tc Th > Tc Carnot’s Principle: • Sadi CARNOT • 1825: • Reflexions sur la Puissance Motrice du Feu The proportion of thermal energy that can be transformed into mechanical motion depends only on the temperatures of the two sources Steam Engines • Any steam engine has a heat source (burner) and a cold source (the atmosphere). Thermal engines are everywhere - in power plants (coal, nuclear, …) - in cars, airplane, boats, - in factories, Entropy: definition • Rudolf CLAUSIUS • 1865: Definition of entropy: d S = d Q/T • 2nd Law of Thermodynamics: Entropy cannot decrease over time Gas are made of molecules Qu i ck Ti me ™a nd a GIF d ec omp res so ra re ne ed ed to s ee th i s pi c tu re. • Clausius showed that gas were made of molecules, explaining the slow diffusion of dust and the origin of viscosity Statistical Thermodynamics: • Ludwig BOLTZMAN • 1872: - Kinetic theory • 1880: Statistical interpretation of entropy: disorder in energy space Statistical Mechanics • Josiah Willard GIBBS • 1880’s: Thermodynamical equilibrium corresponds to maximum of entropy • 1902 : book « Statistical Mechanics » Information theory • Claude E. SHANNON • 1948: « A Mathematical Theory of Communication » -Information theory -Entropy measures the lack of information of a system Second Law of Thermodynamics • Over time, the information contained in an isolated system can only be destroyed • Equivalently, the entropy can only increase MORPHOGENESIS: how does nature produces information ? Conservation Laws • In an isolated system, the Energy, the Momentum, the Angular Momentum, the Electric Charge,…. are conserved. Conservation Laws Angular momentum Conservation Laws • At equilibrium, the only information available on the system are the values of conserved quantities! • Example: elementary particles are characterized by their mass (energy), spin (angular momentum), electric charge… • Electron : m = 9.109x10-31 kg, s = 1/2, e = —1.602 x10-19 C, Out of Equilibrium time E flow E’ • Variations in time or space force transfer of conserved quantities • Transfer of Energy (Heat), Mass, Angular Momentum, Charges, creates current flows. Out of Equilibrium • Transfer of Energy (Heat), creates heat current like in flames and fires. Out of Equilibrium • Transfer of Mass, creates fluid currents like in rivers or streams. Out of Equilibrium • Transfer of Charges, creates electric currents. Out of Equilibrium • Transfer of Angular Momentum creates vortices like this hurricane seen from a satellite. Out of Equilibrium • Pattern Formation A shallow horizontal liquid heated from below exhibits instabilities and formation of rolls and patterns, as a consequence of fluids equations Out of Equilibrium Explosions produce interstellar clouds Collapses produces stars The Sun, the Moon, The Planets, and the Stars have been used as sources of information: measure of time, localization on Earth Beating the 2nd Principle • Without variations in time and space the only information contained in an isolated system is provided by conservation laws • Motion and heterogeneities allow Nature to create a large quantity of information. • All macroscopic equations (fluids, flame,…) describing it are given by conservation laws CODING INFORMATION the art of symbols Signs • Signs can be visual color, shape, design Signs • Signs can be a sound ring, noise, applause musical, speech • Signs can be a smell Signs • Signs can be a smell Signs Signs • Signs can be a smell plants can warn their neighbors with phenols Signs • Signs can be a smell female insects can attract males with pheromones Writings Writings • More than 80,000 characters are used to code the Chinese language Writings • Ancient Egyptians used hieroglyphs to code sounds and words Writings • Japanese language is also using the 96 Hiragana character coding syllables Writings • the Phoenicians and the Greeks found the alphabet simpler to code elementary sounds with 23 characters bgdezhq iklmnopr st uf cyw Writings • Modern numbers are coded with 10 digits created by Indians and transmitted to Europeans through the Arabs Writings • George BOOLE (18151864) used only two characters to code logical operations 0 1 Writings • John von NEUMANN (1903-1957) developed the concept of programming using also binary system to code all possible information 0 1 Writings • Nature uses 4 molecules Writings • Nature uses 4 molecules to code Writings • Nature uses 4 molecules to code the genetic heredity Writings • Proteins uses 20 amino acids to code their functions in the cell molecule of Tryptophan, one of the 20 amino acids Unit of information • Following Shannon (1948) the unit is the bit A system contains N-bits of information if it contains 2N possible characters TRANSMITTING INFORMATION redundancy Transmitting 0 coding 1 • Coding theory uses redundancy to transmit binary bits of information Transmitting 0 000 coding 1 111 • Coding theory uses redundancy to transmit binary bits of information Transmitting 0 000 coding 1 111 Transmission • Coding theory uses redundancy to transmit binary bits of information Transmitting 0 • Coding theory uses redundancy to transmit binary bits of information 000 Transmission 010 111 errors (2nd Principle) 110 coding 1 Transmitting 0 • Coding theory uses redundancy to transmit binary bits of information 000 Transmission 010 111 errors (2nd Principle) 110 coding 1 Reconstruction Transmitting 0 000 Transmission 111 errors (2nd Principle) • Coding theory uses redundancy to transmit binary bits of information 010 Reconstruction 000 110 at reception (correction) 111 coding 1 Transmitting • Humans use also redundancy to make sure they receive the correct information Transmitting • Humans use also redundancy to make sure they receive the correct information Transmitting say it again ! • Humans use also redundancy to make sure they receive the correct information Transmitting • A cell is a big factory designed to duplicate the information contained in the DNA Transmitting • Prior to the cell fission the DNA molecule is unzipped Transmitting • Prior to the cell fission the DNA molecule is unzipped by another protein Transmitting • A cell is a big factory designed to duplicate the information contained in the DNA QuickTime™ and a GIF decompress or are needed to s ee this picture. Transmitting • A cell is a big factory designed to duplicate the information contained in the DNA mitosis Transmitting • A cell is a big factory designed to duplicate the information contained in the DNA mitosis Transmitting • A cell is a big factory designed to duplicate the information contained in the DNA mitosis Transmitting • A cell is a big factory designed to duplicate the information contained in the DNA mitosis Transmitting • A cell is a big factory designed to duplicate the information contained in the DNA mitosis Transmitting • A cell is a big factory designed to duplicate the information contained in the DNA mitosis Transmitting • A cell is a big factory designed to duplicate the information contained in the DNA mitosis Beating the 2nd Principle • The cell divides before the information contained in the DNA fades away • In this way, cell division and DNA duplication at fast pace, conserve the genetic information for millions of years. THE MAXIMUM ENTROPY PRINCIPLE REVISITED The scary art of extrapolation Equilibrium • A physical system of particle reaches equilibrium when all information but the one that must be conserved have vanished Equilibrium • A physical system of particle reaches equilibrium when all information but the one that must be conserved have vanished In a gas the chaotic motion produced by collisions is responsible for the loss of information Equilibrium • By analogy other systems involving a large number of similar individuals can be treated through statistics and information Equilibrium • By analogy other systems involving a large number of similar individuals can be treated through statistics and information Like bureaucracy Equilibrium • By analogy other systems involving a large number of similar individuals can be treated through statistics and information Like bureaucracy 1837 J. S. MILL in Westm. Rev. XXVIII. 71 That vast net-work of administrative tyranny…that system of bureaucracy, which leaves no free agent in all France, except the man at Paris who pulls the wires. (Oxford English Dictionary) Bureaucracy • China (3rd century BC) Confucius • France (18th century) • USSR (1917-1990) • European Community (1952) The French ENA: National School of Administration Bureaucracy Bureaucracy • Rules Conserved quantities Bureaucracy • Rules Conserved quantities • Individuals particles undiscernable Bureaucracy • Rules Conserved quantities • Individuals particles undiscernable • Removal Shocks of an individual Loss of information Bureaucracy • Rules Conserved quantities • Individuals particles undiscernable • Removal Shocks Loss of information Maximum of entropy No evolution Bureaucracy • A bureaucratic system is stable (its entropy is maximum). • Example: China empire lasted for 2000 years. • It cannot be changed without a major source of instability. • Example: collapse of the USSR COMPUTERS: machines and brains Computers • Alan TURING (1912-1954) • 1936: • Description of a computing machine • Computers execute logical operations • They produce information, memorize them, treat them, Computers • A Turing machine is sequential: operations are time ordered rules states Left-Right tape Computers CPU data & instructions data MEMORY • The von NEUMANN computer repeatedly performs the following cycle of events 1. fetch an instruction from memory. 2. fetch any data required by the instruction from memory. 3. execute the instruction (process the data). 4. store results in memory. 5. go back to step 1. Computers • February 14th 1946 ENIAC the first computer Los Alamos NM • Cellular automata produce patterns as in shells Computers a b b a a b a a b a a b b a computer simulation rule change pattern from layer to layer Computers • Nature has also produced brains • Brain does not seem to follow the von Neumann nor Turing schemes Computers • In brain signals are not binary but activated by thresholds • The operations are not performed sequentially Computers • Brain can learn • It can adapt itself: plasticity • Brain memory is associative: it recognizes patterns by comparison with pre-stored ones TO CONCLUDE Entropy & Information • The Second Law of Thermodynamics leads to global loss of information • Systems out of equilibrium produce information… to the cost of the environment • Information can be coded, transmitted, memorized, hidden, treated. • Life is a way of producing information: genetic code, proteins, chemical signals, pattern formation, neurons, brain. • Machines can produce similar features Is Nature a big computer ? THE END