Recent from talks
Nothing was collected or created yet.
History of physics
View on Wikipedia| Part of a series on |
| Physics |
|---|
Physics is a branch of science in which the primary objects of study are matter and energy. These topics were discussed across many cultures in ancient times by philosophers, but they had no means to distinguish causes of natural phenomena from superstitions.
The Scientific Revolution of the 17th century, especially the discovery of the law of gravity, began a process of knowledge accumulation and specialization that gave rise to the field of physics.
Mathematical advances of the 18th century gave rise to classical mechanics, and the increased used of the experimental method led to new understanding of thermodynamics.
In the 19th century, the basic laws of electromagnetism and statistical mechanics were discovered.
At the beginning of the 20th century, physics was transformed by the discoveries of quantum mechanics, relativity, and atomic theory.
Physics today may be divided loosely into classical physics and modern physics.
Ancient history
[edit]Elements of what became physics were drawn primarily from the fields of astronomy, optics, and mechanics, which were methodologically united through the study of geometry. These mathematical disciplines began in antiquity with the Babylonians and with Hellenistic writers such as Archimedes and Ptolemy. Ancient philosophy, meanwhile, included what was called "Physics".
Greek concept
[edit]The move towards a rational understanding of nature began at least since the Archaic period in Greece (650–480 BCE) with the Pre-Socratic philosophers. The philosopher Thales of Miletus (7th and 6th centuries BCE), dubbed "the Father of Science" for refusing to accept various supernatural, religious or mythological explanations for natural phenomena, proclaimed that every event had a natural cause.[1] Thales also made advancements in 580 BCE by suggesting that water is the basic element, experimenting with the attraction between magnets and rubbed amber and formulating the first recorded cosmologies. Anaximander, developer of a proto-evolutionary theory, disputed Thales' ideas and proposed that rather than water, a substance called apeiron was the building block of all matter. Around 500 BCE, Heraclitus proposed that the only basic law governing the Universe was the principle of change and that nothing remains in the same state indefinitely. He, along with his contemporary Parmenides were among the first scholars to contemplate on the role of time in the universe, a key concept that is still an issue in modern physics.

During the classical period in Greece (6th, 5th and 4th centuries BCE) and in Hellenistic times, natural philosophy developed into a field of study. Aristotle (Greek: Ἀριστοτέλης, Aristotélēs) (384–322 BCE), a student of Plato, promoted the concept that observation of physical phenomena could ultimately lead to the discovery of the natural laws governing them.[citation needed] Aristotle's writings cover physics, metaphysics, poetry, theater, music, logic, rhetoric, linguistics, politics, government, ethics, biology and zoology. He wrote the first work which refers to that line of study as "Physics" – in the 4th century BCE, Aristotle founded the system known as Aristotelian physics. He attempted to explain ideas such as motion (and gravity) with the theory of four elements. Aristotle believed that all matter was made of aether, or some combination of four elements: earth, water, air, and fire. According to Aristotle, these four terrestrial elements are capable of inter-transformation and move toward their natural place, so a stone falls downward toward the center of the cosmos, but flames rise upward toward the circumference. Eventually, Aristotelian physics became popular for many centuries in Europe, informing the scientific and scholastic developments of the Middle Ages. It remained the mainstream scientific paradigm in Europe until the time of Galileo Galilei and Isaac Newton.
Early in Classical Greece, knowledge that the Earth is spherical ("round") was common. Around 240 BCE, as the result of a seminal experiment, Eratosthenes (276–194 BCE) accurately estimated its circumference. In contrast to Aristotle's geocentric views, Aristarchus of Samos (Greek: Ἀρίσταρχος; c. 310 – c. 230 BCE) presented an explicit argument for a heliocentric model of the Solar System, i.e. for placing the Sun, not the Earth, at its centre. Seleucus of Seleucia, a follower of Aristarchus' heliocentric theory, stated that the Earth rotated around its own axis, which, in turn, revolved around the Sun. Though the arguments he used were lost, Plutarch stated that Seleucus was the first to prove the heliocentric system through reasoning.

In the 3rd century BCE, the Greek mathematician Archimedes of Syracuse Greek: Ἀρχιμήδης (287–212 BCE) – generally considered to be the greatest mathematician of antiquity and one of the greatest of all time – laid the foundations of hydrostatics, statics and calculated the underlying mathematics of the lever. A scientist of classical antiquity, Archimedes also developed elaborate systems of pulleys to move large objects with a minimum of effort. The Archimedes' screw underpins modern hydroengineering, and his machines of war helped to hold back the armies of Rome in the First Punic War. Archimedes even tore apart the arguments of Aristotle and his metaphysics, pointing out that it was impossible to separate mathematics and nature and proved it by converting mathematical theories into practical inventions. Furthermore, in his work On Floating Bodies, around 250 BCE, Archimedes developed the law of buoyancy, also known as Archimedes' principle. In mathematics, Archimedes used the method of exhaustion to calculate the area under the arc of a parabola with the summation of an infinite series, and gave a remarkably accurate approximation of pi. He also defined the spiral bearing his name, formulae for the volumes of surfaces of revolution and an ingenious system for expressing very large numbers. He also developed the principles of equilibrium states and centers of gravity, ideas that would influence future scholars like Galileo, and Newton.
Hipparchus (190–120 BCE), focusing on astronomy and mathematics, used sophisticated geometrical techniques to map the motion of the stars and planets, even predicting the times that Solar eclipses would happen. He added calculations of the distance of the Sun and Moon from the Earth, based upon his improvements to the observational instruments used at that time. Another of the early physicists was Ptolemy (90–168 CE) during the time of the Roman Empire. Ptolemy was the author of several scientific treatises, at least three of which were of continuing importance to later Islamic and European science. The first is the astronomical treatise now known as the Almagest (in Greek, Ἡ Μεγάλη Σύνταξις, "The Great Treatise", originally Μαθηματικὴ Σύνταξις, "Mathematical Treatise"). The second is the Geography, which is a thorough discussion of the geographic knowledge of the Greco-Roman world.
Much of the accumulated knowledge of the ancient world was lost. Even of the works of the many respectable thinkers, few fragments survive. Although he wrote at least fourteen books, almost nothing of Hipparchus' direct work survived. Of the 150 reputed Aristotelian works, only 30 exist, and some of those are "little more than lecture notes".[according to whom?]
India and China
[edit]
Important physical and mathematical traditions also existed in ancient Indian and Chinese sciences.
In Indian philosophy, Maharishi Kanada was the first to systematically develop a theory of atomism around 200 BCE[3] though some authors have allotted him an earlier era in the 6th century BCE.[4][5] It was further elaborated by the Buddhist atomists Dharmakirti and Dignāga during the 1st millennium CE.[6] Pakudha Kaccayana, a 6th-century BCE Indian philosopher and contemporary of Gautama Buddha, had also propounded ideas about the atomic constitution of the material world. The Vaisheshika school of philosophers believed that an atom was a mere point in space. It was also first to depict relations between motion and force applied. Indian theories about the atom are greatly abstract and enmeshed in philosophy as they were based on logic and not on personal experience or experimentation.
In Indian astronomy, Aryabhata's Aryabhatiya (499 CE) proposed the Earth's rotation, while Nilakantha Somayaji (1444–1544) of the Kerala school of astronomy and mathematics proposed a semi-heliocentric model resembling the Tychonic system.
The study of magnetism in Ancient China dates to the 4th century BCE (in the Book of the Devil Valley Master).[7] A main contributor to this field was Shen Kuo (1031–1095), a polymath and statesman who was the first to describe the magnetic-needle compass used for navigation, as well as establishing the concept of true north. In optics, Shen Kuo independently developed a camera obscura.[8]
Islamic world
[edit]
In the 7th to 15th centuries, scientific progress occurred in the Muslim world. Many classic works in Indian, Assyrian, Sassanian (Persian) and Greek, including the works of Aristotle, were translated into Arabic.[9] Important contributions were made by Ibn al-Haytham (965–1040), an Arab[10] or Persian[11] scientist, considered to be a founder of modern optics. Ptolemy and Aristotle theorised that light either shone from the eye to illuminate objects or that "forms" emanated from objects themselves, whereas al-Haytham (known by the Latin name "Alhazen") suggested that light travels to the eye in rays from different points on an object. The works of Ibn al-Haytham and al-Biruni (973–1050), a Persian scientist, eventually passed on to Western Europe where they were studied by scholars such as Roger Bacon and Vitello.[12]
Ibn al-Haytham used controlled experiments in his work on optics, although to what extent it differed from Ptolemy is debated.[13][14] Arabic mechanics like Bīrūnī and Al-Khazini developed sophisticated "science of weight", carrying out measurements of specific weights and volumes.[15]
Ibn Sīnā (980–1037), known as "Avicenna", was a polymath from Bukhara (in present-day Uzbekistan) responsible for important contributions to physics, optics, philosophy and medicine. He published his theory of motion in Book of Healing (1020), where he argued that an impetus is imparted to a projectile by the thrower. He viewed it as persistent, requiring external forces such as air resistance to dissipate it.[16][17][18] Ibn Sina made a distinction between 'force' and 'inclination' (called "mayl"), and argued that an object gained mayl when the object is in opposition to its natural motion. He concluded that continuation of motion is attributed to the inclination that is transferred to the object, and that object will be in motion until the mayl is spent. This conception of motion is consistent with Newton's first law of motion, inertia, which states that an object in motion will stay in motion unless it is acted on by an external force.[16] This idea which dissented from the Aristotelian view was later described as "impetus" by John Buridan, who was likely influenced by Ibn Sina's Book of Healing.[19]

Hibat Allah Abu'l-Barakat al-Baghdaadi (c. 1080 – c. 1165) adopted and modified Ibn Sina's theory on projectile motion. In his Kitab al-Mu'tabar, Abu'l-Barakat stated that the mover imparts a violent inclination (mayl qasri) on the moved and that this diminishes as the moving object distances itself from the mover.[20] He also proposed an explanation of the acceleration of falling bodies by the accumulation of successive increments of power with successive increments of velocity.[21] According to Shlomo Pines, al-Baghdaadi's theory of motion was "the oldest negation of Aristotle's fundamental dynamic law [namely, that a constant force produces a uniform motion], [and is thus an] anticipation in a vague fashion of the fundamental law of classical mechanics [namely, that a force applied continuously produces acceleration]."[22] Jean Buridan and Albert of Saxony later referred to Abu'l-Barakat in explaining that the acceleration of a falling body is a result of its increasing impetus.[20]
Ibn Bajjah (c. 1085–1138), known as "Avempace" in Europe, proposed that for every force there is always a reaction force. Ibn Bajjah was a critic of Ptolemy and he worked on creating a new theory of velocity to replace the one theorized by Aristotle. Two future philosophers supported the theories Avempace created, known as Avempacean dynamics. These philosophers were Thomas Aquinas, a Catholic priest, and John Duns Scotus.[23] Galileo went on to adopt Avempace's formula "that the velocity of a given object is the difference of the motive power of that object and the resistance of the medium of motion".[23]
Nasir al-Din al-Tusi (1201–1274), a Persian astronomer and mathematician who died in Baghdad, introduced the Tusi couple an important mathematical theorem and founded the Maragha School of astronomy. Geocentric (but not heliocentric) astronomical models developed by the Maragha School have many striking parallels with models developed by Nicolaus Copernicus. The possibility that Maragha results may have influenced Copernicus has a been investigated in some detail.[24]
Medieval Europe
[edit]Awareness of ancient works re-entered the West through translations from Arabic to Latin. Their re-introduction, combined with Judeo-Islamic theological commentaries, had a great influence on Medieval philosophers such as Thomas Aquinas. Scholastic European scholars, who sought to reconcile the philosophy of the ancient classical philosophers with Christian theology, proclaimed Aristotle the greatest thinker of the ancient world. In cases where they did not directly contradict the Bible, Aristotelian physics became the foundation for the physical explanations of the European Churches. Quantification became a core element of medieval physics.[25]
Based on Aristotelian physics, Scholastic physics described things as moving according to their essential nature. Celestial objects were described as moving in circles, because perfect circular motion was considered an innate property of objects that existed in the uncorrupted realm of the celestial spheres. Motions below the lunar sphere were seen as imperfect, and thus could not be expected to exhibit consistent motion. More idealized motion in the "sublunary" realm could only be achieved through artifice, and prior to the 17th century, many did not view artificial experiments as a valid means of learning about the natural world. Physical explanations in the sublunary realm revolved around tendencies. Stones contained the element earth, and earthly objects tended to move in a straight line toward the centre of the earth (and the universe in the Aristotelian geocentric view) unless otherwise prevented from doing so.[26]
Aristotle's physics was not scrutinized until John Philoponus, who relied on observation rather than verbal argument like Aristotle.[27] Philoponus' criticism of Aristotelian principles of physics served as an inspiration for Galileo Galilei ten centuries later,[28] during the Scientific Revolution. Galileo cited Philoponus substantially in his works when arguing that Aristotelian physics was flawed.[29][30] In the 1300s Jean Buridan, a teacher in the faculty of arts at the University of Paris, developed the concept of impetus. It was a step toward the modern ideas of inertia and momentum.[31]
Scientific Revolution
[edit]During the 16th and 17th centuries, a large advancement of scientific progress known as the Scientific Revolution took place in Europe. Dissatisfaction with older philosophical approaches had begun earlier and had produced other changes in society, such as the Protestant Reformation, but the revolution in science began when natural philosophers began to mount a sustained attack on the Scholastic philosophical programme and supposed that mathematical descriptive schemes adopted from such fields as mechanics and astronomy could actually yield universally valid characterizations of motion and other concepts.
Nicolaus Copernicus
[edit]A breakthrough in astronomy was made by Renaissance astronomer Nicolaus Copernicus (1473–1543) when, in 1543, he gave strong arguments for the heliocentric model of the Solar System, ostensibly as a means to render tables charting planetary motion more accurate and to simplify their production. In heliocentric models of the Solar system, the Earth orbits the Sun along with other bodies in Earth's galaxy, a contradiction according to the Greek-Egyptian astronomer Ptolemy (2nd century CE; see above), whose system placed the Earth at the center of the Universe and had been accepted for over 1,400 years. The Greek astronomer Aristarchus of Samos (c. 310 – c. 230 BCE) had suggested that the Earth revolves around the Sun, but Copernicus's reasoning led to lasting general acceptance of this "revolutionary" idea. Copernicus's book presenting the theory (De revolutionibus orbium coelestium, "On the Revolutions of the Celestial Spheres") was published just before his death in 1543 and, as it is now generally considered to mark the beginning of modern astronomy, is also considered to mark the beginning of the Scientific Revolution.[citation needed] Copernicus's new perspective, along with the accurate observations made by Tycho Brahe, enabled German astronomer Johannes Kepler (1571–1630) to formulate his laws regarding planetary motion that remain in use today.
Galileo Galilei
[edit]
The Italian mathematician, astronomer, and physicist Galileo Galilei (1564–1642) was a supporter of Copernicanism who made numerous astronomical discoveries, carried out empirical experiments and improved the telescope. As a mathematician, Galileo's role in the university culture of his era was subordinated to the three major topics of study: law, medicine, and theology (which was closely allied to philosophy). Galileo, however, felt that the descriptive content of the technical disciplines warranted philosophical interest, particularly because mathematical analysis of astronomical observations – notably, Copernicus's analysis of the relative motions of the Sun, Earth, Moon, and planets – indicated that philosophers' statements about the nature of the universe could be shown to be in error. Galileo also performed mechanical experiments, insisting that motion itself – regardless of whether it was produced "naturally" or "artificially" (i.e. deliberately) – had universally consistent characteristics that could be described mathematically.
Galileo's early studies at the University of Pisa were in medicine, but he was soon drawn to mathematics and physics. At age 19, he discovered (and, subsequently, verified) the isochronal nature of the pendulum when, using his pulse, he timed the oscillations of a swinging lamp in Pisa's cathedral and found that it remained the same for each swing regardless of the swing's amplitude. He soon became known through his invention of a hydrostatic balance and for his treatise on the center of gravity of solid bodies. While teaching at the University of Pisa (1589–1592), he initiated his experiments concerning the laws of bodies in motion that brought results so contradictory to the accepted teachings of Aristotle that strong antagonism was aroused. He found that bodies do not fall with velocities proportional to their weights. The story in which Galileo is said to have dropped weights from the Leaning Tower of Pisa is apocryphal, but he did find that the path of a projectile is a parabola and is credited with conclusions that anticipated Newton's laws of motion (e.g. the notion of inertia). Among these is what is now called Galilean relativity, the first precisely formulated statement about properties of space and time outside three-dimensional geometry.[citation needed]

Galileo has been called the "father of modern observational astronomy",[32] the "father of modern physics", the "father of science",[33] and "the father of modern science".[34] According to Stephen Hawking, "Galileo, perhaps more than any other single person, was responsible for the birth of modern science."[35] As religious orthodoxy decreed a geocentric or Tychonic understanding of the Solar system, Galileo's support for heliocentrism provoked controversy and he was tried by the Inquisition. Found "vehemently suspect of heresy", he was forced to recant and spent the rest of his life under house arrest.
The contributions that Galileo made to observational astronomy include the telescopic confirmation of the phases of Venus; his discovery, in 1609, of Jupiter's four largest moons (subsequently given the collective name of the "Galilean moons"); and the observation and analysis of sunspots. Galileo also pursued applied science and technology, inventing, among other instruments, a military compass. His discovery of the Jovian moons was published in 1610, enabling him to obtain the position of mathematician and philosopher to the Medici court. As such, he was expected to engage in debates with philosophers in the Aristotelian tradition and received a large audience for his own publications such as the Discourses and Mathematical Demonstrations Concerning Two New Sciences (published abroad following his arrest for the publication of Dialogue Concerning the Two Chief World Systems) and The Assayer.[36][37] Galileo's interest in experimenting with and formulating mathematical descriptions of motion established experimentation as an integral part of natural philosophy. This tradition, combining with the non-mathematical emphasis on the collection of "experimental histories" by philosophical reformists such as William Gilbert and Francis Bacon, drew a significant following in the years leading to and following Galileo's death, including Evangelista Torricelli and the participants in the Accademia del Cimento in Italy; Marin Mersenne and Blaise Pascal in France; Christiaan Huygens in the Netherlands; and Robert Hooke and Robert Boyle in England.
Johannes Kepler
[edit]
Johannes Kepler (1571–1630) was a German astronomer, mathematician, astrologer, natural philosopher and a key figure in the 17th century Scientific Revolution, best known for his laws of planetary motion, and his books Astronomia nova, Harmonice Mundi, and Epitome Astronomiae Copernicanae, influencing among others Isaac Newton, providing one of the foundations for his theory of universal gravitation.[38] The variety and impact of his work made Kepler one of the founders of modern astronomy, the scientific method, natural and modern science.[39][40][41]
Kepler was partly driven by his belief that there is an intelligible plan that is accessible through reason.[42] Kepler described his new astronomy as "celestial physics",[43] as "an excursion into Aristotle's Metaphysics",[44] and as "a supplement to Aristotle's On the Heavens",[45] treating astronomy as part of a universal mathematical physics.[46]
René Descartes
[edit]
The French philosopher René Descartes (1596–1650) was well-connected to, and influential within, experimental philosophy networks. Descartes had an agenda, however, which was geared toward replacing the Scholastic philosophical tradition. Questioning the reality interpreted through the senses, Descartes sought to re-establish philosophical explanations by reducing all phenomena to the motion of an invisible sea of "corpuscles". (Notably, he reserved human thought and God from his scheme, holding these to be separate from the physical universe). In proposing this philosophical framework, Descartes supposed that different kinds of motion, such as that of planets versus that of terrestrial objects, were not fundamentally different, but were manifestations of an endless chain of corpuscular motions obeying universal principles. Particularly influential were his explanations for circular astronomical motions in terms of the vortex motion of corpuscles in space (Descartes argued, in accord with the beliefs, if not the methods, of the Scholastics, that a vacuum could not exist), and his explanation of gravity in terms of corpuscles pushing objects downward.[47][48][49]
Descartes, like Galileo, was convinced of the importance of mathematical explanation, and he and his followers were key figures in the development of mathematics and geometry in the 17th century. Cartesian mathematical descriptions of motion held that all mathematical formulations had to be justifiable in terms of direct physical action, a position held by Christiaan Huygens and the German philosopher Gottfried Leibniz, who, while following in the Cartesian tradition, developed his own philosophical alternative to Scholasticism, which he outlined in his 1714 work, the Monadology. Descartes has been dubbed the "Father of Modern Philosophy", and much subsequent Western philosophy is a response to his writings, which are studied closely to this day. In particular, his Meditations on First Philosophy continues to be a standard text at most university philosophy departments. Descartes' influence in mathematics is equally apparent; the Cartesian coordinate system – allowing algebraic equations to be expressed as geometric shapes in a two-dimensional coordinate system – was named after him. He is credited as the father of analytical geometry, the bridge between algebra and geometry, important to the discovery of calculus and analysis.
Christiaan Huygens
[edit]
The Dutch physicist, mathematician, astronomer and inventor Christiaan Huygens (1629–1695) was the leading scientist in Europe between Galileo and Newton. Huygens came from a family of nobility that had an important position in the Dutch society of the 17th century; a time in which the Dutch Republic flourished economically and culturally. This period – roughly between 1588 and 1702 – of the history of the Netherlands is also referred to as the Dutch Golden Age, an era during the Scientific Revolution when Dutch science was among the most acclaimed in Europe. At this time, intellectuals and scientists like René Descartes, Baruch Spinoza, Pierre Bayle, Antonie van Leeuwenhoek, John Locke and Hugo Grotius resided in the Netherlands. It was in this intellectual environment that Christiaan Huygens grew up. Christiaan's father, Constantijn Huygens, was, apart from an important poet, the secretary and diplomat for the Princes of Orange. He knew many scientists of his time because of his contacts and intellectual interests, including René Descartes and Marin Mersenne, and it was because of these contacts that Christiaan Huygens became aware of their work, especially Descartes, whose mechanistic philosophy was going to have a huge influence on Huygens' own work. Descartes was later impressed by the skills Huygens showed in geometry, as was Mersenne, who christened him "the new Archimedes" (which led Constantijn to refer to his son as "my little Archimedes").
A child prodigy, Huygens began his correspondence with Marin Mersenne when he was 17 years old. Huygens became interested in games of chance when he encountered the work of Fermat, Blaise Pascal and Girard Desargues. It was Pascal who encouraged him to write Van Rekeningh in Spelen van Gluck, which Frans van Schooten translated and published as De Ratiociniis in Ludo Aleae in 1657. The book is the earliest known scientific treatment of the subject, and at the time the most coherent presentation of a mathematical approach to games of chance. Two years later Huygens derived geometrically the now standard formulae in classical mechanics for the centripetal- and centrifugal force in his work De vi Centrifuga (1659). Around the same time Huygens' research in horology resulted in the invention of the pendulum clock; a breakthrough in timekeeping and the most accurate timekeeper for almost 300 years. The theoretical research of the way the pendulum works eventually led to the publication of one of his most important achievements: the Horologium Oscillatorium. This work was published in 1673 and became one of the three most important 17th century works on mechanics (the other two being Galileo's Discourses and Mathematical Demonstrations Relating to Two New Sciences (1638) and Newton's Philosophiæ Naturalis Principia Mathematica (1687)[50]). The Horologium Oscillatorium is the first modern treatise in which a physical problem (the accelerated motion of a falling body) is idealized by a set of parameters then analyzed mathematically and constitutes one of the seminal works of applied mathematics.[51][52] It is for this reason, Huygens has been called the first theoretical physicist and one of the founders of modern mathematical physics.[53][54] Huygens' Horologium Oscillatorium influenced the work of Isaac Newton, who admired the work. For instance, the laws Huygens described in the Horologium Oscillatorium are structurally the same as Newton's first two laws of motion.[55]
Five years after the publication of his Horologium Oscillatorium, Huygens described his wave theory of light. Though proposed in 1678, it was not published until 1690 in his Traité de la Lumière. His mathematical theory of light was initially rejected in favour of Newton's corpuscular theory of light, until Augustin-Jean Fresnel adopted Huygens' principle to give a complete explanation of the rectilinear propagation and diffraction effects of light in 1821. Today this principle is known as the Huygens–Fresnel principle.
As an astronomer, Huygens began grinding lenses with his brother Constantijn Jr. to build telescopes for astronomical research. He was the first to identify the rings of Saturn as "a thin, flat ring, nowhere touching, and inclined to the ecliptic," and discovered the first of Saturn's moons, Titan, using a refracting telescope.
Huygens was also the first who brought mathematical rigor to the description of physical phenomena. Because of this, and the fact that he developed institutional frameworks for scientific research on the continent, he has been referred to as "the leading actor in 'the making of science in Europe'"[56]
Isaac Newton
[edit]
Cambridge University physicist and mathematician Sir Isaac Newton (1642–1727) was a fellow of the Royal Society of England, who created a single system for describing the workings of the universe. Newton formulated three laws of motion which formulated the relationship between motion and objects and also the law of universal gravitation, the latter of which could be used to explain the behavior not only of falling bodies on the earth but also planets and other celestial bodies. To arrive at his results, Newton invented one form of an entirely new branch of mathematics: calculus (also invented independently by Gottfried Leibniz), which was to become an essential tool in much of the later development in most branches of physics. Newton's findings were set forth in his Philosophiæ Naturalis Principia Mathematica ("Mathematical Principles of Natural Philosophy"), the publication of which in 1687 marked the beginning of the modern period of mechanics and astronomy.
Newton refuted the Cartesian mechanical tradition that all motions should be explained with respect to the immediate force exerted by corpuscles. Using his three laws of motion and law of universal gravitation, Newton removed the idea that objects followed paths determined by natural shapes and instead demonstrated that all the future motions of any body could be deduced mathematically based on knowledge of their existing motion, their mass, and the forces acting upon them. However, observed celestial motions did not precisely conform to a Newtonian treatment, and Newton, who was also deeply interested in theology, imagined that God intervened to ensure the continued stability of the solar system.

Newton's principles (but not his mathematical treatments) proved controversial with Continental philosophers, who found his lack of metaphysical explanation for movement and gravitation philosophically unacceptable. Beginning around 1700, a bitter rift opened between the Continental and British philosophical traditions, which were stoked by heated, ongoing, and viciously personal disputes between the followers of Newton and Leibniz concerning priority over the analytical techniques of calculus, which each had developed independently. Initially, the Cartesian and Leibnizian traditions prevailed on the Continent (leading to the dominance of the Leibnizian calculus notation everywhere except Britain). Newton himself remained privately disturbed at the lack of a philosophical understanding of gravitation while insisting in his writings that none was necessary to infer its reality. As the 18th century progressed, Continental natural philosophers increasingly accepted the Newtonians' willingness to forgo ontological metaphysical explanations for mathematically described motions.[57][58][59]
Newton built the first functioning reflecting telescope[60] and developed a theory of color, published in Opticks, based on the observation that a prism decomposes white light into the many colours forming the visible spectrum. While Newton explained light as being composed of tiny particles, a rival theory of light which explained its behavior in terms of waves was presented in 1690 by Christiaan Huygens. However, the belief in the mechanistic philosophy coupled with Newton's reputation meant that the wave theory saw relatively little support until the 19th century. Newton also formulated an empirical law of cooling, studied the speed of sound, investigated power series, demonstrated the generalised binomial theorem and developed a method for approximating the roots of a function. His work on infinite series was inspired by Simon Stevin's decimals.[61] Most importantly, Newton showed that the motions of objects on Earth and of celestial bodies are governed by the same set of natural laws, which were neither capricious nor malevolent. By demonstrating the consistency between Kepler's laws of planetary motion and his own theory of gravitation, Newton also removed the last doubts about heliocentrism. By bringing together all the ideas set forth during the Scientific Revolution, Newton effectively established the foundation for modern society in mathematics and science.
Other achievements
[edit]Other branches of physics also received attention during the period of the Scientific Revolution. William Gilbert, court physician to Queen Elizabeth I, described how the earth itself behaves like a giant magnet. Robert Boyle (1627–1691) studied the behavior of gases enclosed in a chamber and formulated the gas law named for him; he also contributed to physiology and to the founding of modern chemistry.
Another factor in the Scientific Revolution was the rise of learned societies and academies in various countries. The earliest of these were in Italy and Germany and were short-lived. More influential were the Royal Society of England (1660) and the Academy of Sciences in France (1666). The former was a private institution in London and included John Wallis, William Brouncker, Thomas Sydenham, John Mayow, and Christopher Wren (who contributed not only to architecture but also to astronomy and anatomy); the latter, in Paris, was a government institution and included as a foreign member the Dutchman Huygens. In the 18th century, important royal academies were established at Berlin (1700) and at St. Petersburg (1724). The societies and academies provided the principal opportunities for the publication and discussion of scientific results during and after the scientific revolution. In 1690, James Bernoulli showed that the cycloid is the solution to the tautochrone problem; and the following year, in 1691, Johann Bernoulli showed that a chain freely suspended from two points will form a catenary, the curve with the lowest possible center of gravity available to any chain hung between two fixed points. He then showed, in 1696, that the cycloid is the solution to the brachistochrone problem.
Early thermodynamics
[edit]
A precursor of the engine was designed by the German scientist Otto von Guericke who, in 1650, designed and built the world's first vacuum pump to create a vacuum as demonstrated in the Magdeburg hemispheres experiment. He was driven to make a vacuum to disprove Aristotle's long-held supposition that 'Nature abhors a vacuum'. Shortly thereafter, Irish physicist and chemist Boyle had learned of Guericke's designs and in 1656, in coordination with English scientist Robert Hooke, built an air pump. Using this pump, Boyle and Hooke noticed the pressure-volume correlation for a gas: PV = k, where P is pressure, V is volume and k is a constant: this relationship is known as Boyle's law. In that time, air was assumed to be a system of motionless particles, and not interpreted as a system of moving molecules. The concept of thermal motion came two centuries later. Therefore, Boyle's publication in 1660 speaks about a mechanical concept: the air spring.[62] Later, after the invention of the thermometer, the property temperature could be quantified. This tool gave Joseph Louis Gay-Lussac the opportunity to derive his law, which led shortly later to the ideal gas law. But, already before the establishment of the ideal gas law, an associate of Boyle's named Denis Papin built in 1679 a bone digester, which is a closed vessel with a tightly fitting lid that confines steam until a high pressure is generated.
Later designs implemented a steam release valve to keep the machine from exploding. By watching the valve rhythmically move up and down, Papin conceived of the idea of a piston and cylinder engine. He did not however follow through with his design. Nevertheless, in 1697, based on Papin's designs, engineer Thomas Savery built the first engine. Although these early engines were crude and inefficient, they attracted the attention of the leading scientists of the time. Hence, prior to 1698 and the invention of the Savery Engine, horses were used to power pulleys, attached to buckets, which lifted water out of flooded salt mines in England. In the years to follow, more variations of steam engines were built, such as the Newcomen Engine, and later the Watt Engine. In time, these early engines would replace horses. Thus, each engine began to be associated with a certain amount of "horse power" depending upon how many horses it had replaced. The main problem with these first engines was that they were slow and clumsy, converting less than 2% of the input fuel into useful work. In other words, large quantities of coal (or wood) had to be burned to yield a small fraction of work output; the need for a new science of engine dynamics was born.
18th-century developments
[edit]During the 18th century, the mechanics founded by Newton was developed by several scientists as more mathematicians learned calculus and elaborated upon its initial formulation. The application of mathematical analysis to problems of motion was known as rational mechanics, or mixed mathematics (and was later termed classical mechanics).
Mechanics
[edit]
In 1714, Brook Taylor derived the fundamental frequency of a stretched vibrating string in terms of its tension and mass per unit length by solving a differential equation. The Swiss mathematician Daniel Bernoulli (1700–1782) made important mathematical studies of the behavior of gases, anticipating the kinetic theory of gases developed more than a century later, and has been referred to as the first mathematical physicist.[63] In 1733, Daniel Bernoulli derived the fundamental frequency and harmonics of a hanging chain by solving a differential equation. In 1734, Bernoulli solved the differential equation for the vibrations of an elastic bar clamped at one end. Bernoulli's treatment of fluid dynamics and his examination of fluid flow was introduced in his 1738 work Hydrodynamica.
Rational mechanics dealt primarily with the development of elaborate mathematical treatments of observed motions, using Newtonian principles as a basis, and emphasized improving the tractability of complex calculations and developing of legitimate means of analytical approximation. A representative contemporary textbook was published by Johann Baptiste Horvath. By the end of the century analytical treatments were rigorous enough to verify the stability of the Solar System solely on the basis of Newton's laws without reference to divine intervention – even as deterministic treatments of systems as simple as the three body problem in gravitation remained intractable.[64] In 1705, Edmond Halley predicted the periodicity of Halley's Comet, William Herschel discovered Uranus in 1781, and Henry Cavendish measured the gravitational constant and determined the mass of the Earth in 1798. In 1783, John Michell suggested that some objects might be so massive that not even light could escape from them.
In 1739, Leonhard Euler solved the ordinary differential equation for a forced harmonic oscillator and noticed the resonance phenomenon. In 1742, Colin Maclaurin discovered his uniformly rotating self-gravitating spheroids. In 1742, Benjamin Robins published his New Principles in Gunnery, establishing the science of aerodynamics. British work, carried on by mathematicians such as Taylor and Maclaurin, fell behind Continental developments as the century progressed. Meanwhile, work flourished at scientific academies on the Continent, led by such mathematicians as Bernoulli and Euler, as well as Joseph-Louis Lagrange, Pierre-Simon Laplace, and Adrien-Marie Legendre. In 1743, Jean le Rond d'Alembert published his Traité de dynamique, in which he introduced the concept of generalized forces for accelerating systems and systems with constraints, and applied the new idea of virtual work to solve dynamical problem, now known as D'Alembert's principle, as a rival to Newton's second law of motion. In 1747, Pierre Louis Maupertuis applied minimum principles to mechanics. In 1759, Euler solved the partial differential equation for the vibration of a rectangular drum. In 1764, Euler examined the partial differential equation for the vibration of a circular drum and found one of the Bessel function solutions. In 1776, John Smeaton published a paper on experiments relating power, work, momentum and kinetic energy, and supporting the conservation of energy. In 1788, Lagrange presented his equations of motion in Mécanique analytique, in which the whole of mechanics was organized around the principle of virtual work. In 1789, Antoine Lavoisier stated the law of conservation of mass. The rational mechanics developed in the 18th century received expositions in both Lagrange's Mécanique analytique and Laplace's Traité de mécanique céleste (1799–1825).
Thermodynamics and static electricity
[edit]
During the 18th century, thermodynamics was developed through the theories of weightless "imponderable fluids", such as heat ("caloric"), electricity, and phlogiston (which was rapidly overthrown as a concept following Antoine Lavoisier's identification of oxygen gas late in the century). Assuming that these concepts were real fluids, their flow could be traced through a mechanical apparatus or chemical reactions. This tradition of experimentation led to the development of new kinds of experimental apparatus, such as the Leyden jar; and new kinds of measuring instruments, such as the calorimeter, and improved versions of old ones, such as the thermometer. Experiments also produced new concepts, such as the University of Glasgow experimenter Joseph Black's notion of latent heat and Philadelphia intellectual Benjamin Franklin's characterization of electrical fluid as flowing between places of excess and deficit (a concept later reinterpreted in terms of positive and negative charges). Franklin also showed that lightning is electricity in 1752.
The accepted theory of heat in the 18th century viewed it as a kind of fluid, called caloric; although this theory was later shown to be erroneous, a number of scientists adhering to it nevertheless made important discoveries useful in developing the modern theory, including Joseph Black (1728–1799) and Henry Cavendish (1731–1810). Opposed to this caloric theory, which had been developed mainly by the chemists, was the less accepted theory dating from Newton's time that heat is due to the motions of the particles of a substance. This mechanical theory gained support in 1798 from the cannon-boring experiments of Count Rumford (Benjamin Thompson), who found a direct relationship between heat and mechanical energy.
While it was recognized early in the 18th century that finding absolute theories of electrostatic and magnetic force akin to Newton's principles of motion would be an important achievement, none were forthcoming. This impossibility only slowly disappeared as experimental practice became more widespread and more refined in the early years of the 19th century in places such as the newly established Royal Institution in London. Meanwhile, the analytical methods of rational mechanics began to be applied to experimental phenomena, most influentially with the French mathematician Joseph Fourier's analytical treatment of the flow of heat, as published in 1822.[65][66][67] Joseph Priestley proposed an electrical inverse-square law in 1767, and Charles-Augustin de Coulomb introduced the inverse-square law of electrostatics in 1798.
At the end of the century, the members of the French Academy of Sciences had attained clear dominance in the field.[59][68][69][70] At the same time, the experimental tradition established by Galileo and his followers persisted. The Royal Society and the French Academy of Sciences were major centers for the performance and reporting of experimental work. Experiments in mechanics, optics, magnetism, static electricity, chemistry, and physiology were not clearly distinguished from each other during the 18th century, but significant differences in explanatory schemes and, thus, experiment design were emerging. Chemical experimenters, for instance, defied attempts to enforce a scheme of abstract Newtonian forces onto chemical affiliations, and instead focused on the isolation and classification of chemical substances and reactions.[71]
19th century
[edit]Mechanics
[edit]In 1821, William Hamilton began his analysis of Hamilton's characteristic function. In 1835, he stated Hamilton's canonical equations of motion.
In 1813, Peter Ewart supported the idea of the conservation of energy in his paper On the measure of moving force. In 1829, Gaspard Coriolis introduced the terms of work (force times distance) and kinetic energy with the meanings they have today. In 1841, Julius Robert von Mayer, an amateur scientist, wrote a paper on the conservation of energy, although his lack of academic training led to its rejection. In 1847, Hermann von Helmholtz formally stated the law of conservation of energy.
Electromagnetism
[edit]
In 1800, Alessandro Volta invented the electric battery (known as the voltaic pile) and thus improved the way electric currents could also be studied. A year later, Thomas Young demonstrated the wave nature of light – which received strong experimental support from the work of Augustin-Jean Fresnel – and the principle of interference. In 1820, Hans Christian Ørsted found that a current-carrying conductor gives rise to a magnetic force surrounding it, and within a week after Ørsted's discovery reached France, André-Marie Ampère discovered that two parallel electric currents will exert forces on each other. In 1821, Michael Faraday built an electricity-powered motor, while Georg Ohm stated his law of electrical resistance in 1826, expressing the relationship between voltage, current, and resistance in an electric circuit.
In 1831, Faraday (and independently Joseph Henry) discovered the reverse effect, the production of an electric potential or current through magnetism – known as electromagnetic induction; these two discoveries are the basis of the electric motor and the electric generator, respectively.
In 1873, James Clerk Maxwell published A Treatise on Electricity and Magnetism, which described the transmission of energy in wave form through a "luminiferous ether", and suggested that light was such a wave. This was confirmed in 1888 when Helmholtz student Heinrich Hertz generated and detected electromagnetic radiation in the laboratory.[72][73][74][75]
Laws of thermodynamics
[edit]
In the 19th century, the connection between heat and mechanical energy was established quantitatively by Julius Robert von Mayer and James Prescott Joule, who measured the mechanical equivalent of heat in the 1840s. In 1849, Joule published results from his series of experiments (including the paddlewheel experiment) which show that heat is a form of energy, a fact that was accepted in the 1850s. The relation between heat and energy was important for the development of steam engines, and in 1824 the experimental and theoretical work of Sadi Carnot was published. Carnot captured some of the ideas of thermodynamics in his discussion of the efficiency of an idealized engine. Sadi Carnot's work provided a basis for the formulation of the first law of thermodynamics – a restatement of the law of conservation of energy – which was stated around 1850 by William Thomson, later known as Lord Kelvin, and Rudolf Clausius. Lord Kelvin, who had extended the concept of absolute zero from gases to all substances in 1848, drew upon the engineering theory of Lazare Carnot, Sadi Carnot, and Émile Clapeyron as well as the experimentation of James Prescott Joule on the interchangeability of mechanical, chemical, thermal, and electrical forms of work to formulate the first law.
Kelvin and Clausius also stated the second law of thermodynamics, which was originally formulated in terms of the fact that heat does not spontaneously flow from a colder body to a warmer one. Other formulations followed quickly (for example, the second law was expounded in Thomson and Peter Guthrie Tait's influential work Treatise on Natural Philosophy) and Kelvin in particular understood some of the law's general implications. The second Law – the idea that gases consist of molecules in motion – had been discussed in some detail by Daniel Bernoulli in 1738, but had fallen out of favor, and was revived by Clausius in 1857. In 1850, Hippolyte Fizeau and Léon Foucault measured the speed of light in water and found that it is slower than in air, in support of the wave model of light. In 1852, Joule and Thomson demonstrated that a rapidly expanding gas cools, later named the Joule–Thomson effect or Joule–Kelvin effect. Hermann von Helmholtz put forward the idea of the heat death of the universe in 1854, the same year that Clausius established the importance of dQ/T (Clausius's theorem) (though he did not yet name the quantity).
Statistical mechanics
[edit]
In 1860, James Clerk Maxwell worked out the mathematics of the distribution of velocities of the molecules of a gas, known today as the Maxwell–Boltzmann distribution.
The atomic theory of matter had been proposed again in the early 19th century by the chemist John Dalton and became one of the hypotheses of the kinetic-molecular theory of gases developed by Clausius and James Clerk Maxwell to explain the laws of thermodynamics.
The kinetic theory in turn led to a revolutionary approach to science, the statistical mechanics of Ludwig Boltzmann (1844–1906) and Josiah Willard Gibbs (1839–1903), which studies the statistics of microstates of a system and uses statistics to determine the state of a physical system. Interrelating the statistical likelihood of certain states of organization of these particles with the energy of those states, Clausius reinterpreted the dissipation of energy to be the statistical tendency of molecular configurations to pass toward increasingly likely, increasingly disorganized states (coining the term "entropy" to describe the disorganization of a state). The statistical versus absolute interpretations of the second law of thermodynamics set up a dispute that would last for several decades (producing arguments such as "Maxwell's demon"), and that would not be held to be definitively resolved until the behavior of atoms was firmly established in the early 20th century.[76][77] In 1902, James Jeans found the length scale required for gravitational perturbations to grow in a static nearly homogeneous medium.
Other developments
[edit]In 1822, botanist Robert Brown discovered Brownian motion: pollen grains in water undergoing movement resulting from their bombardment by the fast-moving atoms or molecules in the liquid.
In 1834, Carl Jacobi discovered his uniformly rotating self-gravitating ellipsoids (the Jacobi ellipsoid).
In 1834, John Russell observed a nondecaying solitary water wave (soliton) in the Union Canal near Edinburgh, Scotland, and used a water tank to study the dependence of solitary water wave velocities on wave amplitude and water depth. In 1835, Gaspard Coriolis examined theoretically the mechanical efficiency of waterwheels, and deduced the Coriolis effect. In 1842, Christian Doppler proposed the Doppler effect.
In 1851, Léon Foucault showed the Earth's rotation with a huge pendulum (Foucault pendulum).
There were important advances in continuum mechanics in the first half of the century, namely formulation of laws of elasticity for solids and discovery of Navier–Stokes equations for fluids.
20th century: birth of modern physics
[edit]
(1867–1934) received Nobel prizes in physics (1903) and chemistry (1911).
At the end of the 19th century, physics had evolved to the point at which classical mechanics could cope with highly complex problems involving macroscopic situations; thermodynamics and kinetic theory were well established; geometrical and physical optics could be understood in terms of electromagnetic waves; and the conservation laws for energy and momentum (and mass) were widely accepted. So profound were these and other developments that it was generally accepted that all the important laws of physics had been discovered and that, henceforth, research would be concerned with clearing up minor problems and particularly with improvements of method and measurement.
However, around 1900 serious doubts arose about the completeness of the classical theories – the triumph of Maxwell's theories, for example, was undermined by inadequacies that had already begun to appear – and their inability to explain certain physical phenomena, such as the energy distribution in blackbody radiation and the photoelectric effect, while some of the theoretical formulations led to paradoxes when pushed to the limit. Prominent physicists such as Hendrik Lorentz, Emil Cohn, Ernst Wiechert and Wilhelm Wien believed that some modification of Maxwell's equations might provide the basis for all physical laws. These shortcomings of classical physics were never to be resolved and new ideas were required. At the beginning of the 20th century, a major revolution shook the world of physics, which led to a new era, generally referred to as modern physics.[78]
Radiation experiments
[edit]
In the 19th century, experimenters began to detect unexpected forms of radiation: Wilhelm Röntgen caused a sensation with his discovery of X-rays in 1895; in 1896, Henri Becquerel discovered that certain kinds of matter emit radiation on their own accord. In 1897, J. J. Thomson discovered the electron, and new radioactive elements found by Marie and Pierre Curie raised questions about the supposedly indestructible atom and the nature of matter. Marie and Pierre coined the term "radioactivity" to describe this property of matter, and isolated the radioactive elements radium and polonium. Ernest Rutherford and Frederick Soddy identified two of Becquerel's forms of radiation with electrons and the element helium. Rutherford identified and named two types of radioactivity and in 1911 interpreted experimental evidence as showing that the atom consists of a dense, positively charged nucleus surrounded by negatively charged electrons. Classical theory, however, predicted that this structure should be unstable. Classical theory had also failed to explain successfully two other experimental results that appeared in the late 19th century. One of these was the demonstration by Albert A. Michelson and Edward W. Morley – known as the Michelson–Morley experiment – which showed there did not seem to be a preferred frame of reference, at rest with respect to the hypothetical luminiferous ether, for describing electromagnetic phenomena. Studies of radiation and radioactive decay continued to be a preeminent focus for physical and chemical research through the 1930s, when the discovery of nuclear fission by Lise Meitner and Otto Frisch opened the way to the practical exploitation of what came to be called "atomic" energy.
Albert Einstein's theory of relativity
[edit]
In 1905, a 26-year-old German physicist named Albert Einstein (then a patent clerk in Bern, Switzerland) showed how measurements of time and space are affected by motion between an observer and what is being observed. Einstein's radical theory of relativity revolutionized science. Although Einstein made many other important contributions to science, the theory of relativity alone is one of the greatest intellectual achievements of all time. Although the concept of relativity was not introduced by Einstein, he recognised that the speed of light in vacuum is constant, i.e., the same for all observers, and an absolute upper limit to speed. This does not impact a person's day-to-day life since most objects travel at speeds much slower than light speed. For objects travelling near light speed, however, the theory of relativity shows that clocks associated with those objects will run more slowly and that the objects shorten in length according to measurements of an observer on Earth. Einstein also derived the equation, E = mc2, which expresses the equivalence of mass and energy.
Special relativity
[edit]Einstein argued that the speed of light was a constant in all inertial reference frames and that electromagnetic laws should remain valid independent of reference frame – assertions which rendered the ether "superfluous" to physical theory, and that held that observations of time and length varied relative to how the observer was moving with respect to the object being measured (what came to be called the "special theory of relativity"). It also followed that mass and energy were interchangeable quantities according to the equation E=mc2. In another paper published the same year, Einstein asserted that electromagnetic radiation was transmitted in discrete quantities ("quanta"), according to a constant that the theoretical physicist Max Planck had posited in 1900 to arrive at an accurate theory for the distribution of blackbody radiation – an assumption that explained the strange properties of the photoelectric effect.
The special theory of relativity is a formulation of the relationship between physical observations and the concepts of space and time. The theory arose out of contradictions between electromagnetism and Newtonian mechanics and had great impact on both those areas. The original historical issue was whether it was meaningful to discuss the electromagnetic wave-carrying "ether" and motion relative to it and also whether one could detect such motion, as was unsuccessfully attempted in the Michelson–Morley experiment. Einstein demolished these questions and the ether concept in his special theory of relativity. However, his basic formulation does not involve detailed electromagnetic theory. It arises out of the question: "What is time?" Newton, in the Principia (1686), had given an unambiguous answer: "Absolute, true, and mathematical time, of itself, and from its own nature, flows equably without relation to anything external, and by another name is called duration." This definition is basic to all classical physics.
Einstein had the genius to question it, and found that it was incomplete. Instead, each "observer" necessarily makes use of his or her own scale of time, and for two observers in relative motion, their time-scales will differ. This induces a related effect on position measurements. Space and time become intertwined concepts, fundamentally dependent on the observer. Each observer presides over his or her own space-time framework or coordinate system. There being no absolute frame of reference, all observers of given events make different but equally valid (and reconcilable) measurements. What remains absolute is stated in Einstein's relativity postulate: "The basic laws of physics are identical for two observers who have a constant relative velocity with respect to each other."
Special relativity had a profound effect on physics: started as a rethinking of the theory of electromagnetism, it found a new symmetry law of nature, now called Poincaré symmetry, that replaced Galilean symmetry.
Special relativity exerted another long-lasting effect on dynamics. Although initially it was credited with the "unification of mass and energy", it became evident that relativistic dynamics established a distinction between rest mass, which is an invariant (observer independent) property of a particle or system of particles, and the energy and momentum of a system. The latter two are separately conserved in all situations but not invariant with respect to different observers. The term mass in particle physics underwent a semantic change, and since the late 20th century it almost exclusively denotes the rest (or invariant) mass.
General relativity
[edit]
By 1916, Einstein was able to generalize this further, to deal with all states of motion including non-uniform acceleration, which became the general theory of relativity. In this theory, Einstein also specified a new concept, the curvature of space-time, which described the gravitational effect at every point in space. The curvature of space-time replaced Newton's universal law of gravitation. According to Einstein, gravitational force in the normal sense is an illusion caused by the geometry of space. The presence of a mass causes a curvature of space-time in the vicinity of the mass, and this curvature dictates the space-time path that all freely-moving objects follow. It was also predicted from this theory that light should be subject to gravity – all of which was verified experimentally. This aspect of relativity explained the phenomena of light bending around the sun, predicted black holes as well as properties of the Cosmic microwave background radiation – a discovery rendering fundamental anomalies in the classic Steady-State hypothesis. For his work on relativity, the photoelectric effect and blackbody radiation, Einstein received the Nobel Prize in 1921.
The gradual acceptance of Einstein's theories of relativity and the quantized nature of light transmission, and of Niels Bohr's model of the atom created as many problems as they solved, leading to a full-scale effort to reestablish physics on new fundamental principles. Expanding relativity to cases of accelerating reference frames (the "general theory of relativity") in the 1910s, Einstein posited an equivalence between the inertial force of acceleration and the force of gravity, leading to the conclusion that space is curved and finite in size, and the prediction of such phenomena as gravitational lensing and the distortion of time in gravitational fields.
Quantum mechanics
[edit]
Although relativity resolved the electromagnetic phenomena conflict demonstrated by Michelson and Morley, a second theoretical problem was the explanation of the distribution of electromagnetic radiation emitted by a black body; experiment showed that at shorter wavelengths, toward the ultraviolet end of the spectrum, the energy approached zero, but classical theory predicted it should become infinite. This glaring discrepancy, known as the ultraviolet catastrophe, was solved by the new theory of quantum mechanics. Quantum mechanics is the theory of atoms and subatomic systems. Approximately the first 30 years of the 20th century represent the time of the conception and evolution of the theory. The basic ideas of quantum theory were introduced in 1900 by Max Planck (1858–1947), who was awarded the Nobel Prize for Physics in 1918 for his discovery of the quantified nature of energy. The quantum theory (which previously relied in the "correspondence" at large scales between the quantized world of the atom and the continuities of the "classical" world) was accepted when the Compton Effect established that light carries momentum and can scatter off particles, and when Louis de Broglie asserted that matter can be seen as behaving as a wave in much the same way as electromagnetic waves behave like particles (wave–particle duality).
In 1905, Einstein used the quantum theory to explain the photoelectric effect, and in 1913 the Danish physicist Niels Bohr used the same constant to explain the stability of Rutherford's atom as well as the frequencies of light emitted by hydrogen gas. The quantized theory of the atom gave way to a full-scale quantum mechanics in the 1920s. New principles of a "quantum" rather than a "classical" mechanics, formulated in matrix-form by Werner Heisenberg, Max Born, and Pascual Jordan in 1925, were based on the probabilistic relationship between discrete "states" and denied the possibility of causality. Quantum mechanics was extensively developed by Heisenberg, Wolfgang Pauli, Paul Dirac, and Erwin Schrödinger, who established an equivalent theory based on waves in 1926; but Heisenberg's 1927 "uncertainty principle" (indicating the impossibility of precisely and simultaneously measuring position and momentum) and the "Copenhagen interpretation" of quantum mechanics (named after Bohr's home city) continued to deny the possibility of fundamental causality, though opponents such as Einstein would metaphorically assert that "God does not play dice with the universe".[79] The new quantum mechanics became an indispensable tool in the investigation and explanation of phenomena at the atomic level. Also in the 1920s, the Indian scientist Satyendra Nath Bose's work on photons and quantum mechanics provided the foundation for Bose–Einstein statistics, the theory of the Bose–Einstein condensate.
The spin–statistics theorem established that any particle in quantum mechanics may be either a boson (statistically Bose–Einstein) or a fermion (statistically Fermi–Dirac). It was later found that all fundamental bosons transmit forces, such as the photon that transmits electromagnetism.
Fermions are particles "like electrons and nucleons" and are the usual constituents of matter. Fermi–Dirac statistics later found numerous other uses, from astrophysics (see Degenerate matter) to semiconductor design.
Division into classical and modern
[edit]
The conceptual differences between physics theories discussed in the 19th century and those that were most historically prominent in the first decades of the 20th century lead to a characterization of the earlier sciences as "classical physics" while the work based on quantum and relativity theories became known as "modern physics". Initially applied to mechanics, as in "classical mechanics", the divide eventually came to characterize quantum and relativistic effects.[80]: 411 This characterization was driven initially by physicists like Max Planck and Hendrik Lorentz, established scientists who nevertheless saw issues that established theories could not explain. Their involvement and contributions to the 1911 Solvay Conference lead to the introduction of this split as a concept.[81]: 558
This division is reflected in the titles of many physics textbooks. For example, the preface of Goldstein's Classical mechanics explains why the topic is still relevant for physics students.[82] In Concepts of Modern Physics Arthur Beiser starts with a definition of modern physics:[83]
Modern physics began in 1900 with Max Planck’s discovery of the role of energy quantization in blackbody radiation, a revolutionary idea soon followed by Albert Einstein’s equally revolutionary theory of relativity and quantum theory of light.
Kenneth Krane's Modern physics begins a text on quantum and relativity theories with a few pages on deficiencies of classical physics.[84]: 3 E.T. Whittaker's two-volume History of the Theories of Aether and Electricity subtitles volume one The Classical Theories and volume two The Modern Theories (1900–1926).[85]
Contemporary physics
[edit]Quantum field theory
[edit]

As the philosophically inclined continued to debate the fundamental nature of the universe, quantum theories continued to be produced, beginning with Paul Dirac's formulation of a relativistic quantum theory in 1928. However, attempts to quantize electromagnetic theory entirely were stymied throughout the 1930s by theoretical formulations yielding infinite energies. This situation was not considered adequately resolved until after World War II, when Julian Schwinger, Richard Feynman and Sin-Itiro Tomonaga independently posited the technique of renormalization, which allowed for an establishment of a robust quantum electrodynamics (QED).[86]
Meanwhile, new theories of fundamental particles proliferated with the rise of the idea of the quantization of fields through "exchange forces" regulated by an exchange of short-lived "virtual" particles, which were allowed to exist according to the laws governing the uncertainties inherent in the quantum world. Notably, Hideki Yukawa proposed that the positive charges of the nucleus were kept together courtesy of a powerful but short-range force mediated by a particle with a mass between that of the electron and proton. This particle, the "pion", was identified in 1947 as part of what became a slew of particles discovered after World War II. Initially, such particles were found as ionizing radiation left by cosmic rays, but increasingly came to be produced in newer and more powerful particle accelerators.[87]
Outside particle physics, significant advances of the time were:
- the invention of the laser (1964 Nobel Prize in Physics);
- the theoretical and experimental research of superconductivity, especially the invention of a quantum theory of superconductivity by Vitaly Ginzburg and Lev Landau (1962 Nobel Prize in Physics) and, later, its explanation via Cooper pairs (1972 Nobel Prize in Physics). The Cooper pair was an early example of quasiparticles.
Unified field theories
[edit]Einstein deemed that all fundamental interactions in nature can be explained in a single theory. Unified field theories were numerous attempts to "merge" several interactions. One of many formulations of such theories (as well as field theories in general) is a gauge theory, a generalization of the idea of symmetry. Eventually the Standard Model (see below) succeeded in unification of strong, weak, and electromagnetic interactions. All attempts to unify gravitation with something else failed.
Particle physics and the Standard Model
[edit]

When parity was broken in weak interactions by Chien-Shiung Wu in her experiment, a series of discoveries were created thereafter.[89] The interaction of these particles by scattering and decay provided a key to new fundamental quantum theories. Murray Gell-Mann and Yuval Ne'eman brought some order to these new particles by classifying them according to certain qualities, beginning with what Gell-Mann referred to as the "Eightfold Way". While its further development, the quark model, at first seemed inadequate to describe strong nuclear forces, allowing the temporary rise of competing theories such as the S-Matrix, the establishment of quantum chromodynamics in the 1970s finalized a set of fundamental and exchange particles, which allowed for the establishment of a "standard model" based on the mathematics of gauge invariance, which successfully described all forces except for gravitation, and which remains generally accepted within its domain of application.[79]
The Standard Model, based on the Yang–Mills theory[90] groups the electroweak interaction theory and quantum chromodynamics into a structure denoted by the gauge group SU(3)×SU(2)×U(1). The formulation of the unification of the electromagnetic and weak interactions in the standard model is due to Abdus Salam, Steven Weinberg and, subsequently, Sheldon Glashow. Electroweak theory was later confirmed experimentally (by observation of neutral weak currents),[91][92][93][94] and distinguished by the 1979 Nobel Prize in Physics.[95]
Since the 1970s, fundamental particle physics has provided insights into early universe cosmology, particularly the Big Bang theory proposed as a consequence of Einstein's general theory of relativity. However, starting in the 1990s, astronomical observations have also provided new challenges, such as the need for new explanations of galactic stability ("dark matter") and the apparent acceleration in the expansion of the universe ("dark energy").
While accelerators have confirmed most aspects of the Standard Model by detecting expected particle interactions at various collision energies, no theory reconciling general relativity with the Standard Model has yet been found, although supersymmetry and string theory were believed by many theorists to be a promising avenue forward. The Large Hadron Collider, however, which began operating in 2008, has failed to find any evidence that is supportive of supersymmetry and string theory.[96]
Cosmology
[edit]Cosmology may be said to have become a serious research question with the publication of Einstein's General Theory of Relativity in 1915 although it did not enter the scientific mainstream until the period known as the "Golden age of general relativity".
About a decade later, in the midst of what was dubbed the "Great Debate", Edwin Hubble and Vesto Slipher discovered the expansion of universe in the 1920s measuring the redshifts of Doppler spectra from galactic nebulae. Using Einstein's general relativity, Georges Lemaître and George Gamow formulated what would become known as the Big Bang theory. A rival, called the steady state theory, was devised by Fred Hoyle, Thomas Gold, Jayant Narlikar and Hermann Bondi.
Cosmic microwave background radiation was verified in the 1960s by Arno Allan Penzias and Robert Woodrow Wilson, and this discovery favoured the big bang at the expense of the steady state scenario. Later work was by George Smoot et al. (1989), among other contributors, using data from the Cosmic Background explorer (CoBE) and the Wilkinson Microwave Anisotropy Probe (WMAP) satellites refined these observations. The 1980s (the same decade of the COBE measurements) also saw the proposal of inflation theory by Alan Guth.
Recently the problems of dark matter and dark energy have risen to the top of the cosmology agenda.
Higgs boson
[edit]
On July 4, 2012, physicists working at CERN's Large Hadron Collider announced that they had discovered a new subatomic particle greatly resembling the Higgs boson, a potential key to an understanding of why elementary particles have mass and indeed to the existence of diversity and life in the universe.[97] For now, some physicists are calling it a "Higgslike" particle.[97] Joe Incandela, of the University of California, Santa Barbara, said, "It's something that may, in the end, be one of the biggest observations of any new phenomena in our field in the last 30 or 40 years, going way back to the discovery of quarks, for example."[97] Michael Turner, a cosmologist at the University of Chicago and the chairman of the physics center board, said:
"This is a big moment for particle physics and a crossroads – will this be the high water mark or will it be the first of many discoveries that point us toward solving the really big questions that we have posed?"
— Michael Turner, University of Chicago[97]
Peter Higgs was one of six physicists, working in three independent groups, who, in 1964, invented the notion of the Higgs field ("cosmic molasses"). The others were Tom Kibble of Imperial College, London; Carl Hagen of the University of Rochester; Gerald Guralnik of Brown University; and François Englert and Robert Brout, both of Université libre de Bruxelles.[97]
Although they have never been seen, Higgslike fields play an important role in theories of the universe and in string theory. Under certain conditions, according to the strange accounting of Einsteinian physics, they can become suffused with energy that exerts an antigravitational force. Such fields have been proposed as the source of an enormous burst of expansion, known as inflation, early in the universe and, possibly, as the secret of the dark energy that now seems to be accelerating the expansion of the universe.[97]
Physical sciences
[edit]With increased accessibility to and elaboration upon advanced analytical techniques in the 19th century, physics was defined as much, if not more, by those techniques than by the search for universal principles of motion and energy, and the fundamental nature of matter. Fields such as acoustics, geophysics, astrophysics, aerodynamics, plasma physics, low-temperature physics, and solid-state physics joined optics, fluid dynamics, electromagnetism, and mechanics as areas of physical research. In the 20th century, physics also became closely allied with such fields as electrical, aerospace and materials engineering, and physicists began to work in government and industrial laboratories as much as in academic settings. Following World War II, the population of physicists increased dramatically, and came to be centered on the United States, while, in more recent decades, physics has become a more international pursuit than at any time in its previous history.
Articles on the history of physics
[edit]On branches of physics
[edit]- History of astronomy (timeline)
- History of condensed matter (timeline)
- History of computational physics (timeline)
- History of electromagnetic theory (timeline)
- History of geophysics
- History of gravity, spacetime and cosmology
- History of classical mechanics (timeline)
- History of nuclear physics
- History of quantum mechanics (timeline)
- Atomic theory
- History of molecular theory
- History of quantum field theory
- History of quantum information (timeline)
- History of subatomic physics (timeline)
- History of thermodynamics (timeline)
On specific discoveries
[edit]- Discovery of cosmic microwave background radiation
- History of graphene
- First observation of gravitational waves
- Subatomic particles (timeline)
Historical periods
[edit]See also
[edit]Notes
[edit]- ^ Click the image to see further details.
References
[edit]- ^ "This shift from ecclesiastical reasoning to scientific reasoning marked the beginning of scientific methodology." Singer, C., A Short History of Science to the 19th Century, Streeter Press, 2008, p. 35.
- ^ Miyajima, Kazuhiko (1998), "Projection Methods in Chinese, Korean and Japanese Star Maps", Highlights of Astronomy, 11 (2): 712–715, doi:10.1017/s1539299600018554
- ^ Oliver Leaman, Key Concepts in Eastern Philosophy. Routledge, 1999, p. 269.
- ^ Chattopadhyaya 1986, pp. 169–170.
- ^ Choudhury 2006, p. 202
- ^ Stcherbatsky 1962 (1930). Vol. 1. p. 19.
- ^ Li Shu-hua, "Origine de la Boussole 11. Aimant et Boussole", Isis, Vol. 45, No. 2. (Jul., 1954), p.175
- ^ Joseph Needham, Volume 4, Part 1, p. 98.
- ^ Robinson, Francis, ed. (1996). The Cambridge Illustrated History of the Islamic World. Cambridge University Press. pp. 228–229.
- ^ Esposito (2000)، The Oxford History of Islam، Oxford University Press,، p. 192. “Ibn al-Haytham (d. 1039), known in the West as Alhazan, was an Arab mathematician, astronomer, and physicist. His optical compendium, Kitab al-Manazir, is the greatest medieval work on optics”.
- ^ (Child, Shuter & Taylor 1992, p. 70), (Dessel, Nehrich & Voran 1973, p. 164), Understanding History by John Child, Paul Shuter, David Taylor, p. 70. "Alhazen, a Persian scientist, showed that the eye saw light from other objects. This started optics, the science of light. The Arabs also studied astronomy, the study of the stars."
- ^ Glick, Livesey & Wallis (2005, pp. 89–90)
- ^ Smith, Mark (2015). From Sight to Light: The Passage from Ancient to Modern Optics. The University of Chicago Press. p. 225. Bibcode:2014fslp.book.....S.
The same holds for Alhacen's methodology. It may look modern because of its strong empirical bias and reliance on controlled experiments, but Ptolemy's approach was no less empirical, and it, too, was based on controlled experiments. In addition, Alhacen's two most modern-looking experiments are based on physically unobtainable precision in equipment design and observation, so we are left to doubt that he actually carried them out as described— except, of course, in his mind. And these experiments were not new in conception. They were clearly based on equivalent ones in Ptolemy's Optics, although Alhacen had to reformulate them in significant and creative ways to accommodate the testing of light rays rather than visual rays.
- ^ Darrigol, Olivier (2012). A History of Optics from Greek Antiquity to the Nineteenth Century. Oxford University Press. p. 20.
- ^ Lindberg, David; Shank, Michael (2013). The Cambridge History of Science, Volume 2, Medieval Science. pp. 984–1108.
- ^ a b Espinoza, Fernando (2005). "An analysis of the historical development of ideas about motion and its implications for teaching". Physics Education. 40 (2): 141. Bibcode:2005PhyEd..40..139E. doi:10.1088/0031-9120/40/2/002. S2CID 250809354.
- ^ Nasr, Seyyed Hossein; Razavi, Mehdi Amin (1996). The Islamic intellectual tradition in Persia. Routledge. p. 72. ISBN 978-0-7007-0314-2.
- ^ Sayili, Aydin (1987). "Ibn Sīnā and Buridan on the Motion of the Projectile". Annals of the New York Academy of Sciences. 500 (1): 477–482. Bibcode:1987NYASA.500..477S. doi:10.1111/j.1749-6632.1987.tb37219.x. S2CID 84784804.
- ^ Sayili, Aydin. "Ibn Sina and Buridan on the Motion the Projectile". Annals of the New York Academy of Sciences, vol. 500(1). pp. 477–482.
- ^ a b Gutman, Oliver (2003). Pseudo-Avicenna, Liber Celi Et Mundi: A Critical Edition. Brill Publishers. p. 193. ISBN 90-04-13228-7.
- ^ Crombie, Alistair Cameron, Augustine to Galileo 2, p. 67.
- ^ Pines, Shlomo (1970). "Abu'l-Barakāt al-Baghdādī, Hibat Allah". Dictionary of Scientific Biography. Vol. 1. New York: Charles Scribner's Sons. pp. 26–28. ISBN 0-684-10114-9. (cf. Abel B. Franco (October 2003). "Avempace, Projectile Motion, and Impetus Theory", Journal of the History of Ideas 64 (4), pp. 521–546 [528].)
- ^ a b Gracia, Jorge J. E. (2007-11-26), "Philosophy in the Middle Ages: An Introduction", A Companion to Philosophy in the Middle Ages, Blackwell Publishing Ltd, pp. 1–11, doi:10.1002/9780470996669.ch1, ISBN 9780470996669
- ^ Saliba, George (March 1991). Arabic Sciences and Philosophy. Vol. 1. pp. 67–99. doi:10.1017/S0957423900001429. ISSN 0957-4239.
- ^ Crombie, A. C. (1961). "Quantification in Medieval Physics". Isis. 52 (2): 143–160. doi:10.1086/349467. ISSN 0021-1753. JSTOR 228677.
- ^ Lindberg, David C. (1992). The Beginnings of Western Science. University of Chicago Press. doi:10.7208/chicago/9780226482064.001.0001. ISBN 978-0-226-48231-6.
- ^ "John Philoponus, Commentary on Aristotle's Physics". Archived from the original on 11 January 2016. Retrieved 15 April 2018.
- ^ Galileo (1638). Two New Sciences.
in order to better understand just how conclusive Aristotle's demonstration is, we may, in my opinion, deny both of his assumptions. And as to the first, I greatly doubt that Aristotle ever tested by experiment whether it be true that two stones, one weighing ten times as much as the other, if allowed to fall, at the same instant, from a height of, say, 100 cubits, would so differ in speed that when the heavier had reached the ground, the other would not have fallen more than 10 cubits.
Simp. – His language would seem to indicate that he had tried the experiment, because he says: We see the heavier; now the word see shows that he had made the experiment.
Sagr. – But I, Simplicio, who have made the test can assure[107] you that a cannon ball weighing one or two hundred pounds, or even more, will not reach the ground by as much as a span ahead of a musket ball weighing only half a pound, provided both are dropped from a height of 200 cubits. - ^ Lindberg 1992, p. 162.
- ^ "John Philoponus". The Stanford Encyclopedia of Philosophy. Metaphysics Research Lab, Stanford University. 2018. Archived from the original on 22 April 2018. Retrieved 11 April 2018.
- ^ "John Buridan". The Stanford Encyclopedia of Philosophy. Metaphysics Research Lab, Stanford University. 2018. Archived from the original on 22 April 2018. Retrieved 11 April 2018.
- ^ Singer, Charles (1941), "A Short History of Science to the Nineteenth Century", Nature, 148 (3746), Clarendon Press: 178, Bibcode:1941Natur.148..178C, doi:10.1038/148178a0, page 217.
- ^ Weidhorn, Manfred (2005), The Person of the Millennium: The Unique Impact of Galileo on World History, iUniverse, p. 155, ISBN 0-595-36877-8
- ^ Finocchiaro (2007)[broken anchor].
- ^ "Galileo and the Birth of Modern Science". American Heritage's Invention & Technology. 24: 36. 2009. Retrieved 2020-09-15.
- ^ Drake (1978)
- ^ Biagioli (1993)
- ^ Voelkel, James R. (2001). "Commentary on Ernan McMullin, "The Impact of Newton's Principia on the Philosophy of Science"". Philosophy of Science. 68 (3): 319–326. doi:10.1086/392885. ISSN 0031-8248. JSTOR 3080920. S2CID 144781947.
- ^ "DPMA | Johannes Kepler".
- ^ "Johannes Kepler: His Life, His Laws and Times | NASA". Archived from the original on 24 June 2021. Retrieved 1 September 2023.
- ^ "Molecular Expressions: Science, Optics and You – Timeline – Johannes Kepler".
- ^ Barker and Goldstein. "Theological Foundations of Kepler's Astronomy", Osiris, 16, 2001, pp. 112–113.
- ^ Kepler. New Astronomy, title page, tr. Donohue, pp. 26–27
- ^ Kepler. New Astronomy, p. 48
- ^ Epitome of Copernican Astronomy in Great Books of the Western World, Vol. 15, p. 845
- ^ Stephenson. Kepler's Physical Astronomy, pp. 1–2; Dear, Revolutionizing the Sciences, pp. 74–78
- ^ Shea (1991)
- ^ Garber (1992)
- ^ Gaukroger (2002)
- ^ Bell, A. E. (30 Aug 1941). "The Horologium Oscillatorium of Christian Huygens". Nature. 148 (3748): 245–248. Bibcode:1941Natur.148..245B. doi:10.1038/148245a0. S2CID 4112797. Retrieved 14 November 2013.
- ^ Yoder, Joella G. (1988). Unrolling Time: Christiaan Huygens and the Mathematization of Nature. Cambridge: Cambridge University Press. ISBN 978-0-521-34140-0.
- ^ Bruce, I. (2007). Christian Huygens: Horologium Oscillatorium. Translated and annotated by Ian Bruce.
- ^ Dijksterhuis, F. J. (2008) Stevin, Huygens and the Dutch republic. Nieuw archief voor wiskunde, 5, pp. 100–107.[1]
- ^ Andriesse, C. D. (2005) Huygens: The Man Behind the Principle. Cambridge University Press. Cambridge: 6.
- ^ Iliffe, Rob; Smith, George E. (2016). The Cambridge Companion to Newton. Cambridge University Press. p. 75. ISBN 9781107015463.
- ^ Aldersey-Williams, H. (2020). Dutch Light: Christiaan Huygens and the Making of Science in Europe. Pan Macmillan. p. 24. ISBN 978-1-5098-9332-4. Retrieved 28 August 2021.
- ^ Hall (1980)
- ^ Bertolini Meli (1993)
- ^ a b Guicciardini (1999)
- ^ Wilson, Raymond N. (2013). "1.1 Period 1608–1672". Reflecting Telescope Optics I: Basic Design Theory and its Historical Development. Springer. pp. 1–10. ISBN 978-3-662-03227-5.
- ^ Błaszczyk, Piotr; Katz, Mikhail; Sherry, David (2012), "Ten misconceptions from the history of analysis and their debunking", Foundations of Science, 18: 43–74, arXiv:1202.4153, Bibcode:2012arXiv1202.4153B, doi:10.1007/s10699-012-9285-8, S2CID 119134151
- ^ New Experiments physico-mechanicall, Touching the Spring of the Air and its Effects (1660). [2]
- ^ Darrigol (2005)
- ^ Bos (1980)
- ^ Heilbron (1979)
- ^ Buchwald (1989)
- ^ Golinski (1999)
- ^ Greenberg (1986)
- ^ Guicciardini (1989)
- ^ Garber (1999)
- ^ Ben-Chaim (2004)
- ^ Buchwald (1985)
- ^ JungnickelMcCormmach (1986)
- ^ Hunt (1991)
- ^ Buchwald (1994)
- ^ Smith & Wise (1989)
- ^ Smith (1998)
- ^ Agar (2012)
- ^ a b Kragh (1999)
- ^ Kragh, Helge (2015). "The "new physics"". In Saler, Michael T. (ed.). The fin-de-siècle world. The Routledge worlds. London: Routledge. ISBN 978-0-415-67413-3.
- ^ Staley, Richard (December 2005). "On the Co-Creation of Classical and Modern Physics". Isis. 96 (4): 530–558. doi:10.1086/498592. ISSN 0021-1753. PMID 16536154.
- ^ Goldstein, Herbert (1980). Classical mechanics. Addison-Wesley series in physics (2 ed.). Reading, Mass: Addison-Wesley Pub. Co. ISBN 978-0-201-02918-5.
- ^ Beiser, Arthur (2003). Concepts of modern physics (International ed., 6. ed.). Boston: McGraw-Hill. ISBN 978-0-07-244848-1.
- ^ Krane, Kenneth S. (2020). Modern physics (4 ed.). Hoboken, New Jersey: John Wiley & Sons, Inc. ISBN 978-1-119-49548-2.
- ^ Whittaker, Edmund T. (1989). A history of the theories of aether & electricity. 2: The modern theories, 1900 - 1926 (Repr ed.). New York: Dover Publ. ISBN 978-0-486-26126-3.
- ^ Schweber (1994)
- ^ Galison (1997)
- ^ "Chien-Shiung Wu". Encyclopedia Britannica. 25 May 2023.
- ^ "Antimatter". 2021-03-01.
- ^ "theory of everything".
- ^ Hasert, F. J.; Faissner, H.; Krenz, W.; Von Krogh, J.; Lanske, D.; Morfin, J.; Schultze, K.; Weerts, H.; Bertrand-Coremans, G. H.; Lemonne, J.; Sacton, J. (1973-09-03). "Search for elastic muon-neutrino electron scattering". Physics Letters B. 46 (1): 121–124. Bibcode:1973PhLB...46..121H. doi:10.1016/0370-2693(73)90494-2. ISSN 0370-2693.
- ^ Hasert, F. J.; Kabe, S.; Krenz, W.; Von Krogh, J.; Lanske, D.; Morfin, J.; Schultze, K.; Weerts, H.; Bertrand-Coremans, G. H.; Sacton, J.; Van Doninck, W. (1973-09-03). "Observation of neutrino-like interactions without muon or electron in the gargamelle neutrino experiment". Physics Letters B. 46 (1): 138–140. Bibcode:1973PhLB...46..138H. doi:10.1016/0370-2693(73)90499-1. ISSN 0370-2693.
- ^ F. J. Hasert et al. Nuclear Physics B73, 1(1974); Paper presented at the London Conference 1974, no. 1013.
- ^ The discovery of the weak neutral currents, CERN courier, 2004-10-04, retrieved 2008-05-08
- ^ The Nobel Prize in Physics 1979, Nobel Foundation, archived from the original on 2004-08-03, retrieved 2008-09-10
- ^ Woit, Peter (20 October 2013). "Last Links For a While". Not Even Wrong. Retrieved 2 November 2013.
- ^ a b c d e f Overbye, Dennis (4 July 2012). "Physicists Find Particle That Could Be the Higgs Boson". The New York Times.
Sources
[edit]- Agar, Jon (2012), Science in the Twentieth Century and Beyond, Cambridge: Polity Press, ISBN 978-0-7456-3469-2.
- Hardie, R. P.; Gaye, R. K. (2014-09-01), "PHYSICS", Complete Works of Aristotle, Volume 1, Princeton University Press, pp. 315–446, doi:10.2307/j.ctt5vjv4w.12, ISBN 978-1-4008-3584-3
{{citation}}: CS1 maint: work parameter with ISBN (link) - Ben-Chaim, Michael (2004), Experimental Philosophy and the Birth of Empirical Science: Boyle, Locke and Newton, Aldershot: Ashgate, ISBN 0-7546-4091-4, OCLC 53887772.
- Bertolini Meli, Domenico (1993), Equivalence and Priority: Newton versus Leibniz, New York: Oxford University Press.
- Biagioli, Mario (1993), Galileo, Courtier: The Practice of Science in the Culture of Absolutism, Chicago, Illinois: University of Chicago Press, ISBN 0-226-04559-5, OCLC 185632037.
- Bos, Henk (1980), "Mathematics and Rational Mechanics", in Rousseau, G. S.; Porter, Roy (eds.), The Ferment of Knowledge: Studies in the Historiography of Eighteenth Century Science, vol. 25, New York: Cambridge University Press, pp. 323–324, PMC 1139043.
- Buchwald, Jed (1985), From Maxwell to Microphysics: Aspects of Electromagnetic Theory in the Last Quarter of the Nineteenth Century, Chicago, Illinois: University of Chicago Press, ISBN 0-226-07882-5, OCLC 11916470.
- Buchwald, Jed (1989), The Rise of the Wave Theory of Light: Optical Theory and Experiment in the Early Nineteenth Century, Chicago, Illinois: University of Chicago Press, ISBN 0-226-07886-8, OCLC 18069573.
- Buchwald, Jed (1994), The Creation of Scientific Effects: Heinrich Hertz and Electric Waves, Chicago, Illinois: University of Chicago Press, ISBN 0-226-07888-4, OCLC 29256963.
- Chattopadhyaya, Debiprasad (1986). History of science and technology in ancient India : the beginnings. Firma KLM Pvt. Limited. ISBN 81-7102-053-4. OCLC 45345319.
- Child, John; Shuter, Paul; Taylor, David (1992). Understanding history. Oxford, England: Heinemann Educational. ISBN 0435312111. OCLC 27338645.
- Darrigol, Olivier (2005), Worlds of Flow: A History of Hydrodynamics from the Bernoullis to Prandtl, New York: Oxford University Press, ISBN 0-19-856843-6, OCLC 237027708
- Dessel, Norman F.; Nehrich, Richard B.; Voran, Glenn I. (1973). Science and human destiny. New York: McGraw-Hill. ISBN 9780070165809.
- Drake, Stillman (1978), Galileo at Work: His Scientific Biography, Chicago, Illinois: University of Chicago Press, ISBN 0-226-16226-5, OCLC 185633608.
- Galison, Peter (1997), Image and Logic: A Material Culture of Microphysics, Chicago, Illinois: University of Chicago Press, ISBN 0-226-27917-0, OCLC 174870621.
- Garber, Daniel (1992), Descartes' Metaphysical Physics, Chicago, Illinois: University of Chicago Press.
- Garber, Elizabeth (1999), The Language of Physics: The Calculus and the Development of Theoretical Physics in Europe, 1750–1914, Boston, Massachusetts: Birkhäuser Verlag.
- Gaukroger, Stephen (2002), Descartes' System of Natural Philosophy, New York: Cambridge University Press.
- Glick, Thomas F.; Livesey, Steven John; Wallis, Faith (2005), Medieval Science, Technology, and Medicine: An Encyclopedia, Routledge, ISBN 0-415-96930-1, OCLC 218847614
- Greenberg, John (1986), "Mathematical Physics in Eighteenth-Century France", Isis, 77: 59–78, doi:10.1086/354039, S2CID 144216491.
- Golinski, Jan (1999), "Science as Public Culture: Chemistry and Enlightenment in Britain, 1760–1820", Medical History, 37 (4), New York: Cambridge University Press: 468–469, doi:10.1017/S002572730005897X, PMC 1036800.
- Guicciardini, Niccolò (1989), The Development of Newtonian Calculus in Britain, 1700–1800, New York: Cambridge University Press.
- Guicciardini, Niccolò (1999), Reading the Principia: The Debate on Newton's Methods for Natural Philosophy from 1687 to 1736, New York: Cambridge University Press.
- Hall, A. Rupert (1980), Philosophers at War: The Quarrel between Newton and Leibniz, New York: Cambridge University Press.
- Heilbron, J. L. (1979), Electricity in the 17th and 18th Centuries, Berkeley, California: University of California Press.
- Hunt, Bruce (1991), The Maxwellians, Ithaca, New York: Cornell University Press.
- Jungnickel, Christa; McCormmach, Russell (1986). Intellectual Mastery of Nature: Theoretical Physics from Ohm to Einstein. Chicago, Illinois: University of Chicago Press..
- Kragh, Helge (1999), Quantum Generations: A History of Physics in the Twentieth Century, Princeton, New Jersey: Princeton University Press.
- Choudhury, Sarojakanta (2006). Educational philosophy of Dr. Sarvepalli Radhakrishnan. Deep & Deep Publications. ISBN 81-7629-766-6. OCLC 224913142.
- Rashed, R.; Morelon, Régis (1996), Encyclopedia of the History of Arabic Science, vol. 2, Routledge, ISBN 0-415-12410-7, OCLC 34731151.
- Schweber, Silvan (1994), QED and the Men Who Made It: Dyson, Feynman, Schwinger, and Tomonaga, Princeton, New Jersey: Princeton University Press.
- Shea, William (1991), The Magic of Numbers and Motion: The Scientific Career of René Descartes, Canton, Massachusetts: Science History Publications.
- Smith, Crosbie (1998), The Science of Energy: A Cultural History of Energy Physics in Victorian Britain, Chicago, Illinois: University of Chicago Press.
- Smith, Crosbie; Wise, M. Norton (1989), Energy and Empire: A Biographical Study of Lord Kelvin, New York: Cambridge University Press.
Further reading
[edit]- Buchwald, Jed Z. and Robert Fox, eds. The Oxford Handbook of the History of Physics (2014) 976 pp.; excerpt.
- Byers, Nina; Williams, Gary (2006). Out of the Shadows: Contributions of Twentieth-Century Women to Physics. Cambridge University Press. ISBN 0-521-82197-5.
- Cropper, William H. (2004). Great Physicists: The Life and Times of Leading Physicists from Galileo to Hawking. Oxford University Press. ISBN 0-19-517324-4.
- Dear, Peter (2001). Revolutionizing the Sciences: European Knowledge and Its Ambitions, 1500–1700. Princeton, New Jersey: Princeton University Press. ISBN 0-691-08859-4. OCLC 46622656.
- Gamow, George (1988). The Great Physicists from Galileo to Einstein. Dover Publications. ISBN 0-486-25767-3.
- Heilbron, John L. (2005). The Oxford Guide to the History of Physics and Astronomy. Oxford University Press. ISBN 0-19-517198-5.
- Nye, Mary Jo (1996). Before Big Science: The Pursuit of Modern Chemistry and Physics, 1800–1940. New York: Twayne. ISBN 0-8057-9512-X. OCLC 185866968.
- Segrè, Emilio (1984). From Falling Bodies to Radio Waves: Classical Physicists and Their Discoveries. New York: W. H. Freeman. ISBN 0-7167-1482-5. OCLC 9943504.
- Segrè, Emilio (1980). From X-Rays to Quarks: Modern Physicists and Their Discoveries. San Francisco, California: W. H. Freeman. ISBN 0-7167-1147-8. OCLC 237246197.
- Simonyi, Károly (1978). A Cultural History of Physics. Boca Raton, Florida: A K Peters/CRC Press (published 2012). doi:10.1201/9781032697697. ISBN 978-1568813295.
- Weaver, Jefferson H., ed. (1987). The World of Physics. Simon and Schuster. ISBN 0-671-49931-9. A selection of 56 articles, written by physicists. Commentaries and notes by Lloyd Motz and Dale McAdoo.
- de Haas, Paul, "Historic Papers in Physics (20th Century)".
External links
[edit]History of physics
View on GrokipediaAncient physics
Near Eastern and Egyptian foundations
The earliest foundations of physics in the Near East and Egypt emerged from practical observations and empirical techniques aimed at understanding and harnessing natural phenomena, particularly in astronomy, engineering, and measurement. In Mesopotamia, particularly among the Babylonians, astronomy developed as a tool for timekeeping and prediction, rooted in systematic observations recorded on clay tablets dating back to around 2000 BCE. These tablets, such as those from the city of Babylon, document celestial events including planetary positions and lunar phases, enabling the creation of predictive models. Babylonian astronomers employed a sexagesimal (base-60) numerical system, inherited from the Sumerians around 3000 BCE, which facilitated precise calculations for angles and time divisions still used today in degrees and minutes.[5][6][7] A key achievement was the development of lunar calendars, which reconciled the lunar month of approximately 29.5 days with the solar year through intercalary months added periodically to maintain seasonal alignment. This lunisolar system supported agricultural planning and religious festivals. By the 8th century BCE, Babylonians identified the Saros cycle, a period of about 18 years (223 synodic months) during which lunar and solar eclipses recur in a predictable pattern, with systematic records beginning around 747 BCE. These predictive models, derived from centuries of observations, allowed forecasts of eclipse timings with accuracies of 1-2 hours by the 3rd century BCE. Mesopotamian cosmology envisioned a flat Earth disk floating on primordial waters, enclosed by a solid dome-like heaven supported by pillars or mountains, where celestial bodies moved along fixed paths. Early concepts of balance and motion appeared in practical applications, such as standardized weights and measures for trade, where equilibrium in scales reflected notions of proportional forces and stability in weighing systems established by the 3rd millennium BCE.[8][9][10] In ancient Egypt, physical knowledge advanced through engineering feats and hydraulic management, driven by the Nile's annual floods. Imhotep, serving Pharaoh Djoser around 2650 BCE, is credited as the first named architect and engineer, designing the Step Pyramid at Saqqara using stacked mastabas of limestone blocks, innovating stone construction techniques like ramps and levers to achieve heights over 200 feet. Pyramid building later evolved with precise alignment to cardinal directions, employing basic surveying tools for leveling and orientation. For flood prediction, Egyptians used nilometers—graduated stone markers or wells along the Nile—to measure rising water levels, correlating them with historical data to anticipate inundation heights and durations from as early as the Old Kingdom (ca. 2686–2181 BCE). This empirical hydrology ensured agricultural fertility by guiding irrigation basin designs. Mathematical texts like the Rhind Papyrus (ca. 1650 BCE) demonstrate practical geometry for land measurement, including methods to calculate areas of triangles and circles using approximations (e.g., circle area as (8/9 diameter)^2), applied to re-survey fields after floods. These techniques emphasized utility over abstraction, laying groundwork for later scientific traditions including Greek astronomy.[11][12][13][14]Greek natural philosophy
Greek natural philosophy marked a pivotal transition from mythological explanations of the natural world to rational inquiry, influenced briefly by Near Eastern astronomical observations that encouraged systematic speculation about cosmic order.[15] This shift began with the Pre-Socratic philosophers in the 6th century BCE, who sought underlying principles (archai) to explain the origins and transformations of all things without invoking divine intervention. Thales of Miletus (ca. 585 BCE) proposed water as the primary substance from which all matter arises and to which it returns, viewing it as the source of nourishment and change in the cosmos.[15] Anaximander, his successor, introduced the apeiron—an indefinite, eternal, and boundless substance—as the origin of opposites like hot and cold, generating the ordered universe through processes of separation and return governed by justice.[15] Heraclitus emphasized flux and constant change, positing fire as the fundamental element symbolizing transformation, with the logos (rational principle) underlying the unity of opposites in a dynamic cosmos.[15] Philosophical debates intensified around the nature of reality and change, exemplified by Parmenides' argument for immutability, where true being is eternal, indivisible, and unchanging, rendering sensory perceptions of motion and alteration illusory.[16] In contrast, the Pythagoreans integrated numerology into natural philosophy, discovering that musical harmonics arise from simple whole-number ratios—such as 2:1 for the octave and 3:2 for the fifth—revealing numbers as the essence of cosmic harmony and order.[17] Addressing Parmenides' denial of change, atomism emerged with Leucippus and Democritus (ca. 400 BCE), who theorized that the universe consists of indivisible atoms moving in a void, differing only in shape, size, and arrangement; this discrete model explained multiplicity and motion without infinite divisibility, distinguishing it from continuous substances like those in earlier elemental theories.[18] Aristotle (ca. 350 BCE) synthesized these ideas into a comprehensive system, positing four elements—earth, water, air, and fire—each seeking its natural place (earth at the center, fire at the periphery) due to inherent tendencies toward rest and order.[19] His physics incorporated teleology, where natural motions and changes serve purposeful ends, and explained projectile motion through antiperistasis, the medium's circular displacement sustaining the object's path after the initial impetus.[19] Archimedes (ca. 250 BCE) advanced practical applications, formulating the buoyancy principle in hydrostatics—that a submerged body experiences an upward force equal to the weight of displaced fluid—and the law of the lever, stating equilibrium when moments (force times distance from fulcrum) balance, laying foundations for statics and mechanics.[20]Indian and Chinese traditions
The Vaisheshika school of Indian philosophy, founded by Kanada around 600 BCE, developed an early form of atomism positing that the universe consists of indivisible particles called paramāṇu (atoms), which combine to form larger substances through motion and inherent qualities.[21] This system categorized reality into six padārthas (categories), including dravya (substance, such as the four elements and atoms) and karma (motion or action), explaining physical phenomena like combination and separation without invoking a void, in contrast to parallel developments in Greek thought.[22] These ideas integrated mechanics with metaphysics, viewing atoms as eternal and motion as a fundamental attribute driving cosmic processes. In the 5th century CE, astronomer Aryabhata advanced cosmological models in his Aryabhatiya (499 CE), incorporating elements suggestive of heliocentrism by treating planetary motions relative to the Sun while asserting the Earth's axial rotation to account for the apparent daily motion of celestial bodies.[23] This rotation model, where the Earth spins like a potter's wheel, challenged prevailing geocentric views and enabled precise calculations of eclipses and planetary positions, laying groundwork for later Indian astronomy.[24] The Sulba Sutras, Vedic texts dating to around 800 BCE, applied geometric principles to construct sacrificial altars, employing base-10 (decimal) numbering for measurements and approximations like the Pythagorean theorem to ensure precise proportions, such as √2 ≈ 1.4142135 for diagonal calculations.[25] These works marked early advancements in applied mathematics, using decimal systems for scaling altar designs from squares to circles without explicit zero notation, though zero as a placeholder emerged later in Indian numeral systems. Indian cosmology emphasized vast cyclical time scales, with yugas (epochs) forming a mahayuga of 4.32 million years and a kalpa (day of Brahma) spanning 4.32 billion years, integrating notions of creation, preservation, and dissolution influenced by atomic combinations and motions.[26] This framework viewed physical changes as recurring patterns tied to ethical and cosmic order, contrasting linear Western timelines. The Surya Siddhanta, an influential astronomical treatise from around the 4th–5th century CE, provided detailed calculations for solar, lunar, and planetary positions, including sine tables and eclipse predictions based on epicycle models, serving as a practical handbook for centuries.[27] In ancient China, the Mohist school, led by Mozi around 400 BCE, explored optics through experiments on light reflection in mirrors and the pinhole camera principle, while also analyzing levers and pulleys in the Mozi text to explain mechanical equilibrium and force distribution.[28] These investigations linked sensory perception with practical engineering, emphasizing empirical observation in philosophy. Chinese correlative cosmology revolved around wuxing (five phases or elements)—wood, fire, earth, metal, and water—which interacted through cycles of generation and conquest to explain natural transformations, including seasonal changes and material properties.[29] This system influenced early understandings of motion and balance in the physical world, viewing forces as interdependent rather than isolated. By around 200 BCE, Chinese scholars utilized lodestone (magnetite) devices for divination, marking the earliest known magnetic compasses that aligned with Earth's field, predating navigational uses.[30] The yin-yang duality, formalized in texts like the I Ching by the Warring States period (ca. 400 BCE), conceptualized cosmic balance as complementary opposites driving change, with implications for motion (yang as active force) and stability (yin as receptive), shaping views of natural forces.[31] In 1088 CE, polymath Shen Kuo documented magnetic declination in his Dream Pool Essays, noting that compass needles deviated from true north by about 4 degrees, providing the first recorded observation of this geomagnetic variation and advancing navigational science.[30] Zhang Heng's seismoscope, invented in 132 CE, detected distant earthquakes up to 500 km away using a bronze vessel with dragon heads and toad mouths to indicate direction via dropped balls, relying on inertial mechanics.[32]Islamic synthesis
During the Islamic Golden Age, scholars synthesized Greek natural philosophy and Indian mathematical traditions, advancing physics through empirical observation, mathematical rigor, and innovative instrumentation. This period, spanning roughly the 8th to 13th centuries, saw the preservation of ancient knowledge alongside original contributions that emphasized experimentation and precise measurement. Muhammad ibn Musa al-Khwarizmi (c. 780–850 CE) developed systematic algebraic methods in his treatise Al-Kitab al-Mukhtasar fi Hisab al-Jabr wal-Muqabala (c. 820 CE), which provided tools for solving equations applicable to mechanical problems, such as those in surveying and engineering. His work laid foundational techniques for later applications in physics, including calculations of forces and motions in practical devices. Building on Indian arithmetic, al-Khwarizmi promoted the adoption of Hindu-Arabic numerals (0–9) across the Islamic world, facilitating complex computations essential for astronomical and physical analyses.[33][34] In optics, Ibn al-Haytham (Alhazen, 965–1040 CE) revolutionized the field with his Kitab al-Manazir (Book of Optics, 1021 CE), where he experimentally demonstrated the camera obscura to explain image formation and formulated early laws of refraction through controlled tests on light passing between media. His emphasis on hypothesis testing and quantitative verification marked a shift toward modern scientific methodology in physics.[35][36] Commentaries on Aristotle's natural philosophy further refined concepts of motion and the soul-body relation. Avicenna (Ibn Sina, 980–1037 CE) critiqued Aristotelian views in his Al-Shifa (The Book of Healing), introducing the "floating man" thought experiment to argue for the soul's independence from sensory input and bodily perception, influencing later discussions on self-motion and inertia. Averroes (Ibn Rushd, 1126–1198 CE), in his extensive commentaries on Aristotle's Physics, challenged aspects of projectile motion by questioning the role of the medium as a continuous mover, proposing instead that an impressed force sustains the object's path, prefiguring impetus theory.[37][38] Astronomical advancements included refinements to the astrolabe by Al-Biruni (973–1055 CE), who around 1000 CE enhanced its design for precise angular measurements, enabling him to calculate Earth's circumference at approximately 39,375 km using trigonometric methods and observations from a mountain peak. His work integrated geometry and empirical data to determine planetary positions and geodetic properties.[39] Early experimental physics emerged in mechanical engineering, as seen in Ismail al-Jazari's (1136–1206 CE) The Book of Knowledge of Ingenious Mechanical Devices (1206 CE), which detailed automata and water clocks using crankshafts, cams, and floats to demonstrate principles of force, balance, and periodic motion. Concepts of impetus—an internal force propelling objects—were explored by philosophers like Avempace (Ibn Bajjah, d. 1138 CE) and Avicenna, serving as precursors to Jean Buridan's later formulation in Europe.[40][41] Specific advances extended to medical physics, where hospitals (bimaristans) incorporated precise balances and scales for weighing pharmaceuticals, applying principles of equilibrium and leverage to ensure accurate dosing in treatments. These institutions, such as the 9th-century Baghdad hospital, integrated physical instrumentation with clinical practice, enhancing empirical medicine.[42]Medieval and Renaissance transitions
European scholasticism
European scholasticism emerged as a dominant intellectual framework in medieval universities, integrating Aristotelian natural philosophy with Christian doctrine to explain the physical world within a theological context. The scholastic method, characterized by dialectical reasoning and quaestiones disputatae, sought to reconcile pagan philosophy with revealed truth, emphasizing qualitative descriptions over quantitative analysis. This approach framed physics—understood as natural philosophy—as a branch of divine wisdom, where natural causes operated under God's ultimate governance. Influences from translated Islamic texts, such as those by Avicenna and Averroes, facilitated the adoption of Aristotelian concepts in Europe during the 12th and 13th centuries.[43] Thomas Aquinas (c. 1225–1274) exemplified this synthesis in his Summa Theologica, where he harmonized Aristotle's physics with Christian theology around 1270. Aquinas argued that natural theology could demonstrate God's existence through observable motions and causes in the physical universe, positing an unmoved mover as the prime cause of all change. In his treatment of natural philosophy, he adapted Aristotelian notions of motion, elements, and celestial incorruptibility to affirm creation's order as evidence of divine providence, rejecting any conflict between faith and reason. This framework dominated university curricula, positioning physics as a tool for understanding God's rational design rather than empirical experimentation.[44][45] A key advancement in scholastic mechanics was the theory of impetus, refined by Jean Buridan (c. 1295–1363) around 1340 at the University of Paris. Buridan proposed that a projectile continues moving due to an "impetus" impressed upon it by the initial force, rather than requiring continuous propulsion from the air or another agent, as Aristotle had suggested. This qualitative explanation addressed projectile motion by attributing sustained velocity to an internal motive quality that gradually diminishes due to resistance and gravity, marking a shift toward internalized causes of motion while remaining compatible with theological views of divine conservation. Buridan's ideas influenced later thinkers, bridging Aristotelian dynamics with emerging kinematic insights.[46] At Oxford's Merton College, the "Oxford Calculators" in the mid-14th century advanced quantitative reasoning in natural philosophy, building on nominalist principles associated with William of Ockham (c. 1287–1347). Ockham's nominalism emphasized observable particulars over universal essences, encouraging precise analysis of motion without metaphysical excess. Around 1330, scholars like William Heytesbury formulated the mean speed theorem, stating that for uniformly accelerated motion, the distance traveled equals that of uniform motion at the average of initial and final velocities. This theorem, derived through logical and geometric arguments, provided a rule for calculating distances in accelerated fall—such as a body reaching the midpoint speed covering the same distance as constant motion at half the final speed—laying groundwork for later kinematics while adhering to scholastic qualitative constraints.[47] Scholastic cosmology retained the geocentric Ptolemaic model, adapted to Christian doctrine, envisioning the universe as a series of concentric celestial spheres carrying planets and stars in uniform circular motion. Earth, composed of the four elements, occupied the immutable center, surrounded by ethereal spheres driven by intelligences or God's will, ensuring the heavens' perfection and incorruptibility. This hierarchical structure, detailed in commentaries on Aristotle's De Caelo, underscored divine order, with sublunary changes contrasting celestial eternity, and was taught as integral to natural philosophy.[48] The University of Paris served as a central hub for scholastic natural philosophy from the early 13th century, where curricula mandated Aristotelian texts like Physics and De Caelo alongside theological integration. Faculties structured teaching around lectures, disputations, and quaestiones on topics such as motion, elements, and cosmology, fostering debates that refined impetus and acceleration concepts. Paris's arts faculty emphasized logical analysis over observation, training generations in a physics subordinated to theology, which spread across European universities.[48]Revival of ancient texts
The fall of Constantinople in 1453 marked a pivotal event in the revival of ancient Greek texts, as Byzantine scholars fled to Italy with invaluable manuscripts preserved in their libraries. This exodus accelerated the flow of classical knowledge into Western Europe, introducing works on mathematics, astronomy, and natural philosophy that had been largely inaccessible during the medieval period. Scholars such as Cardinal Bessarion and John Argyropoulos settled in Italian cities like Florence and Venice, where they taught Greek and facilitated translations, bridging the gap between Byzantine scholarship and Renaissance humanism. Building on the foundations of medieval scholasticism, this influx enriched European intellectual life by providing direct access to original sources rather than Latin translations filtered through Arabic intermediaries.[49][50] A key figure in this translation movement was Marsilio Ficino, who in 1484 published the first complete Latin translation of Plato's extant works, including dialogues that explored natural philosophy and cosmology. Sponsored by the Medici family in Florence, Ficino's efforts at the Platonic Academy emphasized the harmony between Platonic thought and Christian theology, inspiring a broader humanist engagement with ancient ideas. This work not only revived interest in Plato's concepts of the universe but also encouraged the study of related scientific texts, laying groundwork for empirical inquiries into nature.[51] The invention of Johannes Gutenberg's printing press around 1455 further amplified this revival by enabling the rapid and widespread dissemination of translated and edited ancient texts. While initially used for the Gutenberg Bible, the press soon produced editions of scientific works, such as those by Archimedes—whose mathematical treatises on levers and hydrostatics were printed in Latin in 1544—and Ptolemy, whose astronomical systems appeared in printed forms like Regiomontanus's 1496 Epitome of the Almagest. This technological innovation reduced costs and errors in copying, allowing scholars across Europe to access reliable versions of these foundational texts, which spurred advancements in geometry and celestial modeling.[52][53] Among the key figures benefiting from this environment was Johannes Regiomontanus, whose 1474 publication of Ephemerides provided astronomical tables calculated using advanced trigonometry, drawing on rediscovered Greek methods. Regiomontanus's earlier De Triangulis (1464) systematized plane and spherical trigonometry, treating it as an independent discipline essential for astronomy, and his printed works exemplified how revived texts fueled mathematical progress. Similarly, the rediscovery of Hero of Alexandria's Pneumatica in Renaissance Italy, through Byzantine manuscripts translated and printed in the late 16th century, reintroduced pneumatic devices and automata, inspiring engineers to explore principles of pressure and mechanics described in the original Greek. These recoveries highlighted the Renaissance shift toward studying ancient empirical descriptions as precursors to modern experimentation.[54][55]Pre-revolutionary experiments
The late Renaissance period marked a shift toward empirical investigation in physics, as scholars and artisans increasingly relied on hands-on experimentation to explore natural phenomena, laying groundwork for the Scientific Revolution. These efforts, often conducted outside formal academic institutions, emphasized practical mechanics, instrumentation, and observations that challenged Aristotelian traditions. Inspired briefly by the revival of ancient texts such as those of Archimedes, experimenters focused on tangible demonstrations rather than abstract philosophy. Leonardo da Vinci (1452–1519) exemplified early mechanical studies through his detailed notebooks, where he analyzed friction as a force opposing motion, noting its dependence on load but independence from the apparent contact area in sketches of sliding blocks and wheels around 1500 CE. He also examined gears and levers in designs for machines like cranes and mills, applying principles of mechanical advantage to optimize force transmission. Da Vinci's investigations into flight involved aerodynamic concepts, such as the role of air resistance in bird wings and ornithopter mechanisms, based on observations of natural motion. Additionally, his anatomical dissections integrated physics with biology, exploring levers in muscle-skeleton systems to understand human movement as a mechanical process.[56][57][58] In the realm of magnetism and related forces, English physician William Gilbert published De Magnete in 1600, presenting systematic experiments that established magnetism as a distinct natural property of the Earth, which he modeled as a giant lodestone. Gilbert's terrella—a spherical magnet—allowed him to replicate compass behavior and demonstrate magnetic poles, distinguishing it from electrical attraction observed in amber (which he termed "electric" from the Greek elektron). His work rejected alchemical mysticism, emphasizing quantitative observations like dip angles to argue for terrestrial magnetism's role in navigation.[59][60][61] Optical innovations emerged with the telescope's invention in 1608 by Dutch spectacle-maker Hans Lippershey, who patented a refracting device using convex and concave lenses to magnify distant objects up to three times. Initial observations with this instrument, conducted by Lippershey and contemporaries like Jacob Metius, revealed enhanced views of landscapes and ships at sea, prompting further refinements in lens grinding among artisan opticians. These early uses demonstrated the potential for remote sensing, influencing subsequent astronomical applications without immediate theoretical frameworks.[62][63] Galileo Galilei contributed key devices in the late 16th century, including the thermoscope around 1593, an open-tube apparatus filled with water and air that expanded or contracted with temperature changes, providing qualitative indications of heat variations. His pendulum studies, initiated after observing chandeliers in Pisa Cathedral around 1581 and refined by 1602, revealed the isochronous property—where swing periods remain nearly constant regardless of amplitude for small angles—through timed experiments with bobs of varying lengths. These investigations highlighted periodicity in oscillatory motion, applicable to timekeeping and mechanics.[64][65] Artisan contributions extended empirical physics into practical domains like mining and warfare. In mining, Italian metallurgist Vannoccio Biringuccio's De la pirotechnia (1540) detailed advancements in ore extraction, including gunpowder blasting and water-powered pumps, which improved efficiency in deep shafts and addressed hydrostatic pressures. These techniques, developed by guild craftsmen in regions like Tuscany, integrated mechanical principles to manage subterranean forces. In ballistics, Florentine artisans and engineers produced cannons and studied trajectories during conflicts, employing empirical trials to calibrate ranges and elevations, as seen in the production of artillery munitions that transformed siege warfare. Such hands-on innovations by non-academic practitioners underscored the interplay between technology and physical observation.[66][67]Scientific Revolution
Heliocentric astronomy
The heliocentric model marked a pivotal shift in astronomical thought during the early Scientific Revolution, proposing the Sun as the center of the planetary system rather than Earth. In 1543, Nicolaus Copernicus published De revolutionibus orbium coelestium, articulating a heliocentric hypothesis where Earth and other planets orbit the Sun, thereby simplifying celestial mechanics by eliminating the Ptolemaic equant—a device that allowed non-uniform motion in epicycle models.[68] This approach explained retrograde motion not as actual planetary reversals but as an optical illusion arising from Earth's faster orbital speed overtaking slower outer planets.[68] Copernicus retained circular orbits with epicycles, yet his system reduced the complexity of geocentric frameworks and aligned planetary periods with their distances from the Sun.[69] Building on Copernicus's ideas, Georg Joachim Rheticus, a Lutheran mathematician, played a crucial role in disseminating the theory through his Narratio prima (1540), an introductory abstract that outlined the heliocentric framework and urged publication of Copernicus's full work.[70] However, the model faced significant opposition from the Catholic Church and Protestant reformers, who viewed it as conflicting with literal interpretations of Scripture, such as passages in Psalms and Joshua implying a stationary Earth and moving Sun.[71] Figures like Martin Luther criticized the theory in 1539 as contrary to biblical accounts, while John Calvin expressed reservations in sermons, though evidence of direct condemnation remains debated.[71] This resistance delayed widespread acceptance, framing heliocentrism initially as a mathematical hypothesis rather than physical reality.[72] Tycho Brahe's meticulous observations from the 1570s to 1590s provided the empirical foundation needed to refine heliocentric models, conducted at his Uraniborg observatory on the island of Hven without telescopes, achieving unprecedented precision through large, fixed instruments like mural quadrants and sextants.[73] These naked-eye measurements, accurate to within arcminutes, cataloged planetary positions—particularly Mars—over two decades, surpassing prior data in detail and reliability.[74] Brahe's work, though adhering to a geo-heliocentric system, supplied the high-quality dataset essential for subsequent advancements.[73] Johannes Kepler, utilizing Brahe's observations, formulated three laws of planetary motion that solidified the heliocentric paradigm. In 1609, Kepler's first law stated that planets orbit the Sun in ellipses with the Sun at one focus, departing from circular assumptions and fitting Mars's path precisely.[75] His second law, also published in 1609, asserted that a line connecting a planet to the Sun sweeps out equal areas in equal times, implying variable orbital speeds—faster near the Sun and slower farther away.[75] By 1619, in Harmonices Mundi, Kepler introduced the third law: the square of a planet's orbital period is proportional to the cube of its semi-major axis , expressed as revealing a harmonic relation among planetary orbits independent of mass.[76] These laws, derived mathematically from observational data, provided a quantitative framework for heliocentric astronomy, influencing future celestial mechanics.[75]Experimental mechanics
The experimental mechanics of the Scientific Revolution marked a pivotal departure from Aristotelian qualitative descriptions of motion, emphasizing precise measurements and mathematical quantification to investigate forces and terrestrial dynamics. Pioneered by figures like Galileo Galilei, this approach involved controlled experiments to test hypotheses about acceleration and periodicity, replacing appeals to natural places or teleological causes with empirical data. This methodological shift facilitated the development of mechanics as a quantitative science, grounded in repeatable observations rather than philosophical deduction.[77] Galileo conducted inclined plane experiments around 1604 to study the motion of falling bodies, using bronze balls rolling down grooved wooden ramps to slow the descent and measure distances with a water clock for timing. These trials revealed that the distance traveled by a uniformly accelerating body is proportional to the square of the time elapsed, expressed in modern notation as , where is distance, is time, and is gravitational acceleration. Complementing this, Galileo's pendulum studies from 1603–1609 demonstrated isochronism—the property that the period of oscillation depends only on the length of the pendulum, not its amplitude or bob mass—providing a reliable method for timing experiments and challenging Aristotelian views on natural circular motion.[77][78][79] In his seminal work Discorsi e Dimostrazioni Matematiche intorno a Due Nuove Scienze (1638), Galileo synthesized these findings into the foundations of kinematics and the strength of materials, detailing theorems on projectile motion, uniform acceleration, and structural resistance derived from his terrestrial experiments. The text's dialogues illustrate how empirical data from inclined planes and pendula underpin mathematical models of force and resistance, establishing mechanics as an experimental discipline. Galileo's telescope, refined in 1609, further advanced experimental precision through observations like the 1610 discovery of Jupiter's four moons and the phases of Venus, which, while primarily astronomical, informed mechanical understandings of orbital dynamics by supporting heliocentric models through quantifiable positional data.[78][77][80] Evangelista Torricelli extended this experimental ethos in 1643 by inventing the mercury barometer, a device that measured atmospheric pressure by observing the height of a mercury column in a sealed tube inverted in a dish of mercury. This innovation not only quantified the "weight" of air—demonstrating it supported the column up to about 76 cm at sea level—but also created a measurable vacuum above the mercury, refuting Aristotelian horror vacui and advancing hydrostatics as a branch of experimental mechanics. Torricelli's work, building on Galileo's legacy, underscored the role of instrumentation in revealing invisible forces acting on Earth.[81][82]Newtonian synthesis
Isaac Newton's Philosophiæ Naturalis Principia Mathematica, published in 1687, marked a pivotal unification of terrestrial and celestial mechanics, laying the groundwork for classical physics.[83] In this seminal work, Newton formulated three laws of motion that describe the fundamental principles governing physical interactions. The first law states that an object remains at rest or in uniform motion unless acted upon by an external force, establishing the concept of inertia.[84] The second law quantifies the relationship between force, mass, and acceleration as , where force is the product of mass and acceleration.[83] The third law asserts that for every action, there is an equal and opposite reaction, ensuring conservation in interactions.[85] Complementing these, Newton proposed the law of universal gravitation, which posits that every particle attracts every other with a force proportional to the product of their masses and inversely proportional to the square of the distance between them, expressed as , where is the gravitational constant.[83] This synthesis extended gravitational attraction from apples to planets, unifying diverse phenomena under a single mathematical framework.[84] To derive the trajectories of celestial bodies, Newton developed the method of fluxions, an early form of calculus, during the 1660s, which he applied extensively in the Principia for calculating orbits and perturbations.[86] His fluxional notation facilitated the analysis of instantaneous rates of change, essential for modeling continuous motion in gravitational fields.[87] This innovation sparked a bitter priority dispute with Gottfried Wilhelm Leibniz, who independently developed a differential and integral calculus with superior notation; the controversy, fueled by national rivalries, persisted into the early 18th century and divided European mathematical communities.[88] Beyond mechanics, Newton's contributions extended to optics, detailed in his Opticks published in 1704. Through experiments with prisms, he demonstrated that white light decomposes into a spectrum of colors, refuting the prevailing view that color arises from modification of white light and instead showing it as an inherent property of light rays.[89] Newton advocated a corpuscular theory of light, proposing that light consists of small particles traveling in straight lines, which explained phenomena like refraction and dispersion based on varying particle velocities in media.[90] These findings not only advanced understanding of vision and color but also influenced instrument design, such as his reflecting telescope.[91] Newton applied his gravitational theory to specific astronomical problems, notably developing a lunar theory that accounted for the Moon's irregular orbit due to perturbations from the Sun and Earth's equatorial bulge.[92] In Book III of the Principia, he analyzed comet trajectories, treating them as elliptical orbits under inverse-square gravitation, which enabled predictions of their paths and return periods; this work inspired Edmund Halley's later successful forecast of the comet named after him.[93] Such applications demonstrated the predictive power of Newton's framework, bridging observation and theory in astronomy.[94] Philosophically, Newton's approach emphasized empirical rigor and mathematical deduction, encapsulated in his famous query from the 1713 edition of the Principia: hypotheses non fingo ("I frame no hypotheses"), rejecting speculative causes in favor of deriving laws directly from phenomena.[95] This stance promoted methodological naturalism, insisting that natural philosophy explain the universe through observable, quantifiable mechanisms without invoking supernatural agents, profoundly shaping scientific inquiry by prioritizing evidence over conjecture.[95] Newton's synthesis built upon the empirical foundations of Galileo and Kepler, integrating kinematics and planetary laws into a cohesive system.[83]18th-century classical foundations
Celestial mechanics
Celestial mechanics in the 18th century advanced the application of Newtonian gravity to the intricate motions within the solar system, focusing on perturbations and multi-body interactions to predict planetary and lunar paths with greater precision.[96] Leonhard Euler made significant early contributions to tackling the three-body problem, particularly in the 1740s, by developing approximations for the gravitational interactions among the Sun, Earth, and Moon, which laid groundwork for more accurate orbital calculations.[97] His lunar theory, first detailed in the 1753 publication Theoria motus lunae, provided formulas for lunar tables that supported navigation for over a century, while his 1772 second lunar theory refined these approximations using successive iterations and dual coordinate systems to model the Moon's irregular orbit under solar perturbations.[97] Joseph-Louis Lagrange further refined celestial mechanics through his analytical approach, introducing generalized coordinates that reformulated Newtonian dynamics in terms of variational principles, enabling more flexible treatments of complex systems without relying on Cartesian frameworks.[98] In 1772, Lagrange analyzed the restricted three-body problem and identified stable equilibrium points, now known as Lagrange points, which demonstrated the potential for long-term orbital stability in perturbed systems; this work directly explained the positioning and stability of Trojan asteroids sharing Jupiter's orbit at the L4 and L5 points.[99] Pierre-Simon Laplace's multi-volume Mécanique Céleste (1799–1825) synthesized these efforts into a comprehensive framework, employing perturbation theory to quantify small deviations in planetary orbits caused by mutual gravitational influences, thereby confirming the long-term stability of the solar system.[96] Within this work, Laplace also elaborated on the nebular hypothesis, positing that the solar system formed from a rotating cloud of gas that contracted and cooled, ejecting rings of material that coalesced into planets, providing a dynamical explanation for the system's architecture.[100] These theoretical advancements coincided with key observational discoveries that expanded the known solar system and tested the models. In 1781, William Herschel identified Uranus as a planet while surveying the constellation Gemini with a homemade telescope, marking the first planetary discovery since antiquity and doubling the known extent of the system at a distance twice that of Saturn.[101] Central to these developments were analytical tools like series expansions, which Euler, Lagrange, and Laplace employed to approximate solutions for perturbed orbits by expanding gravitational potentials into infinite series of periodic terms, allowing computation of secular variations and close approaches without exact integration of the nonlinear equations.[96]Fluid dynamics and waves
In the 18th century, the study of fluid dynamics emerged as a key extension of Newtonian mechanics, treating fluids as continuous media to analyze motion under pressure, gravity, and velocity. Building on Isaac Newton's early ideas in Principia about fluid resistance, scientists developed mathematical frameworks for steady and unsteady flows, laying groundwork for engineering applications like navigation and instrumentation. This period marked a shift from empirical observations to predictive equations, emphasizing conservation principles without invoking viscosity initially.[102] Daniel Bernoulli's Hydrodynamica (1738) introduced a foundational principle of energy conservation for incompressible fluids along streamlines, stating that the sum of pressure, gravitational potential, and kinetic energy remains constant: where is pressure, is density, is gravity, is height, and is velocity.[103] This equation, derived from integrating Newton's laws along fluid paths, explained phenomena like fluid acceleration under pressure gradients and found early use in ship hydrodynamics to optimize hull shapes for reduced drag.[104] For instance, Bernoulli's work influenced designs for faster vessels by predicting how velocity increases lower pressure around curved surfaces, aiding propulsion efficiency.[105] Leonhard Euler advanced this continuum approach in 1757 with his equations for inviscid flow, formalizing the momentum balance for fluids as a set of partial differential equations: neglecting viscous terms to model ideal fluids.[106] These equations generalized Bernoulli's principle to unsteady, three-dimensional flows and were applied to problems like water flow in channels and ship resistance, where Euler calculated wave-making drag for naval architecture.[102] Parallel developments in wave theory addressed vibrations and propagation, starting with Brook Taylor's 1714 analysis of musical string vibrations. Taylor derived the transverse displacement of a taut string under tension, showing that oscillation frequency depends on tension and linear density, providing a mathematical basis for harmonic motion in continuous media.[107] This work paved the way for Jean le Rond d'Alembert's 1747 wave equation, which governs one-dimensional propagation: where is displacement, is time, is position, and is wave speed.[108] D'Alembert solved it using characteristic methods, revealing waves as superpositions of forward and backward traveling components, applicable to sound in air columns.[109] These theories found practical application in acoustics, particularly organ pipe design, where Euler extended Bernoulli's ideas to model airflow and resonance in cylindrical pipes. Euler's Dissertation De Sono (1727, expanded later) analyzed end corrections and harmonic overtones in open and closed pipes, explaining why organ tones produce rich spectra from vortex shedding at the pipe mouth.[110] This informed 18th-century organ builders in tuning pipes for church instruments, balancing fundamental frequencies with partials for tonal clarity.[111] In atmospheric studies, George Hadley's 1735 paper proposed a meridional circulation cell to explain trade winds, attributing their easterly direction to Earth's rotation deflecting equatorward surface flow.[112] This model, now known as the Hadley cell, described rising air at the equator, poleward drift aloft, and sinking at about 30° latitude, with return flow as trade winds—integrating fluid dynamics with Coriolis effects for global wind patterns.[113] Such insights briefly informed celestial perturbation models by analogy to fluid drag in orbital mechanics.[102]Early heat theories
In the early 18th century, advancements in thermometry laid the groundwork for quantitative studies of heat. Daniel Gabriel Fahrenheit developed the mercury-in-glass thermometer in 1714, introducing a standardized scale that marked 32°F as the freezing point of water and 212°F as its boiling point under standard conditions, enabling more precise temperature measurements than previous alcohol-based devices.[114] This instrument facilitated experiments distinguishing heat as a measurable property, separate from mere sensation. Joseph Black's investigations in the 1760s advanced the understanding of heat by differentiating between sensible heat, which causes perceptible temperature changes, and latent heat, which is absorbed or released during phase transitions without altering temperature. In 1762, Black demonstrated that melting ice requires a significant quantity of heat—later quantified as 334 J/g—without raising its temperature, a discovery made through careful calorimetry using mixtures of substances at different temperatures.[115] He also introduced the concept of specific heats, showing that different materials require varying amounts of heat to achieve the same temperature increase; for instance, water has a specific heat of 1 cal/g·°C, higher than metals like lead at 0.031 cal/g·°C.[116] These distinctions were pivotal in shifting heat from a vague quality to a quantifiable entity in chemical and physical processes. The decline of the phlogiston theory in the late 18th century further integrated heat into chemical frameworks. Proposed earlier by Georg Ernst Stahl, phlogiston was imagined as a fire-like substance released during combustion, but Antoine Lavoisier's experiments from the 1770s onward revealed mass conservation and the role of oxygen, disproving the theory by 1780 as combustion involved gain rather than loss of weight.[117] This paved the way for caloric theory, where Lavoisier in the 1780s conceptualized heat as an indestructible, weightless fluid called "caloric" that flows between bodies to equalize temperatures, analogous to fluids in mechanics.[118] Lavoisier and Pierre-Simon Laplace used ice calorimeters to measure heat capacities, confirming that caloric's quantity in a body determines its temperature and enabling precise assessments of heat transfer in reactions.[119] Challenges to caloric theory emerged through mechanical experiments. In 1798, Benjamin Thompson, Count Rumford, observed during cannon boring in Munich that friction between a blunt steel borer and brass produced enough heat to boil approximately 19 pounds of water, with no evident limit to the heat generated despite minimal material removal—contradicting the idea of caloric being stored or released from matter.[120] Rumford argued that heat arises from motion, not a fluid, as the process converted mechanical work indefinitely into thermal effects without caloric depletion.[121] These instruments and concepts marked the transition from heat as a subtle substance to a form of energy, influencing early thermodynamic principles.19th-century unification
Lagrangian and Hamiltonian mechanics
In the late 18th century, Joseph-Louis Lagrange developed a reformulation of classical mechanics that shifted emphasis from forces to energy principles, culminating in his seminal work Mécanique Analytique published in 1788. This approach generalized Newtonian mechanics by employing variational principles, which posit that the path of a system minimizes or extremizes a quantity known as the action, defined as the integral of the Lagrangian over time. Central to this framework is d'Alembert's principle, extended by Lagrange to handle constraints through generalized coordinates, allowing for the derivation of equations of motion without resolving individual forces. The Lagrangian function itself is expressed as , where represents the kinetic energy and the potential energy of the system. Building on this foundation, William Rowan Hamilton introduced a complementary formulation in 1834, transforming mechanics into a symplectic structure suited for analytical solutions and later quantum developments. Hamilton's principle similarly relies on the action integral but employs canonical coordinates and momenta, leading to the Hamilton-Jacobi equation and the canonical equations of motion: and , where is the Hamiltonian, representing the total energy in phase space—a multidimensional space of positions and momenta. This phase space perspective enabled a deeper understanding of dynamical systems' evolution and conserved quantities. Lagrangian and Hamiltonian methods found immediate applications in rigid body dynamics, where they simplified the treatment of rotations and precessions, as seen in Lagrange's analysis of the Euler angles for top motion. In celestial mechanics, these tools revisited planetary perturbations, offering more efficient computational paths than direct Newtonian integrations, thus aiding 19th-century ephemeris calculations. A key advance came in the 1840s through Carl Gustav Jacob Jacobi, who extended Hamilton's approach by introducing determinants—now known as the Jacobi determinant—for evaluating integrals of motion, facilitating the separation of variables in complex systems and uncovering additional conserved quantities. Philosophically, these developments reinforced determinism in mechanics by highlighting conservation laws—such as energy, momentum, and angular momentum—as Noether-like symmetries inherent to the Lagrangian, emphasizing the predictability of physical systems from initial conditions.Electromagnetic theory
The development of electromagnetic theory in the 19th century marked a pivotal unification of electricity, magnetism, and optics, transforming physics from disparate phenomena into a coherent field theory. This era began with experimental discoveries linking electric currents to magnetic effects, progressed through quantitative laws governing induction and electrochemical processes, and culminated in a mathematical framework that predicted electromagnetic waves traveling at the speed of light. In 1800, Alessandro Volta invented the voltaic pile, the first device to produce a steady electric current from chemical reactions involving alternating zinc and copper discs separated by brine-soaked cardboard, providing a reliable source for subsequent electromagnetic experiments.[122] Building on this, Danish physicist Hans Christian Ørsted discovered in 1820 that an electric current flowing through a wire deflects a nearby compass needle, demonstrating that electricity generates magnetism; specifically, during a lecture, Ørsted observed the needle's perpendicular alignment to the current-carrying wire, with the deflection circling the wire in a manner dependent on current direction.[123] This breakthrough, detailed in Ørsted's pamphlet Experimenta circa effectum conflictus electrici in acum magneticam, established the fundamental connection between the two forces, overturning prior assumptions of their independence.[124] Inspired by Ørsted's finding, André-Marie Ampère rapidly formulated a mathematical description of the magnetic forces between current-carrying wires in 1820–1826. In his seminal work Mémoire sur la théorie mathématique des phénomènes électrodynamiques uniquement déduite de l'expérience, Ampère derived the force law between two current elements, showing that parallel currents attract and antiparallel ones repel, with the force proportional to the product of the currents and inversely proportional to the square of their separation distance, modulated by angular factors.[125] Ampère's electrodynamics treated currents as flows of an incompressible fluid, laying the groundwork for vectorial field descriptions and influencing later theoretical syntheses. Michael Faraday advanced these ideas through experiments on electromagnetic induction, announced in 1831. In his Experimental Researches in Electricity (first series, 1832), Faraday demonstrated that a changing magnetic field induces an electromotive force in a nearby circuit, as shown by a galvanometer deflection when he moved a magnet relative to a coil or varied current in a primary circuit linked to a secondary one via an iron ring; the induced current's direction opposed the change in magnetic flux, per Lenz's later rule.[126] Faraday conceptualized these effects using "lines of force," invisible curves representing magnetic and electric field intensities, where field strength is proportional to line density, providing an intuitive, non-mathematical precursor to vector fields. Faraday also established the laws of electrolysis in 1833–1834, quantifying the relationship between electric current and chemical decomposition. His first law states that the mass of a substance altered at an electrode is directly proportional to the quantity of electricity passed; the second law asserts that for a given quantity of electricity, the mass deposited is proportional to the substance's equivalent weight.[127] These principles introduced the Faraday constant, approximately 96,485 coulombs per mole of electrons, linking electricity to atomic-scale chemical changes and supporting the electrochemical equivalence of elements.[127] The theoretical pinnacle arrived with James Clerk Maxwell's 1865 paper A Dynamical Theory of the Electromagnetic Field, which synthesized prior work into four partial differential equations governing electromagnetic fields. Maxwell's equations are: The first describes electric field divergence from charge density , with as the permittivity of free space; the second indicates no magnetic monopoles, as magnetic flux has zero divergence; the third captures Faraday's induction law via curl of electric field ; and the fourth extends Ampère's law with the displacement current term , where is vacuum permeability and is current density, ensuring consistency in non-steady states.[128] From these equations, Maxwell derived the wave equation for electromagnetic disturbances, propagating at speed , numerically matching the speed of light (approximately 3 \times 10^8 m/s), implying that light is an electromagnetic wave and unifying optics with electromagnetism.[128] This field theory not only explained existing phenomena but also predicted new ones, such as radio waves later confirmed experimentally, and found brief applications in modeling mechanical systems like stressed elastic media.Thermodynamic laws
The formulation of the laws of thermodynamics in the mid-19th century marked a pivotal unification in physics, establishing the principles governing energy transformations in thermal systems and laying the groundwork for modern engineering applications. Building on earlier inquiries into heat engines, scientists articulated the conservation of energy and the directional constraints on thermal processes, resolving longstanding debates about the nature of heat as a form of energy rather than a fluid. These laws provided a rigorous framework for analyzing efficiency in machines, influencing industrial advancements during the era of rapid mechanization.[129] A foundational contribution came from Sadi Carnot, who in 1824 analyzed the ideal operation of heat engines through his theoretical cycle, demonstrating that the maximum efficiency of a reversible engine operating between two temperatures depends solely on the temperature difference. Carnot's efficiency is given by where temperatures are measured on an absolute scale, highlighting that no engine could exceed this limit without violating the principles of reversible processes. This work, though initially rooted in the caloric theory, anticipated key thermodynamic insights by emphasizing the role of temperature gradients in converting heat to work.[130] The first law of thermodynamics, expressing the conservation of energy in thermal systems, emerged from independent efforts in the 1840s to equate mechanical work and heat. Julius Robert von Mayer proposed in 1842 that heat arises from mechanical forces, estimating the mechanical equivalent of heat through observations of blood oxygenation and animal physiology, asserting that forces are indestructible and convertible. Concurrently, James Prescott Joule conducted precise experiments from 1840 to 1847, using paddle wheels to agitate water and measure the heat generated, establishing a numerical value for the mechanical equivalent of heat around 772 foot-pounds per British thermal unit. Hermann von Helmholtz formalized this conservation principle in his 1847 memoir "Über die Erhaltung der Kraft," extending it across mechanical, thermal, electrical, and chemical domains by arguing that a single force underlies all natural phenomena. The law is mathematically stated as where is the change in internal energy, is heat added to the system, and is work done by the system; this form was explicitly articulated by Rudolf Clausius in 1850.[131][132][133][134] Priority for the first law sparked intense debates, particularly between supporters of Mayer and Joule, as Mayer's philosophical and physiological arguments predated Joule's experimental data but lacked quantitative precision, while Joule's measurements gained wider acceptance in Britain. Helmholtz's comprehensive theoretical synthesis helped bridge these views, but recognition was uneven; the Royal Society awarded Copley Medals to Joule in 1870, Mayer in 1871, and Helmholtz in 1873 to acknowledge their collective contributions. These disputes underscored the interplay between empirical evidence and theoretical generalization in establishing the law.[135] The second law addressed the limitations on energy conversion, introducing the concept of irreversibility. In 1850, Clausius formulated it as the impossibility of heat flowing spontaneously from cold to hot bodies without external work, later refining this into the entropy principle. He defined entropy for reversible processes as where is the reversible heat transfer and is the absolute temperature, showing that entropy increases in isolated systems for irreversible processes. William Thomson (Lord Kelvin) independently stated in 1851 that it is impossible to construct a heat engine that, operating cyclically, produces no effect other than extracting heat from a single reservoir and converting it entirely to work, reinforcing the directional nature of thermal processes.[136][134][137] These laws directly spurred improvements in steam engine design, enabling engineers to optimize cycles for higher efficiency by better managing heat input and exhaust. By the late 19th century, applications of thermodynamic principles facilitated compound engines and superheated steam, raising typical efficiencies from under 10% in early designs to around 20%, which powered expanded industrial production and transportation networks.[129]Early 20th-century revolutions
Special relativity
The Michelson-Morley experiment of 1887 aimed to detect the Earth's motion through the hypothetical luminiferous ether, a medium postulated to propagate light waves, but yielded a null result, showing no evidence of such motion.[138] This outcome challenged classical notions and motivated Albert Einstein to develop special relativity in 1905, rejecting the ether entirely in favor of a framework where light's speed is invariant.[139] Einstein's theory rests on two postulates: the principle of relativity, stating that the laws of physics are identical in all inertial frames, and the constancy of the speed of light in vacuum for any observer, regardless of the source's or observer's motion.[139] These resolve inconsistencies between Newtonian mechanics and Maxwell's electromagnetic equations by introducing the concept of spacetime as a unified four-dimensional continuum. From these, Einstein derived the Lorentz transformations, which relate coordinates between inertial frames moving at constant velocity relative to each other: where and is the speed of light.[139] These transformations, building on earlier work by Hendrik Lorentz, ensure the invariance of electromagnetic laws across frames.[140] Key consequences include time dilation, where a clock moving relative to an observer ticks slower by factor , and length contraction, where lengths parallel to the motion shorten by .[139] In a separate 1905 paper, Einstein established mass-energy equivalence via , showing that energy release alters an object's inertial mass, a relation derived from the relativistic transformation of energy in electromagnetic processes.[141] Special relativity's relativistic kinematics revolutionized particle physics, providing formulas for momentum and total energy , essential for describing high-speed particles where Newtonian approximations fail, as verified in accelerators like those at CERN. These principles underpin modern applications, from GPS corrections to nuclear energy calculations, confirming the theory's foundational role in 20th-century physics.[142]General relativity
General relativity, developed by Albert Einstein in 1915, extended the framework of special relativity by incorporating gravity as the curvature of spacetime caused by mass and energy. Building on the insight that gravitational effects are indistinguishable from acceleration in a local frame, Einstein formulated a theory where the geometry of spacetime is dynamically linked to its contents. This geometrization resolved longstanding issues in Newtonian gravity and predicted novel phenomena that could be tested observationally.[143] Central to the theory is the equivalence principle, which posits that the effects of gravity are locally equivalent to those of acceleration, implying that free fall in a gravitational field follows geodesics in curved spacetime. Einstein first articulated this principle in 1907 as a key heuristic for generalizing special relativity, refining it in subsequent works to encompass both inertial and gravitational mass equivalence. The principle underpins the theory's covariance, allowing laws of physics to hold in all coordinate systems. From this foundation, Einstein derived the field equations, which relate spacetime curvature to the distribution of matter and energy: Here, is the Einstein tensor, derived from the Riemann curvature tensor that quantifies the intrinsic geometry of spacetime, while represents the stress-energy tensor. These equations, presented by Einstein on November 25, 1915, to the Prussian Academy of Sciences, marked the culmination of his efforts to achieve general covariance.[144] The theory's predictions provided immediate tests of its validity. It accurately explained the anomalous precession of Mercury's perihelion, accounting for the 43 arcseconds per century discrepancy unexplained by Newtonian mechanics, as Einstein calculated in 1915 using the field equations for the Schwarzschild metric. Gravitational deflection of light was confirmed during the 1919 solar eclipse expeditions led by Arthur Eddington, where starlight passing near the Sun was observed to shift by approximately 1.75 arcseconds, matching Einstein's prediction. Additionally, Einstein foresaw gravitational waves—ripples in spacetime propagating at the speed of light—in a 1916 paper, deriving their linear approximation from the field equations for weak fields.[145][146][147] A pivotal exact solution to the vacuum field equations was obtained by Karl Schwarzschild in 1916, describing the spacetime geometry around a spherically symmetric, non-rotating mass: This metric revealed features like event horizons, later interpreted as black holes, though Schwarzschild's work focused on stellar interiors. Cosmologically, general relativity implied dynamic universes; Einstein's 1917 inclusion of a cosmological constant aimed for a static model, but solutions by Alexander Friedmann in 1922 demonstrated expanding universes consistent with the equations, foreshadowing the observed cosmic expansion.[148][149]Quantum postulates
The development of quantum postulates in the early 20th century marked a profound shift from classical physics, introducing the concept of discrete energy levels to explain phenomena that continuous models could not accommodate. In 1900, Max Planck addressed the problem of blackbody radiation, where classical theory predicted an infinite energy density at short wavelengths, known as the ultraviolet catastrophe. To resolve this discrepancy with experimental observations, Planck hypothesized that the energy of electromagnetic oscillators in a blackbody is quantized, emitted or absorbed only in discrete packets proportional to the frequency: , where is Planck's constant and is the frequency.[150][151] This revolutionary assumption, initially viewed as a mathematical expedient, yielded the correct spectral distribution law for blackbody radiation, fitting data across all wavelengths and laying the groundwork for quantum theory.[152] Building on Planck's idea, Albert Einstein extended the quantization to light itself in 1905, proposing that electromagnetic radiation consists of discrete light quanta, later termed photons, each carrying energy . This hypothesis explained the photoelectric effect, where light ejects electrons from a metal surface only above a threshold frequency, regardless of intensity, and with maximum kinetic energy given by , where is the work function.[153][154] Classical wave theory failed to account for the frequency dependence and instantaneous emission, but Einstein's photon model predicted that below the threshold, no electrons are emitted, aligning with experiments by Philipp Lenard.[154] This work not only substantiated the particle-like nature of light but also provided a direct verification of Planck's constant through measurable electron energies.[153] Key experiments supported these postulates by revealing atomic discreteness. In 1909, Robert Millikan's oil-drop experiment measured the charge of electrons as quantized multiples of a fundamental unit C, confirming the elementary nature of electric charge and enabling precise determinations of from photoelectric data.[155][156] Ernest Rutherford's 1911 gold-foil scattering experiment further probed atomic structure, showing that alpha particles mostly pass through atoms undeflected but occasionally scatter at large angles, indicating a tiny, dense, positively charged nucleus surrounded by mostly empty space.[157][158] This nuclear model contradicted the diffuse plum-pudding atom and set the stage for quantized orbits. Niels Bohr integrated these findings in his 1913 atomic model, positing stable electron orbits around the nucleus where angular momentum is quantized: , with an integer, , the electron mass, its velocity, and the orbital radius.[159] Transitions between these discrete levels emit or absorb photons of energy , explaining hydrogen's spectral lines as jumps between levels, such as the Balmer series.[159] This semi-classical approach resolved the instability of Rutherford's model, where orbiting electrons would classically radiate energy and spiral inward, by forbidding radiation in stationary states.[160] The particle attributes of photons were solidified by Arthur Compton's 1923 scattering experiment, where X-rays incident on graphite electrons shifted wavelength by , with the electron mass, the speed of light, and the scattering angle—evidence of momentum conservation in photon-electron collisions.[161][162] This inelastic scattering, treated as billiard-ball-like impacts, refuted pure wave descriptions and confirmed photons as relativistic particles with both energy and momentum, compatible with special relativity for such interactions.[163] These postulates collectively established the dual wave-particle nature at the quantum scale, transforming physics from continuous to discrete foundations.Mid-20th-century consolidations
Quantum electrodynamics
Quantum electrodynamics (QED) represents the quantum field-theoretic description of electromagnetic interactions between charged particles, particularly electrons, and photons, achieving full consistency with both quantum mechanics and special relativity. Developed in the 1930s as an extension of earlier quantum postulates, QED initially suffered from mathematical infinities in perturbation theory calculations, which threatened its predictive power. These challenges were resolved in the mid-1940s through the independent yet convergent efforts of Sin-Itiro Tomonaga, Julian Schwinger, and Richard Feynman, who introduced renormalization techniques to systematically eliminate divergences and yield finite, observable predictions. Their work culminated in the 1965 Nobel Prize in Physics for "their fundamental work in quantum electrodynamics, with deep-ploughing consequences for the physics of elementary particles." A foundational step toward QED was Paul Dirac's 1928 derivation of the relativistic wave equation for the electron, known as the Dirac equation: This first-order equation in both space and time incorporated special relativity into quantum mechanics, naturally accounting for electron spin and predicting negative-energy solutions interpreted as antimatter.[164] The positron, the electron's antiparticle, was experimentally discovered by Carl David Anderson in 1932 through cloud chamber observations of cosmic rays, providing direct confirmation of Dirac's prediction and validating the existence of particle-antiparticle pairs.[165] Building on this, the full QED framework in the 1930s quantized the electromagnetic field alongside Dirac fields, but calculations of processes like electron-photon scattering produced infinite results due to virtual particle loops. The renormalization revolution began with Tomonaga's 1943 (published 1946) covariant perturbation theory, which preserved Lorentz invariance while handling interactions along worldlines, followed by Schwinger's 1948 canonical formalism emphasizing gauge symmetry.[166][167] Feynman complemented these with his 1948 path-integral approach and diagrammatic representation of amplitudes, enabling intuitive visualization of virtual particle exchanges and renormalization as mass and charge redefinitions. A key triumph was the theoretical prediction and explanation of the Lamb shift—the small energy splitting between the 2S_{1/2} and 2P_{1/2} states in hydrogen, experimentally measured by Willis Lamb and Robert Retherford in 1947 using microwave spectroscopy. Bethe's 1947 non-relativistic calculation, refined relativistically by the trio, matched the observation to within 10%, attributing the shift to virtual photon emissions and absorptions.[168] The QED Lagrangian density encodes these interactions with local U(1) gauge invariance: where is the covariant derivative coupling the Dirac field to the photon field , and is the electromagnetic field strength; gauge invariance ensures physical observables are independent of the arbitrary choice of . Virtual particles, such as electron-positron pairs and photons, arise in Feynman diagrams as intermediate states contributing to corrections. One precise prediction is the anomalous magnetic moment of the electron, , where Schwinger's 1948 one-loop calculation yielded , with the electron's gyromagnetic ratio deviating from the Dirac value of 2.[169] Modern experiments verify QED's prediction to over 12 decimal places, with the latest (2023) measurement , confirming the theory's accuracy without discrepancies.[170]Nuclear and particle physics
In 1911, Ernest Rutherford conducted the gold foil experiment, which revealed the nuclear structure of the atom by demonstrating that most alpha particles passed through a thin gold foil undeflected, while a small fraction were scattered at large angles, indicating a dense, positively charged nucleus at the atom's center. This discovery overturned the plum pudding model and laid the foundation for nuclear physics. In 1932, James Chadwick discovered the neutron through bombardment experiments with beryllium using alpha particles from polonium, identifying an electrically neutral particle of approximately the same mass as the proton, which explained the stability of atomic nuclei without excessive electrostatic repulsion. The discovery of nuclear fission occurred in 1938 when Otto Hahn and Fritz Strassmann chemically identified barium as a product of neutron-bombarded uranium, with Lise Meitner and Otto Frisch providing the theoretical interpretation of the nucleus splitting into lighter elements, releasing enormous energy. This led to the recognition of chain reactions, where fission products emit neutrons that induce further fissions, as demonstrated by Enrico Fermi's team in 1942 with the first controlled chain reaction in a graphite-moderated uranium pile at the University of Chicago. These developments directly contributed to the Manhattan Project in the 1940s, a U.S.-led effort that produced the first atomic bombs using enriched uranium and plutonium, culminating in their use during World War II and ushering in the nuclear age. Particle accelerators emerged as essential tools for probing nuclear and subatomic structures. In 1930, Ernest O. Lawrence invented the cyclotron, a device using a magnetic field and alternating electric voltage to accelerate charged particles in a spiral path, enabling higher energies than straight-line accelerators and facilitating discoveries like artificial radioactivity. By the mid-20th century, accelerators had revealed a menagerie of elementary particles beyond protons and neutrons. In 1964, Murray Gell-Mann proposed the quark model to organize these particles into families based on quantum numbers, postulating three types of quarks (up, down, strange) as fundamental constituents held together by the strong force, which simplified the particle zoo and predicted new particles later confirmed experimentally. Key milestones in nuclear and particle physics included the first detection of neutrinos in 1956 by Clyde Cowan and Frederick Reines, who observed antineutrinos from a nuclear reactor interacting with protons to produce positrons and neutrons, confirming the particle postulated by Pauli in 1930 to conserve energy and momentum in beta decay. In the 1950s, initial experiments achieved controlled nuclear fusion, such as the 1951 stellarator concept by Lyman Spitzer and early Z-pinch devices that briefly confined hot plasmas to fuse deuterium nuclei, marking the start of fusion research for energy production despite challenges with stability. Conservation laws played a crucial role in understanding particle interactions: baryon number, conserved in all known processes since the 1930s to account for the difference between protons and neutrons in nuclei, and lepton number, introduced in the 1950s to track electrons, muons, and neutrinos in weak decays, ensuring balance in reactions like beta decay. These principles, along with quantum electrodynamics for electron interactions, guided the classification of particles and forces in nuclear processes.Standard Model foundations
The foundations of the Standard Model emerged in the late 1960s and 1970s as a gauge theory unifying the electromagnetic, weak, and strong nuclear interactions, building on experimental data from particle accelerators that revealed patterns in quark and lepton behaviors. This framework posits that the fundamental forces are mediated by gauge bosons and respects a local gauge symmetry under the group , where governs the strong force, the weak isospin, and the weak hypercharge.[171] The model's Lagrangian incorporates these symmetries through Yang-Mills gauge fields, fermion matter fields in chiral representations, and a scalar Higgs field to break electroweak symmetry, generating masses for particles while preserving the strong force's chiral symmetry.[172] The electroweak sector, unifying electromagnetism and the weak force, was developed through the Glashow-Weinberg-Salam model in the late 1960s. Sheldon Glashow proposed an gauge structure in 1961, incorporating parity-violating weak interactions, but initial formulations suffered from non-renormalizability.[173] Steven Weinberg and Abdus Salam independently extended this in 1967–1968 by incorporating the Higgs mechanism for mass generation, predicting intermediate vector bosons: the charged and neutral .[174] These particles were discovered in 1983 at CERN's Super Proton Synchrotron by the UA1 and UA2 collaborations, confirming the model's predictions with masses around 80 GeV for and 91 GeV for , and earning Carlo Rubbia and Simon van der Meer the 1984 Nobel Prize. Quantum chromodynamics (QCD) describes the strong interaction within the Standard Model, treating quarks as carrying "color" charge under and mediated by eight massless gluons. Formulated in the early 1970s, QCD resolved issues with earlier strong force models by demonstrating asymptotic freedom, where the coupling strength decreases at short distances, allowing perturbative calculations for high-energy processes.[175] This key property was independently discovered in 1973 by David Gross and Frank Wilczek, and David Politzer, using non-Abelian gauge theory renormalization group analysis, and was awarded the 2004 Nobel Prize.[176] QCD's success in explaining deep inelastic scattering data and hadron spectroscopy solidified its role, predicting quark confinement at low energies due to the increasing coupling at long distances.[177] The Higgs mechanism provides the mass generation in the Standard Model through spontaneous symmetry breaking of the electroweak gauge symmetry. Proposed in 1964 by François Englert and Robert Brout, and independently by Peter Higgs, this mechanism introduces a scalar doublet field whose vacuum expectation value breaks to , endowing the and bosons with mass while keeping the photon massless; fermions acquire mass via Yukawa couplings.[178] The 2013 Nobel Prize recognized Englert and Higgs for this theoretical breakthrough, which resolved the unitarity issues in electroweak scattering and enabled the model's renormalizability, as shown by Gerardus 't Hooft and Martinus Veltman in 1971.[179] Key experimental validations underpinned these theoretical advances. The violation of parity conservation in weak decays, proposed by Tsung-Dao Lee and Chen-Ning Yang in 1956 and confirmed by Chien-Shiung Wu's cobalt-60 beta decay experiment in 1957, motivated the chiral structure of electroweak interactions.[180] Additionally, the discovery of the tau lepton in 1975 by Martin Perl and collaborators at SLAC's MARK I detector extended the lepton generations to three, aligning with the quark sector and supporting the family's replication in the model; Perl received the 1995 Nobel Prize for this finding.[181] These milestones, combined with precise electroweak measurements, established the Standard Model's predictive power by the 1980s.[172]Late 20th to 21st-century frontiers
Cosmological models
In the aftermath of general relativity's formulation, early 20th-century cosmologists sought models consistent with Einstein's equations that described the universe on large scales. Alexander Friedmann derived solutions in 1922 indicating an expanding universe from a singular origin, assuming homogeneity and isotropy, which Georges Lemaître independently expanded in 1927 by incorporating observational data to suggest a dynamic cosmos originating from a "primeval atom." These efforts culminated in the Friedmann-Lemaître-Robertson-Walker (FLRW) metric, formalized in the 1930s by Howard Robertson and Arthur Walker, providing a mathematical framework for a homogeneous, isotropic spacetime with a scale factor that evolves over time, enabling descriptions of either expansion or contraction depending on matter and energy content. An alternative to these evolving models emerged in 1948 with the steady-state theory proposed by Hermann Bondi, Thomas Gold, and Fred Hoyle, which posited a universe in perpetual expansion but maintaining constant density through continuous creation of matter, avoiding a singular beginning while aligning with general relativity via a modified perfect cosmological principle. This theory gained traction amid debates over expansion until Edwin Hubble's 1929 observations provided empirical support for a dynamic universe, revealing a linear relation between galactic recession velocities and distances , expressed as where is the Hubble constant, indicating uniform expansion from a common origin. The Big Bang model, building on the FLRW framework and Hubble's law, gained decisive evidence in 1965 when Arno Penzias and Robert Wilson serendipitously detected uniform microwave radiation at 2.7 K, interpreted as the cooled remnant of the hot, dense early universe, matching predictions from Ralph Alpher and Robert Herman's 1948 work on primordial nucleosynthesis involving Standard Model particles like protons and neutrons. This cosmic microwave background (CMB) decisively favored the Big Bang over steady-state cosmology, as the latter struggled to explain such a relic without ad hoc adjustments. Further refinements addressed inconsistencies like the horizon problem—why distant regions appear thermally uniform—and the flatness problem—why the universe's density is finely tuned to critical value—through Alan Guth's 1980 inflation theory, proposing a brief phase of exponential expansion driven by a scalar field, rapidly increasing the universe's scale factor by over 60 e-folds and homogenizing causal regions.[182] Observations of galactic dynamics introduced additional components to cosmological models. In 1933, Fritz Zwicky inferred the presence of unseen "dark matter" in the Coma Cluster, as the observed velocities of galaxies required far more mass than visible stars and gas could provide, based on the virial theorem applied to cluster dynamics.[183] By the late 20th century, dark matter's gravitational influence was evident in galaxy rotation curves and large-scale structure formation within Big Bang simulations. The model's completeness was upended in 1998 when two teams, led by Saul Perlmutter and Adam Riess, analyzed Type Ia supernovae as standard candles and found distant explosions dimmer than expected, implying accelerated expansion driven by a dominant "dark energy" component, comprising about 70% of the universe's energy density and consistent with a cosmological constant in the FLRW equations.[184] Precision measurements bolstered these developments, with the Cosmic Background Explorer (COBE) satellite detecting tiny temperature anisotropies in the CMB in 1992, at the level of 30 μK on degree scales, as reported by George Smoot and colleagues, providing the first evidence of primordial density fluctuations that seeded galaxy formation under inflation and dark matter's influence.[185] These observations refined parameters like the universe's age (about 13.8 billion years) and curvature (nearly flat), integrating dark matter (around 25% density) and dark energy into a Lambda-CDM model that remains the standard framework for interpreting cosmic evolution.Unified field attempts
Efforts to unify the fundamental forces of nature, particularly gravity with the quantum forces described by the Standard Model, have been a central pursuit in theoretical physics since the early 20th century. One of the earliest attempts was the Kaluza-Klein theory, proposed by Theodor Kaluza in 1921, which extended general relativity to five dimensions to incorporate electromagnetism as a geometric effect arising from the extra dimension.[186] In 1926, Oskar Klein refined this idea by suggesting that the fifth dimension is compactified into a small circle, with its radius on the order of the Planck length, making it imperceptible at macroscopic scales and explaining charge quantization through quantum mechanics. This framework laid the groundwork for higher-dimensional theories but struggled to include the strong and weak nuclear forces, limiting its scope to gravity and electromagnetism. In the 1970s, grand unified theories (GUTs) emerged to merge the strong, weak, and electromagnetic forces into a single gauge group at high energies, while treating gravity separately. The minimal SU(5) model, introduced by Howard Georgi and Sheldon Glashow in 1974, embeds the Standard Model symmetries into the SU(5) group, predicting proton decay with lifetimes around to years and the existence of magnetic monopoles as topological solitons. These monopoles, independently discovered by Gerard 't Hooft and Alexander Polyakov in 1974 within non-Abelian gauge theories, arise as stable configurations when a gauge symmetry is spontaneously broken, carrying magnetic charge and potentially explaining aspects of cosmology like inflation. However, experimental searches, such as those by the Super-Kamiokande detector, have set lower limits on proton lifetimes exceeding years, challenging the simplest GUT predictions without fine-tuning.[187] String theory, revitalized in the 1980s as a candidate for quantum gravity, posits that fundamental particles are one-dimensional vibrating strings at the Planck scale, approximately meters, naturally incorporating gravity through closed string modes. The incorporation of supersymmetry in the mid-1980s led to five consistent superstring theories in ten dimensions, resolving anomalies and unifying bosons and fermions. In 1995, Edward Witten proposed M-theory as an 11-dimensional unification of these frameworks, incorporating strong-weak coupling dualities and branes, which has driven much of the subsequent development despite lacking direct experimental tests. Concurrently, loop quantum gravity, pioneered by Abhay Ashtekar in the late 1980s through new variables reformulating general relativity, quantizes spacetime into discrete loops, addressing the Wheeler-DeWitt equation—a timeless constraint from canonical quantum gravity originally formulated by Bryce DeWitt in 1967. This approach predicts a granular structure to spacetime at the Planck scale, with area and volume operators having discrete spectra, but it remains background-independent and non-perturbative, differing fundamentally from string theory's perturbative methods. Despite these advances, unified field theories face significant challenges, including the hierarchy problem, which questions why the electroweak scale ( GeV) is so much smaller than the Planck scale ( GeV) without enormous cancellations in quantum corrections. Proposed solutions like supersymmetry or extra dimensions aim to stabilize this disparity, yet no experimental evidence has confirmed such mechanisms, such as superpartners or proton decay.[187] Similarly, string theory and loop quantum gravity predict phenomena at inaccessible energy scales, with ongoing debates over their falsifiability and the absence of observable signatures in current colliders or cosmological data. These efforts continue to evolve, motivated by the need for a consistent quantum theory of gravity.Recent experimental breakthroughs
In 2012, the ATLAS and CMS experiments at the Large Hadron Collider (LHC) announced the discovery of a new particle consistent with the Higgs boson predicted by the Standard Model, with a mass of approximately 125 GeV/c² observed in proton-proton collisions.[188] Subsequent measurements confirmed its spin-0 nature, negative parity, and coupling strengths to other particles aligning closely with Standard Model expectations, solidifying its role in electroweak symmetry breaking.[189] By the mid-2010s, LHC data had excluded alternative interpretations, such as supersymmetric extensions, at high confidence levels, marking a major validation of the Standard Model while prompting searches for subtle deviations.[190] The direct detection of gravitational waves in 2015 by the Advanced LIGO observatories provided the first observational confirmation of general relativity's predictions for strong-field dynamics. The signal, GW150914, originated from the merger of two black holes with masses of about 36 and 29 solar masses, releasing energy equivalent to three solar masses in gravitational radiation and confirming the existence of binary black hole systems.[191] Over the following decade, LIGO and Virgo detected dozens more events, including neutron star mergers like GW170817, which also verified multimessenger astronomy by correlating gravitational waves with electromagnetic counterparts, further testing general relativity in extreme regimes. Neutrino oscillation experiments in the late 1990s and early 2000s established that neutrinos possess non-zero masses, hinting at physics beyond the Standard Model's original massless assumption. The Super-Kamiokande experiment's 1998 observation of atmospheric neutrino deficits indicated oscillations driven by a mass-squared difference of approximately 2.5 × 10^{-3} eV², with nearly maximal mixing. This was corroborated by the Sudbury Neutrino Observatory (SNO) in 2001, which resolved the solar neutrino problem by detecting flavor conversions in solar neutrinos, yielding a mass-squared difference of about 7.5 × 10^{-5} eV² and confirming the full three-flavor oscillation framework. These findings necessitated neutrino mass terms in particle physics models, influencing extensions like seesaw mechanisms, though the absolute mass scale remains undetermined below 0.1 eV from ongoing experiments. Advances in quantum computing gained momentum with experimental demonstrations of quantum advantage in the late 2010s. In 2019, Google's Sycamore processor, a 53-qubit superconducting device, performed a random circuit sampling task in 200 seconds that would take the fastest classical supercomputer an estimated 10,000 years, achieving quantum supremacy for this contrived problem.[192] This milestone highlighted scalable qubit coherence and error mitigation, paving the way for entanglement-based experiments; for instance, subsequent Bell inequality tests with up to 20 entangled photons verified quantum nonlocality over distances exceeding 1,200 km, advancing quantum networks. By the mid-2020s, hybrid quantum-classical algorithms had been applied to molecular simulations, though practical fault-tolerant computing remains a future goal amid challenges like decoherence. In the 2020s, precision measurements continued to probe the Standard Model's limits without uncovering major new particles. The Fermilab Muon g-2 experiment's 2021 Run-1 results reported a muon's magnetic anomaly (g-2)/2 = 0.00116592061(41), deviating from Standard Model predictions by 4.2 standard deviations, suggesting possible new physics contributions from hypothetical particles or forces.[193] Follow-up data through 2025 provided the final measurement a_μ = 0.001165920705(14), which, with updated Standard Model predictions incorporating improved lattice QCD calculations, agrees within uncertainties, confirming consistency with the theory and resolving prior tensions.[194] Meanwhile, the James Webb Space Telescope (JWST), operational since 2022, delivered cosmological data revealing unexpectedly massive and mature galaxies at redshifts z > 10, within 500 million years of the Big Bang, challenging galaxy formation models and prompting refinements to dark matter simulations.[195] LHC Run 3, ongoing through 2025, has accumulated over 100 fb^{-1} of data at 13.6 TeV without evidence for supersymmetric particles or leptoquarks above TeV scales, shifting emphasis to high-precision Higgs and electroweak measurements that indirectly test unified theories.[196]References
- https://en.wikisource.org/wiki/On_the_Magnet
