In physics, Planck's law (also Planck radiation law ) describes the spectral density of electromagnetic radiation emitted by a black body in thermal equilibrium at a given temperature T, when there is no net flow of matter or energy between the body and its environment.
At the end of the 19th century, physicists were unable to explain why the observed spectrum of black-body radiation, which by then had been accurately measured, diverged significantly at higher frequencies from that predicted by existing theories. In 1900, German physicist Max Planck heuristically derived a formula for the observed spectrum by assuming that a hypothetical electrically charged oscillator in a cavity that contained black-body radiation could only change its energy in a minimal increment, E, that was proportional to the frequency of its associated electromagnetic wave. While Planck originally regarded the hypothesis of dividing energy into increments as a mathematical artifice, introduced merely to get the correct answer, other physicists including Albert Einstein built on his work, and Planck's insight is now recognized to be of fundamental importance to quantum theory.
The law
Every physical body spontaneously and continuously emits electromagnetic radiation and the spectral radiance of a body, Bν, describes the spectral emissive power per unit area, per unit solid angle and per unit frequency for particular radiation frequencies. The relationship given by Planck's radiation law, given below, shows that with increasing temperature, the total radiated energy of a body increases and the peak of the emitted spectrum shifts to shorter wavelengths. According to Planck's distribution law, the spectral energy density (energy distribution) at given temperature is given by (SI units):
]]>Special relativity
In physics, the special theory of relativity, or special relativity for short, is a scientific theory of the relationship between space and time. In Albert Einstein's 1905 treatment, the theory is presented as being based on just two postulates:
i) The laws of physics are invariant (identical) in all inertial frames of reference (that is, frames of reference with no acceleration).
ii) The speed of light in vacuum is the same for all observers, regardless of the motion of light source or observer.
Origins and significance
Special relativity was described by Albert Einstein in a paper published on 26 September 1905 titled "On the Electrodynamics of Moving Bodies". Maxwell's equations of electromagnetism appeared to be incompatible with Newtonian mechanics, and the Michelson–Morley null result failed to detect the Earth's motion against the hypothesized luminiferous aether. These led to the development of the Lorentz transformations, which adjust distances and times for moving objects. Special relativity corrects the hitherto laws of mechanics to handle situations involving all motions and especially those at a speed close to that of light (known as relativistic velocities). Today, special relativity is proven to be the most accurate model of motion at any speed when gravitational and quantum effects are negligible. Even so, the Newtonian model is still valid as a simple and accurate approximation at low velocities (relative to the speed of light), for example, everyday motions on Earth.
Special relativity has a wide range of consequences that have been experimentally verified. They include the relativity of simultaneity, length contraction, time dilation, the relativistic velocity addition formula, the relativistic Doppler effect, relativistic mass, a universal speed limit, mass–energy equivalence, the speed of causality and the Thomas precession. It has, for example, replaced the conventional notion of an absolute universal time with the notion of a time that is dependent on reference frame and spatial position. Rather than an invariant time interval between two events, there is an invariant spacetime interval. Combined with other laws of physics, the two postulates of special relativity predict the equivalence of mass and energy, as expressed in the mass–energy equivalence formula
, where c is the speed of light in a vacuum. It also explains how the phenomena of electricity and magnetism are related.A defining feature of special relativity is the replacement of the Galilean transformations of Newtonian mechanics with the Lorentz transformations. Time and space cannot be defined separately from each other (as was previously thought to be the case). Rather, space and time are interwoven into a single continuum known as "spacetime". Events that occur at the same time for one observer can occur at different times for another.
Until several years later when Einstein developed general relativity, which introduced a curved spacetime to incorporate gravity, the phrase "special relativity" was not used. A translation sometimes used is "restricted relativity"; "special" really means "special case". Some of the work of Albert Einstein in special relativity is built on the earlier work by Hendrik Lorentz and Henri Poincaré. The theory became essentially complete in 1907, with Hermann Minkowski's papers on spacetime.
The theory is "special" in that it only applies in the special case where the spacetime is "flat", that is, where the curvature of spacetime (a consequence of the energy–momentum tensor and representing gravity) is negligible. To correctly accommodate gravity, Einstein formulated general relativity in 1915. Special relativity, contrary to some historical descriptions, does accommodate accelerations as well as accelerating frames of reference.
Just as Galilean relativity is now accepted to be an approximation of special relativity that is valid for low speeds, special relativity is considered an approximation of general relativity that is valid for weak gravitational fields, that is, at a sufficiently small scale (e.g., when tidal forces are negligible) and in conditions of free fall. But general relativity incorporates non-Euclidean geometry to represent gravitational effects as the geometric curvature of spacetime. Special relativity is restricted to the flat spacetime known as Minkowski space. As long as the universe can be modeled as a pseudo-Riemannian manifold, a Lorentz-invariant frame that abides by special relativity can be defined for a sufficiently small neighborhood of each point in this curved spacetime.
Galileo Galilei had already postulated that there is no absolute and well-defined state of rest (no privileged reference frames), a principle now called Galileo's principle of relativity. Einstein extended this principle so that it accounted for the constant speed of light, a phenomenon that had been observed in the Michelson–Morley experiment. He also postulated that it holds for all the laws of physics, including both the laws of mechanics and of electrodynamics.
General Relativity
General relativity, also known as the general theory of relativity and Einstein's theory of gravity, is the geometric theory of gravitation published by Albert Einstein in 1915 and is the current description of gravitation in modern physics. General relativity generalises special relativity and refines Newton's law of universal gravitation, providing a unified description of gravity as a geometric property of space and time or four-dimensional spacetime. In particular, the curvature of spacetime is directly related to the energy and momentum of whatever matter and radiation are present. The relation is specified by the Einstein field equations, a system of second order partial differential equations.
Newton's law of universal gravitation, which describes classical gravity, can be seen as a prediction of general relativity for the almost flat spacetime geometry around stationary mass distributions. Some predictions of general relativity, however, are beyond Newton's law of universal gravitation in classical physics. These predictions concern the passage of time, the geometry of space, the motion of bodies in free fall, and the propagation of light, and include gravitational time dilation, gravitational lensing, the gravitational redshift of light, the Shapiro time delay and singularities/black holes. So far, all tests of general relativity have been shown to be in agreement with the theory. The time-dependent solutions of general relativity enable us to talk about the history of the universe and have provided the modern framework for cosmology, thus leading to the discovery of the Big Bang and cosmic microwave background radiation. Despite the introduction of a number of alternative theories, general relativity continues to be the simplest theory consistent with experimental data.
Reconciliation of general relativity with the laws of quantum physics remains a problem, however, as there is a lack of a self-consistent theory of quantum gravity. It is not yet known how gravity can be unified with the three non-gravitational forces: strong, weak and electromagnetic.
Einstein's theory has astrophysical implications, including the prediction of black holes—regions of space in which space and time are distorted in such a way that nothing, not even light, can escape from them. Black holes are the end-state for massive stars. Microquasars and active galactic nuclei are believed to be stellar black holes and supermassive black holes. It also predicts gravitational lensing, where the bending of light results in multiple images of the same distant astronomical phenomenon. Other predictions include the existence of gravitational waves, which have been observed directly by the physics collaboration LIGO and other observatories. In addition, general relativity has provided the base of cosmological models of an expanding universe.
Widely acknowledged as a theory of extraordinary beauty, general relativity has often been described as the most beautiful of all existing physical theories.
(The Laser Interferometer Gravitational-Wave Observatory (LIGO) is a large-scale physics experiment and observatory designed to detect cosmic gravitational waves and to develop gravitational-wave observations as an astronomical tool. Two large observatories were built in the United States with the aim of detecting gravitational waves by laser interferometry. These observatories use mirrors spaced four kilometers apart which are capable of detecting a change of less than one ten-thousandth the charge diameter of a proton.
The initial LIGO observatories were funded by the United States National Science Foundation (NSF) and were conceived, built and are operated by Caltech and MIT. They collected data from 2002 to 2010 but no gravitational waves were detected.
The Advanced LIGO Project to enhance the original LIGO detectors began in 2008 and continues to be supported by the NSF, with important contributions from the United Kingdom's Science and Technology Facilities Council, the Max Planck Society of Germany, and the Australian Research Council. The improved detectors began operation in 2015. The detection of gravitational waves was reported in 2016 by the LIGO Scientific Collaboration (LSC) and the Virgo Collaboration with the international participation of scientists from several universities and research institutions. Scientists involved in the project and the analysis of the data for gravitational-wave astronomy are organized by the LSC, which includes more than 1000 scientists worldwide, as well as 440,000 active Einstein@Home users as of December 2016.)
]]>In quantum field theory, the Casimir effect is a physical force acting on the macroscopic boundaries of a confined space which arises from the quantum fluctuations of the field. It is named after the Dutch physicist Hendrik Casimir, who predicted the effect for electromagnetic systems in 1948.
In the same year, Casimir together with Dirk Polder described a similar effect experienced by a neutral atom in the vicinity of a macroscopic interface which is referred to as the Casimir–Polder force. Their result is a generalization of the London–van der Waals force and includes retardation due to the finite speed of light. Since the fundamental principles leading to the London–van der Waals force, the Casimir and the Casimir–Polder force, respectively, can be formulated on the same footing, the distinction in nomenclature nowadays serves a historical purpose mostly and usually refers to the different physical setups.
It was not until 1997 that a direct experiment by S. Lamoreaux quantitatively measured the Casimir force to within 5% of the value predicted by the theory.
The Casimir effect can be understood by the idea that the presence of macroscopic material interfaces, such as conducting metals and dielectrics, alters the vacuum expectation value of the energy of the second-quantized electromagnetic field. Since the value of this energy depends on the shapes and positions of the materials, the Casimir effect manifests itself as a force between such objects.
Any medium supporting oscillations has an analogue of the Casimir effect. For example, beads on a string as well as plates submerged in turbulent water or gas illustrate the Casimir force.
In modern theoretical physics, the Casimir effect plays an important role in the chiral bag model of the nucleon; in applied physics it is significant in some aspects of emerging microtechnologies and nanotechnologies.
Physical properties
The typical example is of two uncharged conductive plates in a vacuum, placed a few nanometers apart. In a classical description, the lack of an external field means that there is no field between the plates, and no force would be measured between them. When this field is instead studied using the quantum electrodynamic vacuum, it is seen that the plates do affect the virtual photons which constitute the field, and generate a net force – either an attraction or a repulsion depending on the specific arrangement of the two plates. Although the Casimir effect can be expressed in terms of virtual particles interacting with the objects, it is best described and more easily calculated in terms of the zero-point energy of a quantized field in the intervening space between the objects. This force has been measured and is a striking example of an effect captured formally by second quantization.
The treatment of boundary conditions in these calculations has led to some controversy. In fact, "Casimir's original goal was to compute the van der Waals force between polarizable molecules" of the conductive plates. Thus it can be interpreted without any reference to the zero-point energy (vacuum energy) of quantum fields.
Because the strength of the force falls off rapidly with distance, it is measurable only when the distance between the objects is extremely small. On a submicron scale, this force becomes so strong that it becomes the dominant force between uncharged conductors. In fact, at separations of 10 nm – about 100 times the typical size of an atom – the Casimir effect produces the equivalent of about 1 atmosphere of pressure (the precise value depending on surface geometry and other factors).
]]>Moseley's law is an empirical law concerning the characteristic x-rays emitted by atoms. The law had been discovered and published by the English physicist Henry Moseley in 1913-1914. Until Moseley's work, "atomic number" was merely an element's place in the periodic table and was not known to be associated with any measurable physical quantity.[3] In brief, the law states that the square root of the frequency of the emitted x-ray is approximately proportional to the atomic number.
History
The historic periodic table was roughly ordered by increasing atomic weight, but in a few famous cases the physical properties of two elements suggested that the heavier ought to precede the lighter. An example is cobalt having a weight of 58.9 and nickel having an atomic weight of 58.7.
Henry Moseley and other physicists used x-ray diffraction to study the elements, and the results of their experiments led to organizing the periodic table by proton count.
Apparatus
Since the spectral emissions for the lighter elements would be in the soft X-ray range (absorbed by air), the spectrometry apparatus had to be enclosed inside a vacuum. Details of the experimental setup are documented in the journal articles "The High-Frequency Spectra of the Elements" Part I and Part II.
Results
Moseley found that the
lines (in Siegbahn notation) were indeed related to the atomic number, Z.Following Bohr's lead, Moseley found that for the spectral lines, this relationship could be approximated by a simple formula, later called Moseley's Law.
where:
*
is the frequency of the observed x-ray emission lineFick's laws of diffusion describe diffusion and were derived by Adolf Fick in 1855. They can be used to solve for the diffusion coefficient, D. Fick's first law can be used to derive his second law which in turn is identical to the diffusion equation.
A diffusion process that obeys Fick's laws is called normal or Fickian diffusion; otherwise, it is called anomalous diffusion or non-Fickian diffusion.
History
In 1855, physiologist Adolf Fick first reported his now well-known laws governing the transport of mass through diffusive means. Fick's work was inspired by the earlier experiments of Thomas Graham, which fell short of proposing the fundamental laws for which Fick would become famous. Fick's law is analogous to the relationships discovered at the same epoch by other eminent scientists: Darcy's law (hydraulic flow), Ohm's law (charge transport), and Fourier's Law (heat transport).
Fick's experiments (modeled on Graham's) dealt with measuring the concentrations and fluxes of salt, diffusing between two reservoirs through tubes of water. It is notable that Fick's work primarily concerned diffusion in fluids, because at the time, diffusion in solids was not considered generally possible. Today, Fick's Laws form the core of our understanding of diffusion in solids, liquids, and gases (in the absence of bulk fluid motion in the latter two cases). When a diffusion process does not follow Fick's laws (which happens in cases of diffusion through porous media and diffusion of swelling penetrants, among others), it is referred to as non-Fickian.
]]>In thermodynamics, Carnot's theorem, developed in 1824 by Nicolas Léonard Sadi Carnot, also called Carnot's rule, is a principle that specifies limits on the maximum efficiency that any heat engine can obtain.
Carnot's theorem states that all heat engines operating between the same two thermal or heat reservoirs can't have efficiencies greater than a reversible heat engine operating between the same reservoirs. A corollary of this theorem is that every reversible heat engine operating between a pair of heat reservoirs is equally efficient, regardless of the working substance employed or the operation details. Since a Carnot heat engine is also a reversible engine, the efficiency of all the reversible heat engines is determined as the efficiency of the Carnot heat engine that depends solely on the temperatures of its hot and cold reservoirs.
The maximum efficiency (i.e., the Carnot heat engine efficiency) of a heat engine operating between cold and hot reservoirs, denoted as H and C respectively, is the ratio of the temperature difference between the reservoirs to the hot reservoir temperature, expressed in the equation
Carnot's theorem is a consequence of the second law of thermodynamics. Historically, it was based on contemporary caloric theory, and preceded the establishment of the second law.
]]>Brewster's angle (also known as the polarization angle) is an angle of incidence at which light with a particular polarization is perfectly transmitted through a transparent dielectric surface, with no reflection. When unpolarized light is incident at this angle, the light that is reflected from the surface is therefore perfectly polarized. This special angle of incidence is named after the Scottish physicist Sir David Brewster (1781–1868).
Explanation
When light encounters a boundary between two media with different refractive indices, some of it is usually reflected as shown in the figure above. The fraction that is reflected is described by the Fresnel equations, and depends on the incoming light's polarization and angle of incidence.
The Fresnel equations predict that light with the p polarization (electric field polarized in the same plane as the incident ray and the surface normal at the point of incidence) will not be reflected if the angle of incidence is
where
is the refractive index of the initial medium through which the light propagates (the "incident medium"), and is the index of the other medium. This equation is known as Brewster's law, and the angle defined by it is Brewster's angle.The physical mechanism for this can be qualitatively understood from the manner in which electric dipoles in the media respond to p-polarized light. One can imagine that light incident on the surface is absorbed, and then re-radiated by oscillating electric dipoles at the interface between the two media. The polarization of freely propagating light is always perpendicular to the direction in which the light is travelling. The dipoles that produce the transmitted (refracted) light oscillate in the polarization direction of that light. These same oscillating dipoles also generate the reflected light. However, dipoles do not radiate any energy in the direction of the dipole moment. If the refracted light is p-polarized and propagates exactly perpendicular to the direction in which the light is predicted to be specularly reflected, the dipoles point along the specular reflection direction and therefore no light can be reflected.
With simple geometry this condition can be expressed as
,where
is the angle of reflection (or incidence) and is the angle of refraction.Using Snell's law,
one can calculate the incident angle
at which no light is reflected:.Solving for
givesThe phenomenon of light being polarized by reflection from a surface at a particular angle was first observed by Étienne-Louis Malus in 1808. He attempted to relate the polarizing angle to the refractive index of the material, but was frustrated by the inconsistent quality of glasses available at that time. In 1815, Brewster experimented with higher-quality materials and showed that this angle was a function of the refractive index, defining Brewster's law.
Brewster's angle is often referred to as the "polarizing angle", because light that reflects from a surface at this angle is entirely polarized perpendicular to the plane of incidence ("s-polarized"). A glass plate or a stack of plates placed at Brewster's angle in a light beam can, thus, be used as a polarizer. The concept of a polarizing angle can be extended to the concept of a Brewster wavenumber to cover planar interfaces between two linear bianisotropic materials. In the case of reflection at Brewster's angle, the reflected and refracted rays are mutually perpendicular.
For magnetic materials, Brewster's angle can exist for only one of the incident wave polarizations, as determined by the relative strengths of the dielectric permittivity and magnetic permeability.This has implications for the existence of generalized Brewster angles for dielectric metasurfaces.
]]>The Huygens–Fresnel principle (named after Dutch physicist Christiaan Huygens and French physicist Augustin-Jean Fresnel) states that every point on a wavefront is itself the source of spherical wavelets, and the secondary wavelets emanating from different points mutually interfere. The sum of these spherical wavelets forms a new wavefront. As such, the Huygens-Fresnel principle is a method of analysis applied to problems of luminous wave propagation both in the far-field limit and in near-field diffraction as well as reflection.
History
In 1678, Huygens proposed that every point reached by a luminous disturbance becomes a source of a spherical wave; the sum of these secondary waves determines the form of the wave at any subsequent time. He assumed that the secondary waves travelled only in the "forward" direction and it is not explained in the theory why this is the case. He was able to provide a qualitative explanation of linear and spherical wave propagation, and to derive the laws of reflection and refraction using this principle, but could not explain the deviations from rectilinear propagation that occur when light encounters edges, apertures and screens, commonly known as diffraction effects. The resolution of this error was finally explained by David A. B. Miller in 1991. The resolution is that the source is a dipole (not the monopole assumed by Huygens), which cancels in the reflected direction.
In 1818, Fresnel showed that Huygens's principle, together with his own principle of interference could explain both the rectilinear propagation of light and also diffraction effects. To obtain agreement with experimental results, he had to include additional arbitrary assumptions about the phase and amplitude of the secondary waves, and also an obliquity factor. These assumptions have no obvious physical foundation but led to predictions that agreed with many experimental observations, including the Poisson spot.
Poisson was a member of the French Academy, which reviewed Fresnel's work. He used Fresnel's theory to predict that a bright spot ought to appear in the center of the shadow of a small disc, and deduced from this that the theory was incorrect. However, Arago, another member of the committee, performed the experiment and showed that the prediction was correct. (Lisle had observed this fifty years earlier. This was one of the investigations that led to the victory of the wave theory of light over then predominant corpuscular theory.
In antenna theory and engineering, the reformulation of the Huygens–Fresnel principle for radiating current sources is known as surface equivalence principle.
Huygens' principle as a microscopic model
The Huygens–Fresnel principle provides a reasonable basis for understanding and predicting the classical wave propagation of light. However, there are limitations to the principle, namely the same approximations done for deriving the Kirchhoff's diffraction formula and the approximations of near field due to Fresnel. These can be summarized in the fact that the wavelength of light is much smaller than the dimensions of any optical components encountered.
Kirchhoff's diffraction formula provides a rigorous mathematical foundation for diffraction, based on the wave equation. The arbitrary assumptions made by Fresnel to arrive at the Huygens–Fresnel equation emerge automatically from the mathematics in this derivation.
A simple example of the operation of the principle can be seen when an open doorway connects two rooms and a sound is produced in a remote corner of one of them. A person in the other room will hear the sound as if it originated at the doorway. As far as the second room is concerned, the vibrating air in the doorway is the source of the sound.
Modern physics interpretations
Not all experts agree that the Huygens' principle is an accurate microscopic representation of reality. For instance, Melvin Schwartz argued that "Huygens' principle actually does give the right answer but for the wrong reasons".
This can be reflected in the following facts:
* The microscopic mechanics to create photons and of emission, in general, is essentially acceleration of electrons.
* The original analysis of Huygens included amplitudes only. It includes neither phases nor waves propagating at different speeds (due to diffraction within continuous media), and therefore does not take into account interference.
* The Huygens analysis also does not include polarization for light which imply a vector potential, where instead sound waves can be described with a scalar potential and there is no unique and natural translation between the two.
* In the Huygens description, there is no explanation of why we choose only the forward-going (retarded wave or forward envelope of wave fronts) versus the backward-propagating advanced wave (backward envelope).
* In the Fresnel approximation there is a concept of non-local behavior due to the sum of spherical waves with different phases that comes from the different points of the wave front, and non local theories are subject of many debates (e.g., not being Lorentz covariant) and of active research.
* The Fresnel approximation can be interpreted in a quantum probabilistic manner but is unclear how much this sum of states (i.e., wavelets on the wavefront) is a complete list of states that are meaningful physically or represents more of an approximation on a generic basis like in the linear combination of atomic orbitals (LCAO) method.
The Huygens' principle is essentially compatible with quantum field theory in the far field approximation, considering effective fields in the center of scattering, considering small perturbations, and in the same sense that quantum optics is compatible with classical optics, other interpretations are subject of debates and active research.
The Feynman model where every point in an imaginary wave front as large as the room is generating a wavelet, shall also be interpreted in these approximations and in a probabilistic context, in this context remote points can only contribute minimally to the overall probability amplitude.
Quantum field theory does not include any microscopic model for photon creation and the concept of single photon is also put under scrutiny on a theoretical level.
Mathematical expression of the principle
Consider the case of a point source located at a point
, vibrating at a frequency f. The disturbance may be described by a complex variable known as the complex amplitude. It produces a spherical wave with wavelength , wavenumber . Within a constant of proportionality, the complex amplitude of the primary wave at the point Q located at a distance from is:.]]>Bell's theorem is a term encompassing a number of closely related results in physics, all of which determine that quantum mechanics is incompatible with local hidden-variable theories given some basic assumptions about the nature of measurement. "Local" here refers to the principle of locality, the idea that a particle can only be influenced by its immediate surroundings, and that interactions mediated by physical fields can only occur at speeds no greater than the speed of light. "Hidden variables" are hypothetical properties possessed by quantum particles, properties that are undetectable but still affect the outcome of experiments. In the words of physicist John Stewart Bell, for whom this family of results is named, "If [a hidden-variable theory] is local it will not agree with quantum mechanics, and if it agrees with quantum mechanics it will not be local."
The term is broadly applied to a number of different derivations, the first of which was introduced by John Stewart Bell in a 1964 paper titled "On the Einstein Podolsky Rosen Paradox". Bell's paper was a response to a 1935 thought experiment that Albert Einstein, Boris Podolsky and Nathan Rosen proposed, arguing that quantum physics is an "incomplete" theory. By 1935, it was already recognized that the predictions of quantum physics are probabilistic. Einstein, Podolsky and Rosen presented a scenario that involves preparing a pair of particles such that the quantum state of the pair is entangled, and then separating the particles to an arbitrarily large distance. The experimenter has a choice of possible measurements that can be performed on one of the particles. When they choose a measurement and obtain a result, the quantum state of the other particle apparently collapses instantaneously into a new state depending upon that result, no matter how far away the other particle is. This suggests that either the measurement of the first particle somehow also interacted with the second particle at faster than the speed of light, or that the entangled particles had some unmeasured property which pre-determined their final quantum states before they were separated. Therefore, assuming locality, quantum mechanics must be incomplete, as it cannot give a complete description of the particle's true physical characteristics. In other words, quantum particles, like electrons and photons, must carry some property or attributes not included in quantum theory, and the uncertainties in quantum theory's predictions would then be due to ignorance or unknowability of these properties, later termed "hidden variables".
Bell carried the analysis of quantum entanglement much further. He deduced that if measurements are performed independently on the two separated particles of an entangled pair, then the assumption that the outcomes depend upon hidden variables within each half implies a mathematical constraint on how the outcomes on the two measurements are correlated. This constraint would later be named the Bell inequality. Bell then showed that quantum physics predicts correlations that violate this inequality. Consequently, the only way that hidden variables could explain the predictions of quantum physics is if they are "nonlocal", which is to say that somehow the two particles were able to interact instantaneously no matter how widely the two particles are separated.
Multiple variations on Bell's theorem were put forward in the following years, introducing other closely related conditions generally known as Bell (or "Bell-type") inequalities. The first rudimentary experiment designed to test Bell's theorem was performed in 1972 by John Clauser and Stuart Freedman. More advanced experiments, known collectively as Bell tests, have been performed many times since. Often, these experiments have had the goal of "closing loopholes", that is, ameliorating problems of experimental design or set-up that could in principle affect the validity of the findings of earlier Bell tests. To date, Bell tests have consistently found that physical systems obey quantum mechanics and violate Bell inequalities; which is to say that the results of these experiments are incompatible with any local hidden variable theory.
The exact nature of the assumptions required to prove a Bell-type constraint on correlations has been debated by physicists and by philosophers. While the significance of Bell's theorem is not in doubt, its full implications for the interpretation of quantum mechanics remain unresolved.
Additional Information
According to a scientist named Bell, for local hidden variables, there is no corporal concept that can propagate the quantum mechanics calculations.
The Physicist, John Stewart Bell brought up a theorem, and later on, it was named as Bell’s theorem. There are some microscopic properties present inside a particle known as hidden variables.
In general, for microscopes (existing microscopes), it is very difficult to keep an eye on them and study their behavior.
A Physicist known as Heisenberg stated the Uncertainty Principle. According to his principle, there is no existence of variables outside the context of observation.
Quantum mechanics had no scope in the inacceptable situation after EPR. The EPR quantum mechanics was imperfect in the logic that it botched to version for some rudiments of physical reality. Also, it has desecrated the principle of the finite propagation speed of physical effects.
Bell's Theorem Formula
A formula has been developed by John Stewart Bell to measure the subatomic phenomenon.
The formula is:
P (X = Y) + P (Y = Z) + P (Z = X) ≥ 1
Here,
X, Y, and Z = Photon measurement variables
Mathematically,
P (X = Y) is the probability of X = Y (applicable)
Bell’s Theorem Experiment
Two observers, commonly known as Alice and Bob, have taken part in the EPR assumed experiment. The EPR performs in liberated quantities of spin on a pair of electrons, equipped at a source in an unusual state known as a spin-singlet state.
Bell has found a definite inference of EPR. Alice calculated the spin in the x-direction along with Bob's measurement in the x-direction. It was measured with confidence whereas instantaneously before the measurement of Alice, the result of Bob resulted statistically merely.
Through the experiment, he depicted a statement which stated that “when the spin is in the x-direction but not an element of physical reality, the properties pass from Alice to Bob rapidly.
According to Bell, the entangled particles communicate with each other faster than light. If we believe about anything that can go faster than light, we are wrong as per scientists since nothing goes faster than light.
Bell’s Theorem Proof states about the particle getting twisted out.
What is Bell's Inequality?
We can learn and explain Bell’s inequality with the help of quantum mechanics. The behavior of electrons inside a magnetic field is quite interesting. We can study them with the help of quantum mechanics.
The result of the electron inside the magnetic field sets them apart as half of the electron goes towards the right side, and the other half of the electrons go towards the left.
Again, the electrons which are situated at the right side are sent towards another magnetic field, which is perpendicular to the left.
Also, they get separated in a different way that few of them go down, and few of them go up. This randomness of electrons can be studied with the help of Bell’s theorem.
]]>Fermi–Dirac statistics (F–D statistics) is a type of quantum statistics that applies to the physics of a system consisting of many non-interacting, identical particles that obey the Pauli exclusion principle. A result is the Fermi–Dirac distribution of particles over energy states. It is named after Enrico Fermi and Paul Dirac, each of whom derived the distribution independently in 1926 (although Fermi derived it before Dirac). Fermi–Dirac statistics is a part of the field of statistical mechanics and uses the principles of quantum mechanics.
F–D statistics applies to identical and indistinguishable particles with half-integer spin (1/2, 3/2, etc.), called fermions, in thermodynamic equilibrium. For the case of negligible interaction between particles, the system can be described in terms of single-particle energy states. A result is the F–D distribution of particles over these states where no two particles can occupy the same state, which has a considerable effect on the properties of the system. F–D statistics is most commonly applied to electrons, a type of fermion with spin 1/2.
A counterpart to F–D statistics is Bose–Einstein statistics (B–E statistics), which applies to identical and indistinguishable particles with integer spin (0, 1, 2, etc.) called bosons. In classical physics, Maxwell–Boltzmann statistics (M–B statistics) is used to describe particles that are identical and treated as distinguishable. For both B–E and M–B statistics, more than one particle can occupy the same state, unlike F–D statistics.
]]>In physics and electromagnetism, Gauss's law, also known as Gauss's flux theorem, (or sometimes simply called Gauss's theorem) is a law relating the distribution of electric charge to the resulting electric field. In its integral form, it states that the flux of the electric field out of an arbitrary closed surface is proportional to the electric charge enclosed by the surface, irrespective of how that charge is distributed. Even though the law alone is insufficient to determine the electric field across a surface enclosing any charge distribution, this may be possible in cases where symmetry mandates uniformity of the field. Where no such symmetry exists, Gauss's law can be used in its differential form, which states that the divergence of the electric field is proportional to the local density of charge.
The law was first formulated by Joseph-Louis Lagrange in 1773, followed by Carl Friedrich Gauss in 1835, both in the context of the attraction of ellipsoids. It is one of Maxwell's four equations, which forms the basis of classical electrodynamics. Gauss's law can be used to derive Coulomb's law, and vice versa.
Qualitative description
In words, Gauss's law states that
The net electric flux through any hypothetical closed surface is equal to
times the net electric charge enclosed within that closed surface. The closed surface is also referred to as Gaussian Surface.Gauss's law has a close mathematical similarity with a number of laws in other areas of physics, such as Gauss's law for magnetism and Gauss's law for gravity. In fact, any inverse-square law can be formulated in a way similar to Gauss's law: for example, Gauss's law itself is essentially equivalent to the inverse-square Coulomb's law, and Gauss's law for gravity is essentially equivalent to the inverse-square Newton's law of gravity.
The law can be expressed mathematically using vector calculus in integral form and differential form; both are equivalent since they are related by the divergence theorem, also called Gauss's theorem. Each of these forms in turn can also be expressed two ways: In terms of a relation between the electric field E and the total electric charge, or in terms of the electric displacement field D and the free electric charge.
]]>The Curie–Weiss law describes the magnetic susceptibility
of a ferromagnet in the paramagnetic region above the Curie point:where C is a material-specific Curie constant, T is the absolute temperature, and
is the Curie temperature, both measured in kelvin. The law predicts a singularity in the susceptibility at . Below this temperature, the ferromagnet has a spontaneous magnetization. The name is given after Pierre Curie and Pierre-Ernest Weiss.Brief summary of related concepts
The magnetic moment of a magnet is a quantity that determines the torque it will experience in an external magnetic field. A loop of electric current, a bar magnet, an electron, a molecule, and a planet all have magnetic moments.
The magnetization or magnetic polarization of a magnetic material is the vector field that expresses the density of permanent or induced magnetic moments. The magnetic moments can originate from microscopic electric currents caused by the motion of electrons in individual atoms, or the spin of the electrons or the nuclei. Net magnetization results from the response of a material to an external magnetic field, together with any unbalanced magnetic moment that may be present even in the absence of the external magnetic field, for example, in sufficiently cold iron. The latter is called spontaneous magnetization. Other materials that share this property with iron, like Nickel and magnetite, are called ferromagnets. The threshold temperature below which a material is ferromagnetic is called the Curie temperature and varies between materials.
Limitations
In many materials, the Curie–Weiss law fails to describe the susceptibility in the immediate vicinity of the Curie point, since it is based on a mean-field approximation. Instead, there is a critical behavior of the form
with the critical exponent γ. However, at temperatures
the expression of the Curie–Weiss law still holds true, but with replaced by a temperature Θ that is somewhat higher than the actual Curie temperature. Some authors call Θ the Weiss constant to distinguish it from the temperature of the actual Curie point.]]>Graham's law of effusion (also called Graham's law of diffusion) was formulated by Scottish physical chemist Thomas Graham in 1848. Graham found experimentally that the rate of effusion of a gas is inversely proportional to the square root of the molar mass of its particles. This formula can be written as:
where:
is the rate of effusion for the first gas. (volume or number of moles per unit time).Graham's law states that the rate of diffusion or of effusion of a gas is inversely proportional to the square root of its molecular weight. Thus, if the molecular weight of one gas is four times that of another, it would diffuse through a porous plug or escape through a small pinhole in a vessel at half the rate of the other (heavier gases diffuse more slowly). A complete theoretical explanation of Graham's law was provided years later by the kinetic theory of gases. Graham's law provides a basis for separating isotopes by diffusion—a method that came to play a crucial role in the development of the atomic bomb.
Graham's law is most accurate for molecular effusion which involves the movement of one gas at a time through a hole. It is only approximate for diffusion of one gas in another or in air, as these processes involve the movement of more than one gas.
In the same conditions of temperature and pressure, the molar mass is proportional to the mass density. Therefore, the rates of diffusion of different gases are inversely proportional to the square roots of their mass densities.
.]]>In optics, Lambert's cosine law says that the radiant intensity or luminous intensity observed from an ideal diffusely reflecting surface or ideal diffuse radiator is directly proportional to the cosine of the angle
between the direction of the incident light and the surface normal; . The law is also known as the cosine emission law or Lambert's emission law. It is named after Johann Heinrich Lambert, from his Photometria, published in 1760.A surface which obeys Lambert's law is said to be Lambertian, and exhibits Lambertian reflectance. Such a surface has the same radiance when viewed from any angle. This means, for example, that to the human eye it has the same apparent brightness (or luminance). It has the same radiance because, although the emitted power from a given area element is reduced by the cosine of the emission angle, the solid angle, subtended by surface visible to the viewer, is reduced by the very same amount. Because the ratio between power and solid angle is constant, radiance (power per unit solid angle per unit projected source area) stays the same.
]]>Hagen–Poiseuille equation
In nonideal fluid dynamics, the Hagen–Poiseuille equation, also known as the Hagen–Poiseuille law, Poiseuille law or Poiseuille equation, is a physical law that gives the pressure drop in an incompressible and Newtonian fluid in laminar flow flowing through a long cylindrical pipe of constant cross section. It can be successfully applied to air flow in lung alveoli, or the flow through a drinking straw or through a hypodermic needle. It was experimentally derived independently by Jean Léonard Marie Poiseuille in 1838 and Gotthilf Heinrich Ludwig Hagen, and published by Poiseuille in 1840–41 and 1846. The theoretical justification of the Poiseuille law was given by George Stokes in 1845.
The assumptions of the equation are that the fluid is incompressible and Newtonian; the flow is laminar through a pipe of constant circular cross-section that is substantially longer than its diameter; and there is no acceleration of fluid in the pipe. For velocities and pipe diameters above a threshold, actual fluid flow is not laminar but turbulent, leading to larger pressure drops than calculated by the Hagen–Poiseuille equation.
Poiseuille's equation describes the pressure drop due to the viscosity of the fluid; other types of pressure drops may still occur in a fluid (see a demonstration here). For example, the pressure needed to drive a viscous fluid up against gravity would contain both that as needed in Poiseuille's law plus that as needed in Bernoulli's equation, such that any point in the flow would have a pressure greater than zero (otherwise no flow would happen).
Another example is when blood flows into a narrower constriction, its speed will be greater than in a larger diameter (due to continuity of volumetric flow rate), and its pressure will be lower than in a larger diameter (due to Bernoulli's equation). However, the viscosity of blood will cause additional pressure drop along the direction of flow, which is proportional to length traveled[4] (as per Poiseuille's law). Both effects contribute to the actual pressure drop.
Equation
In standard fluid-kinetics notation:
where:
is the pressure difference between the two ends,The equation does not hold close to the pipe entrance.:
The equation fails in the limit of low viscosity, wide and/or short pipe. Low viscosity or a wide pipe may result in turbulent flow, making it necessary to use more complex models, such as the Darcy–Weisbach equation. The ratio of length to radius of a pipe should be greater than one forty-eighth of the Reynolds number for the Hagen–Poiseuille law to be valid. If the pipe is too short, the Hagen–Poiseuille equation may result in unphysically high flow rates; the flow is bounded by Bernoulli's principle, under less restrictive conditions, by
because it is impossible to have negative (absolute) pressure (not to be confused with gauge pressure) in an incompressible flow.
]]>