Hubbry Logo
OrthogonalityOrthogonalityMain
Open search
Orthogonality
Community hub
Orthogonality
logo
8 pages, 0 posts
0 subscribers
Be the first to start a discussion here.
Be the first to start a discussion here.
Orthogonality
Orthogonality
from Wikipedia

Orthogonality is a term with various meanings depending on the context.

The line segments AB and CD are perpendicular to each other.

In mathematics, orthogonality is the generalization of the geometric notion of perpendicularity. Although many authors use the two terms perpendicular and orthogonal interchangeably, the term perpendicular is more specifically used for lines and planes that intersect to form a right angle, whereas orthogonal is used in generalizations, such as orthogonal vectors or orthogonal curves.[1][2]

The term is also used in other fields like physics, art, computer science, statistics, and economics.

Etymology

[edit]

The word comes from the Ancient Greek ὀρθός (orthós), meaning "upright",[3] and γωνία (gōnía), meaning "angle".[4]

The Ancient Greek ὀρθογώνιον (orthogṓnion) and Classical Latin orthogonium originally denoted a rectangle.[5] Later, they came to mean a right triangle. In the 12th century, the post-classical Latin word orthogonalis came to mean a right angle or something related to a right angle.[6]

Mathematics

[edit]

In mathematics, orthogonality is the generalization of the geometric notion of perpendicularity to linear algebra of bilinear forms.

Two elements u and v of a vector space with bilinear form are orthogonal when . Depending on the bilinear form, the vector space may contain null vectors, non-zero self-orthogonal vectors, in which case perpendicularity is replaced with hyperbolic orthogonality.

In the case of function spaces, families of functions are used to form an orthogonal basis, such as in the contexts of orthogonal polynomials, orthogonal functions, and combinatorics.

Orthogonality and rotation of coordinate systems compared between left: Euclidean space through circular angle ϕ, right: in Minkowski spacetime through hyperbolic angle ϕ (red lines labelled c denote the worldlines of a light signal, a vector is orthogonal to itself if it lies on this line).[7]

Physics

[edit]

Optics

[edit]

In optics, polarization states are said to be orthogonal when they propagate independently of each other, as in vertical and horizontal linear polarization or right-handed and left-handed circular polarization.

Special relativity

[edit]

In special relativity, a time axis determined by a rapidity of motion is hyperbolic-orthogonal to a space axis of simultaneous events, also determined by the rapidity. The theory features relativity of simultaneity.

Hyperbolic orthogonality

[edit]
Euclidean orthogonality is preserved by rotation in the left diagram; hyperbolic orthogonality with respect to hyperbola (B) is preserved by hyperbolic rotation in the right diagram.

In geometry, given a pair of conjugate hyperbolas, two conjugate diameters are hyperbolically orthogonal. This relationship of diameters was described by Apollonius of Perga and has been modernized using analytic geometry. Hyperbolically orthogonal lines appear in special relativity as temporal and spatial directions that show the relativity of simultaneity.

Keeping time and space axes hyperbolically orthogonal, as in Minkowski space, gives a constant result when measurements are taken of the speed of light.

Quantum mechanics

[edit]

In quantum mechanics, a sufficient (but not necessary) condition that two eigenstates of a Hermitian operator, and , are orthogonal is that they correspond to different eigenvalues. This means, in Dirac notation, that if and correspond to different eigenvalues. This follows from the fact that Schrödinger's equation is a Sturm–Liouville equation (in Schrödinger's formulation) or that observables are given by Hermitian operators (in Heisenberg's formulation).[citation needed]

Art

[edit]

In art, the perspective (imaginary) lines pointing to the vanishing point are referred to as "orthogonal lines". The term "orthogonal line" often has a quite different meaning in the literature of modern art criticism. Many works by painters such as Piet Mondrian and Burgoyne Diller are noted for their exclusive use of "orthogonal lines" — not, however, with reference to perspective, but rather referring to lines that are straight and exclusively horizontal or vertical, forming right angles where they intersect. For example, an essay of the Thyssen-Bornemisza Museum states that "Mondrian [...] dedicated his entire oeuvre to the investigation of the balance between orthogonal lines and primary colours."[8]

Computer science

[edit]

Orthogonality in programming language design is the ability to use various language features in arbitrary combinations with consistent results.[9] This usage was introduced by Van Wijngaarden in the design of Algol 68:

The number of independent primitive concepts has been minimized in order that the language be easy to describe, to learn, and to implement. On the other hand, these concepts have been applied “orthogonally” in order to maximize the expressive power of the language while trying to avoid deleterious superfluities.[10]

Orthogonality is a system design property which guarantees that modifying the technical effect produced by a component of a system neither creates nor propagates side effects to other components of the system. Typically this is achieved through the separation of concerns and encapsulation, and it is essential for feasible and compact designs of complex systems. The emergent behavior of a system consisting of components should be controlled strictly by formal definitions of its logic and not by side effects resulting from poor integration, i.e., non-orthogonal design of modules and interfaces. Orthogonality reduces testing and development time because it is easier to verify designs that neither cause side effects nor depend on them.

Orthogonal instruction set

[edit]

An instruction set is said to be orthogonal if it lacks redundancy (i.e., there is only a single instruction that can be used to accomplish a given task)[11] and is designed such that instructions can use any register in any addressing mode. This terminology results from considering an instruction as a vector whose components are the instruction fields. One field identifies the registers to be operated upon and another specifies the addressing mode. An orthogonal instruction set uniquely encodes all combinations of registers and addressing modes.[12]

Telecommunications

[edit]

In telecommunications, multiple access schemes are orthogonal when an ideal receiver can completely reject arbitrarily strong unwanted signals from the desired signal using different basis functions. One such scheme is time-division multiple access (TDMA), where the orthogonal basis functions are nonoverlapping rectangular pulses ("time slots").

Orthogonal frequency-division multiplexing

[edit]

Another scheme is orthogonal frequency-division multiplexing (OFDM), which refers to the use, by a single transmitter, of a set of frequency multiplexed signals with the exact minimum frequency spacing needed to make them orthogonal so that they do not interfere with each other. Well known examples include (a, g, and n) versions of 802.11 Wi-Fi; WiMAX; ITU-T G.hn, DVB-T, the terrestrial digital TV broadcast system used in most of the world outside North America; and DMT (Discrete Multi Tone), the standard form of ADSL.

In OFDM, the subcarrier frequencies are chosen[how?] so that the subcarriers are orthogonal to each other, meaning that crosstalk between the subchannels is eliminated and intercarrier guard bands are not required. This greatly simplifies the design of both the transmitter and the receiver. In conventional FDM, a separate filter for each subchannel is required.

Statistics, econometrics, and economics

[edit]

When performing statistical analysis, independent variables that affect a particular dependent variable are said to be orthogonal if they are uncorrelated,[13] since the covariance forms an inner product. In this case the same results are obtained for the effect of any of the independent variables upon the dependent variable, regardless of whether one models the effects of the variables individually with simple regression or simultaneously with multiple regression. If correlation is present, the factors are not orthogonal and different results are obtained by the two methods. This usage arises from the fact that if centered by subtracting the expected value (the mean), uncorrelated variables are orthogonal in the geometric sense discussed above, both as observed data (i.e., vectors) and as random variables (i.e., density functions). One econometric formalism that is alternative to the maximum likelihood framework, the Generalized Method of Moments, relies on orthogonality conditions. In particular, the Ordinary Least Squares estimator may be easily derived from an orthogonality condition between the explanatory variables and model residuals.

Taxonomy

[edit]

In taxonomy, an orthogonal classification is one in which no item is a member of more than one group, that is, the classifications are mutually exclusive.

Chemistry and biochemistry

[edit]

In chemistry and biochemistry, an orthogonal interaction occurs when there are two pairs of substances and each substance can interact with their respective partner, but does not interact with either substance of the other pair. For example, DNA has two orthogonal pairs: cytosine and guanine form a base-pair, and adenine and thymine form another base-pair, but other base-pair combinations are strongly disfavored. As a chemical example, tetrazine reacts with transcyclooctene and azide reacts with cyclooctyne without any cross-reaction, so these are mutually orthogonal reactions, and so, can be performed simultaneously and selectively.[14]

Organic synthesis

[edit]

In organic synthesis, orthogonal protection is a strategy allowing the deprotection of functional groups independently of each other.

Bioorthogonal chemistry

[edit]
The term bioorthogonal chemistry refers to any chemical reaction that can occur inside of living systems without interfering with native biochemical processes.[15][16][17] The term was coined by Carolyn R. Bertozzi in 2003.[18][19] Since its introduction, the concept of the bioorthogonal reaction has enabled the study of biomolecules such as glycans, proteins,[20] and lipids[21] in real time in living systems without cellular toxicity. A number of chemical ligation strategies have been developed that fulfill the requirements of bioorthogonality, including the 1,3-dipolar cycloaddition between azides and cyclooctynes (also termed copper-free click chemistry),[22] between nitrones and cyclooctynes,[23] oxime/hydrazone formation from aldehydes and ketones,[24] the tetrazine ligation,[25] the isocyanide-based click reaction,[26] and most recently, the quadricyclane ligation.[27]

Supramolecular chemistry

[edit]

In supramolecular chemistry the notion of orthogonality refers to the possibility of two or more supramolecular, often non-covalent, interactions being compatible; reversibly forming without interference from the other.

Analytical chemistry

[edit]

In analytical chemistry, analyses are "orthogonal" if they make a measurement or identification in completely different ways, thus increasing the reliability of the measurement. Orthogonal testing thus can be viewed as "cross-checking" of results, and the "cross" notion corresponds to the etymologic origin of orthogonality. Orthogonal testing is often required as a part of a new drug application.

System reliability

[edit]

In the field of system reliability orthogonal redundancy is that form of redundancy where the form of backup device or method is completely different from the prone to error device or method. The failure mode of an orthogonally redundant back-up device or method does not intersect with and is completely different from the failure mode of the device or method in need of redundancy to safeguard the total system against catastrophic failure.

Neuroscience

[edit]

In neuroscience, a sensory map in the brain which has overlapping stimulus coding (e.g. location and quality) is called an orthogonal map.

Philosophy

[edit]

In philosophy, two topics, authors, or pieces of writing are said to be "orthogonal" to each other when they do not substantively cover what could be considered potentially overlapping or competing claims. Thus, texts in philosophy can either support and complement one another, they can offer competing explanations or systems, or they can be orthogonal to each other in cases where the scope, content, and purpose of the pieces of writing are entirely unrelated.[example needed]

Gaming

[edit]

In board games such as chess which feature a grid of squares, 'orthogonal' is used to mean "in the same row/'rank' or column/'file'". This is the counterpart to squares which are "diagonally adjacent".[28] In the ancient Chinese board game Go a player can capture the stones of an opponent by occupying all orthogonally adjacent points.

Law

[edit]

In law, orthogonality can refer to interests in a proceeding that are not aligned, but also bear no correlation or effect on each other, so as not to create a conflict of interest.

Other examples

[edit]

Stereo vinyl records encode both the left and right stereo channels in a single groove. The V-shaped groove in the vinyl has walls that are 90 degrees to each other, with variations in each wall separately encoding one of the two analogue channels that make up the stereo signal. The cartridge senses the motion of the stylus following the groove in two orthogonal directions: 45 degrees from vertical to either side.[29] A pure horizontal motion corresponds to a mono signal, equivalent to a stereo signal in which both channels carry identical (in-phase) signals.

See also

[edit]

References

[edit]
Revisions and contributorsEdit on WikipediaRead on Wikipedia
from Grokipedia
Orthogonality is a fundamental geometric and algebraic concept denoting the perpendicularity of two lines, planes, or curves at their intersection point, or more abstractly, the independence between elements in a vector space or inner product space where their inner product equals zero. In Euclidean geometry, orthogonal lines form a 90-degree angle, a property central to coordinate systems and projections. This idea generalizes to higher dimensions and non-Euclidean settings, underpinning theorems in linear algebra and analysis. In linear algebra, two vectors u\mathbf{u} and v\mathbf{v} in Rn\mathbb{R}^n are orthogonal if their uv=0\mathbf{u} \cdot \mathbf{v} = 0, implying they are when visualized in space. A set of vectors is orthogonal if every pair is orthogonal, and orthonormal if additionally each has unit length; such bases simplify computations like projections and decompositions. Orthogonal complements of subspaces partition the ambient space, with applications in solving systems of equations and problems. In physics, orthogonality manifests in coordinate systems where axes are mutually perpendicular, facilitating calculations in and . In quantum mechanics, wavefunctions are orthogonal if their integral product vanishes, signifying incompatible or mutually exclusive states, a cornerstone for superposition and measurement postulates. This extends to eigenfunctions of commuting operators, ensuring non-interfering observables. Beyond and physics, orthogonality informs design principles in , where it describes systems or languages allowing independent feature combinations without unintended interactions—for instance, in instruction sets or programming paradigms where modifying one aspect does not affect others. In , orthogonal variables or contrasts are uncorrelated, enabling efficient experimental designs and variance decomposition. These applications highlight orthogonality's role in promoting simplicity, stability, and modularity across disciplines.

Origins and General Concept

Etymology

The term "orthogonality" originates from , combining ὀρθός (orthós), meaning "straight," "right," or "upright," with γωνία (gōnía), meaning "angle," to literally denote "right-angled." This etymological root reflects the geometric notion of perpendicularity at its core. The adjective form "orthogonal" evolved through orthogōnālis and orthogōnius, both signifying "right-angled," before entering as orthogonal in the sense of pertaining to right angles. In English, "orthogonal" first appeared in 1571 in a mathematical context, in Thomas Digges's "A Geometrical Practise, named Pantometria," where it described right angles. The noun "orthogonality," specifically denoting the property or state of being orthogonal, emerged later in the , with its earliest documented use in 1872 within Philosophical Transactions of the Royal Society, marking a shift toward more abstract mathematical applications. Although the concept of right angles—implicitly orthogonal—was foundational in (c. 300 BCE), where lines were defined as those forming equal adjacent angles, the Greek term itself was absent; Euclid relied on descriptive phrases rather than the compound word. By the 19th century, mathematicians like incorporated "orthogonal" into advanced geometric frameworks, such as in his 1827 Disquisitiones generales circa superficies curvas, where he discussed orthogonal coordinate systems on curved surfaces, establishing the term's modern mathematical connotation. In the , "orthogonal" extended beyond to non-spatial senses, such as unrelatedness in and independence in , influenced by its perpendicular origin but applied to abstract structures like vector spaces and functions.

Fundamental Principles

Orthogonality serves as a foundational relation in abstract mathematical and scientific contexts, denoting a form of or non-interference between elements, where their interactions yield a null effect under a defined metric. This concept generalizes the geometric idea of perpendicularity, extending it to diverse structures beyond physical lines or planes, such that two elements are orthogonal if they do not influence or overlap in their contributions to a system. In essence, orthogonality embodies , ensuring that the properties or behaviors of one element remain unaltered by the presence or variation of another. A key property of orthogonality is its promotion of decomposability and simplicity within complex systems, as orthogonal elements can be analyzed or modified independently without propagating effects across the whole. This mirrors the behavior of perpendicular lines in , which intersect at a but maintain distinct directions thereafter, providing a intuitive analogy for the abstract principle. Such facilitates efficient representations and computations, as seen in the of bases or frameworks where orthogonal components span the without . Understanding orthogonality requires a preliminary grasp of elements like vectors within a structured , where the space defines the framework for assessing relations such as alignment or separation. In and scientific , this principle manifests in non-interfering components—for instance, modular systems where altering one subsystem leaves others unaffected—enhancing reliability and across disciplines. This broad applicability underscores orthogonality's role as a prerequisite for more specialized interpretations in various fields.

Mathematics

Geometric and Vector Orthogonality

In Euclidean geometry, two lines are orthogonal if they intersect at a right angle of 90 degrees, and this concept extends to planes that intersect such that their normal vectors are perpendicular. Orthogonality in this context captures the idea of perpendicularity, fundamental to constructing geometric figures like rectangles and cubes. For vectors in Euclidean space, two vectors u\vec{u}
Add your contribution
Related Hubs
User Avatar
No comments yet.