Hubbry Logo
Stereophonic soundStereophonic soundMain
Open search
Stereophonic sound
Community hub
Stereophonic sound
logo
8 pages, 0 posts
0 subscribers
Be the first to start a discussion here.
Be the first to start a discussion here.
Stereophonic sound
Stereophonic sound
from Wikipedia

see caption
How stereophonic sound systems work. The main diagram shows a simplified situation in nature. The inset shows the electronic simulation. Notably, such electronic systems require more than one speaker.
Time difference in a stereophonic recording of a car going past

Stereophonic sound, commonly shortened to stereo, is a method of sound reproduction that recreates a multi-directional, 3-dimensional audible perspective. This is usually achieved by using two independent audio channels through a configuration of two loudspeakers (or stereo headphones) in such a way as to create the impression of sound heard from various directions, as in natural hearing.

Because the multi-dimensional perspective is the crucial aspect, the term stereophonic also applies to systems with more than two channels or speakers, such as quadraphonic and surround sound. Binaural sound systems are also stereophonic.

Stereo sound has been in common use since the 1970s in entertainment media such as broadcast radio, recorded music, television, video cameras, cinema, computer audio, and the Internet.

Etymology

[edit]

The word stereophonic derives from the Greek στερεός (stereós, "firm, solid")[1] + φωνή (phōnḗ, "sound, tone, voice").[2]

Description

[edit]
Two microphones set up to record a piano simultaneously, which creates a stereo sound

Stereo sound systems can be divided into two forms: the first is true or natural stereo, in which a live sound is captured, with any natural reverberation present, by an array of microphones. The signal is then reproduced over multiple loudspeakers to recreate, as closely as possible, the live sound.

Secondly artificial or pan stereo, in which a single-channel (mono) sound is reproduced over multiple loudspeakers. By varying the relative amplitude of the signal sent to each speaker, an artificial direction (relative to the listener) can be suggested. The control that is used to vary this relative amplitude of the signal is known as a pan-pot (panoramic potentiometer). By combining multiple pan-potted mono signals, a complete, yet entirely artificial, sound field can be created.

In technical usage, true stereo means sound recording and sound reproduction that uses stereographic projection to encode the relative positions of objects and events recorded.[citation needed]

During two-channel stereo recording, two microphones are placed in strategically chosen locations relative to the sound source, with both recording simultaneously. The two recorded channels will be similar, but each will have distinct time-of-arrival and sound-pressure-level information. During playback, the listener's brain uses those subtle differences in timing and sound level to triangulate the positions of the recorded objects. Since each microphone records each wavefront at a slightly different time, the wavefronts are out of phase; as a result, constructive and destructive interference can occur if both tracks are played back on the same speaker. This phenomenon is known as phase cancellation. Coincident-pair microphone arrangements produce stereo recordings with minimal phase difference between channels.[3]

History

[edit]
Diagram of Clément Ader's théatrophone prototype at the Opera during the World Exhibition in Paris (1881)

Early work

[edit]
"Wandering Dragon Plays with Phoenix" Part One of Twelve, possibly the earliest accidental stereo, made as a field recording by Berthold Laufer for Franz Boaz in 1901

Clément Ader demonstrated the first two-channel audio system in Paris in 1881, with a series of telephone transmitters connected from the stage of the Paris Opera to a suite of rooms at the Paris Electrical Exhibition, where listeners could hear a live transmission of performances through receivers for each ear. Scientific American reported:

Every one who has been fortunate enough to hear the telephones at the Palais de l'Industrie has remarked that, in listening with both ears at the two telephones, the sound takes a special character of relief and localization which a single receiver cannot produce... This phenomenon is very curious, it approximates to the theory of binauricular audition, and has never been applied, we believe, before to produce this remarkable illusion to which may almost be given the name of auditive perspective.[4]

This two-channel telephonic process was commercialized in France from 1890 to 1932 as the Théâtrophone, and in England from 1895 to 1925 as the Electrophone. Both were services available by coin-operated receivers at hotels and cafés, or by subscription to private homes.[5]

There have been cases in which two recording lathes (for the sake of producing two simultaneous masters) were fed from two separate microphones; when both masters survive, modern engineers have been able to synchronize them to produce stereo recordings from a time before intentional stereophonic recording technology existed.[6]

Modern stereophonic sound

[edit]

Modern stereophonic technology was invented in the 1930s by British engineer Alan Blumlein at EMI, who patented stereo records, stereo films, and also surround sound.[7] In early 1931, Blumlein and his wife were at a local cinema. The sound reproduction systems of the early talkies invariably only had a single set of speakers – which could lead to the somewhat disconcerting effect of the actor being on one side of the screen whilst his voice appeared to come from the other. Blumlein declared to his wife that he had found a way to make the sound follow the actor across the screen. The genesis of these ideas is uncertain, but he explained them to Isaac Shoenberg in the late summer of 1931. His earliest notes on the subject are dated September 25, 1931, and his patent had the title "Improvements in and relating to Sound-transmission, Sound-recording and Sound-reproducing Systems". The application was dated December 14, 1931, and was accepted on June 14, 1933, as UK patent number 394,325.[8] The patent covered many ideas in stereo, some of which are used today and some not. Some 70 claims include:

  • A shuffling circuit, which aimed to preserve the directional effect when sound from a spaced pair of microphones was reproduced via stereo headphones instead of a pair of loudspeakers;
  • The use of a coincident pair of velocity microphones with their axes at right angles to each other, which is still known as a Blumlein pair;
  • Recording two channels in the single groove of a record using the two groove walls at right angles to each other and 45 degrees to the vertical;
  • A stereo disc-cutting head;
  • Using hybrid transformers to matrix between left and right signals and sum and difference signals;

Blumlein began binaural experiments as early as 1933, and the first stereo discs were cut later the same year, twenty-five years before that method became the standard for stereo phonograph discs. These discs used the two walls of the groove at right angles in order to carry the two channels. In 1934, Blumlein recorded Mozart's Jupiter Symphony conducted by Sir Thomas Beecham at Abbey Road Studios in London using his vertical-lateral technique.[7] Much of the development work on this system for cinematic use did not reach completion until 1935. In Blumlein's short test films (most notably, "Trains at Hayes Station", which lasts 5 minutes 11 seconds, and, "The Walking & Talking Film"), his original intent of having the sound follow the actor was fully realized.[9]

In the United States, Harvey Fletcher of Bell Laboratories was also investigating techniques for stereophonic recording and reproduction. One of the techniques investigated was the wall of sound, which used an enormous array of microphones hung in a line across the front of an orchestra. Up to 80 microphones were used, and each fed a corresponding loudspeaker, placed in an identical position, in a separate listening room. Several stereophonic test recordings, using two microphones connected to two styli cutting two separate grooves on the same wax disc, were made with Leopold Stokowski and the Philadelphia Orchestra at Philadelphia's Academy of Music in March 1932. The first (made on March 12, 1932), of Scriabin's Prometheus: Poem of Fire, is the earliest known surviving intentional stereo recording.[10] The performance was part of an all-Russian program including Mussorgsky's Pictures at an Exhibition in the Ravel orchestration, excerpts of which were also recorded in stereo.[11]

Bell Laboratories gave a demonstration of three-channel stereophonic sound on April 27, 1933, with a live transmission of the Philadelphia Orchestra from Philadelphia to Constitution Hall in Washington, D.C. over multiple Class A telephone lines. Leopold Stokowski, normally the orchestra's conductor, was present in Constitution Hall to control the sound mix. Five years later, the same system would be expanded onto multichannel film recording and used from the concert hall in Philadelphia to the recording labs at Bell Labs in New Jersey in order to record Walt Disney's Fantasia (1940) in what Disney called Fantasound.[citation needed]

Bell Labs binaural demonstration at World's Fair

Later that same year, Bell Labs also demonstrated binaural sound, at the Chicago World's Fair in 1933 using a dummy with microphones instead of ears.[12] The two signals were sent out over separate AM station bands.[13]

Carnegie Hall demonstration

[edit]

Utilizing selections recorded by the Philadelphia Orchestra, under the direction of Leopold Stokowski, intended for but not used in Walt Disney's Fantasia, the Carnegie Hall demonstration by Bell Laboratories on April 9 and 10, 1940, used three huge speaker systems. Synchronization was achieved by making the recordings in the form of three motion picture soundtracks recorded on a single piece of film with a fourth track being used to regulate volume expansion. This was necessary due to the limitations of dynamic range on optical motion picture film of the period; however, the volume compression and expansion were not fully automatic, but were designed to allow manual studio enhancement; i.e., the artistic adjustment of overall volume and the relative volume of each track in relation to the others. Stokowski, who was always interested in sound reproduction technology, personally participated in the enhancement of the sound at the demonstration.

The speakers produced sound levels of up to 100 decibels, and the demonstration held the audience "spellbound, and at times not a little terrified", according to one report.[14] Sergei Rachmaninoff, who was present at the demonstration, commented that it was "marvellous" but "somehow unmusical because of the loudness." "Take that Pictures at an Exhibition", he said. "I didn't know what it was until they got well into the piece. Too much 'enhancing', too much Stokowski."

Motion picture era

[edit]

In 1937, Bell Laboratories in New York City gave a demonstration of two-channel stereophonic motion pictures, developed by Bell Labs and Electrical Research Products, Inc.[15] Once again, conductor Leopold Stokowski was on hand to try out the new technology, recording onto a special proprietary nine-track sound system at the Academy of Music in Philadelphia, during the making of the movie One Hundred Men and a Girl for Universal Pictures in 1937, after which the tracks were mixed down to one for the final soundtrack.[16][17] A year later, MGM started using three tracks instead of one to record the musical selections of movie soundtracks, and very quickly upgraded to four. One track was used for dialogue, two for music, and one for sound effects. The very first two-track recording MGM made (although released in mono) was "It Never Rains But What It Pours" by Judy Garland, recorded on June 21, 1938, for the movie Love Finds Andy Hardy.

In the early 1940s, composer-conductor Alfred Newman directed the construction of a sound stage equipped for multichannel recording for 20th Century Fox studios. Several soundtracks from this era still exist in their multichannel elements, some of which have been released on DVD, including How Green Was My Valley, Anna and the King of Siam, The Day the Earth Stood Still and Sun Valley Serenade which, along with Orchestra Wives, feature the only stereophonic recordings of the Glenn Miller Orchestra as it was during its heyday of the Swing Era.

Fantasound

[edit]

Walt Disney began experimenting with multichannel sound in the early 1930s as noted above.[18] The first commercial motion picture to be exhibited with stereophonic sound was Walt Disney's Fantasia, released in November 1940, for which a specialized sound process (Fantasound) was developed. As in the Carnegie Hall demonstrations six months earlier, Fantasound used a separate film containing four optical soundtracks. Three of the tracks were used to carry left, center and right audio, while the fourth track carried three tones which individually controlled the volume level of the other three.[19][20] The film was not initially a financial success, however, after two months of road-show exhibition in selected cities, its soundtrack was remixed into mono sound for general release. It was not until its 1956 re-release that stereo sound was restored to the film.

Cinerama

[edit]

A Cinerama demonstration film by Lowell Thomas and Mike Todd titled This is Cinerama was released on September 30, 1952. The format was a widescreen process featuring three separate 35 mm motion picture films plus a separate sound film running in synchronization with one another at 26 fps, adding one picture panel each to the viewer's left and right at 45-degree angles, in addition to the usual front and center panel.

The Cinerama audio soundtrack technology, developed by Hazard E. Reeves, utilized seven discrete sound tracks on full-coat magnetic 35 mm film. The system featured five main channels behind the screen, two surround channels in the rear of the theater, plus a sync track to interlock the four machines, which were specially outfitted with aircraft servo-motors made by Ampex.

The advent of multitrack magnetic tape and film recording of this nature made high-fidelity synchronized multichannel recording more technically straightforward, though costly. By the early 1950s, all of the major studios were recording on 35 mm magnetic film for mixing purposes, and many of these so-called individual angles still survive, allowing for soundtracks to be remixed into stereo or even surround.

In April 1953, while This is Cinerama was still playing only in New York City, most moviegoing audiences heard stereophonic sound for the first time with House of Wax, an early 3-D film starring Vincent Price and produced by Warner Bros. Unlike the 4-track mag release-print stereo films of the period which featured four thin strips of magnetic material running down the length of the film, inside and outside the sprocket holes, the sound system developed for House of Wax, dubbed WarnerPhonic, was a combination of a 35 mm fully coated magnetic film that contained the audio tracks for left, center and right speakers, interlocked with the two dual-strip Polaroid system projectors, one of which carried a mono optical surround track and one that carried a mono backup track use in the event anything should go wrong.

Only two other films featured this unique hybrid WarnerPhonic sound: the 3-D production of The Charge at Feather River, and Island in the Sky. Unfortunately, as of 2012, the stereo magnetic tracks to both these films are considered lost forever. In addition, a large percentage of 3-D films carried variations on three-track magnetic sound: It Came from Outer Space; I, the Jury; The Stranger Wore a Gun; Inferno; Kiss Me, Kate; and many others.

Widescreen

[edit]

Inspired by Cinerama, the movie industry moved quickly to create simpler and cheaper widescreen systems, the first of which, Todd-AO, was developed by Broadway promoter Michael Todd with financial backing from Rodgers and Hammerstein, to use a single 70 mm film running at 30 frames per second with 6 magnetic soundtracks, for their screen presentation of Oklahoma!. Major Hollywood studios immediately rushed to create their own unique formats, such as MGM's Camera 65, Paramount Pictures' VistaVision and Twentieth Century-Fox Film Corporation's CinemaScope, the latter of which used up to four separate magnetic soundtracks.

VistaVision took a simplified, low-cost approach to stereophonic sound; its Perspecta system featured only a monaural track, but through subaudible tones, it could change the direction of the sound to come from the left, right or both directions at once.

Because of the standard 35 mm-size film, CinemaScope and its stereophonic sound was capable of being retrofitted into existing theaters. CinemaScope 55 was created by the same company in order to use a larger form of the system (55 mm instead of 35 mm) to allow for greater image clarity onscreen, and was supposed to have had 6-track stereo instead of four. However, because the film needed a new, specially designed projector, the system proved impractical, and the two films made in the process, Carousel and The King and I, were released in 35 mm CinemaScope reduction prints. To compensate, the premiere engagement of Carousel used a six-track magnetic full-coat in an interlock, and a 1961 re-release of The King and I, featured the film printed down to 70 mm with a six-channel soundtrack.

Eventually, 50 complete sets of combination 55/35 mm projectors and penthouse reproducers were completed and delivered by Century and Ampex, respectively, and 55 mm release print sound equipment was delivered by Western Electric. Several samples of 55 mm sound prints can be found in the Sponable Collection at the Film and Television Archives at Columbia University. The subsequently abandoned 55/35 mm Century projector eventually became the Century JJ 70/35MM projector.

Todd-AO

[edit]

After this disappointing experience with their proprietary CinemaScope 55 mm system, Fox purchased the Todd-AO system and re-engineered it into a more modern 24 fps system with new 65 mm self-blimped production cameras (Mitchell BFC, "Blimped Fox Camera"), new 65 mm MOS cameras (Mitchell FC, "Fox Camera") and new Super Baltar lenses in a wide variety of focal lengths, first employed on South Pacific. Essentially, although Todd-AO was also available to others, the format became Fox's premier origination and presentation apparatus, replacing the CinemaScope 55 mm system. Current DVDs of the two CinemaScope feature titles were transferred from the original 55 mm negatives, often including the separate 35 mm films as extras for comparison.

Back to mono

[edit]

Beginning in 1957, films recorded in stereo (except for those shown in Cinerama or Todd-AO) carried an alternate mono track for theaters not ready or willing to re-equip for stereo.[21] From then until about 1975, when Dolby Stereo was used for the first time in films, most motion pictures – even some from which stereophonic soundtrack albums were made, such as Zeffirelli's Romeo and Juliet – were still released in monaural sound,[22] stereo being reserved almost exclusively for expensive musicals such as West Side Story,[23] My Fair Lady[24] and Camelot,[25] or epics such as Ben-Hur[26] and Cleopatra.[27] Stereo was also reserved for dramas with a strong reliance on sound effects or music, such as The Graduate.[28]

Dolby Stereo

[edit]

The Westrex Stereo Variable-Area system was developed in 1977 for Star Wars, and was no more expensive to manufacture in stereo than it was for mono. The format employs the same Western Electric/Westrex/Nuoptix RA-1231 recorder, and coupled with QS quadraphonic matrixing technology licensed to Dolby Labs from Sansui, this system can produce the same left, center, right and surround sound of the original CinemaScope system of 1953 by using a single standard-width optical track. This important development, marketed as Dolby Stereo, finally brought stereo sound to so-called flat (non-anamorphic) widescreen films, most commonly projected at aspect ratios of 1.75:1 or 1.85:1.

70 mm projection

[edit]

Producers often took advantage of the six magnetic soundtracks available for 70 mm film release prints, and productions shot in either 65 mm or to save money, in 35 mm and then blown up to 70 mm. In these instances, the 70 mm prints would be mixed for stereo, while the 35 mm reduction prints would be remixed for mono.

Some films shot in 35 mm, such as Camelot, featured four-track stereophonic sound and were then blown up to 70 mm so that they could be shown on a giant screen with six-track stereophonic sound. Unfortunately however, many of these presentations were only pseudo stereo, utilizing a somewhat artificial six-track panning method. A process known somewhat derogatorily as the Columbia Spread was often used to synthesize Left Center and Right Center from a combination of Left and Center and Right and Center, respectively, or, for effects, the effect could be panned anywhere across the five-stage speakers using a one-in/five-out pan pot. Dolby, who did not approve of this practice, which results in loss of separation, instead used the left center and right center channels for LFE (low-frequency effects), utilizing the bass units of the otherwise redundant intermediate front speakers, and later the unused HF capacity of these channels to provide for stereo surround in place of the mono surround.

Dolby Digital

[edit]

Dolby Stereo was succeeded by Dolby Digital 5.1 in the cinema, which retained the Dolby Stereo 70 mm 5.1 channel layout, and more recently with the introduction of digital cinema, Dolby Surround 7.1 and Dolby Atmos in 2010 and 2012 respectively.

Modern home audio and video

[edit]
Domestic stereo system, having two speakers

The progress of stereophonic sound was paced by the technical difficulties of recording and reproducing two or more channels in synchronization with one another and by the economic and marketing issues of introducing new audio media and equipment. A stereo system can cost up to twice as much as a monophonic system since a stereo system contains two preamplifiers, two amplifiers, and two speaker systems. In addition, the user would need an FM stereo tuner to upgrade any tape recorder to a stereo model and to have their phonograph fitted with a stereo cartridge. In the early days, it was unclear whether consumers would think the sound was so much better to be worth twice the price.

Stereo experiments on disc

[edit]

Lateral and vertical recording

[edit]

Thomas Edison had been recording in a hill-and-dale (vertically modulated) format on his cylinders and discs since 1877, and Berliner had been recording in a side-to-side (lateral) format since shortly thereafter. Each format developed on its own trajectory until the late 1920s, when electric recording on disc, utilizing a microphone, surpassed acoustic recording, which required a loud performance into what amounted to a megaphone in reverse.

At that time, AM radio had been around for roughly a decade, and broadcasters were looking for better materials from which to make phonograph records as well as a better format in which to record them to play over the narrow and thus inherently noisy radio channel. As radio had been playing the same shellac discs available to the public, it was found that, even though the playback system was now electric rather than acoustic, the surface noise on the disc would mask the music after just a few plays.

The development of acetate, bakelite, and vinyl, and the production of radio broadcast transcriptions, helped to solve this. Once these considerably quieter compounds were developed, it was discovered that the rubber-idler-wheel-driven turntables of the period had a great deal of low-frequency rumble – but only in the lateral plane. So, even though with all other factors being equal, the vertical plane of recording on disc had the higher fidelity, it was decided to record vertically to produce higher-fidelity recordings on these new materials, for two reasons, the increase in fidelity by avoiding the lateral rumble and to create incompatibility with home phonographs which, with their lateral-only playback systems, would only produce silence from a vertically modulated disc.

After 33+13 RPM recording had been perfected for the movies in 1927, the speed of radio program transcriptions was reduced to match, once again to inhibit playback of the discs on normal home consumer equipment. Even though the stylus size remained the same as consumer records at either 3 mils (76 μm) or 2.7 mils (69 μm), the disc size was increased from 12 inches (30 cm) to the same 16 inches (41 cm) as those used in early talking pictures in order to create further incompatibility. Now, not only could the records not be played on home equipment due to incompatible recording format and speed, they would not even fit on the player, which suited the copyright holders.

Two-channel high fidelity and other experiments

[edit]

An experimental format in the 1920s split the signal into two parts, bass and treble, and recorded the treble on its own track near the edge of the disc in a lateral format, minimizing high-frequency distortion, and recorded the bass on its own track in a vertical fashion to minimize rumble. The overhead in this scheme limited the playing time to slightly longer than a single, even at 33+13 RPM on a 12-inch disc.

Another failed experiment in the late 1920s and early '30s involved recording the left channel on one side of the disc and recording the right channel on the other side of the disc. These were manufactured on twin film-company recording lathes which ran in perfect sync with one another, and were capable of counter-clockwise as well as conventional clockwise recording. Each master was run separately through the plating process, lined up to match, and subsequently mounted in a press. The dual-sided stereo disc was then played vertically, first in a system that featured two tonearms on the same post facing one another. The system had trouble keeping the two tonearms in their respective synchronous revolutions.

Five years later, Bell Labs was experimenting with a two-channel lateral-vertical system, where the left channel was recorded laterally and the right channel was recorded vertically, still utilizing a standard 3 mil 78 RPM groove, over three times larger than the modern LP stylus of the late 20th century. In this system all the low-frequency rumble was in the left channel and all the high-frequency distortion was in the right channel. Over a quarter of a century later, it was decided to tilt the recording head 45 degrees off to the right side so that both the low-frequency rumble and high-frequency distortion were shared equally by both channels, producing the 45/45 system we know today.

Emory Cook

[edit]

In 1952, Emory Cook (1913–2002), who already had become famous by designing new feedback disk-cutter heads to improve sound from tape to vinyl, took the two-channel high-fidelity system described above and developed a binaural[note 1] record out of it. This consisted of two separate channels cut into two separate groups of grooves running next to each other, one running from the edge of the disc to halfway through and the other starting at the halfway point and ending up towards the label. He used two lateral grooves with a 500 Hz crossover in the inner track to try and compensate for the lower fidelity and high-frequency distortion on the inner track.

Each groove needed its own monophonic needle and cartridge on its own branch of the tonearm, and each needle was connected to a separate amplifier and speaker. This setup was intended to demonstrate Cook's cutter heads at a New York audio fair. It was not intended to promote the binaural process, but soon afterward, the demand for such recordings and the equipment to play them grew, and Cook's company, Cook Records, began to produce such records commercially. Cook recorded a vast array of sounds, ranging from railroad sounds to thunderstorms. By 1953, Cook had a catalog of about 25 stereo records for sale to audiophiles.[29]

Magnetic tape recording

[edit]

The first stereo recordings using magnetic tape were made in Germany in the early 1940s using Magnetophon recorders. Around 300 recordings were made of various symphonies, most of which were seized by the Red Army at the end of World War II. The recordings were of relatively high fidelity, thanks to the discovery of AC bias. A 1944 recording of Anton Bruckner's Symphony No. 8 directed by Herbert von Karajan and the Orchester der Berliner Staatsoper and a 1944 or 1945 recording of Walter Gieseking playing Beethoven's Piano Concerto No. 5 (with anti-aircraft fire audible in the background)[30] are the only recordings still known to exist.[citation needed]

In the US, stereo magnetic tape recording was demonstrated on standard 1/4-inch tape for the first time in 1952, using two sets of recording and playback heads, upside-down and offset from one another.[31] A year later, Remington Records began recording a number of its sessions in stereo, including performances by Thor Johnson and the Cincinnati Symphony Orchestra.[32]

Later in 1952, more experimental stereo recordings were conducted with Leopold Stokowski and a group of New York studio musicians at RCA Victor Studios in New York City. In February 1954, the label also recorded a performance of Berlioz' masterpiece The Damnation of Faust by the Boston Symphony Orchestra under the direction of Charles Munch, the success of which led to the practice of regularly recording sessions in stereo.[citation needed]

Shortly afterward, RCA Victor recorded the last two NBC Blue Network broadcast concerts by famed conductor Arturo Toscanini and the NBC Symphony Orchestra, on stereophonic magnetic tape, however, they were never officially released, though they have long been available on pirated LPs and CDs.[citation needed] In the UK, Decca Records began recording sessions in stereo in mid-1954, and by that time even smaller labels in the US such as Concertapes, Bel Canto and Westminster along with major labels such as RCA Victor began releasing stereophonic recordings on two-track prerecorded reel-to-reel magnetic tape, priced at twice or three times the cost of monaural recordings, which retailed for around $2.95 to $3.95 apiece for a standard monaural LP. Even two-track monaural tape which had to be flipped over halfway through and carried exactly the same information as the monaural LP – but without the crackles and pops – were being sold for $6.95.[33]

Stereophonic sound came to at least a select few living rooms of the mid-1950s.[34]

Stereo on disc

[edit]
Label and sleeve from Audio Fidelity Records' second stereo demonstration record, c. 1958

In November 1957, the small Audio Fidelity Records label released the first mass-produced stereophonic disc. Sidney Frey, founder and president, had Westrex engineers, owners of one of the two rival stereo disk-cutting systems, cut a disk for release before any of the major record labels could do so.[35][36] Side 1 featured the Dukes of Dixieland, and Side 2 featured railroad and other sound effects designed to engage and envelop the listener. This demonstration disc was introduced to the public on December 13, 1957, at the Times Auditorium in New York City.[37] Only 500 copies of this initial demonstration record were pressed and three days later, Frey advertised in Billboard Magazine that he would send a free copy to anyone in the industry who wrote to him on company letterhead.[38][39] The move generated such a great deal of publicity[40] that early stereo phonograph dealers were forced to demonstrate on Audio Fidelity Records.

Also in December 1957, Bel Canto Records, another small label, produced its own stereophonic demonstration disc on multicolored vinyl[41] so that stereo dealers would have more than one choice for demonstration. With the supplied special turntables featuring a clear platter lighted from underneath to show off the color as well as the sound, the stunt worked even better for Bel Canto, whose roster of jazz, easy listening and lounge music, pressed onto their trademark Caribbean-blue vinyl sold well throughout 1958 and early into 1959.

When Audio Fidelity released its stereophonic demonstration disc, there was no affordable magnetic cartridge on the market capable of playing it. After the release of other demonstration discs and the respective libraries from which they were culled, the other spur to the popularity of stereo discs was the reduction in price of a stereo cartridge, for playing the discs–from $250 to $29.95 in June 1958.[42] The first four mass-produced stereophonic discs available to the buying public were released in March 1958 – Johnny Puleo and his Harmonica Gang Volume 1 (AFSD 5830), Railroad – Sounds of a Vanishing Era (AFSD 5843), Lionel – Lionel Hampton and his Orchestra (AFSD 5849) and Marching Along with the Dukes of Dixieland Volume 3 (AFSD 5851). By the end of March, the company had four more stereo LPs available, interspersed with several Bel Canto releases.[43]

Although both monaural as well as stereo LP records were manufactured for the first ten years of stereo on disc, the major record labels issued their last monaural albums in 1968, relegating the format to 45 RPM singles, flexidiscs and radio promotional materials which continued until 1975.[44][45][46] In a sense, it was a sudden changeover: to discourage mono sales, in 1966 the record labels had generally eliminated the $1-cheaper pricing for mono LP. Also, even in 1967 stereo LP sales accounted for only 38.6% of the industry total, far outweighed by mono.[47]

Broadcasting

[edit]

Radio

[edit]

Early experimentation

[edit]

The earliest approach to stereo (then commonly called binaural) radio used two separate transmissions to individually send the left and right audio channels, which required listeners to operate two receivers in order to hear the stereo effect. In 1924 Franklin M. Doolittle was issued US patent 1,513,973[48] for the use of dual radio transmissions to create stereo reception. That same year Doolittle began a year-long series of test transmissions, using his medium wave broadcasting station, WPAJ in New Haven, Connecticut, which was temporarily authorized to concurrently operate a second transmitter. Left and right audio was distributed to the two transmitters by dual microphones, placed about 7 inches (18 cm) apart in order to mimic the distance between a person's ears.[49][50] Doolittle ended the experiments primarily because a lack of available frequencies on the congested AM broadcast band which meant that it was not practical for stations to occupy two frequencies,[51] plus it was cumbersome and expensive for listeners to operate two radio receivers.[51]

In 1925 it was reported that additional experimental stereo transmissions had been conducted in Berlin, again with two mediumwave transmissions.[52] In December of that year the BBC's long wave station, 5XX in Daventry, Northamptonshire, participated in the first British stereo broadcast – a concert from Manchester, conducted by Sir Hamilton Harty – with 5XX transmitting nationally the right channel, and local BBC stations broadcasting the left channel on mediumwave.[53] The BBC repeated the experiment in 1926, using 2LO in London and 5XX at Daventry. On June 12, 1946, a similar experimental broadcast using two stations was conducted in Holland, which was mistakenly thought to be the first in Europe and possibly the world.[54]

1952 saw a renewed interest in the United States in stereo broadcasting, still using two stations for the two channels, in part in reaction to the development of two-channel tape recordings. The Federal Communications Commission's (FCC) duopoly rule limited station owners to one AM station per market. But many station owners now had access to a co-owned FM station, and most of these tests paired AM and FM stations. On May 18 KOMO and KOMO-FM in Seattle, Washington conducted an experimental broadcast,[55] and four days later Chicago AM radio station WGN and its sister FM station, WGNB, collaborated on an hourlong stereophonic demonstration.[56] On October 23, 1952, two Washington, D.C. FM stations, WGMS-FM and WASH, conducted their own demonstration.[57] Later that month New York City's WQXR, paired with WQXR-FM, initiated its first stereophonic broadcast, which was relayed to WDRC[note 2] and WDRC-FM.[58][59] By 1954, WQXR was broadcasting all of its live musical programs in stereophonic sound, using its AM and FM stations for the two audio channels.[60] Rensselaer Polytechnic Institute began a weekly series of live stereophonic broadcasts in November 1952 using two AM stations, WHAZ in conjunction with a very low-powered local carrier current station, which meant the stereo listening area did not extend beyond the college campus.[61]

The revived dual transmitter tests were of limited success, because they still required two receivers, and with AM-FM pairings the sound quality of the AM transmissions was generally significantly inferior to the FM signals.

FM standards

[edit]
HH Scott Model 350, c. 1961: the first FM stereo tuner sold in the US

The Zenith-GE pilot-tone stereo system is used throughout the world by FM broadcasting stations.

It was eventually determined that the bandwidth assigned to individual FM stations was sufficient to support stereo transmissions from a single transmitter. In the United States, the FCC oversaw comparison tests, conducted by the National Stereophonic Radio Committee, of six proposed FM standards. These tests were conducted by KDKA-FM in Pittsburgh, Pennsylvania during July and August 1960.[62] In April 1961 the FCC adopted stereophonic FM technical standards, largely based on a Zenith-General Electric proposal, with licensed regular stereophonic FM radio broadcasting set to begin in the United States on June 1, 1961.[63] At midnight in their respective time zones on June 1, General Electric's WGFM in Schenectady, New York, Zenith's WEFM in Chicago, and KMLA in Los Angeles became the first three stations to begin broadcasting using the new stereo standards.[64]

Following experimental FM stereo transmissions in the London area in 1958 and regular Saturday morning demonstration transmissions using TV sound and medium wave (AM) radio to provide the two channels, the first regular BBC transmissions using an FM stereo signal began on the BBC Third Programme network on August 28, 1962.[65]

In Sweden, Televerket invented a different stereo broadcasting system called the Compander System. It had a high level of channel separation and could even be used to broadcast two separate mono signals – for example for language studies (with two languages at the same time). But tuners and receivers with the pilot-tone system were sold so people in southern Sweden could listen to, for example, Danish radio. At last Sweden (the Televerket) decided to start broadcasting in stereo according to the pilot-tone system in 1977.

AM standards

[edit]

Very few stations transmit in AM stereo. This is in part due to the limited audio quality afforded by the format, and the scarcity of AM stereo receivers. Various modulation schemes are used for AM stereo, of which the best-known is Motorola's C-QUAM, the official method for most countries in the world where AM stereo transmission is available. There has been experimental AM adoption of digital HD Radio, which also allows the transmission of stereo sound on AM stations;[66][67] HD Radio's lack of compatibility with C-QUAM along with other interference issues has hindered HD Radio's use on the AM dial.

Television

[edit]

A December 11, 1952, closed-circuit television performance of Carmen from the Metropolitan Opera House in New York City to 31 theaters across the United States, included a stereophonic sound system developed by RCA.[68] The first several shows of the 1958–59 season of The Plymouth Show (also called The Lawrence Welk Show) on the ABC (America) network were broadcast with stereophonic sound in 75 media markets, with one audio channel broadcast via television and the other over the ABC radio network.[69][70] By the same method, NBC Television and the NBC Radio Network offered stereo sound for two three-minute segments of The George Gobel Show on October 21, 1958.[71] On January 30, 1959, ABC's Walt Disney Presents made a stereo broadcast of The Peter Tchaikovsky Story – including scenes from Disney's latest animated feature, Sleeping Beauty – by using ABC-affiliated AM and FM stations for the left and right audio channels.[72]

After the advent of FM stereo broadcasts in 1962, a small number of music-oriented TV shows were broadcast with stereo sound using a process called simulcasting, in which the audio portion of the show was carried over a local FM stereo station.[73] In the 1960s and 1970s, these shows were usually manually synchronized with a reel-to-reel tape recording mailed to the FM station (unless the concert or music originated locally). In the 1980s, satellite delivery of both television and radio programs made this fairly tedious process of synchronization unnecessary. One of the last of these simulcast programs was Friday Night Videos on NBC.

The BBC made extensive use of simulcasting between 1974 and around 1990. The first such transmission was in 1974 when the BBC broadcast a recording of Van Morrison's London Rainbow Concert simultaneously on BBC2 TV and Radio 2. After that it was used for many other music programs, live and recorded, including the annual BBC Promenade concerts and the Eurovision Song Contest. The advent of NICAM stereo sound with TV rendered this unnecessary.

Cable TV systems used simulcasting to deliver stereo programs for many years. One of the first stereo cable stations was The Movie Channel, though the most popular cable TV station that drove up the usage of stereo simulcasting was MTV.

Japanese television began stereo broadcasts in 1978,[74] and regular transmissions with stereo sound came in 1982.[75] By 1984, about 12% of the programming, or about 14 or 15 hours per station per week were broadcast in stereo. West Germany's second television network, ZDF, began offering stereo programs in 1984.[74]

In 1979, The New York Times reported, "What has prompted the [television] industry to embark on establishing high-fidelity [sound] standards now, according to engineering executives involved in the project, is chiefly the rapid march of the new television technologies, especially those that are challenging broadcast television, such as the video disk."[76]

For analog TV (PAL and NTSC), various modulation schemes are used in different parts of the world to broadcast more than one sound channel. These are sometimes used to provide two mono sound channels that are in different languages, rather than stereo. Multichannel television sound is used mainly in the Americas. NICAM is widely used in Europe, except in Germany, where Zweikanalton is used. The EIAJ FM/FM subcarrier system is used in Japan. For digital TV, MP2 audio streams are widely used within MPEG-2 program streams. Dolby Digital is the audio standard used for digital TV in North America, with the capability for anywhere between 1 and 6 discrete channels.

Multichannel Television Sound (MTS) is the method of encoding three additional audio channels into an NTSC-format audio carrier. It was adopted by the FCC as the United States standard for stereo television transmission in 1984. Sporadic network transmission of stereo audio began on NBC on July 26, 1984, with The Tonight Show Starring Johnny Carson – although at the time, only the network's New York City flagship station, WNBC-TV, had stereo broadcast capability.[77] Regular stereo transmission of programs began in 1985. ABC and CBS followed suit in 1986 and 1987.

Recording methods

[edit]

A-B technique: time-of-arrival stereophony

[edit]
A-B stereo microphone placement

The A-B technique uses two omnidirectional microphones some distance apart and equidistant from the source. The technique captures time-of-arrival stereo information as well as some level difference information – especially if employed in close proximity to the source. At a distance of about 60 cm (24 in) between microphones, the time-of-arrival difference for a signal reaching the first microphone and then the other one from the side is approximately 1.75 ms. If you increase the distance between the microphones, you effectively decrease the pickup angle. At a 70 cm (28 in) distance, it is approximately equivalent to the pickup angle of the near-coincident ORTF setup.[citation needed]

This technique can produce phase issues when the stereo signal is mixed to mono.

X-Y technique: intensity stereophony

[edit]
X-Y stereo microphone placement

Here, two directional microphones are colocated, typically pointing at an angle between 90° and 135° with respect to each other.[78] The stereo effect is achieved through differences in sound pressure level between two microphones. Due to the lack of differences in time-of-arrival/phase ambiguities, the sonic characteristic of X-Y recordings has less sense of space and depth when compared to recordings employing an A-B setup. When two figure-eight microphones are used, facing ±45° with respect to the sound source, the X-Y setup is called a Blumlein pair.

M/S technique: mid/side stereophony

[edit]
Mid/side stereo microphone technique

This coincident technique employs a bidirectional microphone facing sideways and another microphone at an angle of 90°, facing the sound source. The second microphone is generally a variety of cardioid, although Alan Blumlein described the usage of an omnidirectional transducer in his original patent.

The left and right channels are produced through a simple matrix: left = mid + side; right = mid − side (using a polarity-reversed version of the side signal). This configuration produces a completely mono-compatible signal and, if the mid and side signals are recorded (rather than the matrixed left and right), the stereo width can be manipulated after the recording has taken place by adjusting the magnitude of the side signal (a greater amplitude giving a greater perceived stereo field which can be made to exceed the distance between loudspeakers). This makes it especially useful for film-based projects.

If the mid/side technique is incorporated into a self-contained stereo microphone assembly that outputs only the final left and right stereo pair signals, the original mid and side signals may be recovered permitting the above-mentioned manipulation. The mid signal is recovered by adding the left and right signals (the in phase and antiphase side signals cancel - giving a true mono signal in the process). The side signal is recovered by subtracting the right signal from the left (the mid signal is present in both channels and therefore cancels leaving the side).

Near-coincident technique: mixed stereophony

[edit]
ORTF stereo microphone technique

These techniques combine the principles of both A-B and X-Y (coincident pair) techniques. For example, the ORTF stereo technique of the Office de Radiodiffusion Télévision Française (Radio France) calls for a pair of cardioid microphones placed 17 cm apart at a total angle between microphones of 110°, which results in a stereophonic pickup angle of 96° (Stereo Recording Angle, or SRA).[79] In the NOS stereo technique of the Nederlandse Omroep Stichting (Dutch Broadcasting Organization), the total angle between microphones is 90° and the distance is 30 cm, thus capturing time-of-arrival stereo information as well as level information. It is noteworthy that all spaced microphone arrays and all near-coincident techniques use a spacing of at least 17 cm or more. 17 cm roughly equals the human ear distance and therefore provides the same interaural time difference (ITD) or more, depending on the spacing between microphones. Although the recorded signals are generally intended for playback over stereo loudspeakers, reproduction over headphones can provide remarkably good results, depending on the microphone arrangement.

Pseudo-stereo

[edit]
A stereo-widening example:[80]
1. A is a square wave and B is one thrice the frequency.
2. Different amounts of A and B are mixed into the left (L) and right (R) channels.
3. To widen the stereo effect, a fraction of the opposing channel is subtracted from each channel.
4. Normalized results show the signals A and B partly separated.

In the course of restoration or remastering of monophonic records, various techniques of pseudo-stereo, quasi-stereo, or rechanneled stereo have been used to create the impression that the sound was originally recorded in stereo. These techniques first involved hardware methods (see Duophonic) or, more recently, a combination of hardware and software. Multitrack Studio, from Bremmers Audio Design (The Netherlands),[81] uses special filters to achieve a pseudo-stereo effect: the shelve filter directs low frequencies to the left channel and high frequencies to the right channel, and the comb filter adds a small delay in signal timing between the two channels, a delay barely noticeable by ear,[note 3] but contributing to an effect of widening original flattiness of mono recording.[82][83]

The special pseudo-stereo circuit – invented by Kishii and Noro, from Japan – was patented in the United States in 2003,[84] with already previously issued patents for similar devices.[85] Artificial stereo techniques have been used to improve the listening experience of monophonic recordings or to make them more "saleable" in today's market, where people expect stereo. Some critics have expressed concern about the use of these methods.[86]

Binaural recording

[edit]

Engineers make a technical distinction between "binaural" and "stereophonic" recording. Of these, binaural recording is analogous to stereoscopic photography. In binaural recording, a pair of microphones is put inside a model of a human head that includes external ears and ear canals; each microphone is where the eardrum would be. The recording is then played back through headphones, so that each channel is presented independently, without mixing or crosstalk. Thus, each of the listener's eardrums is driven with a replica of the auditory signal it would have experienced at the recording location. The result is an accurate duplication of the auditory spatiality that would have been experienced by the listener had he or she been in the same place as the model head. Because of the inconvenience of wearing headphones, true binaural recordings have remained laboratory and audiophile curiosities. However, "loudspeaker-binaural" listening is possible with Ambiophonics.

Numerous early two-track-stereo reel-to-reel tapes as well as several experimental stereo disc formats of the early 1950s branded themselves as binaural, however they were merely different incarnations of the above-described stereo or two-track mono recording methods (lead vocal or instrument isolated on one channel and orchestra on the other sans lead.)

Playback

[edit]

Stereophonic sound attempts to create an illusion of location for various sound sources (voices, instruments, etc.) within the original recording. The recording engineer's goal is usually to create a stereo "image" with localization information. When a stereophonic recording is heard through loudspeaker systems (rather than headphones), each ear, of course, hears sound from both speakers. The audio engineer may, and often does, use more than two microphones (sometimes many more) and may mix them down to two tracks in ways that exaggerate the separation of the instruments, in order to compensate for the mixture that occurs when listening via speakers.

Descriptions of stereophonic sound tend to stress the ability to localize the position of each instrument in space, but this would only be true in a carefully engineered and installed system, where speaker placement and room acoustics are taken into account. In reality, many playback systems, such as all-in-one boombox units and the like, are incapable of recreating a realistic stereo image. Originally, in the late 1950s and 1960s, stereophonic sound was marketed as seeming "richer" or "fuller-sounding" than monophonic sound, but these sorts of claims were and are highly subjective, and again, dependent on the equipment used to reproduce the sound. In fact, poorly recorded or reproduced stereophonic sound can sound far worse than well-done monophonic sound. Nevertheless, many record companies released stereo "demonstration" records to help promote stereo. These records often included instructions for setting up a stereo system, 'balancing' the speakers, and a variety of ambient recordings to show off the stereo effect.[87] When playing back stereo recordings, the best results are obtained by using two identical speakers, in front of and equidistant from the listener, with the listener located on a center line between the two speakers. In effect, an equilateral triangle is formed, with the angle between the two speakers around 60 degrees as seen from the listener's point of view. Many higher quality multichannel (two-channel and beyond) speaker systems, then and now, include detailed instructions specifying the ideal angles and distances between the speakers and the listening position to maximize the effect based on, often extensive, testing of the particular system's design.

Vinyl records

[edit]
A compilation of LP stereo banners

Although Decca recorded Ernest Ansermet's May 1954 conducting of Antar in stereo, it took four years for the first stereo LPs to be sold.[88] In 1958, the first group of mass-produced stereo two-channel vinyl records was issued, by Audio Fidelity in the US and Pye in Britain, using the Westrex "45/45" single-groove system. Whereas the stylus moves horizontally when reproducing a monophonic disk recording, on stereo records, the stylus moves vertically as well as horizontally. One could envision a system in which the left channel was recorded laterally, as on a monophonic recording, with the right channel information recorded with a "hill and dale" vertical motion; such systems were proposed but not adopted due to their incompatibility with existing phono pickup designs (see below).

In the Westrex system, each channel drives the cutting head at a 45-degree angle to the vertical. During playback, the combined signal is sensed by a left-channel coil mounted diagonally opposite the inner side of the groove and a right-channel coil mounted diagonally opposite the outer side of the groove.[89] The Westrex system provided for the polarity of one channel to be inverted: this way large groove displacement would occur in the horizontal plane and not in the vertical one. The latter would require large up-and-down excursions and would encourage cartridge skipping during loud passages.

The combined stylus motion is, in terms of the vector, the sum and difference of the two stereo channels. Effectively, all horizontal stylus motion conveys the L+R sum signal, and vertical stylus motion carries the L−R difference signal. The advantages of the 45/45 system are that it has greater compatibility with monophonic recording and playback systems.

Even though a monophonic cartridge will technically reproduce an equal blend of the left and right channels, instead of reproducing only one channel, this was not recommended in the early days of stereo due to the larger stylus (1.0 mil or 25 micrometres vs. 0.7 mils or 18 micrometres for stereo) coupled with the lack of vertical compliance of the mono cartridges available in the first ten years of stereo. These factors would result in the stylus 'digging into' the stereo vinyl and carving up the stereo portion of the groove, destroying it for subsequent playback on stereo cartridges. This is why one often notices the banner "PLAY ONLY WITH STEREO CARTRIDGE AND STYLUS" on stereo vinyl issued between 1958 and 1964.

Conversely, and with the benefit of no damage to any type of disc even from the beginning, a stereo cartridge reproduces the lateral grooves of monophonic recording equally through both channels, rather than through one channel. Also, it gives a more balanced sound, because the two channels have equal fidelity as opposed to providing one higher-fidelity laterally recorded channel and one lower-fidelity vertically recorded channel. Overall, this approach may give higher fidelity, because the difference signal is usually of low power, and is thus less affected by the intrinsic distortion of hill and dale-style recording.

Additionally, surface noise tends to be picked up in a greater capacity in the vertical channel; therefore, a mono record played on a stereo system can be in worse shape than the same record in stereo and still be enjoyable. (See Gramophone record for more on lateral and vertical recording.)

Although this system was conceived by Alan Blumlein of EMI in 1931 and was patented in the UK the same year, it was not reduced to practice by the inventor as was a requirement for patenting in the US and elsewhere at that time. (Blumlein was killed in a plane crash while testing radar equipment during WW-II, and he, therefore, never reduced the system to actual practice through both a recording and a reproducing means.) EMI cut the first stereo test discs using this system in 1933, but it was not applied commercially until a quarter of a century later, and by another company entirely (Westrex division of Litton Industries Inc, as the successor to Western Electric Company), and dubbed StereoDisk. Stereo sound provides a more natural listening experience, since the spatial location of the source of a sound is (at least in part) reproduced.

In the 1960s, it was common practice to generate stereo versions of music from monophonic master tapes, which were normally marked "electronically reprocessed" or "electronically enhanced" stereo on track listings. These were generated by a variety of processing techniques to try to separate out various elements; this left noticeable and unsatisfactory artifacts in the sound, typically sounding "phasey". However, as multichannel recording became increasingly available, it has become progressively easier to master or remaster more plausible stereo recordings out of the archived multitrack master tapes.

Compact disc

[edit]

The Red Book CD specification includes two channels by default, and so a mono recording on CD either has one empty channel, or else the same signal on both channels.

Common usage

[edit]
Label for 2.0 sound (stereo)

In common usage, a stereo is a two-channel sound reproduction system, and a stereo recording is a two-channel recording. This is cause for much confusion, since five (or more)-channel home theater systems are not popularly described as stereo, but instead as surround.[clarification needed (see talk)]

Most multichannel recordings are stereo recordings only in the sense that they are stereo "mixes" consisting of a collection of mono and/or true stereo recordings. Modern popular music, in particular, is usually recorded using close miking techniques, which artificially separate signals into several tracks. The individual tracks (of which there may be hundreds) are then "mixed down" into a two-channel recording. The audio engineers determine where each track will be placed in the stereo "image", by using various techniques that may vary from very simple (such as "left-right" panning controls) to more sophisticated and extensively based on psychoacoustic research (such as channel equalization, mid-side processing, and the use of delay to exploit the precedence effect). The end product using this process often bears little or no resemblance to the actual physical and spatial relationship of the musicians at the time of the original performance; indeed, it is not uncommon for different tracks of the same song to be recorded at different times (and even in different studios) and then mixed into a final two-channel recording for commercial release.

Classical music recordings are a notable exception. They are more likely to be recorded without having tracks dubbed in later as in pop recordings, so that the actual physical and spatial relationship of the musicians at the time of the original performance can be preserved on the recording.[citation needed]

Balance

[edit]

Balance can mean the amount of signal from each channel reproduced in a stereo audio recording. Typically, a balance control in its center position will have 0 dB of gain for both channels and will attenuate one channel as the control is turned, leaving the other channel at 0 dB.[90]

See also

[edit]

Notes

[edit]

References

[edit]
[edit]
Revisions and contributorsEdit on WikipediaRead on Wikipedia
from Grokipedia
Stereophonic sound, commonly referred to as stereo, is a method of audio recording and reproduction that employs two independent channels—typically left and right—to capture and recreate the spatial directionality and depth of sound sources, mimicking the natural way human ears perceive auditory perspective through binaural cues. This technique relies on the differences in timing, intensity, and spectral content between the channels to produce an illusion of a three-dimensional soundstage when played back through two or more loudspeakers or stereo . The term "stereophonic" derives from the Greek words (solid or three-dimensional) and phōnē (sound), emphasizing its goal of solid, spatially immersive audio. The foundational principles of stereophonic sound were pioneered in the early 1930s by British engineer , who patented a comprehensive system in 1931 encompassing microphone techniques, disk cutting, and playback methods to achieve directional sound reproduction. Blumlein's innovations, including the use of coincident or spaced pairs to capture inter-channel time and level differences, were demonstrated in experimental recordings starting in 1933. Concurrently, American researchers at Bell Laboratories, such as , explored similar concepts for auditorium sound systems, conducting the first public demonstrations of stereophonic sound in 1933 that highlighted stereo's potential for motion picture audio. Despite early promise, widespread adoption was delayed by economic and technological challenges until the post-World War II era, when stereo gained traction in cinema through formats like Fantasound in Disney's Fantasia (1940), which used multi-channel stereophonic reproduction, and later CinemaScope's four-track magnetic stereo in the 1950s. Commercial consumer availability accelerated in 1957 with the release of the first stereophonic long-playing (LP) records by Audio Fidelity Records, marking the transition of stereo from experimental to mass-market technology and spurring the development of compatible phonographs, amplifiers, and FM stereo broadcasting standards by 1961. Today, stereophonic sound remains the standard for music recording and playback in consumer audio, , and live sound reinforcement, underpinning advancements in digital formats while serving as the basis for more complex surround systems.

Etymology and Fundamentals

Etymology

The term stereophonic derives from the Greek roots stereós (στερεός), meaning "solid," "firm," or "three-dimensional," and phōnḗ (φωνή), meaning "sound" or "voice." It was coined in 1927 by engineers at to denote audio reproduction that imparts a sense of spatial distribution and depth, drawing an explicit analogy to stereoscopic vision, which uses two slightly offset images to create a three-dimensional visual effect. The term's early application in audio emerged in amid efforts to replicate auditory directionality. In British Patent GB 394325, filed on December 14, 1931, and granted in 1933, Alan D. Blumlein described a two-channel system under the name "binaural" sound transmission, recording, and reproduction, using phase and intensity differences between channels to mimic natural . Laboratories, Blumlein's employer, adopted and refined this approach, producing the first experimental stereophonic disc recordings in 1933 and conducting demonstrations in , including tests for film applications. Over time, the terms "binaural" and "stereophonic" converged, with stereophonic becoming the standard descriptor for two-channel audio systems that exploit basic principles of , such as interaural differences, to evoke three-dimensionality.

Basic Principles of Stereophony

Stereophonic sound reproduction relies on the human auditory system's ability to localize sounds in space through binaural cues. The primary mechanisms include interaural time differences (ITD), which arise from the slight delay in sound arrival between the two ears due to the head's width, typically up to about 0.6 milliseconds for sounds originating from the sides. Interaural level differences (ILD) occur because the head shadows higher-frequency sounds, reducing intensity at the far ear, particularly effective above 1,500 Hz. Additionally, head-related transfer functions (HRTF) describe how the pinnae, head, and torso filter sounds based on direction, providing spectral cues for elevation and front-back discrimination. These perceptual cues enable the to construct a three-dimensional auditory scene, distinguishing sound sources by , , and distance. In natural hearing, ITD is most sensitive for low frequencies where phase differences are unambiguous, while ILD and HRTF dominate for higher frequencies and vertical localization. Stereophonic sound emulates this binaural process using two-channel audio delivered via loudspeakers or , creating an illusion of spatial width, depth, and within a frontal soundstage. By varying the signals fed to the left and right channels, stereo reproduction simulates the directional cues of natural hearing, allowing listeners to perceive sounds as positioned between or beyond the speakers. This approach, as outlined in early theoretical work, aims to preserve the auditory perspective of a live performance through independent left-right channels rather than a single blended signal. A key aspect of ITD in stereophony is the phase difference between channels, given by the equation Δϕ=2πfτ\Delta \phi = 2\pi f \tau where ff is the frequency and τ\tau is the time delay. This relationship highlights the limits of ITD effectiveness: at frequencies above approximately 700 Hz, phase ambiguities arise due to the half-cycle correspondence to the maximum head-related delay, shifting reliance to ILD for localization. Unlike monaural audio, which delivers identical signals to both ears and lacks inherent spatial information, stereophonic systems exploit and phase differences between channels to encode directional cues, enhancing perceived spaciousness and source separation. In mono, sounds appear centered with no width, whereas stereo's differential signaling mimics the interaural disparities of real-world listening, fostering a more immersive experience.

Historical Development

Early Experiments and Concepts

The concept of stereophonic sound, which aims to recreate the spatial qualities of sound through multiple channels, traces its roots to 19th-century experiments in transmitting audio over distances. In 1821, demonstrated his "Enchanted Lyre," an apparatus that conducted sound through a solid wire from a hidden source, creating the illusion of a self-playing instrument and representing an early effort in remote sound transmission. Building on such ideas, French inventor Clément Ader presented the théâtrophone system at the 1881 Paris International Exposition of Electricity, employing approximately 80 transmitters placed across the stage of the to capture performances. These signals were distributed via multiple telephone lines to receivers at the exhibition hall, allowing listeners to perceive a rudimentary sense of spatial audio from left and right perspectives, marking the first known stereophonic transmission. Advancements accelerated in with theoretical and practical innovations in recording and playback. British engineer , working at , filed British Patent 394325 in December 1931 for a "sound-transmission, sound-recording and sound-reproducing system" that utilized twin microphones to capture binaural audio and corresponding twin loudspeakers for reproduction, effectively inventing modern stereo recording. Granted in 1933, the patent detailed techniques for maintaining directional cues, including variable reluctance recording heads for disc-based storage. Between 1933 and 1935, Blumlein constructed experimental equipment and produced test recordings on both mechanical discs and optical film, demonstrating how separate channels could preserve the illusion of sound movement. Concurrently, researchers at Bell Laboratories conducted pioneering multi-channel demonstrations that influenced stereo development. In April 1933, Bell Labs transmitted a live four-channel stereophonic signal from a Philadelphia Orchestra concert conducted by Leopold Stokowski at the Academy of Music in Philadelphia to Constitution Hall in Washington, D.C., via telephone lines. This setup used multiple microphones and speakers to convey auditory perspective, with channels dedicated to different sections of the orchestra, providing audiences a sense of enveloping sound for the first time in a public venue. Early experiments faced significant technical hurdles, particularly in maintaining across channels. Variations in transmission delays over separate wires or recording paths often caused phase discrepancies, distorting spatial imaging and leading to comb-filtering effects when channels were combined. These issues necessitated innovations like precise timing mechanisms in Blumlein's designs and careful in ' setups to align audio wavefronts accurately.

Pre-Commercial Stereo in Cinema and Audio

In the realm of cinema, early pre-commercial stereo efforts emerged prominently in the 1940s with Disney's Fantasound system, developed in collaboration with RCA for the 1940 release of Fantasia. This innovative multi-channel setup utilized three optical sound tracks on 35mm film—designated for left, center, and right channels—combined with a fourth control track to direct effects to surround speakers, creating an immersive stereophonic experience across up to 56 speakers in select theaters. The system aimed to enhance spatial audio depth, allowing sounds like orchestral swells or fantastical effects to move dynamically around the audience, though its deployment was limited to about a dozen U.S. theaters due to the complexity and cost of installation. Building on this foundation, the 1952 introduction of represented a bold advancement in cinema with integrated stereophonic sound. This three-projector process, premiered in This Is Cinerama, employed a seven-track magnetic stripe on a separate 35mm film print, featuring five channels behind the deeply curved screen for directional front imaging and two surround channels to envelop viewers in ambient effects like roller-coaster rumbles or crowd noises. The audio was synchronized via a complex with the projectors, delivering high-fidelity immersion that heightened the panoramic visuals, though the format's technical demands restricted it to specialized venues. By 1955, the format further refined pre-commercial stereo in cinema through its use in the film Oklahoma!. This 70mm system incorporated six magnetic tracks on the print—five for precise left-center-right-screen and fill channels, plus one for rear surround—enabling rich, directional soundscapes that complemented the expansive 2.2:1 and 30 frames-per-second projection. Developed by and the , it marked a shift toward broader theatrical compatibility compared to , with magnetic stripes allowing for superior dynamic range and separation in musical sequences. Shifting to audio media, experimental stereophonic disc recordings gained traction in the early , notably through Emory Cook's binaural efforts. In 1952, Cook Laboratories released the first commercial binaural records, employing a dual-modulation technique that cut the left channel vertically and the right channel laterally into the disc groove, requiring a compatible two-needle cartridge for playback to achieve spatial separation. These 45 RPM and 33⅓ RPM discs, such as demonstrations of clock ticks or train sounds, showcased early but faced playback challenges, limiting widespread adoption. Parallel to disc innovations, magnetic tape emerged as a key tool for professional stereo mixing in the 1950s. RCA began utilizing half-inch, two-track recorders for stereophonic recordings around , enabling engineers to capture and mix live orchestras with left-right separation for enhanced in studio sessions. Similarly, adopted multi-track tape configurations, including two- and four-track setups on machines like the BTR-2, for professional mixing of classical and , facilitating precise panning and overdubs before final mono or mastering. This tape-based workflow, often at 15 or 30 inches per second, provided greater flexibility than optical film methods, influencing broadcast and record production prototypes.

Commercial Introduction and Standardization

The commercial introduction of stereophonic sound in the mid-1950s marked a pivotal shift toward widespread consumer adoption, beginning with phonograph records. In 1955, conducted a notable demonstration of stereo recording techniques using "Stereosonic" sessions at in , showcasing the potential for two-channel audio reproduction to enhance spatial depth in music playback. This event highlighted early engineering efforts to bridge experimental prototypes with market-ready formats. In the , Audio Fidelity Records released the first commercial stereo LP in 1957, such as Dukes of Dixieland, Volume 3, which accelerated interest despite limited playback equipment. By 1958, released commercial stereo LPs in the , such as demonstration records featuring classical and orchestral performances like "The Stars in Stereo," which were pressed using 45/45-degree lateral cutting to encode left and right channels in a single groove. These releases targeted audiophiles and set the stage for broader European distribution. Standardization accelerated in the United States with the (RIAA) reaching an agreement in March 1958 to adopt the 45/45-degree lateral cutting system for both 33⅓ RPM LPs and 45 RPM singles. This consensus resolved competing formats, such as vertical-lateral proposals, by specifying two channels modulated at 90 degrees to each other, each at 45 degrees to the record's horizontal plane, ensuring compatibility with existing mono equipment while minimizing . The agreement facilitated , with major labels like RCA Victor and Columbia issuing stereo discs later that year, though initial rollout was limited by the scarcity of stereo turntables and cartridges. The late 1950s saw a temporary setback in stereo adoption due to compatibility challenges; early stereo records often produced distortion or excessive stylus wear when played on mono phonographs, as the V-shaped grooves required precise diamond-tipped styli not standard in mono setups. This led to a brief "back-to-mono" emphasis, where labels prioritized mono releases for broader , with stereo versions commanding premium prices and limited availability until equipment costs declined. By the early , improved manufacturing and consumer education spurred a stereo revival, with dual-format releases becoming common. Broadcast radio followed suit with formalized standards. The (FCC) approved FM stereo multiplexing in April 1961, authorizing broadcasts to begin on June 1 using the GE-Zenith system, which encoded left and right channels via a 38 kHz subcarrier suppressed 50% and a 19 kHz pilot tone at 8-10% modulation to synchronize receiver decoding without interfering with mono compatibility. This enabled nationwide FM stereo transmission, with stations like WGFM in , pioneering the format. For AM radio, the 1980s introduced stereophonic capabilities through the (Compatible ) system, developed by and adopted by stations despite initial format rivalries; it modulated the carrier's phase and amplitude for stereo while remaining receivable on mono AM receivers, though full standardization by the FCC occurred in 1993. In cinema, the transition to commercial stereo gained momentum with ' introduction of in October 1975, a matrixed four-channel system (left, center, right, and surround) encoded onto standard 35mm optical prints using for improved and reduced hiss. First employed in films like , it allowed theaters to upgrade to without magnetic stripes, revitalizing theatrical audio and influencing over 80% of major releases by the late 1970s.

Transition to Digital Stereo

The transition to digital stereo in the marked a pivotal shift from analog formats, enabling higher fidelity reproduction through (PCM) and techniques. This era began with the introduction of optical media that stored stereo audio as , eliminating many of the and issues inherent in analog systems like vinyl records and . The (CD), co-developed by and , was launched commercially in 1982 as the first widespread digital stereo format. It utilized a 16-bit depth and 44.1 kHz sampling rate for PCM-encoded stereo audio, allowing for over 70 minutes of lossless playback on a 120 mm disc without the surface or wear associated with analog media. This standard quickly became the benchmark for consumer , with the first CD players and titles released in and that year. In the early 1990s, advancements in audio compression extended digital to multi-channel applications while maintaining compatibility with two-channel playback. , originally known as AC-3, was introduced in 1992 by Dolby Laboratories as a perceptual coding system capable of encoding up to 5.1 channels, including a core downmix for legacy systems. This achieved efficient data rates as low as 128 kbps for while preserving and spatial imaging through techniques like . Its deployment in cinema and further popularized digital processing. Digital broadcasting emerged in the 1990s and 2000s, facilitating wireless stereo transmission without the bandwidth limitations of analog FM stereo. (DAB), developed through the European Eureka 147 project, began public trials in 1990 and saw commercial rollout in the UK by 1995, using MPEG-1/2 Layer II compression to deliver CD-quality stereo over terrestrial frequencies with improved signal robustness. In the United States, —introduced by iBiquity Digital Corporation in the early 2000s—employed (IBOC) technology to overlay digital stereo signals on existing AM and FM bands, enabling channels and noise-free reception up to the coverage edge. By the 2010s and into 2025, streaming services advanced digital stereo toward high-resolution formats, surpassing specifications for audiophiles. Tidal's HiFi tier, launched in 2014 and enhanced through 2025, offers stereo playback in HiRes up to 24-bit/192 kHz, capturing extended and beyond human hearing limits in some cases. Spotify followed suit in September 2025 with its Premium lossless tier, providing 24-bit/44.1 kHz streaming for over 100 million tracks, emphasizing detailed via adaptive bitrate delivery. These platforms leverage cloud-based digital processing to ensure precise channel separation and minimal compression artifacts. The primary advantages of digital stereo include significantly reduced noise floors—often below -90 dB signal-to-noise ratio compared to analog's 60-70 dB—and enhanced channel separation through , which maintains left-right isolation exceeding 90 dB without degradation over time. These benefits stem from binary representation and error correction, allowing faithful reproduction of stereo spatial cues even in compressed or transmitted forms.

Recording Techniques

Coincident and Near-Coincident Methods

Coincident microphone techniques, such as the X-Y configuration, employ two directional s with their capsules positioned as closely as possible—ideally touching or overlapping—to capture stereophonic sound primarily through inter-channel intensity level differences (ILD), minimizing inter-channel time differences (ITD) and ensuring phase coherence between the left and right signals. This approach relies on the varying of sound arriving at each microphone based on its directional sensitivity, creating a stereo image without the comb-filtering artifacts that can arise from time delays. A specific variant, the , uses two bidirectional (figure-8 pattern) arranged coincidentally and crossed at 90 degrees to the sound source, capturing intensity differences for a natural and precise stereo image. The X-Y technique specifically uses two matched cardioid-pattern arranged in a coincident setup, with their capsules crossed at an angle typically between 90 and 120 degrees relative to the sound source—often 90 degrees for a narrower image or up to 120 degrees for wider coverage. The are oriented such that their axes form an "X" shape, with each pointing approximately 45 degrees from the center line when using a 90-degree angle, allowing the setup to simulate the ear's amplitude-based localization cues. This method, which evolved from early stereophonic experiments in the mid-20th century, provides a focused and stable stereo image suitable for controlled environments. Near-coincident methods build on the coincident principle by introducing a small physical separation between microphone capsules, typically 17 to 30 cm, to incorporate subtle ITD cues alongside ILD for a more natural sense of depth and spaciousness while preserving much of the phase integrity. The ORTF technique, developed in the 1960s by the French broadcasting organization , positions two cardioid s 17 cm apart (approximating the distance between human ears) at a 110-degree angle, with each microphone angled 55 degrees from the center line. This configuration balances intensity and time differences to produce a realistic stereo perspective, particularly effective for orchestral recordings where precise imaging is essential. Similarly, the NOS technique, devised in the by the Dutch broadcasting foundation , employs two cardioid microphones spaced 30 cm apart at a 90-degree mutual , with each oriented 45 degrees outward from the center. This setup aims for an "equivalence stereo" balance, emulating natural binaural perception with a recording angle of approximately ±40 degrees, and is optimized for broadcast applications requiring consistent tonal balance across frequencies. These methods are widely applied in classical and recording, where they deliver natural imaging and a of ensemble placement without the temporal smearing that can occur in other configurations, allowing engineers to capture large soundstages like orchestras with clarity. Their primary advantages include excellent mono compatibility—due to minimal phase cancellation when channels are summed—and reliable, artifact-free reproduction on systems. However, they may yield a less expansive of room ambience compared to spaced arrays, potentially resulting in a more intimate or "front-stage" perspective that suits focused performances but can feel constrained for highly reverberant spaces.

Spaced and Time-Of-Arrival Methods

Spaced and time-of-arrival methods in stereophonic recording utilize microphone arrays where physical separation between capsules introduces interaural time differences (ITD), mimicking the natural delay between sounds arriving at ears to create a of depth and spatial width in the stereo image. These techniques prioritize the timing variations of sound waves over intensity differences, often employing omnidirectional or bidirectional microphones positioned several centimeters to meters apart, pointed parallel toward the sound source. By capturing these temporal cues, the methods produce an expansive, natural-sounding stereo field particularly suited to ambient and orchestral recordings, though they can introduce phase-related artifacts if not managed carefully. The A-B technique, also known as the spaced pair, employs two omnidirectional microphones separated by 20 to 60 cm and aligned parallel to the sound field, allowing the ITD to generate a broad stereo perspective with enhanced depth and width. This configuration excels in capturing the natural and spatial relationships in environments like concert halls, as the spacing simulates the approximate distance between human ears, producing a realistic auditory scene when played back over loudspeakers. Engineers often adjust the separation based on the source's size—for instance, narrower spacing for intimate ensembles to avoid excessive width—ensuring the temporal disparities emphasize off-center sounds arriving earlier to one microphone than the other. Developed in the early 1950s by engineers including Arthur Haddy and Roy Wallace, the is a three-microphone array designed specifically for orchestral stereophonic recording, featuring two outer omnidirectional microphones spaced approximately 1.8 to 2 meters apart with a central microphone positioned coincidentally between them, often on a T-shaped or triangular bar suspended above the conductor. This setup combines the spacious ITD effects from the outer pair with the centered perspective of the middle microphone, which is typically panned to mono or blended to anchor the image, resulting in a balanced, immersive orchestral soundstage that became a staple in production from 1954 onward. The technique's enduring popularity stems from its ability to preserve the ensemble's natural layout without spot microphones, though modern variants may use hypercardioids for the outer mics to tighten focus. A variant employing figure-8 microphones in a spaced configuration, known as the Faulkner array or , positions two bidirectional capsules 20 cm apart in parallel alignment facing forward, leveraging ITD for while the figure-8 patterns provide inherent front-back rejection to minimize unwanted rear ambience. Invented by recording engineer Tony Faulkner in the late 1970s, this method builds on principles similar to the coincident but introduces spacing to enhance width and depth, making it effective for or solo instruments where controlled directionality is needed without full omnidirectional pickup. The parallel orientation ensures phase coherence in the forward direction, delivering a focused yet expansive with reduced sensitivity to sounds from behind the array. One primary challenge with spaced and time-of-arrival methods is comb filtering, an interference pattern arising from phase differences between the microphones' signals, particularly at high frequencies where wavelengths are short relative to the spacing, leading to frequency-dependent cancellations and a hollow or blurred sound upon stereo summation or mono compatibility checks. This phenomenon becomes pronounced when ITD exceeds about 1.5 ms, causing notches in the response that degrade . To mitigate comb filtering, engineers employ acoustic baffles, such as the Jecklin disc—a 25-30 cm diameter or disc placed between omnidirectional —to diffuse direct sound paths and simulate spacing acoustically without introducing severe phase shifts, thereby preserving temporal cues while smoothing the frequency response.

Mid-Side and Intensity-Based Techniques

The Mid-Side (M/S) technique is a matrixed stereophonic recording method that employs two microphones to capture audio signals, enabling manipulation of the stereo image while maintaining mono compatibility. The mid (M) signal is derived from a forward-facing cardioid or omnidirectional microphone, which records the central, mono-sum component of the sound field. The side (S) signal comes from a bidirectional (figure-8 ) microphone oriented to the mid microphone, capturing the lateral differences that convey width. Decoding the M/S signals to standard left-right stereo channels involves simple matrixing: the left channel combines the mid and side signals additively, while the right channel subtracts the side from the mid. This process is mathematically defined as: L=M+SL = M + S R=MSR = M - S where LL and RR represent the left and right channels, respectively. By adjusting the gain applied to the S signal prior to decoding—for instance, boosting or attenuating it—the perceived stereo width can be widened or narrowed without altering the phase relationships or introducing artifacts. This flexibility makes M/S particularly valuable for broadcasting and location recording, as the mid signal alone provides a robust mono mix if needed. The M/S technique was pioneered by British engineer , who devised and patented it in 1933 as part of his foundational work on stereophonic sound at . Early adoption included experimental applications in the 1950s by institutions like the , which explored matrixed methods during initial stereo broadcast trials to balance artistic imaging with technical reliability. In contemporary workstations (DAWs), M/S processing is routinely implemented via plugins that allow real-time encoding, decoding, and width control, often integrated into mixing workflows for music and . Intensity-based stereophony, in contrast, emphasizes differences between channels to establish spatial positioning, forming the basis for many non-time-dependent stereo creation methods. Unlike approaches relying on inter-channel time , intensity stereophony uses relative level variations—achieved through panning or gain adjustments—to simulate directionality, primarily influencing mid- and high-frequency localization in human hearing. This principle underpins studio mixing practices, where audio sources are positioned in the stereo field by varying their volume in the left and right channels, creating an immersive without physical spacing. The M/S technique inherently operates within intensity stereophony, as the side signal modulates level disparities to control , but intensity methods extend to broader applications like multitrack panning in DAWs. These approaches prioritize precise control over perceived width and balance, making them staples in production for their compatibility and adjustability.

Binaural and Pseudo-Stereo Approaches

Binaural recording techniques aim to capture audio in a way that mimics human hearing, using a dummy head equipped with microphones positioned at ear locations to simulate the head-related transfer function (HRTF). This setup records the acoustic effects of the head, pinnae, and torso on incoming sound waves, producing signals optimized for headphone playback that preserve spatial cues like interaural time differences and level differences. The dummy head, often constructed from materials approximating human tissue density, allows for immersive, three-dimensional sound reproduction when listened to through headphones, as the recordings inherently account for binaural filtering. Pseudo-stereo approaches, in contrast, artificially generate a stereo image from monophonic sources by applying signal processing such as time delays, reverberation, or phase shifting to create perceived spatial separation without true multichannel capture. These methods decorrelate the mono signal into left and right channels, often using all-pass filters or finite impulse response (FIR) designs to widen the soundstage while maintaining compatibility with stereo systems. A notable example from the 1970s is the QS quadraphonic synthesizer, which employed phase-based matrixing to derive surround effects from stereo or mono inputs, enhancing spatial impression through controlled signal manipulation. In modern applications, binaural techniques have advanced significantly in (VR) and (AR) environments during the 2020s, enabling dynamic head-tracked audio rendering for more realistic immersion in gaming, training simulations, and therapeutic scenarios. Pseudo-stereo processing finds use in enhancing legacy mono recordings for contemporary playback, such as remastering archival audio for streaming platforms. However, binaural recordings face limitations in speaker-based reproduction, as direct playback leads to interaural that distorts spatial cues; effective presentation requires crosstalk cancellation filters to mitigate this interference.

Playback and Reproduction

Stereo Speaker Systems and Setup

Stereo speaker systems typically consist of two matched loudspeakers positioned to create a balanced sound field for the listener. The ideal configuration forms an equilateral triangle with the listening position, where the speakers are separated by a distance equal to the distance from each speaker to the listener, resulting in a 60-degree angle at the listening spot. This setup ensures even coverage of the stereo image across the two channels. Additionally, speakers are often toed-in by approximately 30 degrees toward the listener to direct the high-frequency response more precisely and minimize off-axis coloration. Common types of stereo speakers include bookshelf models, which are compact and suitable for stands or shelves, and floorstanding designs that incorporate larger enclosures for enhanced low-frequency output. Bookshelf speakers generally feature two or three drivers—a for midbass and a for highs—while floorstanding units may include multiple for greater bass extension. Both types employ crossover networks, passive circuits within the speaker cabinet that divide the by , directing low frequencies to the and highs to the to optimize driver performance and prevent overlap. In modern active speakers, these crossovers can be implemented digitally for finer control. Room acoustics play a crucial role in stereo reproduction, where the direct sound from the speakers must dominate early reflections from walls, , and to preserve spatial accuracy. Reflections can introduce comb filtering and smear the stereo field, particularly in untreated rooms, while excessive low-frequency buildup from room modes creates uneven bass response. Acoustic treatments such as , placed in room corners to absorb standing waves, help mitigate these modal issues without overly deadening the space. The evolution of stereo speaker systems began in the 1950s with the advent of commercial hi-fi setups featuring two separate speakers driven by a stereo amplifier, coinciding with the introduction of stereo LPs and enabling true two-channel playback in homes. By the late 20th century, passive designs dominated, but the 2020s have seen a shift toward active systems with built-in digital signal processing (DSP), allowing real-time adjustments for crossovers, equalization, and room correction directly within the speakers. This progression has improved efficiency and adaptability, reducing reliance on external components. Stereo provide an alternative playback method, delivering the left and right channels directly to each without inter-channel . This preserves interaural time differences (ITD) and interaural level differences (ILD) cues more accurately than speakers, enhancing the of spatial audio in portable devices, studio monitoring, and as of 2025.

Balance, Imaging, and Listener Perception

In stereophonic playback, balance refers to the distribution of audio signals between the left and right channels to maintain perceptual consistency across the soundstage. Panning laws govern this process, with the equal being a widely adopted method that preserves constant acoustic power by adjusting channel amplitudes using sinusoidal functions, attenuating each channel by 3 dB at the center position relative to the sides to compensate for the coherent of signals from both channels. In contrast, the equal voltage law applies linear amplitude distribution; without compensation, this leads to a perceived +6 dB increase at the center due to additive power from both channels, though it is often adjusted by attenuating the center by 6 dB for constant level. These laws ensure balanced by aligning perceived source positions with intended spatial cues during mixing and reproduction. The sweet spot represents the optimal listening position in a stereo setup, typically forming an with the speakers, where interaural time differences (ITD) and interaural level differences (ILD) cues from the recording align to produce a stable phantom image—a virtual source perceived centrally between the speakers despite no physical channel. Deviations from this position degrade the illusion, as mismatched arrival times and levels at the ears disrupt localization accuracy. Several factors influence the quality of balance, , and overall listener perception in stereo reproduction. Speaker matching is critical, as mismatches in or phase between left and right speakers introduce imbalances that blur the soundstage and weaken phantom images. Room modes, the resonant frequencies of enclosed spaces, can color the low-frequency response unevenly across the listening area, distorting ITD and ILD cues and collapsing stability. Additionally, head movement disrupts these cues by altering the relative path lengths and intensities from each speaker to the ears, causing the perceived positions of sources to shift or smear, particularly for off-center images. To quantify performance in binaural-like reproduction over speakers, the crosstalk cancellation ratio measures the effectiveness of filters in suppressing unwanted signal leakage between channels, ideally achieving high (e.g., 20-30 dB) to preserve spatial without altering tonal balance. This metric evaluates how well the system isolates left and right signals at the listener's ears, directly impacting imaging precision in the sweet spot.

Compatibility with Mono Systems

One fundamental aspect of stereophonic sound compatibility with monophonic systems involves matrixing the left (L) and right (R) channels to produce a mono signal by summing them as M = (L + R)/2, which preserves the core audio content without introducing level boosts or distortions when played on mono equipment. This downmix technique ensures that the primary information in both channels combines coherently, allowing legacy mono receivers to reproduce the sound without needing stereo decoding hardware. In analog disc recording, the 45/45° cutting system developed by Westrex achieves compatibility by modulating the left and right channels at 45-degree angles to the record surface, resulting in a lateral (side-to-side) groove motion that represents the L + R sum and a vertical (up-and-down) motion for the L - R difference. Mono cartridges, designed with low vertical compliance, respond primarily to the lateral motion and ignore the vertical component, effectively reproducing only the summed L + R signal without interference from the stereo difference information. This design allows stereo records to play safely on mono turntables while delivering full monophonic output. For , the pilot-tone stereo system outlined in BS.450 transmits the L + R sum as the main within the 0-15 kHz , while the L - R difference is double-sideband suppressed-carrier modulated onto a 38 kHz subcarrier, accompanied by a 19 kHz pilot tone to enable stereo decoding. Mono radios, lacking the stereo decoder, simply demodulate the L + R main channel and filter out the ultrasonic subcarrier above 15 kHz, rendering it inaudible and ensuring clean monophonic reception without noise or distortion from the stereo components. Despite these mechanisms, compatibility issues can arise from phase cancellation in poorly mixed stereo signals, where out-of-phase elements in the L and R channels destructively interfere during summation to mono, reducing perceived volume or clarity in certain frequencies. To mitigate such problems, international standards like BS.775 provide guidelines for multichannel systems, emphasizing downmix procedures that maintain mono balance, including checks for phase coherence and level preservation during L + R . Mid-side encoding, as a related technique, can further aid compatibility by isolating the mono-compatible mid (L + R) signal from the stereo side (L - R) information.

Formats and Media

Analog Disc Recording and Playback

Analog disc recording for stereophonic sound involves a cutting lathe that uses a heated, chisel-shaped to engrave a V-shaped groove into a lacquer-coated aluminum disc, with the left and right audio channels modulating the groove walls at 45-degree angles relative to the vertical plane in the 45/45 system. This configuration allows the left channel to affect both walls in phase while the right channel affects them out of phase, enabling sum and difference signals to reconstruct the stereo image during playback. To optimize groove spacing and reduce surface noise, the recording applies the curve, which attenuates low frequencies (below 500 Hz) by up to -12 dB at 20 Hz and boosts high frequencies (above 2.1 kHz) by up to +20 dB at 20 kHz during cutting, with inverse equalization applied on playback. The of analog stereo discs typically spans approximately 20 Hz to 20 kHz, matching the range of human hearing, though practical limitations arise from groove geometry and playback mechanics. Inner groove becomes prominent toward the record's center due to the reduced linear velocity at smaller radii, which compresses the available groove space for high-frequency modulations and increases , often exceeding 5% in the upper frequencies on longer sides. Playback requires a stereo phonograph cartridge, commonly a moving magnet design featuring two small magnets attached to the stylus cantilever and two separate coils to generate independent left and right signals from the groove undulations. The stylus tip, typically diamond or sapphire, traces the groove under a precise tracking force of 1 to 2 grams to ensure stable contact without excessive wear or mistracking. In the , experimental approaches to disc recording predated commercial standards, including vertical-lateral dual-track systems where one channel modulated the groove bottom vertically and the other laterally, as explored by inventors like Emory Cook in his binaural records pressed at 78 RPM on or vinyl. These Cook discs required dual-stylus playback heads—one for each track—to achieve spatial effects, influencing later 45/45 adoption, though they suffered from compatibility issues with monaural equipment.

Magnetic Tape Stereo

The implementation of stereophonic sound on analog originated in the late with professional two-track recording on quarter-inch-wide tape operating at 15 inches per second (ips), employing NAB equalization to optimize and compensate for tape characteristics. This approach, pioneered by with models like the 300 series, enabled separate left and right channels on adjacent tracks, facilitating high-fidelity stereo capture for and studio use. By the early , commercial two-track machines became widely available, supporting multitrack origins that evolved into more complex configurations while maintaining quarter-inch tape as the standard width. For consumer applications, stereo magnetic tape gained traction in the 1960s through the compact cassette format, which introduced in 1963 but initially offered in mono; stereo capability arrived in 1966 with dual-head recorders such as the Philips EL3312 (marketed as Norelco 450 in ), allowing simultaneous recording and playback of two channels on a single cassette. In the 1970s, the adoption of B systems significantly enhanced cassette performance by compressing high-frequency signals during recording and expanding them on playback, reducing tape hiss by about 10 dB and making stereo cassettes viable for . Early magnetic tape also played a role in cinema soundtracks, though its stereo applications there predated widespread consumer adoption. Key formats distinguished reel-to-reel and compact cassette systems: professional and enthusiast reel-to-reel tapes typically ran at 7.5 or 15 ips on quarter-inch stock, delivering superior (up to 20 kHz at higher speeds) and compared to the compact cassette's fixed 1.875 ips, which limited high-frequency extension but prioritized portability. Channel separation in both formats relied on precise azimuth alignment, where recording and playback head gaps are oriented to the tape path to minimize between left and right tracks, often achieving 30-40 dB isolation when properly calibrated. Among its advantages, analog offered editability through physical splicing of the tape medium, enabling precise cuts and rearrangements in , a flexibility not matched by contemporaneous disc formats. Additionally, it provided a of approximately 70 dB—higher than the roughly 60 dB of vinyl records—allowing greater contrast between quiet and loud passages without surface noise overwhelming low-level signals.

Broadcasting Standards

Stereophonic sound transmission in (FM) radio was standardized in the United States through a multiplex approved by the () in April 1961, allowing broadcasts to begin on June 1 of that year. This encodes the left (L) and right (R) audio channels by transmitting the sum signal (L+R) on the main channel up to 15 kHz, while the difference signal (L-R) is modulated onto a suppressed subcarrier at 38 kHz, accompanied by a 19 kHz pilot tone to enable decoding in receivers. The setup also reserves a 67 kHz subcarrier for subsidiary communications authorization (SCA) services, such as or transmission, ensuring compatibility with monaural receivers that ignore the subcarrier components. In contrast, stereophonic (AM) radio faced prolonged challenges in achieving a unified standard due to competing proprietary systems introduced in the , including the Kahn-Hazeltine independent sideband system and Motorola's () approach. The FCC initially encouraged market-driven selection without mandating a single method, leading to fragmented adoption among broadcasters and receiver manufacturers. In 1984, the FCC declined to endorse a specific system, but by 1993, amid declining interest and low penetration, it designated Motorola's as the official standard to promote uniformity, though widespread use had already waned and few stations continued stereo AM transmissions thereafter. For analog television, the adopted the (MTS) standard in 1984, developed by and dbx, which added stereo and secondary audio channels to the existing 4.5 MHz FM audio carrier in broadcasts. MTS encodes L+R and L-R similarly to FM stereo but uses dbx compression for , supporting left, right, and a separate audio program () for bilingual or descriptive audio, with monaural compatibility maintained through the main channel. In , the near instantaneous companded audio multiplex (NICAM) system, standardized by the in 1986, served as a digital add-on to analog PAL and video signals, transmitting compressed stereo or dual-channel audio at 728 kbit/s on a 6.5 MHz or 5.85 MHz subcarrier, offering superior noise performance over MTS while falling back to analog FM audio if digital decoding fails. Digital broadcasting standards have since incorporated stereophonic and immersive audio capabilities natively. Digital Audio Broadcasting (DAB), launched in 1995 with initial services in and the , employs (MP2) compression for stereo transmission at bitrates typically around 192 kbit/s per channel, multiplexed within OFDM carriers for robust mobile reception and with mono decoders. In the United States, (NextGen TV), with deployments expanding post-2025, extends beyond stereo to immersive audio formats like , enabling object-based sound rendering for surround and height channels, enhanced dialogue clarity, and personalized audio streams, all while supporting 4K video and IP integration for over-the-air delivery.

Digital Formats and Optical Media

The introduction of the (CD-DA) format in 1982 marked a significant advancement in stereophonic sound storage, utilizing two-channel (PCM) to encode left and right audio channels separately. Each channel is sampled at 44.1 kHz with 16-bit resolution, providing a of approximately 96 dB and a up to 20 kHz, sufficient for human hearing. Error correction is achieved through Cross-Interleave Reed-Solomon Code (CIRC), which mitigates data loss from scratches or defects on the , ensuring reliable playback of stereo content. This format, developed by and , became the standard for consumer ; by 2007, over 200 billion compact discs (including audio CDs, CD-ROMs, and CD-Rs) had been produced worldwide. Higher-resolution optical media emerged in the late to extend stereophonic capabilities beyond CD-DA, including and (SACD). supports up to two-channel stereo at resolutions like 24-bit/192 kHz PCM using (MLP) compression, which preserves full fidelity without data loss while allowing multichannel extensions; it serves as a stereo base that can upscale to . In contrast, SACD employs (DSD) encoding at a 2.8224 MHz sampling rate with 1-bit , offering superior shaping for frequencies above 20 kHz and a exceeding 120 dB in stereo mode. Both formats, released in 1999 by the and Sony/ respectively, prioritize high-fidelity stereo reproduction but saw limited adoption due to compatibility issues and the rise of digital downloads. Digital file formats for stereophonic sound further diversified storage options, encompassing both uncompressed and compressed variants for optical and non-optical media. Uncompressed formats like and Audio Interchange File Format (AIFF) store two-channel PCM data directly, typically at CD-equivalent 16-bit/44.1 kHz or higher resolutions such as 24-bit/96 kHz, without alteration to the original stereo signal; , standardized by and in 1991, supports interleaving of left and right channels for efficient playback. Lossy compressed formats, including (MPEG-1 Audio Layer III) and (AAC), incorporate joint stereo modes to reduce file size by exploiting inter-channel correlations—such as mid-side encoding—while maintaining perceptual stereo imaging at bitrates as low as 128 kbps. These formats, defined in ISO/IEC standards from the , enabled widespread stereo distribution on optical discs and later digital platforms. In cinema applications, optical media evolved to support stereophonic soundtracks, transitioning from analog to digital systems. The 1960s introduced six-track magnetic stereophonic recording on 70mm film prints, using four tracks for surround channels alongside left and right mains to create immersive fields, as seen in films like (1964). By 1993, Digital Theater Systems (DTS) pioneered digital optical stereophonic delivery with a separate synced to the film projector via timecode, encoding six-channel audio (including base) at 20-bit/44.1 kHz using adaptive differential PCM (ADPCM) for lossless quality; this approach reduced optical track noise and improved synchronization over traditional variable-area soundtracks. DTS's implementation, licensed for over 20,000 films, demonstrated the format's impact on theatrical fidelity.

Applications and Usage

Home Audio and Consumer Electronics

Stereophonic sound became a cornerstone of in the 1960s through integrated stereo receivers that combined AM/FM tuners, power amplifiers, and phono preamplifiers into single units, enabling consumers to enjoy high-fidelity playback from vinyl records and radio broadcasts. These receivers marked a shift from separate components to more convenient all-in-one systems, with early models like the Fisher 500-C () featuring a stereo multiplex FM tuner with high sensitivity for noise rejection and bandwidth up to 20Hz–15kHz, a 35Wpc tube power amplifier, and dual phono inputs with tube-based preamplification providing approximately 20dB gain for magnetic cartridges. Such designs catered to the growing popularity of stereo LPs, offering balanced left and right channel separation for immersive listening in living rooms. The 1980s saw a boom in with the widespread adoption of (CD) players, which delivered pristine digital stereo audio free from the surface noise of analog media, alongside the rise of mini-component systems that compacted receivers, CD players, and speakers into space-saving shelves. The first consumer , Sony's CDP-101, launched in , utilized laser-based optical playback for 16-bit/44.1kHz stereo resolution, revolutionizing home listening by enabling random track access and durable media storage. Manufacturers like Pioneer responded with mini-systems such as the CD3000 and CD1000 series, integrating CD playback with AM/FM tuners and compact amplifiers in sleek, affordable packages that appealed to urban apartments and younger demographics. By the 2000s, home theater systems integrated stereophonic sound via downmix capabilities, allowing multi-channel surround formats like 5.1 to be folded into two-channel stereo for compatibility with existing speaker pairs. Receivers from this era, such as those supporting and SACD formats introduced in 1999, included matrix downmixing algorithms that preserved from front left/right channels while attenuating surround elements to avoid phase issues. This feature ensured seamless playback of stereo content within setups, bridging music listening and movie audio in shared living spaces. Portable stereophonic audio evolved from the TPS-L2, introduced in 1979 as the first battery-powered stereo cassette player with lightweight for personal, on-the-go listening. Priced at around ¥33,000 (about $240 USD), it featured dual headphone jacks and a "hotline" button for conversations without removing , selling over 50,000 units in the first two months and sparking a cultural shift toward individualized stereo experiences. The 2010s brought wireless portability with speakers, exemplified by models like the Jawbone Jambox (2010) and UE Boom (2013), which used 4.0 for low-latency stereo pairing and up to 10 hours of battery life, enabling true wireless left-right channel synchronization across devices. As of 2025, smart speakers continue to advance home stereophonic integration, with systems like 100 allowing stereo pairing of two units for balanced left-right imaging via Trueplay tuning, alongside voice assistants such as Voice Control for hands-free playback commands processed locally on-device. These devices support multi-room stereo synchronization over , blending voice-activated control from assistants like or with high-resolution audio streaming up to 24-bit/48kHz.

Cinema and Theatrical Sound

The introduction of stereophonic sound in cinema during the 1950s aimed to enhance immersion in widescreen formats, with 20th Century Fox pioneering four-track magnetic stereo alongside CinemaScope. This system featured separate left, center, right, and surround channels recorded on magnetic stripes along the film's edge, debuting in the 1953 epic The Robe to create directional audio that complemented the wide aspect ratio. The technology required specialized projectors and theaters equipped with magnetic playback heads, providing a more spatial experience than mono optical tracks by directing sounds across the screen and auditorium. In 1975, Dolby Laboratories introduced , an optically printed matrix system for 35mm films that encoded four channels—left, center, right, and surround—into two optical tracks using and matrixing techniques. This format allowed for high-quality stereo reproduction compatible with existing projectors while enabling surround effects through decoding, marking a practical advancement over magnetic systems. Its widespread adoption began with the 1977 release of Star Wars, where the matrix surround enhanced the film's dynamic soundscape, including spaceship fly-bys and explosions, revolutionizing theatrical audio immersion. The transition to in cinema expanded stereo capabilities within multi-channel frameworks, with debuting in 1992 on as the first theatrical feature to use this compressed 5.1 surround format. In this system, the two primary stereo left and right channels served as mains behind the screen, augmented by center, two surrounds, and a channel for fuller spatial imaging. Encoded digitally on the film print or via separate data tracks, it delivered clearer, higher-fidelity stereo separation without the hiss of analog optical prints. Large-format systems like and OMNIMAX incorporated dual-channel stereo variants tailored for expansive screens, using six-track configurations with left/right mains and additional channels for panoramic audio distribution. These setups, featuring four screen channels and two surrounds, provided stereophonic depth scaled to screens up to seven stories high, ensuring immersive sound that envelops audiences in dome or rectangular theaters.

Modern Implementations and Common Practices

In contemporary streaming platforms, stereophonic sound has evolved through spatial audio implementations that extend traditional two-channel playback into immersive, three-dimensional experiences. launched Spatial Audio with support on June 7, 2021, enabling binaural rendering for to simulate from stereo mixes, with thousands of tracks available at launch for enhanced depth and positioning. followed in October 2021, introducing spatial audio compatible with any via its and Android apps, supporting and 360 Reality Audio formats to deliver object-based stereo panning and height effects in music streaming. These binaural stereo approaches, which process audio to mimic natural head-related transfer functions, allow listeners to perceive sound sources around them, significantly broadening the stereo soundstage without requiring multi-speaker setups. Wireless stereo transmission benefits from advancements in Bluetooth 5.0 and later standards, which improve data rates and stability for high-fidelity two-channel audio over short ranges. Qualcomm's aptX HD codec, integrated into many devices since the mid-2010s and optimized in Bluetooth 5.0+, achieves low-latency stereo playback at up to 576 kbps and 24-bit/48 kHz resolution, minimizing audio-video desync in applications like video viewing on true wireless earbuds. True wireless earbuds, such as the Technics EAH-AZ100 released in 2024, leverage these technologies alongside LDAC codec support to deliver detailed stereo imaging and wide soundstages via Bluetooth, with active noise cancellation preserving spatial cues in mobile environments. Automotive stereo systems in the 2020s incorporate (DSP) for equalization, compensating for vehicle cabin resonances and reflections to maintain balanced stereo separation and across seats. Head-tracking integration, enabled by sensors and DSP algorithms, dynamically adjusts stereo panning based on listener movement, as seen in Fraunhofer IIS's Cingo technology deployed in vehicles since 2023, which renders personalized binaural stereo for during drives. Stereophonic sound enhances accessibility in (VR) and through spatial audio standards like first-order , which encode directional stereo cues for headphone playback, allowing users to localize sounds in immersive environments as of 2025. AI upmixing techniques further support accessibility by converting mono sources to stereo using models that analyze spectral content and generate artificial spatial separation, as demonstrated in tools like UniFab's Audio Upmix AI, which distributes elements into left-right channels while preserving phase coherence. These methods, rooted in neural network-based source separation, enable legacy mono content to achieve modern stereo compatibility without artifacts, promoting inclusive audio experiences in streaming and VR applications.

References

Add your contribution
Related Hubs
User Avatar
No comments yet.