Hubbry Logo
Microphone practiceMicrophone practiceMain
Open search
Microphone practice
Community hub
Microphone practice
logo
8 pages, 0 posts
0 subscribers
Be the first to start a discussion here.
Be the first to start a discussion here.
Microphone practice
Microphone practice
from Wikipedia
A didgeridoo miked with a small phantom powered condenser microphone that clips onto the instrument.

There are a number of well-developed microphone techniques used for recording musical, film, or voice sources or picking up sounds as part of sound reinforcement systems. The choice of technique depends on a number of factors, including:

  • The wish to capture or avoid the collection of extraneous noise. This can be a concern, especially in amplified performances, where audio feedback can be a significant problem. Alternatively, it can be a desired outcome, in situations where ambient noise is useful such as capturing hall reverberation and audience reactions in a live recording.
  • Degree of directionality of pickup: in some settings, such as a home video of a birthday party, the person may wish to pick up all the sounds in the room, making an omnidirectional mic desirable. However, if a TV news crew is filming a reporter at a noisy protest, they may only wish to pick up her voice, making a cardioid mic more desirable.
  • Choice of a signal type: Mono, stereo or multi-channel.
  • Type of sound source: Acoustic instruments produce a sound very different from amplified electric instruments, which are again different from the human voice.
  • Sound pressure levels: a mic that is recording Baroque lute will not face high sound pressure levels, however, a mic being used to record heavy metal drumming or brass instrument may face extreme sound pressure levels, potentially causing distortion in the signal.
  • Situational circumstances: Sometimes a microphone should not be visible, or having a microphone nearby is not appropriate. In scenes for a movie the microphone may be held above, out of the picture frame.
  • Processing: If the signal is destined to be heavily processed, or mixed down, a different type of input may be required.
  • The use of a windscreen for outdoor recording or a pop shield to reduce vocal plosives.

Basic techniques

[edit]
Lloyd Thayer playing National Steel guitar on a radio show.
Peter Brötzmann with a tenor saxophone and two microphones (a Neumann U 87 for the bell, an Electro-Voice RE20 for the body).

There are several classes of microphone placement for recording and amplification.

  • In close miking, a microphone is placed relatively close to an instrument or sound source, within three to twelve inches, producing a dry or non-reverberant sound.[1] This serves to reduce extraneous noise, including room reverberation and is commonly used when attempting to record a number of separate instruments while keeping the signals separate, or when trying to avoid feedback in an amplified performance. Close miking often affects the frequency response of the microphone, especially for directional mics that exhibit bass boost from the proximity effect. Ubiquitous in the tracking of instruments used in pop and popular music, examples of close-mic vocal tracks include many songs on Elliott Smith's Elliott Smith and Either/Or, Lily Allen's "The Fear", the chorus of Fergie's "Glamorous", Imogen Heap's lead on "Hide and Seek", and Madonna's spoken verses on "Erotica".[2]
  • In ambient or distant miking, a microphone – typically a sensitive one – is placed at some distance from the sound source. The goal of this technique is to get a broader, natural mix of the sound source or sources, along with ambient sound, including reverberation from the room or hall. Example include The Jesus and Mary Chain's Psychocandy (excepting the vocals), Robert Plant's vocals on songs from Physical Graffiti, Tom Waits's lead vocals on his junkyard records, and Mick Jagger's lead-vocals on songs from Exile On Main Street.[3]
  • In room miking a distant mic, referred to as the room mic, is used in conjunction with a close mic, the room mic is "typically placed far enough past the critical distance in a room that the room's ambience and reverberations transduce at an equivalent, if not greater, volume than the sound source itself."[4] Ubiquitous in pop, it is the industry standard for tracking rhythm guitars in rock. A celebrated example is the rhythm guitar on Led Zeppelin's "Communication Breakdown"[4] while other examples include John Frusciante's electric guitar parts on BloodSugarSexMagik, Noel Gallagher's lead-guitar on "Champagne Supernova", and Billy Corgan's guitar on "Cherub Rock", "Today", "Bullet With Butterfly Wings", "Zero", and "Fuck You (An Ode to No One)".[5]
  • Accent (or spot) microphone placement. Often, the tonal and ambient qualities will sound very different between a distant- and close-miked pickup. Under certain circumstances, it's difficult to obtain a naturally recorded balance when mixing the two together. For example, if a solo instrument within an orchestra needs an extra mic for added volume and presence, placing the mic too close would result in a pickup that sounds overly present, unnatural and out of context with the distant, overall orchestral pickup. To avoid this pitfall, a compromise in distance may be struck. A microphone that has been placed within a reasonably close range to an instrument or section within a larger ensemble (but not so close as to have an unnatural sound) is known as an accent (or spot) pickup. The amount of accent signal that's introduced into the mix should sound natural relative to the overall pickup, and a good accent mic should only add presence to a solo passage and not stick out as a separate, identifiable pickup.[6][7][8]
  • Instrumental use of microphones has been developed by many experimental composers, musicians and sound artists. They use microphones in unconventional ways, for example by preparing them with objects, moving them around or using contact microphones to colour the sound and be able to amplify otherwise very quiet sounds. Karlheinz Stockhausen used microphone movements by musicians in Mikrophonie I to discover the diverse sounds of a big tam-tam[9] and Pauline Oliveros amplified apple boxes with contact microphones.[10]

When using multiple microphones, respecting a 3-to-1 rule and placing microphones at least three times further from each other than they are from the source they are being used to pick up avoids cancellation and phase issues such as comb filtering when the microphone signals are mixed together.[11]

A parabolic microphone used to capture sounds on the field during a football game.

Parabolic microphones are used to capture sounds on the field during football games. The parabolic dish has been compared metaphorically to a telephoto lens, in the way that it can focus the capture of sound.[12]

Multitrack recording

[edit]

In multitrack recording often each instrument or vocalist is miked separately, with one or more microphones recording to separate tracks. Microphones may also be used to record the overall performance or just the ambiance of the performance space. At a later stage, the channels are combined into two channels for stereo or more for surround sound. Different levels from individual tracks are sent to final channels to position the elements in the stereo or surround sound-stage. The artists need not perform in the same place at the same time, and individual tracks (or sections of tracks) can be re-recorded to correct errors. Multitrack recording permits greater control over the final sound, but recording two channels (stereo recording) is simpler and cheaper, and can give a sound that is more natural.

Stereo recording techniques

[edit]

There are two features of sound that the human brain uses to place objects in the stereo soundfield between the loudspeakers. These are the relative level difference between the two channels Δ L, and the time delay difference in arrival times for the same sound in each channel Δ t. There are several established microphone configurations used in ambient or room stereo recording.

X-Y technique

[edit]
XY stereo

The X-Y technique uses two directional microphones at the same place, and typically rotated 90° or more with respect to each other.[13] A stereo effect is achieved through differences in pickup level between two microphones due to their directionality. Due to the lack of difference in time-of-arrival, the sonic characteristic of X-Y recordings is generally less spacey and has less depth compared to recordings employing an AB setup.

Blumlein pair stereo

When the microphones used are bidirectional and placed facing ±45° with respect to the sound source, the X-Y setup is called a Blumlein pair. The sonic image produced by this configuration is considered by many authorities to create a realistic soundstage.[citation needed]

A further refinement of the Blumlein pair was developed by EMI in 1958, who called it Stereosonic. They added in-phase crosstalk above 700 Hz to better align the mid and treble phantom sources with the bass ones.[14]

A-B technique

[edit]

A-B technique uses two parallel microphones, typically omnidirectional, some distance apart, capturing time-of-arrival stereo information as well as some level (amplitude) difference information, especially if employed close to the sound source(s). At a separation distance of about 50 cm, the time delay for a signal reaching first one and then the other microphone from the side is approximately 1.5 ms. If the distance is increased between the microphones it effectively decreases the pickup angle. At 70 cm distance, it is about equivalent to the pickup angle of the near-coincident ORTF setup.[citation needed]

M/S technique

[edit]
Mid-side stereo

Mid/side coincident technique employs a bidirectional microphone (with a figure of 8 polar pattern) facing sideways (Side) and a cardioid (generally a variety of cardioid, although Alan Blumlein described the usage of an omnidirectional transducer in his original patent) facing the sound source (Mid). The capsules are stacked vertically and brought together as closely as possible to minimize comb filtering caused by differences in arrival time.

The left and right channels are produced through a simple matrix: Left = Mid + Side, Right = Mid − Side ("minus" indicates you add the side signal with the polarity reversed). This configuration produces a completely mono-compatible signal and, if the Mid and Side signals are recorded (rather than the matrixed Left and Right), the stereo width (and with that, the perceived distance of the sound source) can be manipulated after the recording has taken place.

If the mid/side technique is incorporated into a self-contained stereo microphone assembly that outputs only the final left and right stereo pair signals, the original mid and side signals may still be recovered, permitting the above-mentioned manipulation. The mid signal is recovered by adding the left and right signals; The in phase and antiphase side signals cancel, giving a true mono signal in the process. The side signal is recovered by subtracting the right signal from the left; The mid signal is present in both channels and therefore cancels, leaving the side.

Jecklin disk technique

[edit]
Jecklin Disk

The Jecklin disk technique is similar to A-B recording, using two omnidirectional microphones spaced 36 cm apart. A sound-absorbing Jecklin disk with 35 cm diameter is placed between the two microphones. The disk makes the apparent separation between the mics much larger than an equivalent A-B recording.

Choosing a technique

[edit]

If a stereo signal is to be reproduced in mono, out-of-phase parts of the signal will cancel, which may cause the unwanted reduction or loss of some parts of the signal. Mono compatibility can be an important factor in choosing which technique to use.

  • Since the A-B techniques use phase differences to give the stereo image, they are the least compatible with mono.
  • The Jecklin disk has a relatively small distance between the mics. Frequencies coming in sideways can reach only one microphone and will not interfere. Therefore, the Jecklin disk works reasonably well when converted to mono. The Jecklin disk also gives phase shifts and amplitude differences matching well with what a real pair of ears would hear at this position and therefore is well suited for headphone playback.
  • In the X-Y techniques, the microphones would ideally be in exactly the same place, which is not possible – if they are slightly separated left to right, there may be some loss of high frequencies when played back in mono, so they are often separated vertically. This only causes problems with sound from above or below the height of the microphones.
  • The M/S technique is ideal for mono compatibility, since summing Left+Right just gives the Mid signal back.

The equipment for the techniques also varies from bulky to small and convenient. A-B techniques generally use two separate microphone units, often mounted on a bar to define the separation. X-Y microphone capsules can be mounted in one unit or even on the top of a handheld digital recorder.

M/S arrays can be very compact and fit easily into a standard blimp windscreen, which makes boom-operated stereo recordings possible. They provide a variable soundstage width to match a zoom lens, which can be manipulated in post-production. This makes M/S a very popular stereo technique for film location recording. They are often used in small pencil microphones to mount on video cameras, sometimes even coupled to the zoom.

References

[edit]
[edit]
Revisions and contributorsEdit on WikipediaRead on Wikipedia
from Grokipedia
Microphone practice encompasses the principles and techniques for selecting, positioning, and handling microphones to capture high-fidelity audio in applications such as , live sound reinforcement, , and . It focuses on optimizing sound quality by balancing direct source capture with ambient characteristics, while mitigating issues like noise, feedback, and phase interference through informed choices in equipment and setup. Central to microphone practice are the two primary transducer types: dynamic and condenser microphones. Dynamic microphones generate an electrical signal via , with sound waves vibrating a diaphragm attached to a within a ; they are durable, capable of handling high levels (SPLs) up to 150 dB or more, and do not require external power, making them suitable for rugged live environments and loud instruments like or guitar amplifiers. In contrast, condenser microphones employ a mechanism where sound alters the between a fixed backplate and a vibrating diaphragm, requiring (typically 48V) for operation; they provide superior , sensitivity, and frequency range (often 20 Hz to 20 kHz), ideal for capturing nuanced studio recordings of vocals or acoustic instruments. Other types, such as microphones, offer distinct tonal characteristics for specific applications. Polar patterns define a microphone's directional sensitivity and are essential for effective practice, as they influence isolation of the sound source from off-axis noise or reflections. Omnidirectional patterns exhibit uniform sensitivity in all directions, capturing a natural, immersive sound but prone to room ; cardioid patterns, shaped like a heart, offer maximum front-facing pickup with rejection of rear sounds (typically 20 dB or more at 180°), commonly used for vocals and instruments to reduce feedback; hypercardioid and supercardioid provide narrower front lobes with some rear sensitivity for focused live applications; while bidirectional (figure-of-eight) patterns pick up equally from front and back, useful in or interview setups. Overall, proficient microphone practice requires experimentation, acoustic awareness, and iterative adjustment to achieve professional results across diverse scenarios.

Fundamentals of Microphone Use

Microphone Characteristics

Microphones have evolved significantly since the late 19th century, beginning with carbon microphones invented by and improved by around 1877, which were widely used in early due to their simplicity but suffered from and limited fidelity. In 1917, E.C. Wente developed the first precision condenser microphone, marking a shift toward higher accuracy in capture for recording and . By the mid-20th century, tube-based condenser microphones, such as the Neumann U47 introduced in 1947, became staples in studios, imparting a characteristic warm tone through amplification that enhanced the richness of vocals and instruments in iconic recordings by artists like and . Modern electret condenser microphones, pioneered in 1962 by Gerhard Sessler and James West at , offer high performance at lower cost by incorporating a permanently charged material, replacing carbon types that dominated until the and enabling widespread use in professional and consumer applications. The primary types of microphones—dynamic, condenser, and —differ in transduction mechanisms, affecting their sensitivity, , and suitability for various sound sources. Dynamic microphones, using a moving coil in a , exhibit moderate sensitivity and can handle high levels (SPL) up to 150 dB without , making them ideal for loud sources like drums, as exemplified by the Beta 52A for kick drums. Their is wide but often tailored, with boosts in bass or mids for specific applications, though less linear than other types. Condenser microphones, employing a with a vibrating diaphragm, provide high sensitivity and a smooth, extended typically from 20 Hz to 20 kHz, capturing subtle details in vocals and acoustic instruments but requiring and being more susceptible to overload from extreme SPL. microphones, featuring a thin metal ribbon suspended in a , deliver a linear with excellent transient capture and a warm, natural tone, suiting studio vocals and guitars, though their lower output and fragility limit use with high-SPL sources like drums. Polar patterns define a microphone's directional sensitivity, influencing off-axis rejection and the amount of room sound captured. Cardioid patterns, heart-shaped, prioritize sound from the front while rejecting rear sources by about 20 dB at 180° and reducing off-axis noise by roughly 6 dB at 90°, minimizing unwanted room in live or noisy environments. Omnidirectional patterns exhibit equal sensitivity in all directions, offering no off-axis rejection but capturing full room ambiance and natural acoustics, suitable for controlled studio spaces. Figure-8 (bidirectional) patterns accept sound from front and rear equally while strongly rejecting sides, providing good isolation for dual-source setups but incorporating more room sound than cardioids. Key specifications quantify microphone performance, guiding selection for recording practices. Frequency range, often 20 Hz to 20 kHz for human hearing coverage, indicates the spectrum reproduced faithfully, with condensers typically achieving flatter responses across this band. Sensitivity, measured in dBV/Pa at 94 dB SPL and 1 kHz, reflects output voltage per unit pressure; dynamic mics range from -70 dBV/Pa, while condensers and electrets fall between -46 dBV/Pa and -35 dBV/Pa (5–18 mV/Pa), with higher values enabling capture of quieter sources. Self-noise, expressed as equivalent SPL (e.g., 20–30 dB(A)), represents inherent electrical noise; lower levels, such as 29 dB SPL in advanced MEMS models, are crucial for low-level recordings like ambient or classical music, where signal-to-noise ratios exceed 60 dB.

Placement Principles

The proximity effect refers to the exaggerated low-frequency response experienced when using directional microphones, such as cardioid patterns, positioned close to a source. This phenomenon arises from the pressure-gradient principle in these microphones, where the difference in across the diaphragm increases disproportionately at low frequencies as distance decreases, leading to a bass boost that can exceed 10 dB at distances under 15 cm. For vocal recording, typical close placement distances of 2.5–15 cm (1–6 inches) often invoke this effect, allowing performers to control warmth by varying their distance from the . Phase alignment is a critical principle in multi-microphone setups to prevent comb filtering, a form of destructive interference that creates frequency notches when sound waves from the same source arrive at slightly different times at separate . The 3:1 rule provides a practical guideline for minimizing this issue: the distance between two should be at least three times the distance from each to its intended sound source, ensuring that the level of bleed from one source into the adjacent is attenuated by approximately 9–10 dB due to the . This spacing helps maintain signal coherence without excessive phase shifts, though it is a and may require adjustments based on room acoustics. Off-axis response plays a key role in minimizing bleed, or unwanted sound leakage between microphones, by leveraging the directional characteristics of microphones to reject signals arriving from angles outside the primary pickup zone. For instance, cardioid microphones exhibit reduced sensitivity off-axis, often by 10 dB or more at 90–180 degrees, allowing engineers to angle the microphone so that its null point faces potential interfering sources, thereby capturing the desired on-axis while suppressing spill. This technique, combined with precise angling to optimize the direct-to-reverberant ratio, enhances isolation without additional barriers. Acoustic considerations in microphone placement include the boundary effect, which occurs when a microphone is positioned near a reflective surface, such as a or , causing low-frequency reinforcement due to in-phase reflections that effectively double the at the diaphragm. This can result in a 6 dB bass boost near a single surface, escalating to 12 dB in a corner, potentially leading to an overly boomy response unless compensated by positioning at least 1 meter from boundaries or using low-cut filters. To mitigate handling noise and mechanical vibrations that transmit through stands—such as footfalls or stand adjustments—shock mounts are employed, which suspend the microphone in elastic suspensions to dampen resonances below 8 Hz, significantly reducing low-frequency rumble in environments prone to physical disturbances.

Mono Recording Techniques

Close Miking

Close miking is a technique that involves positioning the very close to the source, typically within a few inches, to capture primarily the direct while minimizing environmental reflections and ambient . This approach maximizes the desired signal relative to unwanted sounds, providing greater isolation in mono recording setups. The primary benefits of close miking include achieving high gain before feedback in live settings, as the proximity to the source allows for sufficient signal level without excessive amplification that could trigger acoustic feedback. It also reduces the capture of room ambience, resulting in a cleaner, more controlled recording ideal for loud sources such as guitar amplifiers or percussion, where separation from other instruments is crucial. Additionally, close miking enhances the by increasing the sound pressure level at the capsule, which is particularly useful for quieter or detailed sources. In vocal close miking, the is typically placed 2 to 6 inches from the singer's mouth to capture intimate, direct sound while managing proximity effect, which boosts low frequencies and can be adjusted by slight distance variations. A is essential, positioned 2 to 3 inches in front of the microphone to diffuse bursts from consonants like "p" and "b," preventing and ensuring clarity without altering the vocal . This setup is standard for studio recordings, allowing performers to maintain consistent distance for balanced dynamics. For instrument applications, close miking the often involves placing a dynamic , such as the , 1 to 2 inches above the top head near the edge or rim, angled slightly toward the center to capture the snap and attack while avoiding excessive bleed. On the bottom side, a similar can be positioned 1 to 2 inches below the resonant head to emphasize snare wire rattle, blended in for added texture. For guitar cabinets, the is commonly placed 1 to 6 inches from the speaker cone, off-axis at a 45-degree angle to tame harsh highs and capture a balanced tone with enhanced low-end from proximity effect. Despite its advantages, close miking has drawbacks, including an overly dry sound that lacks natural , often requiring artificial reverb to be added during mixing to restore spatial depth. Additionally, sources producing sound pressure levels exceeding the microphone's limits, such as over 140 dB from loud drums or amps, can cause , necessitating robust dynamic microphones designed for high SPL handling. The technique may also exaggerate tonal imbalances due to proximity effect, where low frequencies are boosted within 6 inches, potentially requiring EQ correction.

Distant and Room Miking

Distant miking refers to the placement of at least 2 meters (approximately 6.5 feet) from the sound source, allowing the capture of both direct sound and the surrounding room acoustics for a more natural blend. This approach is especially effective for acoustic instruments and ensembles, where it helps maintain tonal balance and incorporates environmental ambience without isolating the source excessively. Omnidirectional polar patterns are commonly used in such setups to fully embrace the room's reflections. In room miking configurations, microphones are positioned to emphasize the space's contribution to the sound. For , an overhead setup often involves placing microphones 2–3 meters (6.5–10 feet) from the instrument, directed toward the strings to blend the direct tone with room reverb for a spacious result. Similarly, for orchestral sections, multiple microphones—such as directional models positioned 1–1.5 meters from groups of 3–4 musicians—can capture sectional balance while integrating ambient details from the hall. The influence of acoustic treatment on distant miking is significant, as it shapes how the room contributes to the recording. Dead rooms, with minimal reverb (RT60 under 0.3 seconds), provide clarity but can sound unnaturally dry for acoustic sources, while live rooms with longer decay times add warmth and depth. Absorbers, such as panels distributed on walls and ceilings, are used to control reverb time, targeting an ideal RT60 of 0.5–1 second in project studios to balance intimacy and naturalness without excessive muddiness. Distant and room miking present challenges like increased bleed from nearby sources and a higher due to greater sensitivity to environmental sounds and reflections. These issues can lead to comb filtering or masked details if not addressed. Solutions include strategic acoustic treatment to minimize unwanted reflections and the application of high-pass filters at 80–100 Hz to attenuate low-frequency rumble and bleed without significantly altering the desired tone.

Multitrack Recording Practices

Instrument-Specific Miking

In multitrack recording, drums require targeted miking to isolate elements while capturing their dynamic interplay. For the kick drum, a common technique places a dynamic inside the drum, positioned off-center at about one-third of the drum's and halfway back into the shell, typically 4–6 inches from the batter head to emphasize attack and low-end punch. Toms are often miked closely with dynamic microphones angled downward near the outer rim of the batter head to minimize bleed and highlight . Overheads, using a spaced pair of condenser microphones positioned above the kit, focus on cymbals to provide air and high-frequency detail without overpowering the close-miked . Electric guitars in multitrack setups benefit from close miking of the to capture tonal nuances from the speaker cone. A dynamic , such as the , is typically positioned at the edge of the cone, about a pinky-width from the grille cloth and halfway between the center and rim, to balance mids and low end while reducing harshness. For , blending a miked amp signal with direct injection (DI) is standard; the amp mic uses a dynamic model placed 1 inch or more from the speaker cone to add cabinet warmth, while the DI provides clean lows, with phase alignment achieved by delaying the DI track to match the mic's acoustic delay. Keyboards like the grand demand a combination of close and room miking for balanced in layered recordings. Close mics, often small-diaphragm condensers, are placed 6 inches above the hammers and 15 inches apart inside the lid to accentuate percussive attack and string clarity. Room mics, such as an XY pair of small-diaphragm condensers 4–8 feet in front at 5–6 feet high, add natural ambience and sustain. For strings like the , a small-diaphragm condenser microphone positioned near the f-hole captures the instrument's resonant body and bow detail effectively in multitrack isolation. Brass instruments in multitrack environments use careful distancing to control harshness and sibilance. A is employed 6–12 inches from the bell to mitigate breath noise and plosives, while the —often a condenser for detail—is placed slightly off-axis at 12–16 inches to soften highs without losing projection. Woodwind instruments require similar attention to placement for clarity, with the typically positioned about 18 inches from the instrument, slightly off-axis to reduce key noise and breath sounds, varying by type (e.g., 1–3 feet for in classical settings). In both cases, the 3:1 rule—ensuring the distance between microphones is at least the distance from each mic to its source—helps minimize bleed and phase issues in multitrack setups.

Overdubbing and Layering Strategies

Overdubbing involves selectively rerecording portions of a track to correct errors or enhance performances, often using the punch-in technique to seamlessly insert new audio without disrupting the existing recording. In this process, engineers drop the performer into record mode at a precise point, typically triggered manually or via , to fix issues like off-pitch notes or timing errors while preserving the original take's momentum. For instance, during the recording of Foo Fighters' , producer frequently employed punching in on drum tracks as an alternative to tape splicing, allowing quick corrections that maintained the session's energy, though audible punch points could sometimes be detected on close listening. To minimize disruptions, performers rely on headphone monitoring, but controlling bleed—the unwanted pickup of monitor mix through the —is essential; closed-back headphones, which seal around the ears to prevent sound leakage, are standard for this purpose, significantly reducing the amount of instrumental bleed captured during vocal or instrument overdubs. Layering doubles and harmonies builds depth by recording multiple takes of the same part, stacked to create a fuller, more robust sound, particularly effective for vocals where consistency in placement ensures tonal uniformity across layers. Engineers position the at the same distance and angle for each take—typically 6 to 12 inches from the singer's mouth for close-miking vocals—to match the and avoid introducing variations that could unbalance the blend. This approach is crucial for harmony stacks, as even slight shifts in position can alter proximity effect or high-frequency capture, leading to mismatched layers that require extensive EQ correction in mixing; for example, maintaining a fixed singer-to- distance helps replicate the natural of the lead vocal in supporting harmonies, promoting cohesive layering without artificial doubling effects. Managing track counts in workflows prevents session overload and maintains clarity, especially when layering sections like drums or guitars where multiple microphones increase the risk of phase interference. Similar sources, such as overhead and close mics on a , are often bused to —auxiliary routing channels—for collective processing like compression or EQ, which streamlines mixing and allows unified phase alignment checks across the group. In sections, phase issues arise from time delays between mics capturing the same transient (e.g., a snare hit), potentially causing cancellation; busing facilitates inverting polarity on individual tracks within the to test and correct these, ensuring the layered elements reinforce rather than detract from the groove. This technique, rooted in analog console practices, remains vital even in digital environments to handle the proliferation of tracks during successive overdubs. Modern and layering have evolved through hybrid analog-digital tracking, combining the warmth of analog preamps and tape saturation with the precision of workstations (DAWs) for virtually unlimited track layering. DAW integration, accelerating in the 1990s with tools like (introduced in 1991), eliminated the physical limitations of tape machines, enabling engineers to overdub and layer indefinitely without bouncing tracks down to free space—a process that previously constrained creativity to 24 or 48 tracks. This shift allowed for complex arrangements, such as intricate vocal stacks or multi-guitar layers, by facilitating non-destructive editing and easy recall of sessions, fundamentally transforming microphone practice from rigid tape-era workflows to flexible, iterative digital processes.

Stereo Recording Techniques

Coincident Pair Methods

Coincident pair methods involve positioning two capsules as closely as possible—ideally at the exact same point in space—to capture audio through interchannel intensity differences rather than time delays, ensuring phase coherence and excellent mono compatibility. This approach minimizes phase cancellation issues that can arise in configurations, providing precise and localization of sources. Typically employing directional polar patterns such as cardioids, these techniques are favored for close to medium-distance recording of centered sources like vocals or , where accurate spatial representation is essential. The X-Y technique, a classic coincident pair method, uses two cardioid microphones angled at 90 degrees to each other, with their capsules intersecting at a single point, often aligned at ±45 degrees relative to the sound source. This setup creates a sharp image with good , suitable for focused recordings such as solo instruments or ensembles where width can be adjusted post-capture via panning. Angles may vary to 135 degrees for broader imaging, but the standard 90-degree configuration balances clarity and spread without introducing comb filtering. A near-coincident variant, the ORTF technique, employs two cardioid microphones spaced 17 cm apart at a 110-degree angle, approximating human ear geometry to enhance natural width and ambience while retaining much of the phase coherence of true coincident pairs. Developed in the early by the French broadcasting organization Office de Radiodiffusion-Télévision Française (ORTF), now part of , it offers superior spaciousness compared to strict X-Y setups, making it ideal for orchestral or live ensemble captures. Similarly, the NOS technique, devised in the by the Dutch broadcasting foundation (NOS), uses cardioids at a 90-degree angle with 30 cm spacing and was widely adopted for 1970s broadcasts to achieve realistic stereo reproduction with minimal artifacts. Key advantages of coincident pair methods include their robustness against phase problems, enabling seamless collapse to mono without loss of center imaging, and precise source localization that supports adjustable stereo width in mixing. For optimal results, microphones should be identical models to ensure level matching within 1 dB, positioned at equal distances from the source, and mounted vertically or horizontally using a stereo bar or adaptor to maintain capsule alignment. Directional patterns like cardioids are preferred to reject off-axis noise, and fine adjustments to angle and distance should prioritize the equilateral triangle principle between mics and listener for balanced playback.

Spaced Pair Methods

Spaced pair methods, also known as A-B techniques, utilize two omnidirectional microphones positioned several feet apart and pointed in parallel to capture a wide stereo image through interaural time differences and level variations. This approach typically involves spacing the microphones 1 to 10 feet apart, with closer distances like 3 feet suitable for more intimate sources to balance width and phase coherence. The technique relies on the natural arrival time disparities of sound waves at each microphone, producing an expansive soundstage that mimics binaural perception. A notable variant that achieves a spacing despite its coincident placement is the , consisting of two figure-8 microphones crossed at 90 degrees. Invented by British engineer in the 1930s as part of his pioneering work on stereo recording, this method was first demonstrated at in 1934, creating a realistic depth and width by leveraging the bidirectional patterns to simulate spatial separation. These methods offer an expansive and natural soundstage, particularly effective for capturing room ambience and low-frequency richness in reverberant environments. However, they are phase-sensitive, with potential comb filtering arising from time-of-arrival differences that can cause cancellations, especially when signals are summed to mono. Such phase issues can result in a hollow or uneven tonal response, though they may be mitigated through time-alignment adjustments in workstations (DAWs) to synchronize the microphone signals. Spaced pair techniques excel in applications like recording grand pianos or orchestras, where microphones are placed at medium distances—such as 1.7 meters from the source for pianos—to capture full transients and spatial depth without close-miking's level inconsistencies. They are ideal for live room captures of ensembles or choirs but should be avoided for very close sources due to uneven amplitude responses across the stereo field. The Blumlein pair similarly suits orchestral and piano recordings, positioned 1 to 3 feet from the instrument or 8 feet in front of larger groups for balanced imaging.

Mid-Side and Specialized Techniques

The Mid-Side (M/S) stereo recording technique employs a coincident pair of microphones: a forward-facing cardioid microphone capturing the "mid" signal, which represents the central audio content, and a figure-8 microphone oriented sideways to capture the "side" signal, encoding left-right differences. This setup, developed by and patented in 1933 as part of his foundational work on systems, allows for flexible that can be adjusted during . The technique was further popularized in the by Danish Holger Lauridsen for creating mono-compatible stereo recordings. Decoding the M/S signals into conventional left-right stereo channels involves a simple matrixing process based on sum and difference operations: L=M+SL = M + S R=MSR = M - S Here, LL is the left channel, RR is the right channel, MM is the mid signal, and SS is the side signal; a scaling factor kk can be applied to SS (e.g., L=M+kSL = M + kS, R=MkSR = M - kS) to control stereo width, enabling narrowing for mono compatibility or widening for enhanced imaging without phase issues. This post-production adjustability ensures 100% mono compatibility, as summing LL and RR yields 2M2M, canceling the side signal while preserving the centered content. The Jecklin disc technique, devised by Swiss engineer Jürg Jecklin in the 1970s, uses two omnidirectional microphones positioned 16-18 cm apart behind a 30 cm diameter acoustically absorbing disc, which creates a baffle effect simulating head-related transfer functions to enhance stereo separation and naturalness. The disc, typically covered in sound-absorbing material like foam or felt, attenuates high frequencies from the sides while allowing omnidirectional capture, producing a binaural-like stereo image suitable for loudspeaker playback. In niche applications, M/S is favored for soundtracks due to its coincident nature and adjustable width, facilitating immersive yet flexible surround mixes in post-production environments. Similarly, the Jecklin disc excels in recording classical ensembles within reverberant halls, such as venues or churches, where its baffle simulates listener perspective to balance direct sound and natural ambiance without excessive comb filtering.

Technique Selection and Applications

Selecting the appropriate microphone technique depends on several key criteria, including the size and nature of the sound source, the characteristics of the recording environment, and the need for mono compatibility. For point sources such as solo vocals or individual instruments, coincident techniques like X-Y are preferred because they provide a focused, centered with minimal phase differences, ensuring precise localization without spreading the source across the stereo field. In contrast, spaced pair methods like A-B are better suited for larger ensembles, such as choirs or orchestras, where the wider microphone separation captures a more natural spatial spread and depth, accommodating the extended sound field of multiple performers. Room acoustics play a critical role in technique selection, as reverberant spaces benefit from methods that can emphasize or control ambience. Mid-side (M/S) recording excels here by allowing adjustment of the side signal's level, which modulates the amount of room reverb and overall width without altering the core mid signal, providing flexibility in dry or lively environments. Mono compatibility is another essential factor, particularly for broadcast or playback scenarios where may collapse to mono; M/S ensures this by design, as the side channels cancel out completely, leaving a clean, uncolored mid signal, while coincident pairs like X-Y also maintain good compatibility due to their lack of time-based phase issues. Spaced pairs, however, can introduce comb filtering in mono sums from inter-channel delays, making them less ideal unless the room is controlled. Genre-specific applications further guide technique choice, tailoring the stereo image to artistic goals. In pop music, the X-Y coincident pair is commonly used for vocals and small ensembles, delivering tight, accurate imaging that highlights lead elements in a controlled studio setting without excessive spaciousness. For classical music, spaced pair A-B techniques are favored for orchestras, creating an open, enveloping sound that replicates the natural hall ambience and instrument placement across a wide stage. In the 2020s, these traditional methods are increasingly hybridized with immersive audio for virtual reality (VR) and spatial formats, such as combining stereo pairs with object-based metadata to enable dynamic positioning in 3D environments like Dolby Atmos or binaural playback. Digital tools aid in technique selection by simulating outcomes before recording. Polar pattern calculators, such as those allowing combination of up to 12 microphones with adjustable angles and phases, help visualize resulting stereo images and mono compatibility, enabling engineers to optimize setups like M/S or X-Y for specific sources. DAW plugins, including evolutions in iZotope RX since version 6 around 2017, facilitate phase preview through modules like Advanced Azimuth, which automatically aligns stereo pairs by detecting transients and suggesting delays, preventing issues in multi-mic recordings. Emerging practices extend techniques into full-sphere audio, with binaural microphones gaining traction for 360-degree recordings that emulate human ear cues in two channels, ideal for VR content where quick capture of immersive atmospheres like live crowds enhances spatial realism without complex arrays. integration complements this by using a single coincident array to record in A-format, convertible to B-format for manipulation of virtual stereo pairs, allowing flexible width and orientation adjustments that bridge traditional with higher-order immersive formats.

Surround Recording Techniques

Surround recording techniques extend stereo methods to multichannel formats like 5.1 or immersive audio, using microphone arrays to capture spatial audio with front, surround, and sometimes height channels, emphasizing phase coherency for seamless integration and compatibility with downmixes. The Double Mid-Side (Double M/S) technique adapts the Mid-Side stereo method for surround sound, employing a forward-facing cardioid for the front mid, a rear-facing cardioid for the rear mid, and a shared figure-8 microphone for the side signal, all in a coincident configuration. Signals are matrixed in post-production: center from front mid, left from front mid plus side, right from front mid minus side, left surround from rear mid plus side, and right surround from rear mid minus side, allowing adjustable width and mono compatibility as side signals cancel in downmixes. This method minimizes phase issues due to its coincident nature but requires matched microphones to avoid frequency response discrepancies between cardioids and figure-8 patterns. It is applied in compact setups for music and film, providing flexibility for immersive mixes in environments like studios or live events. The Optimized Cardioid Triangle (OCT) focuses on the front channels (left, center, right) using a central cardioid microphone positioned 8 cm ahead of two outward-angled higher-order cardioids spaced 40-90 cm apart, creating recording angles from 90° to 160° for high channel separation. An OCT2 variant increases the center offset to 40 cm for greater spaciousness akin to the Decca Tree. As a near-coincident array, it reduces phase cancellations compared to spaced methods, though low-frequency response may require omnidirectional supplements; surround channels need separate arrays. OCT suits orchestral front captures in 5.1 formats, offering precise localization in reverberant halls. The IRT Cross, developed for ambient capture in surround recordings, arranges four cardioid microphones in a 20-25 cm square to feed left, right, left surround, and right surround channels, positioned behind a main front array to enhance diffuse sound. Its spaced design leverages time-of-arrival differences for envelopment but risks echoes if distanced too far from the source, necessitating close integration with primary mics for phase alignment. Less mono-compatible due to potential comb filtering, it excels in classical and live ensemble recordings to add natural ambiance without overpowering direct sound. Other methods include the Wide Cardioid Surround Array (WCSA), using five wide cardioids spaced 60-200 cm for equal timbre and broad sweet spots in 5.1 setups, ideal for symphonic and pop concerts with careful spacing to manage phase decorrelation. The Fukada Tree employs five cardioids in a Decca-like front triplet with omnidirectional outriggers for improved separation in music productions. These techniques prioritize phase coherency through matched patterns and positioning, ensuring compatibility with stereo downmixes and applications in immersive formats like Dolby Atmos.

References

Add your contribution
Related Hubs
User Avatar
No comments yet.