Hubbry Logo
MIDIMIDIMain
Open search
MIDI
Community hub
MIDI
logo
8 pages, 0 posts
0 subscribers
Be the first to start a discussion here.
Be the first to start a discussion here.
MIDI
MIDI
from Wikipedia

Musical Instrument Digital Interface
International standardMIDI Standard
Developed byMMA
IntroducedFebruary 1, 1983; 42 years ago (1983-02-01)
IndustryMusic technology (electronic and digital)
Connector type5‑pin DIN (with modern alternatives: USB, Bluetooth, Ethernet)
Compatible hardwareSerial transmission at 31,250 baud
Physical rangeUp to 15 meters (50 feet)
Superseded byMIDI 2.0
Websitehttps://www.midi.org
Example of music created in MIDI format
Several rack-mounted synthesizers that share a single controller
Using MIDI, a single controller (often a musical keyboard, as pictured here) can play multiple electronic instruments, which increases the portability and flexibility of stage setups. This system fits into a single rack case, but before the advent of MIDI, it would have required four separate full-size keyboard instruments, plus outboard mixing and effects units.

Musical Instrument Digital Interface (/ˈmɪdi/; MIDI) is an American-Japanese technical standard that describes a communication protocol, digital interface, and electrical connectors that connect a wide variety of electronic musical instruments, computers, and related audio devices for playing, editing, and recording music.[1] A single MIDI cable can carry up to sixteen channels of MIDI data, each of which can be routed to a separate device. Each interaction with a key, button, knob or slider is converted into a MIDI event, which specifies musical instructions, such as a note's pitch, timing and velocity. One common MIDI application is to play a MIDI keyboard or other controller and use it to trigger a digital sound module (which contains synthesized musical sounds) to generate sounds, which the audience hears produced by a keyboard amplifier. MIDI data can be transferred via MIDI or USB cable, or recorded to a sequencer or digital audio workstation to be edited or played back.[2]

MIDI also defines a file format that stores and exchanges the data. Advantages of MIDI include small file size, ease of modification and manipulation and a wide choice of electronic instruments and synthesizer or digitally sampled sounds.[3]: 4  A MIDI recording of a performance on a keyboard could sound like a piano or other keyboard instrument; however, since MIDI records the messages and information about their notes and not the specific sounds, this recording could be changed to many other sounds, ranging from synthesized or sampled guitar or flute to full orchestra.

Before the development of MIDI, electronic musical instruments from different manufacturers could generally not communicate with each other. This meant that a musician could not, for example, plug a Roland keyboard into a Yamaha synthesizer module. With MIDI, any MIDI-compatible keyboard (or other controller device) can be connected to any other MIDI-compatible sequencer, sound module, drum machine, synthesizer, or computer, even if they are made by different manufacturers.

MIDI technology was standardized in 1983 by a panel of music industry representatives and is maintained by the MIDI Manufacturers Association (MMA). All official MIDI standards are jointly developed and published by the MMA in Los Angeles, and the MIDI Committee of the Association of Musical Electronics Industry (AMEI) in Tokyo. In 2016, the MMA established The MIDI Association (TMA) to support a global community of people who work, play, or create with MIDI.[4]

History

[edit]

In the early 1980s, there was no standardized means of synchronizing electronic musical instruments manufactured by different companies.[5] Manufacturers had their own proprietary standards to synchronize instruments, such as CV/gate, DIN sync and Digital Control Bus (DCB).[6] Ikutaro Kakehashi, the president of Roland, felt the lack of standardization was limiting the growth of the electronic music industry.[6] In June 1981, he proposed developing a standard to the Oberheim Electronics founder Tom Oberheim,[5] who had developed his own proprietary interface, the Oberheim Parallel Bus.[7]

Kakehashi felt that Oberheim's system was too cumbersome, and spoke to Dave Smith, the president of Sequential Circuits, about creating a simpler, cheaper alternative.[7] While Smith discussed the concept with American companies, Kakehashi discussed it with Japanese companies Yamaha, Korg and Kawai.[5] Representatives from all companies met to discuss the idea in October.[5] Initially, only Sequential Circuits and the Japanese companies were interested.[8]

Dave Smith (right), one of the creators of MIDI

Using Roland's DCB as a basis,[6] Smith and Sequential Circuits engineer Chet Wood devised a universal interface to allow communication between equipment from different manufacturers. Smith and Wood proposed this standard in a paper, Universal Synthesizer Interface,[9] at the Audio Engineering Society show in October 1981.[10][11]: 4  The standard was discussed and modified by representatives of Roland, Yamaha, Korg, Kawai, and Sequential Circuits.[5][12]: 20  Kakehashi favored the name Universal Musical Interface (UMI), pronounced you-me,[7] but Smith felt this was "a little corny".[13] However, he liked the use of instrument instead of synthesizer, and proposed Musical Instrument Digital Interface (MIDI).[13][11]: 4  Robert Moog, the president of Moog Music, announced MIDI in the October 1982 issue of Keyboard.[14]: 276 

At the 1983 Winter NAMM Show, Smith demonstrated a MIDI connection between Prophet 600 and Roland JP-6 synthesizers. The MIDI specification was published in August 1983.[5] The MIDI standard was unveiled by Kakehashi and Smith, who received Technical Grammy Awards in 2013 for their work.[15][16][17] In 1983, the first instruments were released with MIDI, the Roland Jupiter-6 and the Prophet 600. In 1983, the first MIDI drum machine, the Roland TR-909,[18][19] and the first MIDI sequencer, the Roland MSQ-700, were released.[20]

The MIDI Manufacturers Association (MMA) was formed following a meeting of "all interested companies" at the 1984 Summer NAMM Show in Chicago. The MIDI 1.0 Detailed Specification was published at the MMA's second meeting at the 1985 Summer NAMM Show. The standard continued to evolve, adding standardized song files in 1991 (General MIDI) and adapted to new connection standards such as USB and FireWire. In 2016, the MIDI Association was formed to continue overseeing the standard.[8] In 2017, an abridged version of MIDI 1.0 was published as an international standard IEC 63035.[21] An initiative to create a 2.0 standard was announced in January 2019.[22] The MIDI 2.0 standard was introduced at the 2020 Winter NAMM Show.[23]

The BBC cited MIDI as an early example of open-source technology. Smith believed MIDI could only succeed if every manufacturer adopted it, and so "we had to give it away".[24]

Impact

[edit]

MIDI's appeal was originally limited to professional musicians and record producers who wanted to use electronic instruments in the production of popular music. The standard allowed different instruments to communicate with each other and with computers, and this spurred a rapid expansion of the sales and production of electronic instruments and music software.[12]: 21  This interoperability allowed one device to be controlled from another, which reduced the amount of hardware musicians needed.[25] MIDI's introduction coincided with the dawn of the personal computer era and the introduction of samplers and digital synthesizers.[26] The creative possibilities brought about by MIDI technology are credited for helping revive the music industry in the 1980s.[27]

MIDI introduced capabilities that transformed the way many musicians work. MIDI sequencing makes it possible for a user with no notation skills to build complex arrangements.[28] A musical act with as few as one or two members, each operating multiple MIDI-enabled devices, can deliver a performance similar to that of a larger group of musicians.[29] The expense of hiring outside musicians for a project can be reduced or eliminated,[2]: 7  and complex productions can be realized on a system as small as a synthesizer with integrated keyboard and sequencer.

MIDI also helped establish home recording. By performing preproduction in a home environment, an artist can reduce recording costs by arriving at a recording studio with a partially completed song.[2]: 7–8  In 2022, the Guardian wrote that MIDI remained as important to music as USB was to computing, and represented "a crucial value system of cooperation and mutual benefit, one all but thrown out by today's major tech companies in favour of captive markets". In 2005, Smith's MIDI Specification was inducted into the TECnology Hall of Fame, an honor given to "products and innovations that have had an enduring impact on the development of audio technology."[30] As of 2022, Smith's original MIDI design is still in use.[31]

Applications

[edit]

Instrument control

[edit]

MIDI was invented so that electronic or digital musical instruments could communicate with each other and so that one instrument can control another. For example, a MIDI-compatible sequencer can trigger beats produced by a drum sound module. Analog synthesizers that have no digital component and were built prior to MIDI's development can be retrofitted with kits that convert MIDI messages into analog control voltages.[14]: 277  When a note is played on a MIDI instrument, it generates a digital MIDI message that can be used to trigger a note on another instrument.[2]: 20  The capability for remote control allows full-sized instruments to be replaced with smaller sound modules, and allows musicians to combine instruments to achieve a fuller sound, or to create combinations of synthesized instrument sounds, such as acoustic piano and strings.[32] MIDI also enables other instrument parameters (volume, effects, etc.) to be controlled remotely.

Synthesizers and samplers contain various tools for shaping an electronic or digital sound. Filters adjust timbre, and envelopes automate the way a sound evolves over time after a note is triggered.[33] The frequency of a filter and the envelope attack (the time it takes for a sound to reach its maximum level), are examples of synthesizer parameters, and can be controlled remotely through MIDI. Effects devices have different parameters, such as delay feedback or reverb time. When a MIDI continuous controller number (CCN) is assigned to one of these parameters, the device responds to any messages it receives that are identified by that number. Controls such as knobs, switches, and pedals can be used to send these messages. A set of adjusted parameters can be saved to a device's internal memory as a patch, and these patches can be remotely selected by MIDI program changes.[a][34]

Composition

[edit]

MIDI events can be sequenced with computer software, or in specialized hardware music workstations. Many digital audio workstations (DAWs) are specifically designed to work with MIDI as an integral component. MIDI piano rolls have been developed in many DAWs so that the recorded MIDI messages can be easily modified.[35][better source needed] These tools allow composers to audition and edit their work much more quickly and efficiently than did older solutions, such as multitrack recording.[citation needed] Compositions can be programmed for MIDI that are impossible for human performers to play.[36]

Because a MIDI performance is a sequence of commands that create sound, MIDI recordings can be manipulated in ways that audio recordings cannot. It is possible to change the key, instrumentation or tempo of a MIDI arrangement,[37]: 227  and to reorder its individual sections,[38] or even edit individual notes. The ability to compose ideas and quickly hear them played back enables composers to experiment.[39]: 175 

Algorithmic composition programs provide computer-generated performances that can be used as song ideas or accompaniment.[2]: 122 

Some composers may take advantage of the standard, portable set of commands and parameters in MIDI 1.0 and General MIDI (GM) to share musical data files among various electronic instruments. The data composed via the sequenced MIDI recordings can be saved as a standard MIDI file (SMF), digitally distributed, and reproduced by any computer or electronic instrument that also adheres to the same MIDI, GM, and SMF standards. MIDI data files are much smaller than corresponding recorded audio files.[citation needed]

Use with computers

[edit]

The personal computer market stabilized at the same time that MIDI appeared, and computers became a viable option for music production.[14]: 324  In 1983 computers started to play a role in mainstream music production.[40] In the years immediately after the 1983 ratification of the MIDI specification, MIDI features were adapted to several early computer platforms. The Yamaha CX5M introduced MIDI support and sequencing in an MSX system in 1984.[41]

The spread of MIDI on home computers was largely facilitated by Roland Corporation's MPU-401, released in 1984, as the first MIDI-equipped sound card, capable of MIDI sound processing[42] and sequencing.[43][44] After Roland sold MPU sound chips to other sound card manufacturers,[42] it established a universal standard MIDI-to-PC interface.[45] The widespread adoption of MIDI led to computer-based MIDI software being developed.[40] Soon after, a number of platforms began supporting MIDI, including the Apple II, Macintosh, Commodore 64, Amiga, Acorn Archimedes, and IBM PC compatibles.[14]: 325–7  The 1985 Atari ST shipped with MIDI ports as part of the base system.

In 2015, Retro Innovations released the first MIDI interface for a VIC-20, making the computer's four voices available to electronic musicians and retro-computing enthusiasts for the first time.[46] Retro Innovations also makes a MIDI interface cartridge for Tandy Color Computer and Dragon computers.[47]

Chiptune musicians also use retro gaming consoles to compose, produce and perform music using MIDI interfaces. Custom interfaces are available for the Family Computer/Nintendo Entertainment System,[48] Game Boy,[49] Game Boy Advance[50] and Sega Mega Drive/Sega Genesis.[51]

Computer files

[edit]
MIDI files contain sound events such as a finger striking a key, which can be visualized using software such as Synthesia.

A MIDI file is not an audio recording. Rather, it is a set of instructions – for example, for pitch or tempo – and can use a thousand times less disk space than the equivalent recorded audio.[52][53] Due to their tiny filesize, fan-made MIDI arrangements became an attractive way to share music online, before the advent of broadband internet access and multi-gigabyte hard drives.[54] The major drawback to this is the wide variation in quality of users' audio cards, and in the actual audio contained as samples or synthesized sound in the card that the MIDI data only refers to symbolically. Even a sound card that contains high-quality sampled sounds can have inconsistent quality from one sampled instrument to another.[52] Early budget-priced cards, such as the AdLib and the Sound Blaster and its compatibles, used a stripped-down version of Yamaha's frequency modulation synthesis (FM synthesis) technology[55] played back through low-quality digital-to-analog converters. The low-fidelity reproduction[52] of these ubiquitous[55] cards was often assumed to somehow be a property of MIDI itself. This created a perception of MIDI as low-quality audio, while in reality MIDI itself contains no sound,[56] and the quality of its playback depends entirely on the quality of the sound-producing device.[37]: 227 

Standard MIDI File
Filename extensions
.mid .midi .smf (Standard MIDI File) .rmi (RIFF MIDI)
Internet media type
audio/midi audio/x-midi
Type codeMidi
Uniform Type Identifier (UTI)public.midi-audio"midi". Apple Developer Documentation: Uniform Type Identifiers. Apple Inc. Archived from the original on May 22, 2023. Retrieved May 22, 2023.
UTI conformationpublic.audio
Magic number4D 54 68 64 ("MThd" in ASCII; appears at the start of a MIDI file)
Size limitationNo strict limit; practical limits depend on hardware/software handling the file.
Developed byMIDI Manufacturers Association (MMA), International MIDI Association (IMA)
Initial release1990
Latest release
MIDI 2.0
2020
Type of formatMusical performance data (not actual audio)
CompressionUncompressed (but can be embedded in compressed archives)
Container forMIDI data (notes, controllers, tempo, instrument changes)
Contained byRIFF, Standard MIDI File
Extended fromOriginal MIDI 1.0 protocol
Extended toMIDI 2.0 (adds higher resolution and two-way communication)
StandardsMIDI Manufacturers Association (MMA) specifications
Open format?Yes
Free format?Yes
Websitehttps://midi.org/standard-midi-files-specification

The Standard MIDI File (SMF) is a file format that provides a standardized way for music sequences to be saved, transported, and opened in other systems. The standard was developed and is maintained by the MMA, and usually uses a .mid extension.[57] The compact size of these files led to their widespread use in computers, mobile phone ringtones, webpage authoring and musical greeting cards. These files are intended for universal use and include such information as note values, timing and track names. Lyrics may be included as metadata, and can be displayed by karaoke machines.[58]

SMFs are created as an export format of software sequencers or hardware workstations. They organize MIDI messages into one or more parallel tracks and time-stamp the events so that they can be played back in sequence. A header contains the arrangement's track count, tempo and an indicator of which of three SMF formats the file uses. A type 0 file contains the entire performance, merged onto a single track, while type 1 files may contain any number of tracks that are performed synchronously. Type 2 files are rarely used[59] and store multiple arrangements, with each arrangement having its own track and intended to be played in sequence.

RMID files
[edit]

Microsoft Windows bundles SMFs together with Downloadable Sounds (DLS) in a Resource Interchange File Format (RIFF) wrapper, as RMID files with a .rmi extension. RIFF-RMID has been deprecated in favor of Extensible Music Files (XMF).[60]

Software
[edit]

The main advantage of the personal computer in a MIDI system is that it can serve a number of different purposes, depending on the software that is loaded.[2]: 55  Multitasking allows simultaneous operation of programs that may be able to share data with each other.[2]: 65 

Sequencers
[edit]

Sequencing software can be used to manipulate recorded MIDI data with standard computer editing features such as cut, copy and paste and drag and drop. Keyboard shortcuts can be used to streamline workflow, and, in some systems, editing functions may be invoked by MIDI events. The sequencer can set each channel to play a different sound and gives a graphical overview of the arrangement. A variety of editing tools are made available, including a notation display or scorewriter that can be used to create printed parts for musicians. Tools such as looping, quantization, randomization, and transposition simplify the arranging process.

Beat creation is simplified, and groove templates can be used to duplicate another track's rhythmic feel. Realistic expression can be added through the manipulation of real-time controllers. Mixing can be performed, and MIDI can be synchronized with recorded audio and video tracks. Work can be saved, and transported between different computers or studios.[61][62]: 164–6 

Sequencers may take alternate forms, such as drum pattern editors that users can use to create beats by clicking on pattern grids,[2]: 118  and loop sequencers such as ACID Pro, which combine MIDI with prerecorded audio loops whose tempos and keys are matched to each other. Cue-list sequencing is used to trigger dialogue, sound effect, and music cues in stage and broadcast production.[2]: 121 

Notation software
[edit]

With MIDI, notes played on a keyboard can automatically be transcribed to sheet music.[12]: 213  Scorewriting software typically lacks advanced sequencing tools and is optimized for the creation of a neat, professional printout designed for live instrumentalists.[62]: 157  These programs provide support for dynamics and expression markings, chord and lyric display, and complex score styles.[62]: 167  Software is available that can print scores in braille.[63]

Notation programs include Finale, Encore, Sibelius, MuseScore and Dorico. SmartScore software can produce MIDI files from scanned sheet music.[64]

Editors and librarians
[edit]

Users can program their equipment through the patch editor as a computer interface. These became essential with the appearance of complex synthesizers such as the Yamaha FS1R,[65] which contained several thousand programmable parameters, but had an interface that consisted of fifteen tiny buttons, four knobs and a small LCD.[66] Digital instruments typically discourage users from experimentation, due to their lack of the feedback and direct control that switches and knobs would provide,[67]: 393  but patch editors give owners of hardware instruments and effects devices the same editing functionality that is available to users of software synthesizers.[68] Some editors are designed for a specific instrument or effects device, while other, universal editors support a variety of equipment, and ideally can control the parameters of every device in a setup through the use of System Exclusive messages.[2]: 129  System Exclusive messages use the MIDI protocol to send information about the synthesizer's parameters.

Patch librarians have the specialized function of organizing the sounds in a collection of equipment and exchanging entire banks of sounds between an instrument and a computer. In this way the device's limited patch storage is augmented by a computer's much greater disk capacity.[2]: 133  Once transferred to the computer, custom patches can be shared with other owners of the same instrument.[69] Universal editor/librarians that combine the two functions were once common, and included Opcode Systems' Galaxy, eMagic's SoundDiver, and MOTU's Unisyn. Although these older programs have been largely abandoned with the trend toward computer-based synthesis using virtual instruments, several editor/librarians remain available, including Coffeeshopped Patch Base,[70] Sound Quest's Midi Quest, and several editors from Sound Tower. Native Instruments' Kore was an effort to bring the editor/librarian concept into the age of software instruments,[71] but was abandoned in 2011.[72]

Auto-accompaniment programs
[edit]

Programs that can dynamically generate accompaniment tracks are called auto-accompaniment programs. These create a full-band arrangement in a style that the user selects and sends the result to a MIDI sound-generating device for playback. The generated tracks can be used as educational or practice tools, as accompaniment for live performances, or as a songwriting aid.[73]: 42 

Synthesis and sampling
[edit]

Computers can use software to generate sounds, which are then passed through a digital-to-analog converter (DAC) to a power amplifier and loudspeaker system.[12]: 213  The number of sounds that can be played simultaneously (the polyphony) is dependent on the power of the computer's CPU, as are the sample rate and bit depth of playback, which directly affect the quality of the sound.[74] Synthesizers implemented in software are subject to timing issues that are not necessarily present with hardware instruments, whose dedicated operating systems are not subject to interruption from background tasks as desktop operating systems are. These timing issues can cause synchronization problems, and clicks and pops when sample playback is interrupted. Software synthesizers also may exhibit additional latency in their sound generation.[75]

The roots of software synthesis go back as far as the 1950s, when Max Mathews of Bell Labs wrote the MUSIC-N programming language, which was capable of non-real-time sound generation.[76] Reality, by Dave Smith's Seer Systems was an early synthesizer that ran directly on a host computer's CPU. Reality achieved a low latency through tight driver integration, and therefore could run only on Creative Labs soundcards.[77][78] Syntauri Corporation's Alpha Syntauri was another early software-based synthesizer. It ran on the Apple IIe computer and used a combination of software and the computer's hardware to produce additive synthesis.[79] Some systems use dedicated hardware to reduce the load on the host CPU, as with Symbolic Sound Corporation's Kyma System,[76] and the Creamware/Sonic Core Pulsar/SCOPE systems,[80] which power an entire recording studio's worth of instruments, effect units, and mixers.[81] The ability to construct full MIDI arrangements entirely in computer software allows a composer to render a finalized result directly as an audio file.[32]

Game music
[edit]

Until the mid-1990s, floppy disks were the primary distribution medium for IBM PC compatible games. The small size of MIDI files made them a viable means of providing soundtracks. Prior to Windows 95, games typically used either Ad Lib or Sound Blaster audio cards. These cards use FM synthesis, which generates sound through modulation of sine waves. John Chowning, the technique's pioneer, theorized that the technology would be capable of accurate recreation of any sound if enough sine waves were used, but most computer audio cards performed FM synthesis with only two sine waves. Combined with the cards' 8-bit audio, this resulted in a sound described as "artificial"[82] and "primitive".[83]

Wavetable daughterboards that were later available provided audio samples that could be used in place of the FM sound. These were expensive, but often used the sounds from respected MIDI instruments such as the E-mu Proteus.[83] The computer industry moved in the mid-1990s toward wavetable-based soundcards with 16-bit playback but standardized on a 2 MB of wavetable storage, a space too small in which to fit good-quality samples of 128 General MIDI instruments plus drum kits. To make the most of the limited space, some manufacturers stored 12-bit samples and expanded those to 16 bits on playback.[84]

Other applications

[edit]

Despite its association with music devices, MIDI can control any electronic or digital device that can read and process a MIDI command. MIDI has been adopted as a control protocol in a number of non-musical applications. MIDI Show Control uses MIDI commands to direct stage lighting systems and to trigger cued events in theatrical productions. VJs and turntablists use it to cue clips, and to synchronize equipment, and recording systems use it for synchronization and automation. Wayne Lytle, the founder of Animusic, derived a system he dubbed MIDIMotion in order to produce the Animusic series of computer-animated music video albums; Animusic would later design its own animation software specifically for MIDIMotion called Animotion.[85] Apple Motion allows for a similar control of animation parameters through MIDI. The 1987 first-person shooter game MIDI Maze and the 1990 Atari ST puzzle video game Oxyd use MIDI to network computers together.

Devices

[edit]
5-pin DIN MIDI cable plugged in a socket
DIN connector pin numbers

Connectors and interface

[edit]

DIN connector

[edit]

Per the original MIDI 1.0 standard, cables terminate in a 180° five-pin DIN connector (DIN 41524). Typical applications use only three of the five conductors: a ground wire (pin 2), and a balanced pair of conductors (pins 4 and 5) that carry the MIDI signal as an electric current.[86][73]: 41  This connector configuration can only carry messages in one direction, so a second cable is necessary for two-way communication.[2]: 13  Some proprietary applications, such as phantom-powered footswitch controllers, use the spare pins for direct current (DC) power transmission.[87]

Opto-isolators keep MIDI devices electrically separated from their MIDI connections, which prevents ground loops[88]: 63  and protects equipment from voltage spikes.[14]: 277  There is no error detection capability in MIDI, so the maximum cable length is set at 15 meters (49 ft) to limit interference.[89]

TRS minijack connector

[edit]

To save space, some MIDI devices (smaller ones in particular) started using 3.5 mm TRS phone connectors (also known as audio minijack connectors).[90] This became widespread enough that the MIDI Manufacturers' Association standardized the wiring.[91] The MIDI-over-minijack standards document also recommends the use of 2.5 mm connectors over 3.5 mm ones to avoid confusion with audio connectors.[92]

Thru port

[edit]

Most devices do not copy messages from their input to their output port. A third type of port, the thru port, emits a copy of everything received at the input port, allowing data to be forwarded to another instrument[14]: 278  in a daisy-chain arrangement.[93] Not all devices feature thru ports, and devices that lack the ability to generate MIDI data, such as effects units and sound modules, may not include out ports.[67]: 384 

Management devices

[edit]

Each device in a daisy chain adds delay to the system. This can be avoided by using a MIDI thru box, which contains several outputs that provide an exact copy of the box's input signal. A MIDI merger is able to combine the input from multiple devices into a single stream, and allows multiple controllers to be connected to a single device. A MIDI switcher allows switching between multiple devices, and eliminates the need to physically repatch cables. MIDI routers combine all of these functions. They contain multiple inputs and outputs, and allow any combination of input channels to be routed to any combination of output channels. Routing setups can be created using computer software, stored in memory, and selected by MIDI program change commands.[2]: 47–50  This enables the devices to function as standalone MIDI routers in situations where no computer is present.[2]: 62–3 [94] MIDI data processors are used for utility tasks and special effects. These include MIDI filters, which remove unwanted MIDI data from the stream, and MIDI delays, effects that send a repeated copy of the input data at a set time.[2]: 51 

Interfaces

[edit]

A computer MIDI interface's main function is to synchronize communications between the MIDI device and the computer.[93] Some computer sound cards include a standard MIDI connector, whereas others connect by any of various means that include the D-subminiature DA-15 game port, USB, FireWire, Ethernet or a proprietary connection. The increasing use of USB connectors in the 2000s has led to the availability of MIDI-to-USB data interfaces that can transfer MIDI channels to USB-equipped computers. Some MIDI keyboard controllers are equipped with USB jacks, and can be connected directly to computers that run music software.

MIDI's serial transmission leads to timing problems. A three-byte MIDI message requires nearly 1 millisecond for transmission.[95] Because MIDI is serial, it can only send one event at a time. If an event is sent on two channels at once, the event on the second channel cannot transmit until the first one is finished, and so is delayed by 1 ms. If an event is sent on all channels at the same time, the last channel's transmission is delayed by as much as 16 ms. This contributed to the rise of MIDI interfaces with multiple in- and out-ports, because timing improves when events are spread between multiple ports as opposed to multiple channels on the same port.[75] The term MIDI slop refers to audible timing errors that result when MIDI transmission is delayed.[96]

Controllers

[edit]
A Novation Remote 25 two-octave MIDI controller
Smaller MIDI controllers are popular due to their portability. This two-octave unit provides a variety of controls for manipulating various sound design parameters of computer-based or standalone hardware instruments, effects, mixers and recording devices.

There are two types of MIDI controllers: performance controllers that generate notes and are used to perform music,[97] and controllers that may not send notes, but transmit other types of real-time events. Many devices are a combination of the two types.

Keyboards are by far the most common type of MIDI controller.[69] MIDI was designed with keyboards in mind and any controller that is not a keyboard is considered an "alternative" controller.[98] This was seen as a limitation by composers who were not interested in keyboard-based music, but the standard proved flexible, and MIDI compatibility was introduced to other types of controllers, including guitars, and other stringed instruments and drum controllers and wind controllers, which emulate the playing of drum kit and wind instruments, respectively and specialized and experimental controllers.[12]: 23  Nevertheless, some features of the keyboard playing for which MIDI was designed do not fully capture other instruments' capabilities; Jaron Lanier cites the standard as an example of technological "lock-in" that unexpectedly limited what was possible to express.[99] Some of these shortcomings have been addressed in extensions to the protocol.

Software synthesizers offer great power and versatility, but some players feel that division of attention between a MIDI keyboard and a computer keyboard and mouse robs some of the immediacy from the playing experience.[100] Devices dedicated to real-time MIDI control provide an ergonomic benefit and can provide a greater sense of connection with the instrument than an interface that is accessed through a computer. Controllers may be general-purpose devices that are designed to work with a variety of equipment, or they may be designed to work with a specific piece of software. Examples of the latter include Akai's APC40 controller for Ableton Live, and Korg's MS-20ic controller, a reproduction of the control panel on their MS-20 analog synthesizer. The MS-20ic controller includes patch cables that can be used to control signal routing in their virtual reproduction of the MS-20 synthesizer and can also control third-party devices.[101]

Instruments

[edit]
A General MIDI sound module.
A sound module, which requires an external controller (e.g., a MIDI keyboard) to trigger its sounds. These devices are highly portable, but their limited programming interface requires computer-based tools for comfortable access to their sound parameters.

A MIDI instrument contains ports to send and receive MIDI signals, a CPU to process those signals, an interface for user programming, audio circuitry to generate sound, and controllers. The operating system and factory sounds are often stored in a read-only memory (ROM) unit.[2]: 67–70 

A MIDI instrument can also be a stand-alone module (without a piano-style keyboard) consisting of a General MIDI soundboard (GM, GS and XG), onboard editing, including transposing, MIDI instrument selection and adjusting volume, pan, reverb levels and other MIDI controllers. Typically, the MIDI module includes a screen, so the user can view information for the currently selected function.

Synthesizers

[edit]

Synthesizers may employ any of a variety of sound generation techniques. They may include an integrated keyboard or may exist as sound modules that generate sounds when triggered by an external controller, such as a MIDI keyboard. Sound modules are typically designed to be mounted in a 19-inch rack.[2]: 70–72  Manufacturers commonly produce a synthesizer in both standalone and rack-mounted versions, and often offer the keyboard version in a variety of sizes.

Samplers

[edit]

A sampler can record and digitize audio, store it in random-access memory (RAM), and play it back. With a sampler, users typically can edit a sample and save it to a hard disk, apply effects to it, and shape it with the same tools that subtractive synthesizers use. They also may be available in either keyboard or rack-mounted form.[2]: 74–8  Instruments that generate sounds through sample playback, but have no recording capabilities, are known as "ROMplers".

Samplers did not become established as viable MIDI instruments as quickly as synthesizers did due to the expense of memory and processing power at the time.[14]: 295  The first low-cost MIDI sampler was the Ensoniq Mirage, introduced in 1984.[14]: 304  MIDI samplers are typically limited by displays that are too small to use to edit sampled waveforms, although some can be connected to a computer monitor.[14]: 305 

Drum machines

[edit]

Drum machines typically are sample playback devices that specialize in drum and percussion sounds. They commonly contain a sequencer for creating drum patterns and arranging them into a song. There often are multiple audio outputs so that each sound or group of sounds can be routed to a separate output. The individual drum voices may be playable from another MIDI instrument or from a sequencer.[2]: 84 

Workstations and hardware sequencers

[edit]
A button matrix MIDI controller
Yamaha's Tenori-on controller allows arrangements to be built by "drawing" on its array of lighted buttons. The resulting arrangements can be played back using its internal sounds or external sound sources, or recorded in a computer-based sequencer.

Sequencer technology predates MIDI. Analog sequencers use CV/Gate signals to control pre-MIDI analog synthesizers. MIDI sequencers typically are operated by transport features modeled after those of tape decks. They are capable of recording MIDI performances and arranging them into individual tracks using a multitrack recording paradigm. Music workstations combine controller keyboards with an internal sound generator and a sequencer. These can be used to build complete arrangements and play them back using their own internal sounds and function as self-contained music production studios. They commonly include file storage and transfer capabilities.[2]: 103–4 

Effects units

[edit]

Some effects units can be remotely controlled via MIDI. For example, the Eventide H3000 Ultra-harmonizer allows such extensive MIDI control that it is playable as a synthesizer.[14]: 322  The Drum Buddy, a pedal-format drum machine, has a MIDI connection so that it can have its tempo synchronized with a looper pedal or time-based effects such as delay.

Technical specifications

[edit]
8-N-1 asynchronous serial communication of two MIDI bytes. Each 8-bit byte is preceded by a start bit and succeeded by a stop bit for framing purposes, to total 10 bits.[14]: 286  So while the 31,250 baud rate corresponds to 31.25 kbit/s, the net bit rate is 25 kbit/s. Each byte with its frame uses 320 microseconds.[102]

MIDI messages are made up of 8-bit bytes transmitted at 31,250[b] (±1%) baud using 8-N-1 asynchronous serial communication as described in the figure. The first bit of each byte identifies whether the byte is a status byte or a data byte, and is followed by seven bits of information.[2]: 13–14 

A MIDI link can carry sixteen independent channels, numbered 1–16. A device may listen to specific channels and ignore messages on other channels (omni off mode), or it can listen to all channels, effectively ignoring the channel address (omni on).

A device that is polyphonic can sound multiple notes simultaneously, until the device's polyphony limit is reached, or the notes reach the end of their decay envelope, or explicit note-off MIDI commands are received. A device that is monophonic instead terminates any previous note when new note-on commands arrive.

Some receiving devices may be set to all four combinations of omni off/on and mono/poly modes.[2]: 14–18 

Messages

[edit]

A MIDI message is an instruction that controls some aspect of the receiving device. A MIDI message consists of a status byte, which indicates the type of the message, followed by up to two data bytes that contain the parameters.[37] MIDI messages can be channel messages sent on only one of the 16 channels and monitored only by devices on that channel, or system messages that all devices receive. Each receiving device ignores data not relevant to its function.[67]: 384  There are five types of message: Channel Voice, Channel Mode, System Common, System Real-Time, and System Exclusive.[103]

Channel Voice messages transmit real-time performance data over a single channel. Examples include note-on messages which contain a MIDI note number that specifies the note's pitch, a velocity value that indicates how forcefully the note was played, and the channel number; note-off messages that end a note; program change messages that change a device's patch; and control changes that allow adjustment of an instrument's parameters. MIDI notes are numbered from 0 to 127 assigned to C−1 to G9. This extends beyond the 88-note piano range from A0 to C8 and corresponds to a frequency range of 8.175799 to 12543.85 Hz.[c]

System Exclusive messages

[edit]

System Exclusive (SysEx) messages send information about a synthesizer's functions, rather than performance data such as which notes are being played and how loud. Because they can include functionality beyond what the MIDI standard provides, they are a major reason for the flexibility and longevity of the MIDI standard. Manufacturers use them to create proprietary messages that control their equipment more thoroughly than the limitations of standard MIDI messages.[14]: 287 

The MIDI Manufacturers Association issues a unique identification number to MIDI companies.[104] These are included in SysEx messages, to ensure that only the specifically addressed device responds to the message, while all others know to ignore it. Many instruments also include a SysEx ID setting, so a controller can address two devices of the same model independently.[105]

Universal System Exclusive messages are a special class of SysEx messages used for extensions to MIDI that are not intended to be exclusive to one manufacturer.[106]

Implementation chart

[edit]

Devices typically do not respond to every type of message defined by the MIDI specification. The MIDI implementation chart was standardized by the MMA as a way for users to see what specific capabilities an instrument has, and how it responds to messages.[2]: 231  A populated MIDI implementation chart is usually published as part of the documentation for MIDI devices.

Electrical specifications

[edit]

MIDI 1.0's electrical interface is based around a fully isolated current loop[102] along the red and blue lines in the following schematic:

MIDI interconnection schematic

"DIN / TRS" in this schematic indicates that either a DIN connector[d] or a TRS phone connector[e] may be used.[107][108]

To transmit a logic 0 and a start bit, the sender's UART[f] produces a low voltage. This results in a nominal 5 milliamperes[102] current flow sourced from the sender's high voltage supply,[g] which travels rightwards along the red lines though the shielded[h] twisted-pair cable and into the receiver's opto-isolator. The current exits the opto-isolator and returns back leftwards along the blue lines into the sender's UART, which sinks the current.[i] Resistors R1 and R2 limit the current and are equal to provide a balanced impedance. The diode is for protection.[110] This current turns on the opto-isolator's[j] LED and phototransistor, so the receiver's UART can read the signal with the help of pull-up resistor R3 to the receiver's voltage supply. While the supplies in the original specification are 5 volts, the receiver and sender may use different voltage levels.

To transmit a logic 1, a stop bit, and while idle, the sender's UART produces the same high voltage as its voltage supply provides, which results in no current flow. This avoids wasting power when idle.

Extensions

[edit]

MIDI's flexibility and widespread adoption have led to many refinements of the standard, and have enabled its application to purposes beyond those for which it was originally intended.

General MIDI

[edit]
GM Standard Drum Map on the keyboard
General MIDI's Percussion Key Map specifies the percussion sound that a given note triggers. MIDI note numbers shown in parentheses next to their corresponding keyboard note.

MIDI allows the selection of an instrument's sounds through program change messages, but there is no guarantee that any two instruments have the same sound at a given program location.[111] Program #0 may be a piano on one instrument, or a flute on another. The General MIDI (GM) standard was established in 1991, and provides a standardized sound bank that allows a Standard MIDI File created on one device to sound similar when played back on another. GM specifies a bank of 128 sounds arranged into 16 families of eight related instruments, and assigns a specific program number to each instrument.[112] Any given program change selects the same instrument sound on any GM-compatible instrument.[113] Percussion instruments are placed on channel 10, and a specific MIDI note value is mapped to each percussion sound.

The GM standard eliminates variation in note mapping. Some manufacturers had disagreed over what note number should represent middle C, but GM specifies that note number 69 plays A440, which in turn fixes middle C as note number 60.

GM-compliant devices must offer 24-note polyphony.[114] GM-compatible devices are required to respond to velocity, aftertouch, and pitch bend, to be set to specified default values at startup, and to support certain controller numbers such as for sustain pedal, and Registered Parameter Numbers (RPNs).[115]

A simplified version of GM, called GM Lite, is used for devices with limited processing power.[111][116]

GS, XG, and GM2

[edit]

A general opinion quickly formed that the GM's 128-instrument sound set was not large enough. Roland's General Standard, or Roland GS, included additional sounds, drumkits and effects, provided a bank select command that could be used to access them, and used MIDI Non-Registered Parameter Numbers (NRPNs) to access its new features. Yamaha's Extended General MIDI, or Yamaha XG, followed in 1994. XG similarly offered extra sounds, drumkits and effects, but used standard controllers instead of NRPNs for editing, and increased polyphony to 32 voices. Both standards feature backward compatibility with the GM specification but are not compatible with each other.[117] Neither standard has been adopted beyond its creator, but both are commonly supported by music software titles.

Member companies of Japan's AMEI developed the General MIDI Level 2 specification in 1999. GM2 maintains backward compatibility with GM, but increases polyphony to 32 voices, standardizes several controller numbers such as for sostenuto and soft pedal (una corda), RPNs and Universal System Exclusive Messages, and incorporates the MIDI Tuning Standard.[118] GM2 is the basis of the instrument selection mechanism in Scalable Polyphony MIDI (SP-MIDI), a MIDI variant for low-power devices that allows the device's polyphony to scale according to its processing power.[111]

Tuning standard

[edit]

Most MIDI synthesizers use equal temperament tuning. The MIDI tuning standard (MTS), ratified in 1992, allows alternate tunings.[119] MTS allows microtunings that can be loaded from a bank of up to 128 patches, and allows real-time adjustment of note pitches.[120] Manufacturers are not required to support the standard. Those who do are not required to implement all of its features.[119]

Time code

[edit]

A sequencer can drive a MIDI system with its internal clock, but when a system contains multiple sequencers, they must synchronize to a common clock. MIDI timecode (MTC), developed by Digidesign,[121] implements SysEx messages[122] developed specifically for timing purposes, and can translate to and from the SMPTE timecode standard.[14]: 288  MIDI interfaces such as Mark of the Unicorn's MIDI Timepiece can convert SMPTE code to MTC.[123] While MIDI clock is based on tempo, timecode is based on frames and is independent of tempo. MTC, like SMPTE timecode, includes position information and can recover in the event of a dropout.[124]

Machine control

[edit]

MIDI Machine Control (MMC) consists of a set of SysEx commands[125] that operate the transport controls of hardware recording devices. MMC lets a sequencer send Start, Stop, and Record commands to a connected tape deck or hard disk recording system, and to fast-forward or rewind the device to start playback at the same point as the sequencer. No synchronization data is involved, although the devices may synchronize through MTC.[126]

Show control

[edit]
A theatrical event operated by MIDI Show Control
MIDI Show Control is used to cue and synchronize lighting and effects for theatrical events, such as the Waterworld attraction at Universal Studios Hollywood.[127]

MIDI Show Control (MSC) is a set of SysEx commands for sequencing and remotely cueing show control devices such as lighting, music and sound playback, and motion control systems.[128] Applications include stage productions, museum exhibits, recording studio control systems, and amusement park attractions.[127]

Timestamping

[edit]

One solution to MIDI timing problems is to mark MIDI events with the times they are to be played, transmit them beforehand, and store them in a buffer in the receiving device. Sending data beforehand reduces the likelihood that a busy passage overwhelms the transmission link. Once stored in the receiver, the information is no longer subject to timing issues associated with MIDI or USB interfaces and can be played with a high degree of accuracy.[129] MIDI timestamping only works when both hardware and software support it. MOTU's MTS, eMagic's AMT, and Steinberg's Midex 8 had implementations that were incompatible with each other, and required users to own software and hardware manufactured by the same company to work.[75] Timestamping is built into FireWire MIDI interfaces,[130] Mac OS X Core Audio, and Linux ALSA Sequencer.

Sample dump standard

[edit]

An unforeseen capability of SysEx messages was their use for transporting audio samples between instruments. This led to the development of the sample dump standard (SDS), which established a new SysEx format for sample transmission.[14]: 287  SDS was later augmented with a pair of commands that allow the transmission of information about sample loop points, without requiring that the entire sample be transmitted.[131]

Downloadable Sounds

[edit]

The Downloadable Sounds (DLS) specification, ratified in 1997, allows mobile devices and computer sound cards to expand their wave tables with downloadable sound sets.[132] The DLS Level 2 specification followed in 2006, and defined a standardized synthesizer architecture. The Mobile DLS standard calls for DLS banks to be combined with SP-MIDI, as self-contained Mobile XMF files.[133]

MIDI Polyphonic Expression

[edit]

MIDI Polyphonic Expression (MPE) is a method of using MIDI that enables pitch bend, and other dimensions of expressive control, to be adjusted continuously for individual notes.[134] MPE works by assigning each note to its own MIDI channel so that controller messages can be applied to each note individually.[135][134] The specifications were released in November 2017 by AMEI and in January 2018 by the MMA.[136] Instruments like the Continuum Fingerboard, LinnStrument, ROLI Seaboard, Sensel Morph, and Eigenharp let users control pitch, timbre, and other nuances for individual notes within chords.[137]

Alternative hardware transports

[edit]

In addition to using a 31.25 kbit/s current-loop over a DIN connector, the same data can be transmitted over different hardware transports such as USB, FireWire, and Ethernet.

USB and FireWire

[edit]

Members of the USB-IF in 1999 developed a standard for MIDI over USB, the "Universal Serial Bus Device Class Definition for MIDI Devices".[138] MIDI over USB has become increasingly common as other interfaces that had been used for MIDI connections (ISA card, game port, etc.) disappeared from personal computers. Linux, Microsoft Windows, Macintosh OS X, and Apple iOS operating systems include standard class drivers to support devices that use the "Universal Serial Bus Device Class Definition for MIDI Devices".

Apple Computer developed the FireWire interface during the 1990s. It began to appear on digital video (DV) cameras toward the end of the decade, and on G3 Macintosh models in 1999.[139] It was created for use with multimedia applications.[130] Unlike USB, FireWire uses intelligent controllers that can manage their own transmission without attention from the main CPU.[140] As with standard MIDI devices, FireWire devices can communicate with each other with no computer present.[141]

XLR connectors

[edit]

The Octave-Plateau Voyetra-8 synthesizer was an early MIDI implementation using XLR3 connectors in place of the 5-pin DIN. It was released in the pre-MIDI years and later retrofitted with a MIDI interface but kept its XLR connector.[142]

Serial parallel, and joystick port

[edit]

As computer-based studio setups became common, MIDI devices that could connect directly to a computer became available. These typically used the 8-pin mini-DIN connector that was used by Apple for serial ports prior to the introduction of the Blue and White G3 models. MIDI interfaces intended for use as the centerpiece of a studio, such as the Mark of the Unicorn MIDI Time Piece, were made possible by a fast transmission mode that could take advantage of these serial ports' ability to operate at 20 times the standard MIDI speed.[2]: 62–3 [141] Mini-DIN ports were built into some late-1990s MIDI instruments and enabled such devices to be connected directly to a computer.[143] Some devices connected via a PCs' DB-25 parallel port, or through the DA-15 game port found on many PC sound cards.[141]

mLAN

[edit]

Yamaha introduced the mLAN protocol in 1999. It was conceived as a local area network for musical instruments using FireWire as the transport and was designed to carry multiple MIDI channels together with multichannel digital audio, data file transfers, and timecode.[139][140] mLan was used in a number of Yamaha products, notably digital mixing consoles and the Motif synthesizer, and in third-party products such as the PreSonus FIREstation and the Korg Triton Studio.[144] No new mLan products have been released since 2007.

SCSI MIDI Device Interface (SMDI)

[edit]

SCSI MIDI Device Interface (SMDI) was used by some samplers and hard disk recorders in the 1990s (e.g. Kurzweil K2000 and Peavey SP Sample Playback Synthesizer[145]) for fast bidirectional sample transport to hard disk drives and magneto-optical drives.[146][147]

Ethernet and Internet Protocol

[edit]

Computer network implementations of MIDI provide network routing capabilities, and the high-bandwidth channel that earlier alternatives to MIDI, such as ZIPI, were intended to bring. Proprietary implementations have existed since the 1980s, some of which use fiber optic cables for transmission.[2]: 53–4  The Internet Engineering Task Force's RTP-MIDI open specification has gained industry support. Apple has supported this protocol from Mac OS X 10.4 onwards, and a Windows driver based on Apple's implementation exists for Windows XP and newer versions.[148]

Wireless

[edit]

Systems for wireless MIDI transmission have been available since the 1980s.[2]: 44  Several commercially available transmitters allow wireless transmission of MIDI and OSC signals over Wi-Fi and Bluetooth.[149] iOS devices are able to function as MIDI control surfaces, using Wi-Fi and OSC.[150] An XBee radio can be used to build a wireless MIDI transceiver as a do-it-yourself project.[151] Android devices are able to function as full MIDI control surfaces using several different protocols over Wi-Fi and Bluetooth.[152]

MIDI 2.0

[edit]

The MIDI 2.0 standard was unveiled on January 17, 2020, at the Winter NAMM Show in Anaheim, California. Representatives Yamaha, ROLI, Microsoft, Google, and the MIDI Association introduced the update,[153] which enables bidirectional communication while maintaining backward compatibility.[154]

Research on the new MIDI protocol began in 2005.[155][156][157] Prototype devices showcasing wired and wireless connections have been shown privately at NAMM.[156] Licensing and product certification policies have been developed,[158] although no projected release date was announced.[159] Proposed physical layer and transport layer included Ethernet-based protocols such as RTP MIDI and Audio Video Bridging/Time-Sensitive Networking,[160] as well as User Datagram Protocol (UDP)-based transport.[157]

AMEI and MMA announced that complete specifications will be published following interoperability testing of prototype implementations from major manufacturers such as Google, Yamaha, Steinberg, Roland, Ableton, Native Instruments, and ROLI, among others.[161][162][163] In January 2020, Roland announced the A-88mkII controller keyboard that supports MIDI 2.0.[164] MIDI 2.0 includes MIDI Capability Inquiry specification for property exchange and profiles, and the new Universal MIDI Packet format for high-speed transports which supports both MIDI 1.0 and MIDI 2.0 voice messages.

Some devices operating MIDI 1.0 can "retrofit" some 2.0 features. Since its release in early January 2020 by the MIDI Manufacturers Association, more details have yet to come out about the new update. As of 2021 there were five components to MIDI such as; M2-100-U v1.0 MIDI 2.0 Specification Overview, M2-101-UM v1.1 MIDI-CI Specification, M2-102-U v1.0 Common Rules for MIDI-CI Profiles, M2-103-UM v1.0 Common Rules for MIDI-CI PE and M2-104-UM v1.0 UMP and MIDI 2.0 Protocol Specification. Other specifications regarding MIDI 2.0 include; allowing the use of 32,000 controllers and wide range note enhancements. These enhancements are made better through the property exchange.[165] In June 2023 updated and new MIDI 2.0 specifications were released consisting of M2-100-U MIDI 2.0 Specification Overview, Version 1.1, M2-101-UM MIDI Capability Inquiry (MIDI-CI), Version 1.2, M2-102-U Common Rules for MIDI-CI Profiles, Version 1.1, M2-104-UM Universal MIDI Packet (UMP) Format and MIDI 2.0 Protocol, Version 1.1, and M2-116-U MIDI Clip File (SMF2), Version 1.0.[166]

See also

[edit]

Notes

[edit]

References

[edit]
[edit]
Revisions and contributorsEdit on WikipediaRead on Wikipedia
from Grokipedia
MIDI, or Musical Instrument Digital Interface, is a and protocol that enables electronic musical instruments, computers, and other related devices to connect and exchange data describing musical events, such as note pitches, durations, velocities, and control changes, without transmitting audio signals themselves. Developed in the early , it provides a universal language for music performance and production, allowing devices from different manufacturers to interoperate seamlessly through a . Originally using a 5-pin for hardware connections, MIDI operates at a rate of 31.25 kBaud and supports up to 16 independent channels for polyphonic control, facilitating applications from live performances to composition software. The protocol's core consists of MIDI messages—compact binary codes categorized as Channel Voice (for notes and expressions), Channel Mode (for operational settings), and System messages (for synchronization and system-wide functions)—which are transmitted in real-time or stored in Standard MIDI Files (.mid) for playback and editing. This message-based system, akin to digital , allows musicians to control multiple synthesizers from a single keyboard, sequence performances, or integrate with digital audio workstations (DAWs). Since its inception, MIDI has evolved to include modern transports like USB, , and Ethernet, expanding its use beyond traditional instruments to gaming, , and accessibility tools for music creation. MIDI originated from collaborative efforts in 1981–1983 by engineers at companies including Sequential Circuits, , Yamaha, and , with Dave Smith and Chet Wood proposing an initial "Universal Synthesizer Interface" that evolved into the finalized MIDI 1.0 specification by August 1983. The first MIDI-compatible product, the Sequential Circuits Prophet-600 synthesizer, shipped in late 1983, marking the standard's commercial debut and sparking widespread adoption in the music industry. Over four decades, MIDI has become foundational to electronic music, powering everything from consumer keyboards to professional studios, with ongoing support from the MIDI Manufacturers Association (MMA). In 2020, the MMA introduced MIDI 2.0 as an extensible update to the original protocol, enhancing resolution for finer control (e.g., 32-bit values for velocity and position), adding bidirectional communication via MIDI Capability Inquiry (MIDI-CI), and introducing profiles for standardized device behaviors across applications. These advancements enable richer expression, such as per-note pitch bend and controllers, while maintaining full backward compatibility with MIDI 1.0 devices through the Universal MIDI Packet format. As of 2025, MIDI 2.0 support has expanded in operating systems like macOS, Windows, Android, and , alongside new hardware from manufacturers, positioning it for future innovations in interactive music and AI-assisted composition.

Introduction

Definition and Purpose

MIDI, or Musical Instrument Digital Interface, is a and communications protocol that enables the serial transmission of digital data between electronic , computers, and related devices. Developed in 1983, it allows these devices to exchange musical performance information without transmitting actual audio signals, instead using compact messages to describe events such as notes, timing, and instrument parameters. This event-based approach contrasts with audio waveforms, requiring minimal bandwidth and facilitating efficient data handling over simple serial connections. The primary purposes of MIDI include of instruments, of performance timing across multiple devices, storage and retrieval of musical data in formats like Standard MIDI Files, and promotion of among diverse hardware. By standardizing data exchange, MIDI eliminates the need for proprietary cables and interfaces that previously limited connectivity, enabling musicians to orchestrate complex setups with synthesizers, sequencers, and controllers from different manufacturers. For instance, a single keyboard can trigger sounds on remote synthesizers while maintaining precise alignment for live performances or recordings. MIDI emerged in response to widespread incompatibility issues among early 1980s synthesizers from leading companies such as , Yamaha, and , where proprietary systems hindered multi-device integration and broader adoption in music production. This addressed those barriers, fostering a unified that revolutionized electronic music creation by allowing seamless collaboration between instruments and digital tools. Over time, the protocol has evolved, with MIDI 2.0 introducing enhancements like higher resolution and bidirectional communication for even greater expressiveness.

Basic Components and Operation

A MIDI system comprises several essential components that enable the transmission of musical performance data between devices. The MIDI controller serves as the input device, such as a keyboard or pad controller, which detects user actions like pressing a key and generates corresponding MIDI messages to send via its MIDI OUT port. The receiving component is typically a sound module or synthesizer, which processes incoming MIDI data through its MIDI IN port to trigger audio generation, such as synthesizing a piano tone. Transmission occurs via MIDI cables or interfaces that connect these devices, ensuring reliable serial data flow. Additionally, a host computer equipped with sequencing software acts as a central manager, recording MIDI events for editing and playback, or synchronizing multiple devices. The operational process follows a straightforward data flow initiated by user input on the controller. When a plays a note, the controller creates MIDI events—such as a "note on" message to start the sound or "note off" to end it—which are then serialized into compact byte sequences for transmission. These bytes travel serially at a standard rate through the MIDI cable from the controller's output to the receiver's input, where the sound module or interprets them to produce the appropriate audio response, such as varying pitch or based on the event details. This event-driven allows for real-time control, with the host computer optionally capturing the sequence of events for later reproduction. Unlike audio signals, MIDI operates on an event-based protocol, conveying discrete instructions rather than continuous waveforms, which keeps lightweight and versatile across devices. Core message elements include the note number, a value from 0 to 127 representing pitch (e.g., 60 for middle C); , ranging from 0 to 127 to indicate or intensity of the input; and channel assignment, supporting up to 16 independent channels to manage polyphonic or multi-instrument performances simultaneously. This structure enables precise, low-bandwidth communication of musical intent without embedding sound itself. The fundamental data pathway in a MIDI setup can be described as a linear chain: Controller (MIDI OUT)MIDI Cable (MIDI IN)Audio Output. This configuration supports both live performance and sequenced playback, highlighting MIDI's role in separating control from sound generation.

History

Development

The development of MIDI began in the early amid growing frustration with proprietary interfaces that hindered among electronic musical instruments. In June 1981, Ikutaro Kakehashi, founder of , proposed the idea of a universal standard during a meeting at the NAMM in , suggesting collaboration with American manufacturers and specifically recommending Dave Smith of Sequential Circuits. Smith, who had been working on integration solutions, formalized the concept later that year. On October 30, 1981, Smith and engineer Chet Wood presented a paper titled "An Overview of the Proposed Universal Synthesizer Interface" at the (AES) convention in , outlining the initial Universal Synthesizer Interface (USI) with a 19.2 kbps rate and 1/4-inch phone jacks for connectivity. Following the AES presentation, international collaboration intensified to refine the protocol. In December 1981, a conference in involving representatives from , Yamaha, , Kawai, and Sequential Circuits addressed feedback on the USI, criticizing its speed as too slow for musical applications and its connectors as unreliable for use. The group adopted key technical decisions: switching to the 5-pin for its robustness and shielding against noise, and increasing the rate to 31.25 kbps to align with existing technologies like Roland's Digital Control Bus while enabling real-time performance data transmission. By July 1982, further refinements via agreements expanded the protocol to support 16 channels for polyphonic control. These changes balanced simplicity—limiting the protocol to essential note on/off, , and control messages—with functionality for practical music production, while designing for future expansions through reserved message types to ensure . Key milestones marked the protocol's transition from proposal to standard. In January 1983, at the Winter in , the first public MIDI demonstration occurred, successfully linking a Sequential Circuits Prophet-600 synthesizer to a , allowing synchronized note playback and control across devices from different manufacturers. This demo, attended by industry leaders including and representatives from , validated the protocol's viability and garnered commitments from Yamaha, , and others to implement it. The official MIDI 1.0 specification was finalized and published in August 1983, establishing the protocol as an developed through these collaborative efforts, though the formal MIDI Manufacturers Association (MMA) would not incorporate until 1985 to oversee ongoing maintenance. Initial challenges included reconciling differing priorities between U.S. and Japanese firms—such as connector preferences and data throughput—while keeping the serial protocol affordable and implementable on 8-bit microprocessors common in synthesizers.

Adoption and Impact

Following its public demonstration at the 1983 , where a Sequential Circuits Prophet 600 synthesizer successfully interfaced with a , MIDI saw rapid integration by leading manufacturers including Sequential Circuits, , Yamaha, , and Kawai, establishing it as an industry standard within the first year. This collaborative effort, driven by figures like Dave Smith and Ikutaro Kakehashi, ensured broad compatibility across devices without proprietary restrictions. The 1990s marked a significant boom in MIDI's accessibility, fueled by falling prices for hardware and the emergence of user-friendly software such as Steinberg's Cubase, released in 1989 as a MIDI sequencing application for the ST computer. Affordable MIDI interfaces and controllers proliferated, enabling integration with personal computers like the ST and early PCs, which democratized music creation beyond professional studios. MIDI profoundly transformed the music industry by empowering home studios, where musicians could sequence and control multiple synthesizers and drum machines via a single computer or keyboard, drastically lowering production barriers that previously required expensive multitrack tape recorders. It reduced costs for live performances by allowing one controller to trigger diverse sound sources in real time, streamlining setups for electronic acts and minimizing the need for large ensembles of hardware. This shift revolutionized genres like and hip-hop, where precise beat programming and layered electronic textures became staples, fostering innovation in dance-oriented and urban music production. Culturally, MIDI underpinned iconic 1980s tracks, such as those on Depeche Mode's 1984 album , enabling intricate, synchronized electronic arrangements that defined synth-pop's sound. In the , MIDI interfaces bridged instruments with PCs, facilitating hybrid workflows that integrated hardware synths into software environments and expanded creative possibilities for composers and producers. Economically, the MIDI Manufacturers Association (MMA), established in , oversaw protocol maintenance and product certification to guarantee interoperability, which spurred market growth and prevented fragmentation among vendors. By the , MIDI had achieved near-universal prevalence, incorporated into virtually all electronic instruments from keyboards to drum modules, sustaining its economic viability through widespread hardware and software ecosystems. Even amid the ascent of and sample-based tools, MIDI's lightweight data transmission and ensure its enduring role in controlling virtual instruments and live setups.

Applications

Instrument Control

MIDI enables precise real-time control of musical instruments through channel voice messages that trigger and modulate sounds across connected devices. The core functions include Note On and Note Off messages, which initiate and terminate specific notes on a or , allowing a single controller like a keyboard to activate sounds on remote hardware. sensitivity is incorporated into these messages, where the Note On velocity value (ranging from 0 to 127) determines the initial and dynamics of the sound, simulating the force of a key press for expressive performance. Aftertouch, available as polyphonic key pressure (per-note) or channel pressure (overall), further enhances expression by modulating parameters like or filter cutoff in response to continued pressure after the initial strike. Pitch bend messages provide continuous pitch variation, typically over a range of ±2 semitones but configurable by the receiving device, enabling smooth glissandi and microtonal adjustments during live play. Channel assignment allows up to 16 independent channels per MIDI , facilitating polyphonic layering where different instruments or parts are assigned to separate channels for simultaneous control from one source. For instance, a melody line might be routed to channel 1 on a lead , while bass notes are sent to channel 8 on a sub-bass module, ensuring isolated parameter control without interference. In implementations, channel 10 is conventionally dedicated to percussion, where specific note numbers trigger distinct drum sounds rather than pitched instruments, supporting complex rhythmic layering alongside melodic elements. Synchronization is achieved via system real-time messages, including the MIDI Clock signal transmitted at 24 pulses per (PPQ) to maintain precise tempo alignment between controllers, sequencers, and sound generators. Start, Stop, and Continue commands coordinate playback initiation, halting, and resumption, ensuring all devices in a chain—such as a slaved to a keyboard sequencer—operate in without drift. In live performance setups, a can control multiple remote synthesizers by assigning patches to different channels, allowing a performer to layer strings on channel 3 with on channel 5 for instant . Similarly, velocity-sensitive drum pads transmit Note On messages with data to trigger sampled percussion sounds on a dedicated module, replicating acoustic dynamics in electronic .

Composition and Production

In music composition and production, MIDI facilitates file-based workflows through Standard MIDI Files (SMF), a standardized format introduced in that stores sequence data including note events, timing, and instrument assignments, enabling portability across software and hardware. These .mid files can be imported into digital audio workstations (DAWs) such as or , where they serve as editable sequences for building arrangements, allowing producers to record, import, or generate MIDI data non-destructively while preserving the original file integrity. MIDI editing in DAWs provides precise control over performances, with tools for quantization to align notes to a rhythmic grid—such as snapping to 16th notes—correcting timing without altering pitch or , as implemented in Ableton Live's Quantize command or Clip View utilities. Transposition shifts note pitches by semitones or octaves via sliders or keyboard shortcuts, facilitating key changes across tracks, while humanization introduces subtle variations in timing (up to a quarter grid division) and velocity to mimic organic playing and avoid mechanical rigidity. Layering multiple MIDI tracks with virtual instruments further enhances production, where each track can route to different software synthesizers, building complex arrangements from simple sequences. MIDI integrates seamlessly with audio elements in DAWs by triggering software synthesizers through protocols like VST plugins, where MIDI note-on messages generate sounds from virtual instruments such as emulations of acoustic pianos or electronic pads, blending MIDI-driven layers with recorded audio tracks. Additionally, MIDI data can be converted to score notation using specialized software like Dorico, which imports .mid files and renders them as printable with proper staff notation, dynamics, and articulations based on embedded velocity and duration values. As of 2025, MIDI plays a key role in AI-assisted music composition, where tools generate MIDI sequences for import into DAWs. For example, Hooktheory's (released June 2024) uses AI to create chord progressions and melodies in MIDI format, enabling composers to refine machine-generated ideas collaboratively. Other platforms, such as AI and Midigen, produce editable MIDI outputs for genres like EDM and classical, enhancing creativity while sparking debates on authorship and originality in music production. The role of MIDI in production underwent a significant historical shift in the , transitioning from dedicated hardware sequencers—such as those in early MIDI keyboards or rackmount units—to software-based systems within emerging DAWs like Cubase and , driven by advances in personal computing power and that allowed real-time editing and unlimited track counts. This evolution, accelerated by MIDI's interoperability, enabled over nascent connections, fostering collaborative remote production where composers could exchange editable sequences without physical hardware constraints.

Non-Musical Uses

MIDI Time Code (MTC), a synchronization protocol within the MIDI standard, enables precise timing for coordinating non-musical elements in live performances and theatrical productions. MTC translates SMPTE timecode into MIDI messages, allowing devices such as lighting consoles, video playback systems, and pyrotechnic controllers to synchronize events to an external clock source. In theater and live events, MTC facilitates the automation of complex shows by triggering cues for stage lights, projected visuals, and special effects like fog or explosions in alignment with a central timeline. For instance, lighting systems from manufacturers like Electronic Theatre Controls integrate MTC to execute pre-programmed sequences during performances, ensuring seamless integration with audio or narrative elements. Complementing MTC, MIDI Show Control (MSC) extends MIDI's utility for broader show automation in entertainment venues. MSC uses system exclusive messages to command diverse equipment, including dimmers, moving lights, and automated scenery, beyond simple timing. This protocol is widely adopted in theme parks and concerts, where it allows a single controller—often a computer or dedicated console—to orchestrate multiple subsystems for immersive experiences. In practice, MSC commands can cue video servers to play specific clips or activate hydraulic platforms in synchrony, enhancing the reliability of large-scale productions. MIDI Machine Control (MMC) provides commands for remote operation of recording and playback devices, extending MIDI into professional automation workflows. MMC supports transport functions like play, stop, record, and locate, enabling centralized control of tape decks, video recorders, and workstations. In film scoring, MMC automates between scoring software and linear recording media, allowing composers to cue sections of orchestral performances or sound effects without manual intervention. For example, during , a sequencer can issue MMC commands to advance tape positions on multitrack recorders, streamlining the integration of live recordings with visual timelines. This capability reduces errors in time-intensive processes, as seen in studios using MMC-compatible hardware for precise dub sessions. In gaming and interactive applications, MIDI controllers serve as intuitive input devices for rhythm-based video games, bridging physical interaction with digital feedback. Games like and its successors employ specialized controllers that mimic musical instruments, but adaptations allow standard MIDI keyboards or drum pads to interface directly for enhanced playability. For instance, supports MIDI-compatible drum kits, enabling players to use professional electronic percussion for authentic rhythm challenges. Similarly, Synthesia, a piano-focused , relies on MIDI input from keyboards to match on-screen notes, fostering skill development through gamified practice. These integrations highlight MIDI's role in creating responsive, tactile gaming experiences without requiring proprietary hardware. Haptic feedback interfaces leverage MIDI to provide tactile responses in interactive systems, enhancing user immersion in non-musical contexts. By mapping MIDI control change messages to motors or actuators, devices deliver physical sensations synchronized with events, such as in simulations or tools. Research demonstrates that integrating haptics with MIDI controllers improves interaction in touchscreen-based applications, where vibrations simulate button presses or environmental cues. For example, wearable MIDI devices like the via apps such as MIDIWrist use built-in haptics to confirm control activations, aiding users in gaming or remote device operation. This approach extends to modular systems, where haptic modules respond to MIDI for dynamic feedback in interactive installations. Emerging applications of MIDI in the (IoT) involve mapping sensor data to MIDI messages for environmental control and data visualization. Devices convert real-time inputs from , , or motion sensors into MIDI note or control change events, enabling intuitive automation of smart systems. For instance, the IO-Lights controller uses ambient levels to generate MIDI continuous controller values, which can adjust IoT-connected or HVAC systems in response to environmental changes. Similarly, projects like Weather Thingy translate climate data—such as or —into MIDI parameters to trigger actions in connected networks, like modulating building automations during musical or interactive events. This bidirectional use of MIDI facilitates creative IoT integrations, where sensor-driven MIDI signals provide a standardized interface for non-traditional control paradigms.

Hardware

Interfaces and Connectors

The standard physical interface for MIDI transmission is the 5-pin , a 180-degree keyed circular plug defined in the MIDI 1.0 specification. Only three pins are utilized: pin 2 serves as the ground and connection, pin 4 as the (typically +5 V through a , up to 5 mA), and pin 5 as the current sink for the data signal; pins 1 and 3 remain unconnected. This current-loop design, combined with opto-isolators at the receiving end, provides electrical isolation to protect against ground loops and electrical , ensuring reliable low-speed serial data transfer at TTL-compatible voltage levels around 5 V. Standard MIDI does not use RJ45 connectors; the MIDI 1.0 specification defines the use of 5-pin DIN connectors for the physical interface, and RJ45 (Ethernet) connectors are not part of the standard MIDI specification but may be used in non-standard extensions or MIDI over Ethernet/IP implementations (e.g., RTP-MIDI). Many MIDI devices include a Thru port alongside the Out and In ports, which outputs an exact digital copy of the incoming MIDI data received on the In port without processing or delay. This enables daisy-chaining, where multiple devices can be connected in series from a single controller's Out port—such as Out to Device 1 In, Device 1 Thru to Device 2 In, and so on—allowing sequential addressing while minimizing cable requirements. For compact modern devices like portable synthesizers and pedals, the 3.5 mm TRS minijack has emerged as an alternative to the bulky 5-pin DIN, standardized by the MIDI Manufacturers Association (MMA) in 2018 using Type A wiring. In this configuration, the tip connects to DIN pin 5 (current sink), the ring to DIN pin 4 (), and the sleeve to DIN pin 2 (), maintaining compatibility with traditional MIDI electrical characteristics while supporting bidirectional ports in a smaller form factor. Adapters or crossover cables may be needed for legacy Type B implementations, but Type A ensures for new equipment. USB MIDI, introduced in 1999 through collaboration between the and MMA, encapsulates MIDI messages within a class-compliant USB Audio Device Class protocol, allowing direct connection to computers and devices over USB 2.0 (or higher) without proprietary drivers on operating systems supporting the standard. This virtual cable system supports multiple bidirectional MIDI ports per connection, with low latency suitable for real-time performance, and has largely supplanted dedicated DIN interfaces in consumer setups since the early .

Controllers and Input Devices

MIDI controllers and input devices serve as the primary means for musicians to generate MIDI data through tactile or gestural interactions, translating physical actions into digital signals for controlling virtual instruments or external hardware. These devices emerged alongside the MIDI standard in the early , with the first commercial implementations appearing on synthesizers like the Sequential Circuits Prophet-600 in 1982, which featured a five-pin MIDI output for keyboard input. By , at the NAMM trade show, demonstrations showcased keyboard controllers interfacing synthesizers such as the Prophet-600 with the , establishing keyboards as the foundational input method. Today, these devices range from traditional piano-style keyboards to specialized pads and sensors, prioritizing expressiveness through features like velocity sensitivity, which measures the force of key presses or strikes to vary note intensity. Keyboard controllers remain the most prevalent MIDI input devices, designed to replicate the feel of acoustic pianos while enabling polyphonic note transmission across multiple octaves. Velocity-sensitive keys, a standard since the mid-1980s, allow for dynamic by sending varying MIDI values (0-127) based on touch. Drum pad controllers, such as the Akai Professional MPD series, extend this to percussion, offering compact, velocity- and pressure-sensitive surfaces for triggering beats and samples. The MPD218 model, for instance, incorporates 16 MPC-style pads with adjustable sensitivity, six assignable knobs for parameter control, and pad banks for expanded triggering options, making it suitable for beat-making and live . These pads often support aftertouch for sustained modulation, enhancing rhythmic expressivity without requiring full drum kits. Beyond keyboards and pads, alternative input devices cater to diverse playing styles and instruments. Wind controllers, like those in the EWI series, use breath sensors, keys, and touch strips to emulate woodwinds or brass, converting airflow and fingering into MIDI pitch and modulation data for realistic phrasing. Guitar MIDI pickups, such as Roland's GK series, attach to electric guitars to analyze string vibrations and fretting positions, transforming guitar techniques like bends and slides into precise MIDI notes and continuous controller messages. Motion-based inputs further innovate control, with devices like the sensor enabling hand-gesture tracking for parameter automation; integrations via software such as GECO or MidiPaw map finger positions and gestures to MIDI continuous controllers for effects like filter sweeps or volume fades. These alternatives broaden accessibility, allowing non-keyboardists to interface with MIDI ecosystems. Key features in modern MIDI controllers enhance customization and depth. Programmable zones divide the input surface—such as a keyboard—into independent sections, each assignable to specific MIDI channels or instruments for splits (e.g., bass on lower keys, on upper) or layering multiple timbres simultaneously. Expression pedals connect via TRS jacks to provide real-time control over parameters like , wah-wah effects, or modulation depth, often programmable to send specific MIDI continuous controller numbers. These elements, refined since the , allow performers to create complex arrangements from a single device, as seen in controllers supporting up to four zones with independent pedal assignments. The evolution of MIDI controllers reflects advancements in connectivity and integration, transitioning from standalone MIDI keyboards reliant on five-pin DIN cables to versatile USB/MIDI hybrids in the . Early models, like the Fatar-based keyboards from the mid-, focused on basic note input for hardware synths, but by the 2000s, USB adoption enabled class-compliant operation without drivers, streamlining computer integration. Contemporary examples, such as ' Komplete Kontrol series, combine 61- or 88-key velocity-sensitive keyboards with encoders, screens, and deep software mapping for DAWs like , supporting both traditional MIDI and USB protocols for seamless workflow. This progression has democratized music production, with controllers now incorporating wireless options like for mobile setups. Devices typically connect via USB or MIDI interfaces for compatibility with hosts.

Sound Modules and Generators

Sound modules, also known as tone generators or MIDI sound generators, are dedicated hardware devices that receive MIDI to produce audio output without integrated performance controls like keyboards. These units expanded the flexibility of MIDI systems by allowing musicians to separate sound generation from input, enabling compact rack-mounted setups for live and studio use. Synthesizers in this category employ various synthesis techniques to create tones from MIDI note messages, often combining generation with processing elements like filters and envelopes. For instance, the JV-1080, released in 1994, is a prominent sample-and-synthesis (S&S) module that uses PCM as starting points, processed through subtractive synthesis methods including multi-stage filters, envelopes, and low-frequency oscillators for dynamic sound shaping. This approach allows for 64-voice and 16-part multitimbrality, making it a staple in professional recordings for its versatile orchestral and electronic patches. While primarily sample-based, it supports FM-like modulation via its structure, contributing to its widespread adoption in the . Samplers function by loading user-recorded or pre-stored waveforms into and triggering them via MIDI , often with pitch transposition and envelope control to emulate instruments. The S-series, starting with the S612 in 1985, pioneered affordable rack-mount MIDI samplers with 12-bit resolution and up to 48 kHz sampling rates in later models like the S1000 (1988), which offered 16-bit sampling at 44.1 kHz and 16-voice . These devices allowed musicians to capture external audio sources—such as vocals or instruments—and map them across MIDI keyboards, revolutionizing sample-based composition by providing 12 to 32 seconds of for multisampled programs. Drum machines as MIDI modules generate percussive sounds from ROM-based samples or synthesis, triggered by MIDI note-ons typically on channel 10 for compatibility with standards like . Dedicated units like the Yamaha RY30 (1991) combine sample playback with synthesis parameters, offering 64-voice , 80 preset drum kits, and advanced MIDI implementation for sequencing up to 16 parts, including velocity-sensitive triggering and programmable tables for custom mappings. Emulations of classic designs, such as the TR-808 and TR-909, are realized in modern MIDI-compatible modules like the TR-8S, which recreates their analog-modeled kicks, snares, and hi-hats using ACB technology for authentic timbres controllable via MIDI. Workstations integrate sound generation with onboard sequencing, providing comprehensive MIDI environments in a single unit. The series, introduced in 2011, exemplifies this by combining nine synthesis engines—including digital synthesis, sampling, and physical modeling—with a 16-track MIDI sequencer and audio recorder, supporting up to 200,000-note capacity and real-time pattern manipulation. This all-in-one design facilitates full song production directly from MIDI input, with over 21 GB of waveforms for diverse tonal palettes. Sound modules like these are typically controlled by external MIDI controllers, such as keyboards, to initiate note playback and parameter changes.

Supporting Devices

Supporting devices encompass auxiliary hardware that extends the functionality of MIDI systems by managing signal flow, effects, sequencing patterns, and expanding connectivity. These tools are essential for complex setups where multiple instruments and processors need coordinated control without relying on software. Effects units, particularly MIDI-controllable pedals, enable dynamic adjustment of audio in real-time MIDI environments. For instance, the Boss GT-100 is a compact multi-effects processor offering over 100 effects, including delay and reverb, with MIDI input and output ports that allow external controllers to send program changes, continuous controller messages, and system exclusive data for precise parameter control, such as modulating delay time or reverb decay. This integration facilitates seamless synchronization with sequencers or keyboards, enhancing live performance and studio workflows. Management devices handle the routing and processing of MIDI data in intricate configurations, preventing signal conflicts and optimizing transmission. MIDI mergers combine outputs from multiple sources into a single stream, as seen in the MIDI Solutions Merger, which accepts two inputs and distributes merged data to two outputs, ideal for linking several controllers to one . Splitters, like the MIDI Solutions Thru, replicate a single input across multiple outputs to drive several destinations simultaneously, such as distributing clock signals to synchronized synths. Filters, exemplified by the MIDI Solutions Event Processor, selectively process messages by mapping, scaling, or blocking specific events like note velocities or channel assignments, ensuring clean data flow in dense setups. These devices typically use standard 5-pin DIN connectors and support low-latency operation to maintain timing accuracy. Hardware sequencers provide autonomous pattern creation and playback, independent of computers, for driving external MIDI gear. The Akai MPC One serves as a standalone sequencer with a quad-core processor, 16 velocity-sensitive pads, and comprehensive MIDI , supporting multitimbral control over multiple tracks to diverse instruments like synths and modules. It features dedicated MIDI in/out ports alongside USB connectivity, enabling precise note and transmission with features like note repeat and tape stop for creative phrasing. Multi-port USB interfaces act as hubs to overcome limitations of single-port devices, facilitating connections to numerous MIDI peripherals. The MOTU Micro Lite, for example, offers five MIDI inputs and five outputs—totaling 80 channels—powered directly via USB, with driver support for Mac and Windows systems to ensure plug-and-play expansion. This allows users to interface with up to ten hardware devices in a daisy-chain configuration, using standard MIDI connectors for reliable, bus-powered operation in portable or studio environments.

Protocol

Messages and Data Format

MIDI messages are structured as serial streams of 8-bit bytes, transmitted at a fixed rate to ensure between devices. Each message consists of one or more bytes: a status byte (with the most significant bit set to 1, ranging from 0x80 to 0xFF) that identifies the message type and, for channel-specific messages, the MIDI channel (0-15), followed by zero or more data bytes (with the most significant bit set to 0, ranging from 0x00 to 0x7F) that provide parameters such as note numbers or values. To optimize bandwidth, the protocol employs running status, where a repeated status byte can be omitted if it matches the previous message, allowing consecutive data bytes to be sent directly. Messages are categorized into four main types, each serving distinct functions in musical performance and control. Channel Voice messages handle core performance data, including Note On (status 0x90 to 0x9F, followed by note number and ) to trigger sounds, Note Off (0x80 to 0x8F, with velocity typically 0x00), Polyphonic Key Pressure (0xA0 to 0xAF, for aftertouch on individual notes), Control Change (0xB0 to 0xBF, for parameters like modulation or ), Program Change (0xC0 to 0xCF, to select patches), and Channel Pressure (0xD0 to 0xDF, for overall aftertouch). Channel Mode messages (0xB0 to 0xBF with specific controller values) configure channel behavior, such as All Notes Off or Local Control on/off. System Common messages apply globally across all channels, including MIDI Time Code (MTC) quarter-frame messages (0xF1), Song Position Pointer (0xF2), and Song Select (0xF3). System Real-Time messages manage timing and synchronization, such as Timing Clock (0xF8) for metronome pulses, Start (0xFA), Stop (0xFC), and Active Sensing (0xFE) to indicate device activity. System Exclusive (SysEx) messages enable flexible, device-specific communication outside the standard categories, beginning with a status byte of 0xF0 and ending with 0xF7. They encapsulate variable-length data payloads, often used for manufacturer-specific functions like transmitting patch dumps, data, or device settings—for example, a might send a SysEx message with ID 0x41 followed by model-specific parameters. Universal SysEx messages, identified by non-proprietary IDs such as 0x7E for real-time universal or 0x7F for non-real-time, support standardized operations like bulk dumps or master tuning adjustments across compatible devices. Implementation charts standardize device by tabulating supported messages and behaviors in a tabular , typically divided into transmission and reception sections. These charts each message type (e.g., Note On, Control Change) with indicators for basic, mode 1, or mode 2 support, ensuring users can verify compatibility—for instance, a might recognize all Channel Voice messages but omit certain System Common ones. Such charts are recommended in MIDI device documentation to promote ecosystem reliability.

Electrical and Transmission Standards

MIDI employs a current-loop interface for electrical signaling to ensure galvanic isolation between connected devices, preventing ground loops and noise interference. The transmission uses a nominal current of 5 mA, where a logical 0 is represented by current ON (with pin 5 effectively pulled low through the loop) and a logical 1 by current OFF. This setup operates with 5 V ±10% power supply for the original specification, though updates allow for 3.3 V ±5% compatibility by adjusting output resistors. The interface relies on optocouplers at the receiving end for isolation, with recommended devices such as the Sharp PC-900 or HP 6N138, which activate with less than 5 mA and exhibit rise and fall times under 2 microseconds to maintain signal integrity. The serial transmission adheres to an asynchronous format at a fixed rate of 31.25 kbps ±1%, utilizing 8 data bits, no parity, and 1 stop bit (8N1 configuration), with the least significant bit transmitted first. Each byte, including start and stop bits, takes 320 microseconds to transmit, enabling reliable transfer of MIDI messages over the without built-in correction mechanisms like checksums in standard protocol elements. A start bit (logical 0) initiates transmission, followed by the 8 data bits and a stop bit (logical 1). Cable specifications mandate a maximum of 15 meters (50 feet) using shielded twisted-pair wiring, with the shield connected solely to pin 2 at both ends to minimize while avoiding ground connections that could introduce loops. This configuration supports direct connection from one MIDI output to one input without buffering, though chaining multiple devices may accumulate timing errors in optocoupler rise/fall times, potentially degrading signals beyond three units. Reliability in MIDI transmission stems from the low baud rate, which reduces bit error rates, combined with optical isolation to eliminate common-mode noise. The protocol includes Active Sensing messages (0xFE), transmitted at least every 300 ms by active senders, allowing receivers to detect cable disconnection or and reset voices accordingly if no data arrives within that interval; however, no automatic retransmission or parity checks are enforced for core messages.

Extensions

General MIDI and Variants

General MIDI (GM), first published in 1991 by the MIDI Manufacturers Association (MMA) and the Japan MIDI Standards Committee (JMSC), establishes a standardized mapping of instruments to ensure consistent playback of MIDI data across compatible sound generators. The specification requires support for 128 distinct instruments, assigned to program change values from 0 to 127, organized into categories such as , organ, guitar, and orchestral sounds. Additionally, it designates MIDI channel 10 exclusively for percussion, featuring a fixed 24-note to handle rhythmic elements without conflicting with melodic parts. This fixed structure promotes portability, allowing Standard MIDI Files to render predictably on any GM-compliant device without custom reconfiguration. Roland's GS (General Sound) format, introduced in 1991 alongside the SC-55 Sound Canvas module, extends GM by incorporating bank select messages to access instrument variations and enhanced effects control. It adds 98 tonal instruments, 15 percussion sounds, eight drum kits, and adjustable reverb and chorus parameters, enabling more nuanced while remaining fully backward compatible with GM. GS devices respond to standard program changes on the default but unlock expanded options through controller messages, such as scaling effect depths or selecting tonal variations like alternate guitar timbres. This extension became widely adopted in professional and consumer MIDI hardware, bridging basic GM portability with greater creative flexibility. Yamaha's XG (eXtended ) specification, debuted in 1994, builds on GM and GS by introducing further percussion expansions, including multiple specialized drum kits like Rock, , and Analog sets, alongside SFX kits for sound effects. It supports over 600 voices via bank selects, with detailed parameter controls for , pitch, and velocity sensitivity, and incorporates 12 reverb types such as Hall1, Room3, Stage2, and Plate for varied acoustic simulations. XG maintains GM compatibility by defaulting to the original sound set but allows real-time adjustments through system exclusive messages, enhancing expressiveness in complex arrangements without altering core MIDI channel assignments. General MIDI Level 2 (GM2), ratified by the MMA in 1999, refines the original standard with an expanded instrument palette, adding another 128 voices across melodic and percussion categories, plus new registered parameter numbers (RPNs) for functions like fine tuning and modulation depth ranging. It ensures partial with GM1 by prioritizing core mappings and messages, while enabling richer interactions through features like key-based controllers and global SysEx parameters for broader device interoperability.

Specialized Protocols

Several specialized protocols extend the core MIDI standard to address specific needs in timing synchronization, device control, and data transfer, primarily using System Exclusive (SysEx) messages for flexibility. These protocols enable MIDI systems to integrate with non-musical equipment and facilitate efficient exchange of audio data across devices. MIDI Time Code (MTC), introduced in 1986, provides a MIDI-based implementation of for synchronizing MIDI devices with linear media such as audio recorders and video systems. It encodes time in a format of hours:minutes:seconds:frames, transmitted via eight sequential Quarter Frame messages to convey the full time value, or via Full Frame messages for absolute positioning. MTC supports common frame rates like 24, 25, 29.97, and 30 frames per second, allowing precise alignment of musical events with time-based media without relying on tempo-dependent MIDI Clock. This protocol has been widely adopted in professional recording environments for its compatibility with existing SMPTE infrastructure. MIDI Machine Control (MMC), standardized in 1991, defines a set of SysEx commands for of transport functions on audio and video devices, bridging MIDI with traditional linear recording equipment. Commands include play, stop, record, rewind, fast-forward, and locate, enabling MIDI controllers or sequencers to operate tape machines, hard disk recorders, or video decks as if they were MIDI devices. MMC operates in a master-slave configuration, where the controller sends commands and receives status responses, supporting both simple point-to-point connections and more complex networked setups. This protocol enhanced studio workflows by integrating MIDI sequencing with analog-style media handling. The Sample Dump Standard (SDS), adopted in January 1986 by the MIDI Manufacturers Association (MMA) and the Japanese MIDI Standards Committee, specifies a SysEx-based method for transferring data between samplers and other MIDI devices. It supports both non-handshaking (one-way) and handshaking (bidirectional with acknowledgments) modes to ensure reliable transmission over MIDI's limited bandwidth. The process begins with a Dump Header message containing metadata like sample rate, length (up to 16,777,215 words), and loop points, followed by Data Packets of 120 bytes each, with the receiver sending ACK, NAK, WAIT, or CANCEL responses as needed. SDS also includes optional Loop Point Transmit and Request messages for managing up to 16,384 loop points per , making it essential for sharing custom samples in early production. Downloadable Sounds (DLS), ratified by the MMA in 1997, establishes a standardized file format for delivering custom instrument sounds to wavetable synthesizers, particularly for multimedia applications on computers and mobile devices. Level 1 provides a baseline architecture for downloading waveforms, envelopes, and modulation data, ensuring consistent playback across compatible hardware regardless of the sound set's complexity. DLS Level 2, an extension introduced later, adds support for advanced features like layered instruments and enhanced articulations to match evolving multimedia needs. For resource-constrained environments, Scalable Polyphony MIDI (SP-MIDI), developed alongside Mobile DLS, optimizes content delivery by adapting polyphony and voice allocation to the device's capabilities, such as limiting simultaneous notes on low-end phones while preserving full fidelity on capable systems. These formats have enabled portable custom tones and interactive audio in early mobile and web content.

Modern Developments

Alternative Transports

As MIDI technology evolved, alternatives to the original 5-pin DIN emerged to address limitations in speed, distance, and connectivity, particularly for integration with modern and networking environments. These transports encapsulate MIDI messages within other protocols, enabling higher bandwidth, multiple virtual channels, and or networked transmission while preserving the core MIDI data format. USB-MIDI, standardized by the and the MIDI Manufacturers Association (MMA), allows MIDI devices to connect directly to computers and other hosts via USB ports, supporting up to 16 virtual MIDI cables per endpoint for simultaneous multi-device handling. This class-compliant driver model eliminates the need for custom drivers in many cases, providing transfer speeds far exceeding the original MIDI's 31.25 kbps rate, and became the dominant interface for consumer MIDI controllers by the early . FireWire (), specified in the MMA's MIDI Media Adaptation Layer (part of the AM824 protocol), offered networking and high-speed data transfer for digital audio workstations (DAWs) in the pre-2010s era, supporting low-latency connections between multiple devices without a central host; however, its adoption waned with the decline of FireWire hardware. Wireless transports expanded MIDI's mobility, with the MMA's (BLE) MIDI specification, finalized in 2016, enabling cable-free connections between devices like keyboards and tablets over short ranges (up to 30 meters) with latencies under 10 ms in typical setups. For longer-range or networked wireless use, (Real-time Transport Protocol for MIDI, defined in IETF RFC 6295) supports transmission over and Ethernet, including session management and packet loss recovery for stable performance in home studios. Proprietary solutions, such as Roland's WM-1 wireless adaptor, achieve ultra-low latency (as low as 3 ms in fast mode) using custom implementations, allowing seamless integration with /macOS devices without compromising timing-critical applications. Ethernet-based transports facilitate studio-wide networking, with enabling MIDI over IP for multi-room setups and RTP over (AVB, per standards) providing synchronized, low-jitter transmission in professional environments through the MMA's AVBTP payload format. These Ethernet-based transports typically use RJ45 connectors, but are distinct from the standard MIDI physical interface that uses 5-pin DIN connectors; they represent non-standard extensions for MIDI over networks (e.g., RTP-MIDI over Ethernet). Yamaha's mLAN, an IEEE 1394-based protocol for combined audio and MIDI networking, allowed up to 64 channels over distances up to 100 meters but was discontinued around 2008 as FireWire support diminished. Other niche transports include the Musical Data Interface (SMDI), developed in the for high-speed sample transfers between computers and samplers via SCSI buses, offering rates up to 10 MB/s compared to MIDI's slow sample dump standard. Rare adapters have repurposed XLR connectors for MIDI signals—leveraging their 3-pin compatibility for longer cable runs in live sound—and DB-15 ports on legacy PC sound cards, such as models, for basic MIDI I/O in early computing setups.

MIDI 2.0

MIDI 2.0, released in 2020 by the MIDI Association, represents a major evolution of the MIDI protocol, addressing longstanding limitations of the original specification such as unidirectional communication and low-resolution parameter control by introducing bidirectional data flow and high-precision encoding. The protocol maintains full backward compatibility with MIDI 1.0 devices through architectural alignment and translation mechanisms, ensuring seamless integration in existing setups while enabling new capabilities like enhanced expressivity. Key enhancements include the Universal MIDI Packet format, which supports 32-bit resolution for controller values—vastly expanding from the 7-bit limit of MIDI 1.0—and facilitates property exchange via MIDI Capability Inquiry (MIDI-CI) for automatic device discovery and configuration. This bidirectional communication allows devices to negotiate capabilities, exchange profiles, and synchronize settings without manual intervention, streamlining workflows in music production environments. Central to MIDI 2.0's adoption is the Universal Profile, a standardized configuration that promotes plug-and-play across hardware and software by defining common mappings and behaviors. Devices supporting the Universal Profile can automatically detect and adapt to each other, reducing setup complexity while preserving MIDI 1.0 compatibility through fallback modes. For instance, a MIDI 2.0 controller can query a legacy via MIDI-CI to confirm supported features and adjust data transmission accordingly. By 2025, MIDI 2.0 implementation has accelerated with significant platform integrations and hardware releases. version 25H2, released in October 2025, introduced native support for MIDI 2.0, enabling direct high-resolution data handling without third-party drivers. At the NAMM 2025 show in January, the MIDI Association announced Network MIDI 2.0 (UDP), a specification for transmitting MIDI 2.0 data over Ethernet and with low latency and high bandwidth, supporting connections up to 100 meters via cable or 45 meters wirelessly. Yamaha's Montage M series, launched with firmware updates in 2024 and refined in 2025, fully implements MIDI 2.0, including bidirectional USB communication and integration with the Universal Profile for enhanced control in performance and studio settings. Advancements in MIDI 2.0 further boost musical expressivity and efficiency, notably through integration with MIDI Polyphonic Expression (MPE+), which leverages the protocol's expanded data capacity for per-note control of pitch, , and across multiple voices. The protocol supports up to 256 channels organized into 16 groups of 16, allowing complex multi-timbral arrangements without channel conflicts, while jitter-reduction techniques and the Universal MIDI Packet minimize latency in real-time applications. Digital audio workstations like Steinberg's Cubase 14, building on support introduced in version 13, now fully adopt MIDI for high-resolution parameter editing, MPE+ handling, and property exchange, facilitating smoother collaboration in professional production. These developments position MIDI as a foundational upgrade for modern , with ongoing adoptions enhancing its role in live performances and virtual instruments.

References

Add your contribution
Related Hubs
User Avatar
No comments yet.