Recent from talks
Nothing was collected or created yet.
Typesetting
View on Wikipedia

Typesetting is the composition of text for publication, display, or distribution by means of arranging physical type (or sort) in mechanical systems or glyphs in digital systems representing characters (letters and other symbols).[1] Stored types are retrieved and ordered according to a language's orthography for visual display. Typesetting requires one or more fonts (which are widely but erroneously confused with and substituted for typefaces).
One significant effect of typesetting was that authorship of works could be spotted more easily, making it difficult for copiers who have not gained permission.[2]
Pre-digital era
[edit]Manual typesetting
[edit]During much of the letterpress era, movable type was composed by hand for each page by workers called compositors. A tray with many dividers, called a case, contained cast metal sorts, each with a single letter or symbol, but backwards (so they would print correctly). The compositor assembled these sorts into words, then lines, then pages of text, which were then bound tightly together by a frame, making up a form or page. If done correctly, all letters were of the same height, and a flat surface of type was created. The form was placed in a press and inked, and then printed (an impression made) on paper.[3] Metal type read backwards, from right to left, and a key skill of the compositor was their ability to read this backwards text.
Before computers were invented, and thus becoming computerized (or digital) typesetting, font sizes were changed by replacing the characters with a different size of type. In letterpress printing, individual letters and punctuation marks were cast on small metal blocks, known as "sorts," and then arranged to form the text for a page. The size of the type was determined by the size of the character on the face of the sort. A compositor would need to physically swap out the sorts for a different size to change the font size.
During typesetting, individual sorts are picked from a type case with the right hand, and set from left to right into a composing stick held in the left hand, appearing to the typesetter as upside down. As seen in the photo of the composing stick, a lower case 'q' looks like a 'd', a lower case 'b' looks like a 'p', a lower case 'p' looks like a 'b' and a lower case 'd' looks like a 'q'. This is reputed to be the origin of the expression "mind your p's and q's". It might just as easily have been "mind your b's and d's".[3]
A forgotten but important part of the process took place after the printing: after cleaning with a solvent the expensive sorts had to be redistributed into the typecase - called sorting or dissing - so they would be ready for reuse. Errors in sorting could later produce misprints if, say, a p was put into the b compartment.

The diagram at right illustrates a cast metal sort: a face, b body or shank, c point size, 1 shoulder, 2 nick, 3 groove, 4 foot. Wooden printing sorts were used for centuries in combination with metal type. Not shown, and more the concern of the casterman, is the "set", or width of each sort. Set width, like body size, is measured in points.
In order to extend the working life of type, and to account for the finite sorts in a case of type, copies of forms were cast when anticipating subsequent printings of a text, freeing the costly type for other work. This was particularly prevalent in book and newspaper work where rotary presses required type forms to wrap an impression cylinder rather than set in the bed of a press. In this process, called stereotyping, the entire form is pressed into a fine matrix such as plaster of Paris or papier mâché to create a flong, from which a positive form is cast in type metal.
Advances such as the typewriter and computer would push the state of the art even farther ahead. Still, hand composition and letterpress printing have not fallen completely out of use, and since the introduction of digital typesetting, it has seen a revival as an artisanal pursuit. However, it is a small niche within the larger typesetting market.
Hot metal typesetting
[edit]The time and effort required to manually compose the text led to several efforts in the 19th century to produce mechanical typesetting. While some, such as the Paige compositor, met with limited success, by the end of the 19th century, several methods had been devised whereby an operator working a keyboard or other devices could produce the desired text. Most of the successful systems involved the in-house casting of the type to be used, hence are termed "hot metal" typesetting. The Linotype machine, invented in 1884, used a keyboard to assemble the casting matrices, and cast an entire line of type at a time (hence its name). In the Monotype System, a keyboard was used to punch a paper tape, which was then fed to control a casting machine. The Ludlow Typograph involved hand-set matrices, but otherwise used hot metal. By the early 20th century, the various systems were nearly universal in large newspapers and publishing houses.
Phototypesetting
[edit]
Phototypesetting or "cold type" systems first appeared in the early 1960s and rapidly displaced continuous casting machines. These devices consisted of glass or film disks or strips (one per font) that spun in front of a light source to selectively expose characters onto light-sensitive paper. Originally they were driven by pre-punched paper tapes. Later they were connected to computer front ends.
One of the earliest electronic photocomposition systems was introduced by Fairchild Semiconductor. The typesetter typed a line of text on a Fairchild keyboard that had no display. To verify correct content of the line it was typed a second time. If the two lines were identical a bell rang and the machine produced a punched paper tape corresponding to the text. With the completion of a block of lines the typesetter fed the corresponding paper tapes into a phototypesetting device that mechanically set type outlines printed on glass sheets into place for exposure onto a negative film. Photosensitive paper was exposed to light through the negative film, resulting in a column of black type on white paper, or a galley. The galley was then cut up and used to create a mechanical drawing or paste up of a whole page. A large film negative of the page is shot and used to make plates for offset printing.
Digital era
[edit]The next generation of phototypesetting machines to emerge were those that generated characters on a cathode-ray tube display. Typical of the type were the Alphanumeric APS2 (1963),[4] IBM 2680 (1967), I.I.I. VideoComp (1973?), Autologic APS5 (1975),[5] and Linotron 202 (1978).[6] These machines were the mainstay of phototypesetting for much of the 1970s and 1980s. Such machines could be "driven online" by a computer front-end system or took their data from magnetic tape. Type fonts were stored digitally on conventional magnetic disk drives.
Computers excel at automatically typesetting and correcting documents.[7] Character-by-character, computer-aided phototypesetting was, in turn, rapidly rendered obsolete in the 1980s by fully digital systems employing a raster image processor to render an entire page to a single high-resolution digital image, now known as imagesetting.
The first commercially successful laser imagesetter, able to make use of a raster image processor, was the Monotype Lasercomp. ECRM, Compugraphic (later purchased by Agfa) and others rapidly followed suit with machines of their own.
Early minicomputer-based typesetting software introduced in the 1970s and early 1980s, such as Datalogics Pager, Penta, Atex, Miles 33, Xyvision, troff from Bell Labs, TeX from Donald Knuth, and IBM's Script product with CRT terminals, were better able to drive these electromechanical devices, and used text markup languages to describe type and other page formatting information. The descendants of these text markup languages include LaTeX, SGML, XML and HTML.
The minicomputer systems output columns of text on film for paste-up and eventually produced entire pages and signatures of 4, 8, 16 or more pages using imposition software on devices such as the Israeli-made Scitex Dolev. The data stream used by these systems to drive page layout on printers and imagesetters, often proprietary or specific to a manufacturer or device, drove development of generalized printer control languages, such as Adobe Systems' PostScript and Hewlett-Packard's PCL.

Computerized typesetting was so rare that BYTE magazine (comparing itself to "the proverbial shoemaker's children who went barefoot") did not use any computers in production until its August 1979 issue used a Compugraphics system for typesetting and page layout. The magazine did not yet accept articles on floppy disks, but hoped to do so "as matters progress".[8] Before the 1980s, practically all typesetting for publishers and advertisers was performed by specialist typesetting companies. These companies performed keyboarding, editing and production of paper or film output, and formed a large component of the graphic arts industry. In the United States, these companies were located in rural Pennsylvania, New England or the Midwest, where labor was cheap and paper was produced nearby, but still within a few hours' travel time of the major publishing centers.
In 1985, with the new concept of WYSIWYG (for What You See Is What You Get) in text editing and word processing on personal computers, desktop publishing became available, starting with the Apple Macintosh, Aldus PageMaker (and later QuarkXPress) and PostScript and on the PC platform with Xerox Ventura Publisher under DOS as well as Pagemaker under Windows. Improvements in software and hardware, and rapidly lowering costs, popularized desktop publishing and enabled very fine control of typeset results much less expensively than the minicomputer dedicated systems. At the same time, word processing systems, such as Wang, WordPerfect and Microsoft Word, revolutionized office documents. They did not, however, have the typographic ability or flexibility required for complicated book layout, graphics, mathematics, or advanced hyphenation and justification rules (H and J).
By 2000, this industry segment had shrunk because publishers were now capable of integrating typesetting and graphic design on their own in-house computers. Many found the cost of maintaining high standards of typographic design and technical skill made it more economical to outsource to freelancers and graphic design specialists.
The availability of cheap or free fonts made the conversion to do-it-yourself easier, but also opened up a gap between skilled designers and amateurs. The advent of PostScript, supplemented by the PDF file format, provided a universal method of proofing designs and layouts, readable on major computers and operating systems.
QuarkXPress had enjoyed a market share of 95% in the 1990s, but lost its dominance to Adobe InDesign from the mid-2000s onward.[9]
SCRIPT variants
[edit]
IBM created and inspired a family of typesetting languages with names that were derivatives of the word "SCRIPT". Later versions of SCRIPT included advanced features, such as automatic generation of a table of contents and index, multicolumn page layout, footnotes, boxes, automatic hyphenation and spelling verification.[10]
NSCRIPT was a port of SCRIPT to OS and TSO from CP-67/CMS SCRIPT.[11]
Waterloo Script was created at the University of Waterloo (UW) later.[11] One version of SCRIPT was created at MIT and the AA/CS at UW took over project development in 1974. The program was first used at UW in 1975. In the 1970s, SCRIPT was the only practical way to word process and format documents using a computer. By the late 1980s, the SCRIPT system had been extended to incorporate various upgrades.[12]
The initial implementation of SCRIPT at UW was documented in the May 1975 issue of the Computing Centre Newsletter, which noted some the advantages of using SCRIPT:
- It easily handles footnotes.
- Page numbers can be in Arabic or Roman numerals, and can appear at the top or bottom of the page, in the centre, on the left or on the right, or on the left for even-numbered pages and on the right for odd-numbered pages.
- Underscoring or overstriking can be made a function of SCRIPT, thus uncomplicating editor functions.
- SCRIPT files are regular OS datasets or CMS files.
- Output can be obtained on the printer, or at the terminal…
The article also pointed out SCRIPT had over 100 commands to assist in formatting documents, though 8 to 10 of these commands were sufficient to complete most formatting jobs. Thus, SCRIPT had many of the capabilities computer users generally associate with contemporary word processors.[13]
SCRIPT/VS was a SCRIPT variant developed at IBM in the 1980s.
DWScript is a version of SCRIPT for MS-DOS, named after its author, D. D. Williams,[14] but was never released to the public and only used internally by IBM.
Script is still available from IBM as part of the Document Composition Facility for the z/OS operating system.[15]
SGML and XML systems
[edit]The standard generalized markup language (SGML) was based upon IBM Generalized Markup Language (GML). GML was a set of macros on top of IBM Script. DSSSL is an international standard developed to provide a stylesheets for SGML documents.
XML is a successor of SGML. XSL-FO is most often used to generate PDF files from XML files.
The arrival of SGML/XML as the document model made other typesetting engines popular. Such engines include Datalogics Pager, Penta, Miles 33's OASYS, Xyvision's XML Professional Publisher, FrameMaker, and Arbortext. XSL-FO compatible engines include Apache FOP, Antenna House Formatter, and RenderX's XEP. These products allow users to program their SGML/XML typesetting process with the help of scripting languages.
YesLogic's Prince is another one, which is based on CSS Paged Media.
Troff and successors
[edit]During the mid-1970s, Joe Ossanna, working at Bell Laboratories, wrote the troff typesetting program to drive a Wang C/A/T phototypesetter owned by the Labs; it was later enhanced by Brian Kernighan to support output to different equipment, such as laser printers. While its use has fallen off, it is still included with a number of Unix and Unix-like systems, and has been used to typeset a number of high-profile technical and computer books. Some versions, as well as a GNU work-alike called groff, are now open source.
TeX and LaTeX
[edit]
The TeX system, developed by Donald E. Knuth at the end of the 1970s, is another widespread and powerful automated typesetting system that has set high standards, especially for typesetting mathematics. LuaTeX and LuaLaTeX are variants of TeX and of LaTeX scriptable in Lua. TeX is considered fairly difficult to learn on its own, and deals more with appearance than structure. The LaTeX macro package, written by Leslie Lamport at the beginning of the 1980s, offered a simpler interface and an easier way to systematically encode the structure of a document. LaTeX markup is widely used in academic circles for published papers and books. Although standard TeX does not provide an interface of any sort, there are programs that do. These programs include Scientific Workplace and LyX, which are graphical/interactive editors; TeXmacs, while being an independent typesetting system, can also aid the preparation of TeX documents through its export capability.
Other text formatters
[edit]GNU TeXmacs (whose name is a combination of TeX and Emacs, although it is independent from both of these programs) is a typesetting system which is at the same time a WYSIWYG word processor.
SILE borrows some algorithms from TeX and relies on other libraries such as HarfBuzz and ICU, with an extensible core engine developed in Lua.[16][17] By default, SILE's input documents can be composed in a custom LaTeX-inspired markup (SIL) or in XML. Via the adjunction of 3rd-party modules, composition in Markdown or Djot is also possible.[18]
See also
[edit]- Dingbat
- Formula editor
- History of Western typography
- Ligature (typography)
- The Long Short Cut
- Point (typography)
- Prepress
- Printing
- Printing press
- Sort (typesetting)
- Strut (typesetting)
- Symbols – Comprehensive list of typographical symbols
- Technical writing
References
[edit]- ^ Dictionary.com Unabridged. Random House, Inc. 23 December 2009. Dictionary.reference.com
- ^ Murray, Stuart A., The Library: An Illustrated History, ALA edition, Skyhorse, 2009, page 131
- ^ a b Lyons, M. (2001). Books: A Living History. (pp. 59–61).
- ^ Encyclopedia of Computer Science and Technology, 1976
- ^ Encyclopedia of Computer Science and Technology
- ^ Linotype History
- ^ Petru-Ioan Becheru (Oct 2011). "Correcting Romanian typesetting mistakes by using regular expressions". An. Univ. Spiru Haret—ser. matemat.-inform. 7 (2): 31–36. ISSN 1841-7833. 83. Archived from the original on 2020-04-15. Retrieved 2012-04-09.(webpage has a translation button)
- ^ Helmers, Carl (August 1979). "Notes on the Appearance of BYTE..." BYTE. pp. 158–159.
- ^ "How QuarkXPress became a mere afterthought in publishing". Ars Technica. 2014-01-14. Retrieved 2022-08-07.
- ^ U01-0547, "Introduction to SCRIPT," Archived 2009-06-06 at the Wayback Machine is available through PRTDOC.
- ^ a b SCRIPT 90.1 Implementation Guide, June 6, 1990
- ^ A Chronology of Computing at The University of Waterloo
- ^ Glossary of University of Waterloo Computing Chronology
- ^ DWScript – Document Composition Facility for the IBM Personal Computer Version 4.6 Updates, DW-04167, Nov 8th, 1985
- ^ IBM Document Composition Facility (DCF)
- ^ "SILE Website". Retrieved 2023-08-01.
- ^ Simon Cozens (2017). "SILE, A new typesetting system" (PDF). TUGboat, Volume 38 (2017), No. 1. Retrieved 2023-08-01.
- ^ "Markdown and Djot to PDF with SILE". GitHub. Retrieved 2023-07-14.
External links
[edit]- Metal Type – For Those who Remember Hot Metal Typesetting
- TeX Users Group
- Fundamentals of typesetting Archived 2021-06-10 at the Wayback Machine
- The Lout Document Formatting System
- XML page at www.W3C.org
- SGML page at www.xml.org Archived 2019-02-19 at the Wayback Machine
- TYPESET-8, the first low cost computerized type setting system
Typesetting
View on GrokipediaFundamentals
Definition and Principles
Typesetting is the process of composing text for publication, display, or distribution by arranging physical type, digital glyphs, or their equivalents into pages, distinguishing it from the act of writing content or the mechanical reproduction via printing.[7] This arrangement focuses on creating visually coherent and legible layouts that enhance the presentation of written material across various media.[7] The origins of movable type trace back to China around 1040 AD with Bi Sheng's ceramic types,[8] while its development in Europe began in the mid-15th century when Johannes Gutenberg created reusable metal characters around 1440, enabling the efficient arrangement for printing one of the first major Western books, such as the 42-line Bible in 1455.[9] This innovation emphasized core elements like legibility through clear character forms, precise spacing to avoid visual clutter, and hierarchy to guide the reader's eye through the text structure.[9] Over time, these foundations evolved from physical manipulation of type to digital methods, but the underlying goals of clarity and organization persisted.[7] Central to typesetting are several key principles that govern text arrangement. Kerning involves adjusting the space between specific pairs of letters to achieve visual balance, such as reducing the gap between an uppercase "W" and "A" to prevent awkward white space.[10] Leading refers to the vertical space between lines of text, measured from baseline to baseline, which historically used thin lead strips and now influences readability by preventing lines from appearing cramped or overly separated.[10] Alignment determines text positioning, with options including flush left (ragged right) for natural reading flow, justified for uniform edges in formal documents, or centered for symmetrical emphasis.[11] Measure, or line length, optimizes comprehension by limiting lines to 45-75 characters, reducing eye strain and maintaining rhythmic reading pace.[12] The basic workflow of typesetting begins with manuscript preparation, including copyediting for grammar and formatting consistency, followed by layout where text is arranged into pages with applied principles like spacing and alignment.[13] This leads to proofing stages, where drafts are reviewed for errors and refinements, culminating in final output as print-ready files or digital formats integral to book design and broader typography.[13] In book design, typesetting integrates these elements to support narrative flow, while in typography, it ensures typefaces and layouts harmonize for effective visual communication.[7] Typesetting plays a crucial role in enhancing readability by organizing text into clear, navigable structures that minimize cognitive load for readers.[14] It contributes to aesthetics through harmonious layouts that evoke professionalism and visual appeal, making content more engaging without distracting from the message.[14] Ultimately, across print and digital media, typesetting facilitates effective communication by conveying tone, hierarchy, and intent, ensuring the written word reaches audiences with precision and impact.[14]Terminology and Tools
In typesetting, several core terms define the basic elements of character and spacing. The em is a relative unit of measurement equal to the current font size in points; for example, in 12-point type, one em equals 12 points.[15] The en, half the width of an em, serves as a smaller spacing unit, often used for dashes or indents.[15] A pica represents a traditional unit equivalent to 12 points, approximately one-sixth of an inch in both British/American and PostScript systems.[15] The point, the smallest standard measure, equals 1/72 inch in modern digital contexts.[15] A glyph is the fundamental visual form of an individual character, numeral, or symbol within a font.[15] A ligature combines two or more characters into a single glyph to improve readability and aesthetics, such as the joined forms of "fi" or "æ".[15] The baseline is the invisible horizontal line upon which most glyphs in a typeface rest, ensuring consistent alignment across lines of text.[15] Foundational tools facilitate the physical assembly and proofing of type, particularly in manual processes. The composing stick is an adjustable metal tray held in one hand, used to assemble individual pieces of type into lines of specified width, with a movable "knee" to set the measure.[16] Galleys are shallow brass trays, typically 2 feet long and 4–7 inches wide, into which lines of type are slid for temporary holding and proofing before further assembly.[16] The chase functions as a sturdy frame, often iron or wood, to lock assembled type pages securely for printing, enclosing the galleys or forms to prevent shifting.[16] Measurement systems in typesetting evolved from traditional to digital standards, affecting precision in layout. The traditional Didot point, rooted in European conventions, measures 0.376065 mm (or about 0.0148 inch), with 12 Didot points forming one cicero.[17] In contrast, the modern PostScript point, standardized for digital workflows, is exactly 1/72 inch or 0.3528 mm, making it slightly smaller than the Didot point by a factor of approximately 1.066 (1 Didot point ≈ 1.066 PostScript points).[17] This conversion ensures compatibility in desktop publishing, where 1 pica remains 12 points across both systems for consistent scaling.[17] Universal concepts guide text flow and layout integrity regardless of method. Hyphenation rules dictate word breaks to maintain even spacing, requiring at least two letters before and after the hyphen, avoiding more than two consecutive hyphenated lines, and prohibiting breaks in proper nouns or after the first syllable.[18] Widows are short lines (often a single word) at the end of a paragraph or column, isolated at the top of the next page, while orphans are similar short lines at the start of a page or column, detached from the preceding paragraph; both disrupt visual rhythm and are avoided by adjusting spacing or rephrasing.[19] Grid systems consist of horizontal and vertical lines that organize page elements for alignment and consistency, originating in early printed works like the Gutenberg Bible and used to relate text blocks, margins, and spacing without rigid constraints.[20]Historical Methods
Manual Typesetting
Manual typesetting emerged in the mid-15th century through Johannes Gutenberg's development of movable type in Mainz, Germany, around 1450, revolutionizing book production by allowing reusable metal characters to be arranged for printing.[9] Gutenberg's innovation utilized a specialized alloy composed of lead, tin, and antimony, which provided the necessary durability, low melting point for casting, and resistance to wear during repeated pressings.[21] This metal type, cast from individual molds, replaced earlier labor-intensive methods like woodblock carving, enabling the production of works such as the Gutenberg Bible circa 1455.[9] The core process began with compositors selecting individual type sorts—metal pieces bearing letters, punctuation, or spaces—from shallow wooden cases, where uppercase characters occupied the upper case and lowercase the lower case, organized by frequency of use for efficiency.[22] These sorts were assembled line by line in a handheld composing stick, set to the desired measure (line length), with spaces added to justify the text evenly and nicks aligned outward for orientation.[22] Completed lines were slid onto a galley, a rectangular tray, and secured with string or leads; proofing followed by inking the type with hand rollers and pulling impressions on dampened paper using a proof press to detect misalignments or defects.[22] Pages were then imposed on a stone or another galley, surrounded by wooden furniture, and locked securely into a metal chase using expanding quoins to form the complete forme for transfer to the printing press.[22] In England, the practice took root with William Caxton, who established the country's first printing press in Westminster in 1476 after learning the craft in Bruges, producing the first English-language books and adapting continental techniques to local needs.[23] Early printers encountered significant challenges, including acute shortages of type due to the high cost and labor of casting, which often necessitated shared cases among workshops or rapid reuse of sorts between jobs to sustain operations.[24] Despite its precision, manual typesetting proved highly labor-intensive and error-prone, with experienced compositors typically achieving rates of about 1,500 to 2,000 characters per hour under optimal conditions, far slower than later mechanized methods.[25] Common mistakes included inserting type upside down, mixing incompatible fonts from shared cases, or uneven justification, all of which demanded meticulous proofreading to avoid costly reprints.[26] Scalability was severely limited for large print runs, as type had to be distributed back into cases after each job, restricting output to small editions and making mass production impractical without extensive manpower.[24] Artisanal expertise defined the craft, as compositors wielded considerable discretion in aesthetic choices, such as fine-tuning letter and word spacing for visual harmony, selecting appropriate leading between lines, and integrating ornamental sorts like fleurons or rules to elevate the page's design and readability.[22] These decisions, honed through years of apprenticeship, transformed raw text into polished compositions that balanced functionality with artistic intent.[22]Hot-Metal Typesetting
Hot-metal typesetting represented a significant mechanization of the printing process, transitioning from labor-intensive manual methods to automated systems that cast type from molten metal alloys. This era began with the invention of the Linotype machine by Ottmar Mergenthaler in 1886, which produced entire lines of type, known as slugs, directly from keyboard input, revolutionizing newspaper production by enabling faster composition compared to hand-setting individual characters.[27][28] The machine's debut at the New York Tribune demonstrated its potential, casting lines at speeds that far exceeded manual techniques, which had served as precursors by relying on reusable metal sorts assembled by hand.[29] Central to hot-metal typesetting were two primary machines: the Linotype for line casting and the Monotype for individual character casting. The Linotype assembled brass matrices—small molds engraved with characters—into lines via a keyboard mechanism, then poured molten metal to form solid slugs ready for printing. In contrast, the Monotype system, developed by Tolbert Lanston and operational by 1897, separated composition into a keyboard unit that punched perforated paper tape and a caster unit that interpreted the tape to produce discrete type characters and spaces, allowing greater flexibility in spacing and corrections.[30][31] The core process in these machines involved selecting and aligning matrices to form text, followed by casting with a molten alloy typically composed of approximately 84% lead, 12% antimony, and 4% tin to ensure durability and low melting point around 240–250°C. An operator's keyboard input released matrices from magazines into an assembler, where they formed justified lines; a mold wheel then aligned with the matrix assembly as molten metal was injected, solidifying into type upon cooling before ejection as slugs or individual sorts. Excess metal was recycled, and matrices were returned to storage via an elevator mechanism, enabling continuous operation.[32][33][34] Advancements included the Intertype machine, introduced in 1911 as a direct competitor to the Linotype by offering interchangeable parts and matrices while incorporating design improvements for reliability, with widespread adoption in the 1920s among newspapers seeking cost-effective alternatives. For larger display type, the Ludlow Typograph, invented by William I. Ludlow and first commercially used in 1911, combined hand-assembly of matrices with automated casting to produce slugs up to 72 points in size, ideal for headlines and advertising.[35][36] Hot-metal typesetting peaked in the mid-20th century, dominating newspaper production with machines like the Linotype outputting up to six lines per minute, as seen in operations at the New York Times until its transition away from the system in 1978.[37][38] Its decline accelerated in the 1970s due to inherent limitations, including inflexibility for post-composition corrections that required recasting entire lines, and hazardous working conditions from lead fumes emitted during melting—known to cause poisoning via inhalation—and risks of molten metal spills leading to burns.[39][32][40]Phototypesetting
Phototypesetting represented a significant evolution from hot-metal methods, which served as the primary analog precursor for storing and composing type, by employing photographic techniques to project character images onto light-sensitive materials. Early experiments began in the 1920s in Germany with the Uhertype, a manually operated device designed by Hungarian engineer Edmond Uher that used photographic matrices on a rotating disk to expose characters one at a time.[41] Commercial development accelerated after World War II, with Mergenthaler Linotype introducing the Linofilm system in the mid-1950s, following initial testing in 1955-1956.[42] Independently, in France, the Photon machine—initially known as Lumitype—was patented in 1946 by inventors René Higonnet and Louis Moyroud and first commercially available in 1954, marking the debut of a fully automated photocomposition system.[43] The core process of phototypesetting involved generating negative film strips containing type images, which were then exposed onto photosensitive paper or film to create reproducible masters. Light sources, such as stroboscopic flash tubes, projected the character negatives through lenses for size and positioning adjustments, while later innovations incorporated cathode-ray tubes (CRTs) or early lasers to scan and expose the images directly.[42] The exposed material underwent chemical development in a darkroom to produce a positive or negative image suitable for contact printing onto printing plates, often for offset lithography. This photographic workflow allowed for precise control over line lengths, spacing, and justification, typically driven by perforated tape or early magnetic input from keyboards.[42] Several key systems defined the era, advancing from mechanical to electronic exposure methods. The Harris-Intertype Fototronic, introduced in the 1960s, utilized CRT technology for electronic character generation, enabling speeds up to 100 characters per second and supporting up to 480 characters per font disc.[42] In the 1970s, Compugraphic's MPS series, building on CRT-based designs, offered modular phototypesetters for mid-range production, achieving resolutions up to 2,500 dpi in high-end configurations and facilitating integration with early computer interfaces for directory and tabular work.[42] These systems, along with the Photon 900 series (up to 500 characters per second) and Linofilm variants (10-18 characters per second initially, scaling to 100 with enhancements), provided typographic quality comparable to metal type but with greater flexibility.[42] Phototypesetting offered distinct advantages over hot-metal techniques, including a cleaner production environment free from molten lead and associated hazards, as well as simpler corrections through re-exposure rather than recasting.[42] It enabled variable fonts, sizes, and styles without physical inventory limitations, with speeds reaching up to 600 characters per second in advanced models like the Photon ZIP 200, dramatically reducing composition time for complex layouts.[42] In applications, phototypesetting dominated book publishing and advertising from the 1960s through the 1980s, particularly for high-volume runs integrated with offset printing presses.[42] Notable uses included the rapid production of scientific indexes like the National Library of Medicine's Index Medicus (composed in 16 hours using Photon systems) and technical monographs, where it halved processing times compared to traditional methods.[42] Despite its innovations, phototypesetting faced limitations inherent to analog photography, such as delicate film handling that risked damage during transport and storage, necessitating controlled darkroom conditions for development and processing.[42] Enlargements often led to quality degradation due to optical distortions and loss of sharpness in the photographic emulsion, restricting scalability for very large formats without multiple exposures.[42]Early Digital Methods
Computer-Driven Systems
Computer-driven typesetting emerged in the 1960s through the use of mainframe computers to automate text composition and control phototypesetting hardware, marking a shift from purely manual or mechanical processes to digitized workflows. Early systems, such as the PC6 program developed at MIT in 1963–1964, ran on the IBM 7090 mainframe to generate formatted output for devices like the Photon 560 phototypesetter, producing the first computer-generated phototypeset documents, including excerpts from Lewis Carroll's Alice's Adventures in Wonderland.[44] By the 1970s, these capabilities expanded with minicomputer-based setups, including the IBM 1130, which supported high-speed composition for commercial printing applications like newspaper production, with over 272 installations reported by 1972.[44] Key variants of these proprietary systems included RUNOFF, created in 1964 by Jerome H. Saltzer at MIT for the Compatible Time-Sharing System (CTSS) on the IBM 7094. RUNOFF, paired with the TYPSET editor, enabled batch processing of documents using simple dot-commands for pagination, justification, and headers, outputting to line printers or early phototypesetters via magnetic tape.[45][46] This system represented an early milestone in automated text formatting, influencing subsequent tools by demonstrating how computers could handle structured input for reproducible output without real-time interaction. At Bell Laboratories, similar proprietary formatting approaches evolved in the late 1960s to support internal document production on early computers, laying groundwork for more advanced composition drivers.[44] The typical process in these systems relied on offline input methods, such as punch cards or paper/magnetic tape, fed into mainframes or minicomputers for processing. Software interpreted control codes to perform tasks like line justification and hyphenation—often rudimentary, without exception dictionaries in initial versions—before generating driver signals for phototypesetters. Early raster imaging appeared in some setups, using cathode-ray tubes (CRTs) to expose characters onto film, though precision was limited to fixed resolutions like 432 units per inch horizontally. Output was directed to specialized hardware, such as CRT-based phototypesetters, enabling faster production than hot-metal methods but still requiring physical film development.[44][47] Significant milestones in the 1970s included the rise of dedicated Computer-Assisted Typesetting (CAT) systems, which integrated computers directly with phototypesetting equipment for streamlined workflows. The Graphic Systems CAT, introduced in 1972, used punched tape input and film strips with 102 glyphs per font to produce high-resolution output at speeds supporting 15 font sizes from 5 to 72 points. In Europe, companies like Berthold advanced these technologies with the Diatronic system (1967, refined through the 1970s) and the ADS model in 1977, which employed CRT exposure for variable fonts and sizes, dominating high-end markets for book and periodical composition. Integration with minicomputers accelerated adoption; for instance, Digital Equipment Corporation's PDP-11 series powered several large-scale installations, including drivers for Harris phototypesetters like the 7500 model, where PDP-11/45 units handled input processing and output control in newspaper environments during the late 1970s.[47][48][49] Despite their innovations, these systems had notable limitations that constrained widespread use. Operations were predominantly batch-oriented, with jobs submitted via tape or cards and processed sequentially without user interaction, often taking hours for complex documents. Users typically needed programming expertise to embed control codes, as interfaces lacked graphical previews or intuitive editing. Moreover, output was tightly coupled to proprietary hardware, such as specific phototypesetters, leading to incompatibility and high costs for upgrades—exemplified by the need for custom drivers and frequent mechanical repairs in early CRT units.[44][47] These early computer-driven systems played a crucial transitional role by demonstrating the feasibility of digital control in typesetting, particularly through the introduction of computer-managed fonts. They pioneered the handling of bitmap fonts on CRT displays, allowing for scalable character generation independent of mechanical matrices, which set the stage for more standardized, device-agnostic formatting languages in subsequent decades.[47]Markup-Based Systems
Markup-based systems emerged in the 1970s as a means to describe document structure using tags, facilitating portable and programmable typesetting for phototypesetters and early digital outputs.[50] One of the earliest examples is Troff, developed by Joe Ossanna at AT&T Bell Labs in 1973 specifically for driving the Graphic Systems CAT phototypesetter on UNIX systems.[51] Troff used simple markup commands to format text, enabling precise control over spacing, fonts, and layout for high-quality printed output.[51] An extension, nroff, was created around the same time to adapt Troff's markup for terminal and line-printer display, broadening its utility in non-printing environments.[51] Building on these foundations, the Standard Generalized Markup Language (SGML) was formalized as an ISO standard in 1986, providing a meta-language for defining structured documents through descriptive tags that separate content from presentation.[52] SGML emphasized generic coding, allowing documents to be marked up for multiple uses, such as interchange and processing across systems.[52] This approach influenced later developments, including the Extensible Markup Language (XML), a simplified subset of SGML published by the W3C in 1998 to enable structured data exchange on the web.[53] XML uses tags like<p> to denote elements, supporting hierarchical document structures while ensuring interoperability.[53]
A parallel lineage began with TeX, created by Donald Knuth in 1978 to address the need for high-fidelity mathematical typesetting in his multivolume The Art of Computer Programming.[54] TeX employs a programming-like markup syntax with macros for defining complex layouts, compiling source files into device-independent output. In the early 1980s, Leslie Lamport extended TeX with LaTeX, introducing higher-level commands like \documentclass and environments for easier document preparation.[55] LaTeX's macro system abstracts TeX's primitives, allowing users to focus on content while automating formatting.[55]
In markup-based workflows, authors write source code embedded with tags—such as TeX's \section{Title} or XML's <section><title>Title</title></section>—which a processor compiles into final output like PDF or PostScript. This declarative approach excels in version control, as plain-text sources integrate seamlessly with tools like Git, and supports automation through scripts for batch processing.[54] Unlike imperative early computer systems influenced by predecessors like SCRIPT, markup prioritizes structural description over step-by-step instructions.[56]
These systems found widespread applications in specialized domains. LaTeX dominates academic publishing, powering journals from the American Mathematical Society and enabling precise rendering of equations in fields like physics and computer science.[57] For instance, LaTeX is used for its superior handling of technical content in AMS journals.[57] SGML, meanwhile, supported technical documentation in military standards, such as MIL-M-28001A, where it structured interchange of engineering data for defense applications under the CALS initiative.[58]
TeX's unique box-and-glue model underpins its precision, representing page elements as rigid boxes (e.g., glyphs or subformulas) connected by stretchable glue for optimal spacing and line breaking.[59] This algorithmic framework, detailed in Knuth's The TeXbook, ensures consistent hyphenation and justification without what-you-see-is-what-you-get (WYSIWYG) interfaces, prioritizing source fidelity for reproducible results.[60]