Recent from talks
Nothing was collected or created yet.
Electronic data processing
View on WikipediaThis article needs additional citations for verification. (October 2009) |
Electronic data processing (EDP) or business information processing can refer to the use of automated methods to process commercial data. Typically, this uses relatively simple, repetitive activities to process large volumes of similar information. For example: stock updates applied to an inventory, banking transactions applied to account and customer master files, booking and ticketing transactions to an airline's reservation system, billing for utility services. The modifier "electronic" or "automatic" was used with "data processing" (DP), especially c. 1960, to distinguish human clerical data processing from that done by computer.[1][2]
History
[edit]
Herman Hollerith then at the U.S. Census Bureau devised a tabulating system that included cards (Hollerith card, later Punched card), a punch for holes in them representing data, a tabulator and a sorter.[3] The system was tested in computing mortality statistics for the city of Baltimore.[3] In the first commercial electronic data processing Hollerith machines were used to compile the data accumulated in the 1890 U.S. Census of population.[4] Hollerith's Tabulating Machine Company merged with two other firms to form the Computing-Tabulating-Recording Company, later renamed IBM. The punch-card and tabulation machine business remained the core of electronic data processing until the advent of electronic computing in the 1950s (which then still rested on punch cards for storing information).[5]

The first commercial business computer was developed in the United Kingdom in 1951, by the J. Lyons and Co. catering organization.[6] This was known as the 'Lyons Electronic Office' – or LEO for short. It was developed further and used widely during the 1960s and early 1970s. (Lyons formed a separate company to develop the LEO computers and this subsequently merged to form English Electric Leo Marconi and then International Computers Limited.[7] By the end of the 1950s punched card manufacturers, Hollerith, Powers-Samas, IBM and others, were also marketing an array of computers.[8] Early commercial systems were installed exclusively by large organizations. These could afford to invest the time and capital necessary to purchase hardware, hire specialist staff to develop bespoke software and work through the consequent (and often unexpected) organizational and cultural changes.
At first, individual organizations developed their own software, including data management utilities, themselves. Different products might also have 'one-off' bespoke software. This fragmented approach led to duplicated effort and the production of management information needed manual effort.
High hardware costs and relatively slow processing speeds forced developers to use resources 'efficiently'. Data storage formats were heavily compacted, for example. A common example is the removal of the century from dates, which eventually led to the 'millennium bug'.
Data input required intermediate processing via punched paper tape or punched card and separate input to a repetitive, labor-intensive task, removed from user control and error-prone. Invalid or incorrect data needed correction and resubmission with consequences for data and account reconciliation.
Data storage was strictly serial on paper tape, and then later to magnetic tape: the use of data storage within readily accessible memory was not cost-effective until hard disk drives were first invented and began shipping in 1957. Significant developments took place in 1959 with IBM announcing the 1401 computer and in 1962 with ICT (International Computers & Tabulators) making delivery of the ICT 1301. Like all machines during this time the processor together with the peripherals – magnetic tape drives, disks drives, drums, printers and card and paper tape input and output required considerable space in specially constructed air conditioned accommodation.[9] Often parts of the punched card installation, in particular sorters, were retained to present the card input to the computer in a pre-sort form that reduced the processing time involved in sorting large amounts of data.[9]
Data processing facilities became available to smaller organizations in the form of the computer services bureau. These offered processing of specific applications e.g. payroll and were often a prelude to the purchase of customers' own computers. Organizations used these facilities for testing programs while awaiting the arrival of their own machine.
These initial machines were delivered to customers with limited software. The design staff was divided into two groups. Systems analysts produced a systems specification and programmers translated the specification into machine language.
Literature on computers and EDP was sparse and mostly obtained through articles appearing in accountancy publications and material supplied by the equipment manufacturers. The first issue of The Computer Journal published by The British Computer Society appeared in mid 1958. [9] The UK Accountancy Body now named The Association of Chartered Certified Accountants formed an Electronic Data Processing Committee in July 1958 with the purpose of informing its members of the opportunities created by the computer.[9] The Committee produced its first booklet in 1959, An Introduction to Electronic Computers. Also in 1958 The Institute of Chartered Accountants in England and Wales produced a paper Accounting by Electronic Methods.[9] The notes show what may be possible and the potential implications of using a computer.
Progressive organizations attempted to go beyond the straight systems transfer from punched card equipment and unit accounting machines to the computer, to producing accounts to the trial balance stage and integrated management information systems.[9] New procedures redesigned the way paper flowed, changed organizational structures, called for a rethink of the way information was presented to management and challenged the internal control principles adopted by the designers of accounting systems.[10] But the full realization of these benefits had to await the arrival of the next generation of computers
Today
[edit]As with other industrial processes commercial IT has moved in most cases from a custom-order, craft-based industry where the product was tailored to fit the customer; to multi-use components taken off the shelf to find the best-fit in any situation. Mass-production has greatly reduced costs and IT is available to the smallest organization.
LEO was hardware tailored for a single client. Today, Intel Pentium and compatible chips are standard and become parts of other components which are combined as needed. One individual change of note was the freeing of computers and removable storage from protected, air-filtered environments. Microsoft and IBM at various times have been influential enough to impose order on IT and the resultant standardizations allowed specialist software to flourish.
Software is available off the shelf. Apart from products such as Microsoft Office and IBM Lotus, there are also specialist packages for payroll and personnel management, account maintenance and customer management, to name a few. These are highly specialized and intricate components of larger environments, but they rely upon common conventions and interfaces.
Data storage has also been standardized. Relational databases are developed by different suppliers using common formats and conventions. Common file formats can be shared by large mainframes and desktop personal computers, allowing online, real-time input and validation.
In parallel, software development has fragmented. There are still specialist technicians, but these increasingly use standardized methodologies where outcomes are predictable and accessible.[9] At the other end of the scale, any office manager can dabble in spreadsheets or databases and obtain acceptable results (but there are risks, because many do not know what Software testing is). Specialized software is software that is written for a specific task rather for a broad application area. These programs provide facilities specifically for the purpose for which they were designed. At the other end of the scale, any office manager can dabble in spreadsheets or databases and obtain acceptable results.[9]
See also
[edit]References
[edit]- ^ Illingworth, Valerie (11 December 1997). Dictionary of Computing. Oxford Paperback Reference (4th ed.). Oxford University Press. p. 126. ISBN 9780192800466.
- ^ Anthony Ralston. Encyclopedia of Computer Science 4ed. Nature group. p. 502.
- ^ a b "From Herman Hollerith to IBM". Because of Her Story. Smithsonian. Retrieved 22 August 2021.
- ^ "Tabulating Equipment, From Herman Hollerith to IBM". Smithsonian, National Museum of American History. Retrieved July 6, 2019.
- ^ "Herman Hollerith (1860-1929)". Immigrant Entrepreneurship. Retrieved July 6, 2019.
- ^ World’s First Business Computer, LEO, Turns 60 Archived 2014-02-19 at the Wayback Machine, TechWeek Europe
- ^ Bird, Peter (2002). "J. Lyons & Co.: LEO Computers". Retrieved 18 May 2009.
- ^ Goldsmith J A. Choosing your Computer. The Accountant 14 June 1958.
- ^ a b c d e f g h Goyal, Meera; Mathur, Nishit (2020). Information Technology & Its Implications in Business. SBPD Publications. ISBN 978-93-5167-164-0. Retrieved 22 August 2021.
- ^ Mitchell R. baba Control with a Computer. The Accountant 3 November 1962.
Electronic data processing
View on GrokipediaFundamentals
Definition and Scope
Electronic data processing (EDP) refers to the use of electronic devices, primarily computers, to automate the handling of data through operations such as calculation, classification, sorting, and summarization, thereby minimizing human intervention in routine tasks.[8] This process transforms raw data into meaningful information, enabling efficient business and organizational functions like payroll computation and inventory management.[9] The scope of EDP includes various processing modes tailored to different needs: batch processing, where transactions are accumulated and handled in groups with a delay between input and output; real-time processing, which responds to data inputs immediately to influence ongoing activities; and transaction processing, which manages individual events as they occur, often in an event-driven manner.[8] It distinctly separates from manual data handling, which depends on human labor for transcription and computation and is susceptible to errors and slower speeds, while EDP forms a core component within the broader field of information technology that also covers areas like networking, cybersecurity, and software infrastructure.[8][10] The term EDP originated in the 1950s amid the rise of first-generation computers, initially describing the application of these machines for business data handling in sectors such as banking and manufacturing to manage growing transaction volumes.[11][9] By the mid-1950s, systems like the IBM 702 were specifically designed for such commercial purposes, marking a shift from scientific computing to automated administrative tasks.[11] Although now frequently integrated into general data processing terminology, EDP continues to emphasize electronic automation as distinct from mechanical or electromechanical predecessors.[9] Key characteristics of EDP include its high speed, which allows for rapid data transformation far exceeding manual capabilities; accuracy, achieved through mechanical precision that reduces transcription errors; capacity to handle large volumes of data efficiently; and programmability, enabling adaptable operations via software instructions for diverse applications.[8][9] These attributes have made EDP indispensable for scaling operations in data-intensive environments.[8]Core Principles
The input-process-output (IPO) model forms the foundational framework for electronic data processing (EDP) systems, describing the sequential flow of data through a computing environment to transform raw inputs into usable outputs. In this model, input involves capturing and entering data from various sources, such as punched cards, magnetic tapes, or keyboards, into the system for initial validation to ensure completeness and accuracy before further handling.[12] The processing stage applies arithmetic, logical, and control operations to manipulate the data, coordinated by a central control unit that interprets instructions and manages storage access, often using binary representation for efficient computation.[12] Output then delivers the processed results in forms like printed reports, updated files, or display screens, completing the cycle while incorporating error-checking mechanisms, such as batch totals reconciliation, to verify that outputs align with inputs and prevent discrepancies during transmission or storage.[8] Automation in EDP relies on principles that enable systematic, repeatable handling of data volumes without constant human intervention, emphasizing modularity, repeatability, and scalability to adapt workflows to varying demands. Modularity breaks down processing into independent, interchangeable components—like separate programs for data entry, computation, and reporting—allowing easier maintenance and upgrades without disrupting the entire system.[12] Repeatability ensures consistent results by standardizing procedures, such as predefined instruction sequences executed identically across runs, which minimizes variability in outcomes for routine tasks like payroll calculations. Scalability supports expansion by designing systems to handle increased data loads through additional storage or parallel processing units, facilitating growth from small departmental setups to enterprise-wide operations.[12] Key automation modes include batch processing, where transactions are accumulated, validated in groups, and processed sequentially at scheduled intervals to optimize resource use for high-volume, non-urgent tasks, versus online processing, which enables real-time interaction via direct terminal connections for immediate response in time-sensitive applications like inventory updates.[8][12] Maintaining data integrity and basic security in EDP involves foundational controls to safeguard accuracy and reliability throughout the data lifecycle, focusing on prevention of errors or unauthorized alterations without advanced cryptographic methods. Validation rules, such as range checks for numerical fields or format verifications for dates, are applied during input to reject invalid entries and ensure data conforms to predefined business logic before processing.[8] Checksums provide a simple mathematical verification by computing a fixed-size value from data blocks—often a sum or modular arithmetic result—to detect transmission errors or tampering, with recomputation at each stage confirming consistency against the original.[13] Basic auditing trails, including transaction logs and control totals that track input-output balances, enable post-processing reviews to identify anomalies, supporting reconstruction of records and enforcement of segregation of duties between data preparation, processing, and verification roles.[14] These measures collectively reduce risks of corruption in batch or online environments by embedding checks at modular boundaries. Efficiency in EDP systems is evaluated through core metrics that quantify performance in handling data flows, guiding optimizations for cost-effective operations. Throughput measures the rate of successful data units processed per unit time, such as characters per second on storage media (e.g., over 1 million on magnetic drums), indicating overall system capacity under load.[12] Turnaround time assesses the elapsed duration from data submission to result delivery, critical in batch systems where delays from queuing can extend hours but are minimized in online setups for near-instantaneous feedback. Resource utilization tracks the proportion of hardware elements—like CPU or storage—in active use versus idle states, with time-sharing techniques boosting this to over 90% by interleaving tasks and reducing wait times that plague sequential batch runs.[12] These metrics highlight trade-offs, such as higher throughput in batch modes at the expense of longer turnaround, informing scalable designs that balance speed and resource demands.[8]Historical Development
Origins and Early Innovations
The origins of electronic data processing trace back to mechanical and electromechanical systems designed to handle large-scale data tabulation, particularly for census operations. In the late 19th century, Herman Hollerith developed the punch card tabulating machine to efficiently process the 1890 U.S. Census, which involved over 60 million cards encoding demographic data through punched holes read by electrical contacts.[2] This innovation reduced census processing time from nearly a decade in 1880 to just months, marking a pivotal shift toward automated data handling and laying the groundwork for the Computing-Tabulating-Recording Company, which evolved into IBM.[15] Hollerith's machines relied on electromechanical relays and sorters, processing data at speeds up to 80 cards per minute, but remained limited by mechanical components that could not support complex computations.[16] The transition to fully electronic processing began during World War II, driven by military needs for rapid calculations. Parallel developments in Europe included the British Colossus, an electronic machine used for cryptographic processing from 1943 to 1945. In 1945, John Mauchly and J. Presper Eckert at the University of Pennsylvania completed the Electronic Numerical Integrator and Computer (ENIAC), the first general-purpose electronic digital computer, funded by the U.S. Army to compute artillery firing tables.[17] ENIAC used over 17,000 vacuum tubes for arithmetic operations, performing up to 5,000 additions per second—1,000 times faster than electromechanical predecessors—and demonstrated the feasibility of electronic computation for data-intensive tasks.[18] Concurrently, IBM engineers, led by Wallace J. Eckert, advanced the field with the Selective Sequence Electronic Calculator (SSEC) in 1948, which integrated 12,500 vacuum tubes with 21,400 relays to perform scientific calculations and modify stored programs electronically.[19] These innovations by Eckert, Mauchly, and early IBM teams bridged electromechanical tabulation to electronic systems, enabling programmable data processing beyond simple sorting.[20] The culmination of these efforts arrived with the UNIVAC I in 1951, developed by Eckert and Mauchly's company (later acquired by Remington Rand) as the first commercially available electronic data processing system.[5] Delivered to the U.S. Census Bureau, UNIVAC I used magnetic tape for input and output at transfer rates of 7,200 characters per second, succeeding where ENIAC's fixed-function design had limitations.[21] However, early electronic systems faced significant hurdles: vacuum tubes were prone to failure, with ENIAC experiencing significant downtime due to tube burnout from heat and power demands exceeding 150 kilowatts.[22] High development costs, such as ENIAC's $487,000 price tag (equivalent to over $7 million today), restricted access to government and large institutions, while programming required manual reconfiguration via wiring panels and switches, often taking days for setup.[23] These challenges underscored the nascent stage of electronic data processing, yet they spurred innovations in reliability and usability.[24]Post-War Expansion and Milestones
Following World War II, electronic data processing transitioned from military applications to commercial use, marking a pivotal expansion in the mid-20th century. The UNIVAC I, delivered to the U.S. Census Bureau in 1951, achieved widespread public recognition through its role in predicting Dwight D. Eisenhower's landslide victory in the 1952 U.S. presidential election during a CBS broadcast. Despite polls favoring Adlai Stevenson, UNIVAC accurately forecasted the outcome early in the evening based on partial returns, though network executives hesitated to air the prediction immediately due to the unexpected margin, displaying odds as "00 to 1" instead of the full three-digit figure. This event, featuring designers J. Presper Eckert and Harold J. Sweeney alongside Walter Cronkite, introduced computers to the American public and symbolized the shift toward practical, high-profile commercial applications.[25] In 1953, IBM introduced the IBM 701 Electronic Data Processing Machine, its first large-scale commercial electronic computer, which bridged scientific and business computing needs. Announced in 1952, with deliveries starting in 1953, the 701 performed up to 16,000 additions or subtractions per second and supported magnetic tape storage, enabling efficient data handling for defense, engineering, and emerging business tasks. With 19 units installed by 1955, it established IBM's dominance in the mainframe market and paved the way for business-oriented systems like the IBM 702, fostering corporate adoption of electronic processing for accounting and inventory management.[26] Key milestones in the 1950s and 1960s further propelled electronic data processing. IBM's FORTRAN, developed by John Backus and his team, was released in April 1957 for the IBM 704, becoming the first widely used high-level programming language tailored for scientific and engineering computations. By translating algebraic formulas and logical structures into machine code, FORTRAN reduced programming time from weeks to hours, enabling complex numerical simulations and data analysis that were previously infeasible.[27] Storage innovations also advanced rapidly; in the 1960s, IBM's 1311 Disk Storage Drive (1962) introduced removable disk packs holding 2 million characters at 1,500 RPM, while the 2314 facility (1965) offered 29 MB packs for higher-density direct access, complementing magnetic tape systems like the 9-track IBM 2401 (1962) that supported sequential data archiving at speeds up to 200 inches per second. These developments enhanced reliability and capacity, allowing corporations to manage growing volumes of business records.[28] The expansion influenced corporate structures profoundly, leading to the establishment of dedicated data processing departments. In the 1950s, large firms consolidated punched-card operations into centralized computing units using systems like the IBM 650 (over 1,800 installations by 1960), which automated payroll and inventory at costs comparable to manual methods. By the early 1960s, the IBM 1401's affordability (over 10,000 units shipped by 1964) enabled even mid-sized companies to form in-house departments, optimizing operations research and materials management while surpassing traditional equipment revenues by 1962. Concurrently, COBOL emerged in 1959 through the Conference on Data Systems Languages (CODASYL), initiated by the U.S. Department of Defense to create a standardized language for business data processing across diverse hardware. Drawing from Grace Hopper's FLOW-MATIC, COBOL's English-like syntax facilitated readable code for financial and administrative tasks, with its first specification released in 1960 and ANSI standardization in 1968.[29] Globally, electronic data processing spread to Europe, exemplified by the Ferranti Mark 1 in 1951, the world's first commercially available general-purpose stored-program computer. Delivered to the University of Manchester and funded partly by a British government contract, it performed practical computations like weather modeling before any U.S. counterpart, stimulating adoption in academic and industrial settings across the continent. A second unit, renamed FERUT, was sold to the University of Toronto, underscoring early international interest. However, rapid growth precipitated challenges, including the 1960s software crisis, where escalating project costs, delays, and reliability issues—exacerbated by hardware advances outpacing programming methodologies—prompted calls for structured engineering practices at events like the 1968 NATO Software Engineering Conference.[30][31]Key Components and Technologies
Hardware Elements
Electronic data processing (EDP) systems depend on interconnected hardware components that handle computation, data ingress, egress, and persistence, forming the physical foundation for automated information handling from the mid-20th century onward. These elements evolved rapidly, transitioning from bulky, power-intensive designs to more compact and efficient configurations, enabling widespread adoption in business and scientific applications up to the late 20th century. Central to this hardware ecosystem are the central processing unit for core operations, input/output peripherals for data exchange, storage media for retention, and overarching architectures that unify these parts into cohesive platforms. The central processing unit (CPU) constitutes the computational heart of EDP hardware, performing arithmetic operations such as addition and multiplication, as well as logical functions like comparisons and branching, to manipulate data streams. In the 1940s and early 1950s, CPUs relied on vacuum tubes—glass-enclosed electron devices—for signal amplification and switching, as exemplified by the UNIVAC I (1951), which used approximately 5,000 tubes to achieve processing speeds of about 1,000 additions per second but required significant cooling due to heat output. The shift to transistors in the mid-1950s revolutionized CPU design; these solid-state semiconductors, invented in 1947, replaced vacuum tubes by providing faster switching (up to 100 times quicker) with lower power consumption and greater reliability. Early transistorized CPUs appeared in systems like the TRADIC (1954), the U.S. Air Force's first all-transistor computer, which processed data at rates exceeding vacuum-tube predecessors while occupying far less space. By the 1960s, this evolution enabled second-generation EDP machines, such as the IBM 7090 (1959), to handle complex batch processing tasks with transistor-based logic circuits operating at microsecond speeds. Input and output devices bridged the gap between human operators and EDP machinery, allowing data to enter systems in structured formats and results to be rendered for review. Punch cards emerged as a dominant input medium, with each 80-column card encoding up to 80 alphanumeric characters via rectangular holes punched in specific positions, a standard refined by IBM from Herman Hollerith's 1890 census designs. Card readers, integral peripherals, scanned these cards optically or mechanically; for instance, the IBM 2501 reader (introduced in the 1960s for System/360 compatibility) processed up to 1,000 cards per minute, facilitating efficient data loading for payroll and inventory applications. Magnetic tapes supplemented punch cards by offering sequential bulk storage, with the IBM 7-track tape format (1952) storing up to approximately 2.9 million characters per 1,200-foot reel at a density of 200 characters per inch (1,400 bits per inch across 7 tracks) and transfer rates of up to 7,200 characters per second in later models, ideal for archiving transaction logs in early EDP workflows.[32] For output, line printers generated high-volume printed reports; the IBM 1403 (1959), a chain printer using a looping metal type band, produced 600 lines per minute at 120 characters per line, enabling rapid dissemination of financial summaries and data tables in business environments. Storage media provided the persistence necessary for EDP operations, evolving from mechanical to electronic forms to balance capacity, speed, and cost. Magnetic drum memory, a cylindrical rotating surface coated with ferromagnetic material, served as an early auxiliary storage solution in the 1950s, with devices like the one in the ERA 1101 (1950) offering capacities of 8,000 words (about 40 KB) and average access times of 10-20 milliseconds via fixed read/write heads, suitable for buffering intermediate results in codebreaking and data sorting tasks. Magnetic-core memory, comprising tiny ferrite rings wired into matrices, dominated main memory from the 1950s to 1970s; invented by Jay Forrester in 1949 for the Whirlwind computer, it provided non-volatile random access with cycle times of 1-5 microseconds and capacities scaling to 65,536 words (approximately 144 KB) in expanded PDP-1 systems (1960), using 18-bit words, far surpassing drums in speed for real-time arithmetic processing.[33] Early disk storage introduced random-access capabilities for larger datasets; the IBM 305 RAMAC (1956), the first commercial hard disk, stored 5 million 6-bit characters (equivalent to approximately 3.75 MB) across 50 24-inch platters spinning at 1,200 RPM, with average seek times of 600 milliseconds, transforming EDP by enabling direct retrieval of customer records without sequential scanning.[34] System architectures integrated these components into scalable platforms, with mainframes exemplifying comprehensive EDP hardware design. The IBM System/360, announced on April 7, 1964, pioneered a family of compatible processors sharing a single instruction set architecture, encompassing six CPU models with performance ranging from 0.1 to 5 MIPS and main memory capacities from 8 KB to 512 KB, all interfacing via standardized I/O channels to support up to 256 peripherals including tapes and disks. This modular design allowed seamless upgrades without software reconfiguration, consolidating disparate prior IBM lines into a unified EDP ecosystem that processed millions of transactions daily for banking and logistics by the late 1960s.| Storage Type | Invention/Introduction | Typical Capacity | Access Speed | Key Use in EDP |
|---|---|---|---|---|
| Magnetic Drum | 1932 (Tauschek); commercial 1950 (ERA 1101) | 8,000 words (~40 KB) | 10-20 ms average | Auxiliary buffering for sequential data operations[28] |
| Magnetic-Core | 1949 (Forrester, MIT) | 32K words (128 KB) by 1960s | 1-5 μs cycle time | Main memory for fast random-access arithmetic[28] |
| Early Disk (RAMAC) | 1956 (IBM 305) | 5 million 6-bit characters (~3.75 MB) | 600 ms seek | Random retrieval of business records[28] |
