Recent from talks
Contribute something
Nothing was collected or created yet.
Formal methods
View on WikipediaIn computer science, formal methods are mathematically rigorous techniques for the specification, development, analysis, and verification of software and hardware systems.[1] The use of formal methods for software and hardware design is motivated by the expectation that, as in other engineering disciplines, performing appropriate mathematical analysis can contribute to the reliability and robustness of a design.[2]
Formal methods employ a variety of theoretical computer science fundamentals, including logic calculi, formal languages, automata theory, control theory, program semantics, type systems, and type theory.[3]
Uses
[edit]Formal methods can be applied at various points through the development process.
Specification
[edit]Formal methods may be used to give a formal description of the system to be developed, at whatever level of detail desired. Further formal methods may depend on this specification to synthesize a program or to verify the correctness of a system.
Alternatively, specification may be the only stage in which formal methods is used. By writing a specification, ambiguities in the informal requirements can be discovered and resolved. Additionally, engineers can use a formal specification as a reference to guide their development processes.[4]
The need for formal specification systems has been noted for years. In the ALGOL 58 report,[5] John Backus presented a formal notation for describing programming language syntax, later named Backus normal form then renamed Backus–Naur form (BNF).[6] Backus also wrote that a formal description of the meaning of syntactically valid ALGOL programs was not completed in time for inclusion in the report, stating that it "will be included in a subsequent paper." However, no paper describing the formal semantics was ever released.[7]
Synthesis
[edit]Program synthesis is the process of automatically creating a program that conforms to a specification. Deductive synthesis approaches rely on a complete formal specification of the program, whereas inductive approaches infer the specification from examples. Synthesizers perform a search over the space of possible programs to find a program consistent with the specification. Because of the size of this search space, developing efficient search algorithms is one of the major challenges in program synthesis.[8]
Verification
[edit]Formal verification is the use of software tools to prove properties of a formal specification, or to prove that a formal model of a system implementation satisfies its specification.
Once a formal specification has been developed, the specification may be used as the basis for proving properties of the specification, and by inference, properties of the system implementation.
Sign-off verification
[edit]Sign-off verification is the use of a formal verification tool that is highly trusted. Such a tool can replace traditional verification methods (the tool may even be certified).[citation needed]
Human-directed proof
[edit]Sometimes, the motivation for proving the correctness of a system is not the obvious need for reassurance of the correctness of the system, but a desire to understand the system better. Consequently, some proofs of correctness are produced in the style of mathematical proof: handwritten (or typeset) using natural language, using a level of informality common to such proofs. A "good" proof is one that is readable and understandable by other human readers.
Critics of such approaches point out that the ambiguity inherent in natural language allows errors to be undetected in such proofs; often, subtle errors can be present in the low-level details typically overlooked by such proofs. Additionally, the work involved in producing such a good proof requires a high level of mathematical sophistication and expertise.
Automated proof
[edit]In contrast, there is increasing interest in producing proofs of correctness of such systems by automated means. Automated techniques fall into three general categories:
- Automated theorem proving, in which a system attempts to produce a formal proof from scratch, given a description of the system, a set of logical axioms, and a set of inference rules.
- Model checking, in which a system verifies certain properties by means of an exhaustive search of all possible states that a system could enter during its execution.
- Abstract interpretation, in which a system verifies an over-approximation of a behavioural property of the program, using a fixpoint computation over a (possibly complete) lattice representing it.
Some automated theorem provers require guidance as to which properties are "interesting" enough to pursue, while others work without human intervention. Model checkers can quickly get bogged down in checking millions of uninteresting states if not given a sufficiently abstract model.
Proponents of such systems argue that the results have greater mathematical certainty than human-produced proofs, since all the tedious details have been algorithmically verified. The training required to use such systems is also less than that required to produce good mathematical proofs by hand, making the techniques accessible to a wider variety of practitioners.
Critics note that some of those systems are like oracles: they make a pronouncement of truth, yet give no explanation of that truth. There is also the problem of "verifying the verifier"; if the program that aids in the verification is itself unproven, there may be reason to doubt the soundness of the produced results. Some modern model checking tools produce a "proof log" detailing each step in their proof, making it possible to perform, given suitable tools, independent verification.
The main feature of the abstract interpretation approach is that it provides a sound analysis, i.e. no false negatives are returned. Moreover, it is efficiently scalable, by tuning the abstract domain representing the property to be analyzed, and by applying widening operators[9] to get fast convergence.
Techniques
[edit]This section needs expansion. You can help by adding to it. (June 2024) |
Formal methods includes a number of different techniques.
Specification languages
[edit]The design of a computing system can be expressed using a specification language, which is a formal language that includes a proof system. Using this proof system, formal verification tools can reason about the specification and establish that a system adheres to the specification.[10]
Binary decision diagrams
[edit]A binary decision diagram is a data structure that represents a Boolean function.[11] If a Boolean formula expresses that an execution of a program conforms to the specification, a binary decision diagram can be used to determine if is a tautology; that is, it always evaluates to TRUE. If this is the case, then the program always conforms to the specification.[12]
SAT solvers
[edit]A SAT solver is a program that can solve the Boolean satisfiability problem, the problem of finding an assignment of variables that makes a given propositional formula evaluate to true. If a Boolean formula expresses that a specific execution of a program conforms to the specification, then determining that is unsatisfiable is equivalent to determining that all executions conform to the specification. SAT solvers are often used in bounded model checking, but can also be used in unbounded model checking.[13]
Applications
[edit]Formal methods are applied in different areas of hardware and software, including routers, Ethernet switches, routing protocols, security applications, and operating system microkernels such as seL4. There are several examples in which they have been used to verify the functionality of the hardware and software used in data centres. IBM used ACL2, a theorem prover, in the AMD x86 processor development process.[citation needed] Intel uses such methods to verify its hardware and firmware (permanent software programmed into a read-only memory)[citation needed]. Dansk Datamatik Center used formal methods in the 1980s to develop a compiler system for the Ada programming language that went on to become a long-lived commercial product.[14][15]
There are several other projects of NASA in which formal methods are applied, such as Next Generation Air Transportation System[citation needed], Unmanned Aircraft System integration in National Airspace System,[16] and Airborne Coordinated Conflict Resolution and Detection (ACCoRD).[17] B-Method with Atelier B,[18] is used to develop safety automatisms for the various subways installed throughout the world by Alstom and Siemens, and also for Common Criteria certification and the development of system models by ATMEL and STMicroelectronics.
Formal verification has been frequently used in hardware by most of the well-known hardware vendors, such as IBM, Intel, and AMD. There are many areas of hardware, where Intel have used formal methods to verify the working of the products, such as parameterized verification of cache-coherent protocol,[19] Intel Core i7 processor execution engine validation[20] (using theorem proving, BDDs, and symbolic evaluation), optimization for Intel IA-64 architecture using HOL light theorem prover,[21] and verification of high-performance dual-port gigabit Ethernet controller with support for PCI express protocol and Intel advance management technology using Cadence.[22] Similarly, IBM has used formal methods in the verification of power gates,[23] registers,[24] and functional verification of the IBM Power7 microprocessor.[25]
In software development
[edit]In software development, formal methods are mathematical approaches to solving software (and hardware) problems at the requirements, specification, and design levels. Formal methods are most likely to be applied to safety-critical or security-critical software and systems, such as avionics software. Software safety assurance standards, such as DO-178C allows the usage of formal methods through supplementation, and Common Criteria mandates formal methods at the highest levels of categorization.
For sequential software, examples of formal methods include the B-Method, the specification languages used in automated theorem proving, RAISE, and the Z notation.
In functional programming, property-based testing has allowed the mathematical specification and testing (if not exhaustive testing) of the expected behaviour of individual functions.
The Object Constraint Language (and specializations such as Java Modeling Language) has allowed object-oriented systems to be formally specified, if not necessarily formally verified.
For concurrent software and systems, Petri nets, process algebra, and finite-state machines (which are based on automata theory; see also virtual finite state machine or event driven finite state machine) allow executable software specification and can be used to build up and validate application behaviour.
Another approach to formal methods in software development is to write a specification in some form of logic—usually a variation of first-order logic—and then to directly execute the logic as though it were a program. The OWL language, based on description logic, is an example. There is also work on mapping some version of English (or another natural language) automatically to and from logic, as well as executing the logic directly. Examples are Attempto Controlled English, and Internet Business Logic, which do not seek to control the vocabulary or syntax. A feature of systems that support bidirectional English–logic mapping and direct execution of the logic is that they can be made to explain their results, in English, at the business or scientific level.[citation needed]
Semi-formal methods
[edit]Semi-formal methods are formalisms and languages that are not considered fully "formal". It defers the task of completing the semantics to a later stage, which is then done either by human interpretation or by interpretation through software like code or test case generators.[26]
Some practitioners believe that the formal methods community has overemphasized full formalization of a specification or design.[27][28] They contend that the expressiveness of the languages involved, as well as the complexity of the systems being modelled, make full formalization a difficult and expensive task. As an alternative, various lightweight formal methods, which emphasize partial specification and focused application, have been proposed. Examples of this lightweight approach to formal methods include the Alloy object modelling notation,[29] Denney's synthesis of some aspects of the Z notation with use case driven development,[30] and the CSK VDM Tools.[31]
Formal methods and notations
[edit]There are a variety of formal methods and notations available.
Specification languages
[edit]- Abstract State Machines (ASMs)
- A Computational Logic for Applicative Common Lisp (ACL2)
- Actor model
- Alloy
- ANSI/ISO C Specification Language (ACSL)
- Autonomic System Specification Language (ASSL)
- B-Method
- CADP
- Common Algebraic Specification Language (CASL)
- Esterel
- Java Modeling Language (JML)
- Knowledge Based Software Assistant (KBSA)
- Lustre
- mCRL2
- Perfect Developer
- Petri nets
- Predicative programming
- Process calculi
- RAISE
- Rebeca Modeling Language
- SPARK Ada
- Specification and Description Language
- TLA+
- USL
- VDM
- VDM-SL
- VDM++
- Z notation
Model checkers
[edit]- ESBMC[32]
- MALPAS Software Static Analysis Toolset – an industrial-strength model checker used for formal proof of safety-critical systems
- PAT – a free model checker, simulator and refinement checker for concurrent systems and CSP extensions (e.g., shared variables, arrays, fairness)
- SPIN
- UPPAAL
Solvers and competitions
[edit]Many problems in formal methods are NP-hard, but can be solved in cases arising in practice. For example, the Boolean satisfiability problem is NP-complete by the Cook–Levin theorem, but SAT solvers can solve a variety of large instances. There are "solvers" for a variety of problems that arise in formal methods, and there are many periodic competitions to evaluate the state-of-the-art in solving such problems.[33]
- The SAT competition is a yearly competition that compares SAT solvers.[34] SAT solvers are used in formal methods tools such as Alloy.[35]
- CASC is a yearly competition of automated theorem provers.
- SMT-COMP is a yearly competition of SMT solvers, which are applied to formal verification.[36]
- CHC-COMP is a yearly competition of solvers of constrained Horn clauses, which have applications to formal verification.[37]
- QBFEVAL is a biennial competition of solvers for true quantified Boolean formulas, which have applications to model checking.[38][39]
- SV-COMP is an annual competition for software verification tools.[40]
- SyGuS-COMP is an annual competition for program synthesis tools.[41]
Organizations
[edit]See also
[edit]References
[edit]- ^ Butler, R. W. (2001-08-06). "What is Formal Methods?". Retrieved 2006-11-16.
- ^ Holloway, C. Michael. "Why Engineers Should Consider Formal Methods" (PDF). 16th Digital Avionics Systems Conference (27–30 October 1997). Archived from the original (PDF) on 16 November 2006. Retrieved 2006-11-16.
{{cite journal}}: Cite journal requires|journal=(help) - ^ Monin, pp.3-4
- ^ Utting, Mark; Reeves, Steve (August 31, 2001). "Teaching formal methods lite via testing". Software Testing, Verification and Reliability. 11 (3): 181–195. doi:10.1002/stvr.223.
- ^ Backus, J.W. (1959). "The Syntax and Semantics of the Proposed International Algebraic Language of Zürich ACM-GAMM Conference". Proceedings of the International Conference on Information Processing. UNESCO.
- ^ Knuth, Donald E. (1964), Backus Normal Form vs Backus Naur Form. Communications of the ACM, 7(12):735–736.
- ^ O'Hearn, Peter W.; Tennent, Robert D. (1997). Algol-like Languages.
- ^ Gulwani, Sumit; Polozov, Oleksandr; Singh, Rishabh (2017). "Program Synthesis". Foundations and Trends in Programming Languages. 4 (1–2): 1–119. doi:10.1561/2500000010.
- ^ A. Cortesi and M. Zanioli, Widening and Narrowing Operators for Abstract Interpretation. Computer Languages, Systems and Structures. Volume 37(1), pp. 24–42, Elsevier, ISSN 1477-8424 (2011).
- ^ Bjørner, Dines; Henson, Martin C. (2008). Logics of Specification Languages. pp. VII–XI.
- ^ Bryant, Randal E. (2018). "Binary Decision Diagrams". In Clarke, Edmund M.; Henzinger, Thomas A.; Veith, Helmut; Bloem, Roderick (eds.). Handbook of Model Checking. p. 191.
- ^ Chaki, Sagar; Gurfinkel, Arie (2018). "BDD-Based Symbolic Model Checking". In Clarke, Edmund M.; Henzinger, Thomas A.; Veith, Helmut; Bloem, Roderick (eds.). Handbook of Model Checking. p. 191.
- ^ Prasad, Mukul R; Biere, Armin; Gupta, Aarti (January 25, 2005). "A survey of recent advances in SAT-based formal verification". International Journal on Software Tools for Technology Transfer. 7 (2): 156–173. doi:10.1007/s10009-004-0183-4.
- ^ Bjørner, Dines; Gram, Christian; Oest, Ole N.; Rystrøm, Leif (2011). "Dansk Datamatik Center". In Impagliazzo, John; Lundin, Per; Wangler, Benkt (eds.). History of Nordic Computing 3: IFIP Advances in Information and Communication Technology. Springer. pp. 350–359.
- ^ Bjørner, Dines; Havelund, Klaus. "40 Years of Formal Methods: Some Obstacles and Some Possibilities?". FM 2014: Formal Methods: 19th International Symposium, Singapore, May 12–16, 2014. Proceedings (PDF). Springer. pp. 42–61.
- ^ Gheorghe, A. V., & Ancel, E. (2008, November). Unmanned aerial systems integration to National Airspace System. In Infrastructure Systems and Services: Building Networks for a Brighter Future (INFRA), 2008 First International Conference on (pp. 1-5). IEEE.
- ^ Airborne Coordinated Conflict Resolution and Detection, http://shemesh.larc.nasa.gov/people/cam/ACCoRD/ Archived 2016-03-05 at the Wayback Machine
- ^ "Atelier B". www.atelierb.eu.
- ^ C. T. Chou, P. K. Mannava, S. Park, "A simple method for parameterized verification of cache coherence protocols", Formal Methods in Computer-Aided Design, pp. 382–398, 2004.
- ^ Formal Verification in Intel Core i7 Processor Execution Engine Validation, http://cps-vo.org/node/1371 Archived 2015-05-03 at the Wayback Machine, accessed at September 13, 2013.
- ^ J. Grundy, "Verified optimizations for the Intel IA-64 architecture", In Theorem Proving in Higher Order Logics, Springer Berlin Heidelberg, 2004, pp. 215–232.
- ^ E. Seligman, I. Yarom, "Best known methods for using Cadence Conformal LEC", at Intel.
- ^ C. Eisner, A. Nahir, K. Yorav, "Functional verification of power gated designs by compositional reasoning[dead link]", Computer Aided Verification, Springer Berlin Heidelberg, pp. 433–445.
- ^ P. C. Attie, H. Chockler, "Automatic verification of fault-tolerant register emulations", Electronic Notes in Theoretical Computer Science, vol. 149, no. 1, pp. 49–60.
- ^ K. D. Schubert, W. Roesner, J. M. Ludden, J. Jackson, J. Buchert, V. Paruthi, B. Brock, "Functional verification of the IBM POWER7 microprocessor and POWER7 multiprocessor systems", IBM Journal of Research and Development, vol. 55, no 3.
- ^ X2R-2, deliverable D5.1.
- ^ Daniel Jackson and Jeannette Wing, "Lightweight Formal Methods", IEEE Computer, April 1996
- ^ Vinu George and Rayford Vaughn, "Application of Lightweight Formal Methods in Requirement Engineering" Archived 2006-03-01 at the Wayback Machine, Crosstalk: The Journal of Defense Software Engineering, January 2003
- ^ Daniel Jackson, "Alloy: A Lightweight Object Modelling Notation", ACM Transactions on Software Engineering and Methodology (TOSEM), Volume 11, Issue 2 (April 2002), pp. 256-290
- ^ Richard Denney, Succeeding with Use Cases: Working Smart to Deliver Quality, Addison-Wesley Professional Publishing, 2005, ISBN 0-321-31643-6.
- ^ Sten Agerholm and Peter G. Larsen, "A Lightweight Approach to Formal Methods" Archived 2006-03-09 at the Wayback Machine, In Proceedings of the International Workshop on Current Trends in Applied Formal Methods, Boppard, Germany, Springer-Verlag, October 1998
- ^ "ESBMC". esbmc.org.
- ^ Bartocci, Ezio; Beyer, Dirk; Black, Paul E.; Fedyukovich, Grigory; Garavel, Hubert; Hartmanns, Arnd; Huisman, Marieke; Kordon, Fabrice; Nagele, Julian; Sighireanu, Mihaela; Steffen, Bernhard; Suda, Martin; Sutcliffe, Geoff; Weber, Tjark; Yamada, Akihisa (2019). "TOOLympics 2019: An Overview of Competitions in Formal Methods". In Beyer, Dirk; Huisman, Marieke; Kordon, Fabrice; Steffen, Bernhard (eds.). Tools and Algorithms for the Construction and Analysis of Systems. Lecture Notes in Computer Science. Cham: Springer International Publishing. pp. 3–24. doi:10.1007/978-3-030-17502-3_1. ISBN 978-3-030-17502-3.
- ^ Froleyks, Nils; Heule, Marijn; Iser, Markus; Järvisalo, Matti; Suda, Martin (2021-12-01). "SAT Competition 2020". Artificial Intelligence. 301 103572. doi:10.1016/j.artint.2021.103572. ISSN 0004-3702.
- ^ Cornejo, César (2021-01-27). "SAT-based arithmetic support for alloy". Proceedings of the 35th IEEE/ACM International Conference on Automated Software Engineering. ASE '20. New York, NY, USA: Association for Computing Machinery. pp. 1161–1163. doi:10.1145/3324884.3415285. ISBN 978-1-4503-6768-4.
- ^ Barrett, Clark; Deters, Morgan; de Moura, Leonardo; Oliveras, Albert; Stump, Aaron (2013-03-01). "6 Years of SMT-COMP". Journal of Automated Reasoning. 50 (3): 243–277. doi:10.1007/s10817-012-9246-5. ISSN 1573-0670.
- ^ Fedyukovich, Grigory; Rümmer, Philipp (2021-09-13). "Competition Report: CHC-COMP-21". Electronic Proceedings in Theoretical Computer Science. 344: 91–108. arXiv:2008.02939. doi:10.4204/EPTCS.344.7. ISSN 2075-2180.
- ^ Shukla, Ankit; Biere, Armin; Pulina, Luca; Seidl, Martina (November 2019). "A Survey on Applications of Quantified Boolean Formulas". 2019 IEEE 31st International Conference on Tools with Artificial Intelligence (ICTAI). IEEE. pp. 78–84. doi:10.1109/ICTAI.2019.00020. ISBN 978-1-7281-3798-8.
- ^ Pulina, Luca; Seidl, Martina (2019-09-01). "The 2016 and 2017 QBF solvers evaluations (QBFEVAL'16 and QBFEVAL'17)". Artificial Intelligence. 274: 224–248. doi:10.1016/j.artint.2019.04.002. ISSN 0004-3702.
- ^ Beyer, Dirk (2022). "Progress on Software Verification: SV-COMP 2022". In Fisman, Dana; Rosu, Grigore (eds.). Tools and Algorithms for the Construction and Analysis of Systems. Lecture Notes in Computer Science. Vol. 13244. Cham: Springer International Publishing. pp. 375–402. doi:10.1007/978-3-030-99527-0_20. ISBN 978-3-030-99527-0.
- ^ Alur, Rajeev; Fisman, Dana; Singh, Rishabh; Solar-Lezama, Armando (2017-11-28). "SyGuS-Comp 2017: Results and Analysis". Electronic Proceedings in Theoretical Computer Science. 260: 97–115. arXiv:1611.07627. doi:10.4204/EPTCS.260.9. ISSN 2075-2180.
Further reading
[edit]- Jonathan P. Bowen and Michael G. Hinchey, Formal Methods. In Allen B. Tucker, Jr. (ed.), Computer Science Handbook, 2nd edition, Section XI, Software Engineering, Chapter 106, pages 106-1 – 106-25, Chapman & Hall / CRC Press, Association for Computing Machinery, 2004.
- Hubert Garavel (editor) and Susanne Graf. Formal Methods for Safe and Secure Computer Systems[permanent dead link]. Bundesamt für Sicherheit in der Informationstechnik, BSI study 875, Bonn, Germany, December 2013.
- Garavel, Hubert; ter Beek, Maurice H.; van de Pol, Jaco (29 August 2020). "The 2020 Expert Survey on Formal Methods". Formal Methods for Industrial Critical Systems: 25 International Conference, FMICS 2020 (PDF). Lecture Notes in Computer Science (LNCS). Vol. 12327. Springer. pp. 3–69. doi:10.1007/978-3-030-58298-2_1. ISBN 978-3-030-58297-5. S2CID 221381022.* Michael G. Hinchey, Jonathan P. Bowen, and Emil Vassev, Formal Methods. In Philip A. Laplante (ed.), Encyclopedia of Software Engineering, Taylor & Francis, 2010, pages 308–320.
- Marieke Huisman, Dilian Gurov, and Alexander Malkis, Formal Methods: From Academia to Industrial Practice – A Travel Guide, arXiv:2002.07279, 2020.
- Gleirscher, Mario; Marmsoler, Diego (9 September 2020). "Formal methods in dependable systems engineering: a survey of professionals from Europe and North America". Empirical Software Engineering. 25 (6). Springer Nature: 4473–4546. arXiv:1812.08815. doi:10.1007/s10664-020-09836-5.
- Jean François Monin and Michael G. Hinchey, Understanding formal methods, Springer, 2003, ISBN 1-85233-247-6.
External links
[edit]- Archival material
- Formal method keyword on Microsoft Academic Search via Archive.org
- Evidence on Formal Methods uses and impact on Industry supported by the DEPLOY Archived 2012-06-08 at the Wayback Machine project (EU FP7) in Archive.org
Formal methods
View on GrokipediaOverview
Definition
Formal methods refer to the application of rigorous mathematical techniques to the specification, development, and verification of software and hardware systems, with a particular emphasis on discrete mathematics, logic, and automata theory.[7][8] These techniques enable the creation of unambiguous descriptions of system behavior, ensuring that designs meet intended requirements through formal analysis rather than ad hoc processes.[9] At their core, formal methods rely on abstract models that represent system properties mathematically, precise semantics that define the meaning of these models without ambiguity, and exhaustive analysis methods that explore all possible behaviors systematically, in contrast to selective testing approaches.[10][11] This foundation allows for the derivation of properties such as safety and liveness directly from the model, providing a structured pathway from high-level specifications to implementation.[12] Unlike empirical methods, which depend on testing to provide probabilistic assurance of correctness by sampling system executions, formal methods seek mathematical certainty through techniques like proofs of correctness that guarantee adherence to specifications under all conditions.[9][13] This distinction underscores formal methods' role in achieving complete verification, where testing can only falsify but not prove absence of errors.[14] Key mathematical foundations include first-order logic, which formalizes statements using predicates, variables, and quantifiers to express properties over domains, and state transition systems, which model computational processes as sets of states connected by transitions triggered by inputs or events.[15] These prerequisites provide the logical and structural basis for constructing and analyzing formal specifications.[16]Importance and benefits
Formal methods provide provable correctness for software and hardware systems by enabling mathematical proofs that verify the absence of certain errors, such as infinite loops or deadlocks, which is essential for ensuring system reliability in complex environments.[11] This approach allows developers to demonstrate that a system meets its specifications under all possible conditions, offering a level of assurance unattainable through testing alone, which can only show the presence of errors but not their absence.[17] Early error detection is another key benefit, as formal techniques identify inconsistencies and ambiguities in requirements and designs during initial phases, preventing costly rework later. In safety-critical industries, formal methods play a crucial role in achieving compliance with stringent standards, such as DO-178C for aviation software, where they supplement traditional verification to provide evidence of correctness for high-assurance levels.[18] Similarly, in automotive systems, ISO 26262 recommends formal methods for ASIL C and D classifications to verify functional safety requirements, ensuring that electronic control units behave predictably in fault-prone scenarios.[19] These applications facilitate certification by regulators, reducing the risk of failures that could lead to loss of life or property damage. Quantitative impacts underscore the value of formal methods in error avoidance; for instance, the 1996 Ariane 5 Flight 501 failure, caused by inadequate requirements capture and design faults, resulted in a $370 million loss and a one-year program delay, but proof-based formal engineering could have prevented it through rigorous specification and verification.[20] Case studies from NASA and the U.S. Army demonstrate cost savings in long-term maintenance: in one Army project using the SCADE tool, formal analysis detected 73% of defects early, yielding a net savings of $213,000 (5% of project cost) by avoiding expensive late fixes.[21] While formal methods require high upfront investment—typically adding 10-20% to initial system costs due to specification development and tool expertise—these expenses are amortized through reduced testing (by 50-66%) and maintenance in complex, high-stakes systems, where traditional methods falter.[22] This trade-off is particularly favorable for projects involving reusable components or regulatory compliance, where long-term reliability outweighs short-term overhead.[23]History
Origins and early developments
The origins of formal methods trace back to foundational work in mathematical logic during the mid-20th century, particularly Alan Turing's 1936 paper "On Computable Numbers, with an Application to the Entscheidungsproblem," which introduced the concept of a universal computing machine and proved the undecidability of the halting problem, establishing fundamental limits on what can be mechanically computed.[24] This work laid the groundwork for understanding computability in algorithmic terms, influencing later efforts to rigorously specify and verify computational processes. Complementing Turing's contributions, Alonzo Church developed lambda calculus in the 1930s as a formal system for expressing functions and computation, providing an alternative model equivalent to Turing machines.[25] Together with Turing's results, Church's framework supported the Church-Turing thesis, posited around 1936, which asserts that any effectively calculable function can be computed by a Turing machine, thus unifying notions of effective computation in logic and early computer science.[26] In the 1960s, as computing shifted toward practical programming languages, formal methods began to influence program semantics and design. C. A. R. Hoare's 1969 paper "An Axiomatic Basis for Computer Programming" introduced axiomatic semantics, using preconditions and postconditions to formally reason about program correctness, enabling proofs of partial correctness for imperative programs. Concurrently, Edsger W. Dijkstra advanced structured programming in the late 1960s, advocating for disciplined control structures like sequence, selection, and iteration to replace unstructured jumps, as exemplified in his 1968 critique of the GOTO statement and subsequent writings on program derivation. These developments emphasized mathematical rigor in software construction, bridging theoretical logic with engineering practice to mitigate errors in increasingly complex systems. The emergence of formal methods as a distinct field in the 1970s was driven by growing concerns over software reliability amid the "software crisis," highlighted at the 1968 NATO Conference on Software Engineering in Garmisch, Germany, where experts like Friedrich L. Bauer and others discussed the need for systematic, engineering-like approaches to combat project overruns and failures in large-scale systems such as OS/360. This motivation spurred the development of key specification methods in the UK, including the Vienna Development Method (VDM) originated by Cliff B. Jones and colleagues at the IBM Vienna Laboratory in the early 1970s, providing a rigorous framework for stepwise refinement and data abstraction in software design.[27] Similarly, Tony Hoare introduced Communicating Sequential Processes (CSP) in his 1978 paper, offering a mathematical model for specifying patterns of interaction in concurrent systems.[28] Robin Milner developed Logic for Computable Functions (LCF) in the mid-1970s at the University of Edinburgh, an interactive theorem-proving system that laid the foundation for mechanized reasoning about functional programs.[29] These efforts marked the transition from theoretical foundations to practical mechanized reasoning, setting the stage for rigorous software analysis without delving into later refinements. Early formal verification tools also emerged, including the Boyer-Moore theorem prover, developed by Robert S. Boyer and J Strother Moore starting in the early 1970s as an automated system for proving theorems in a computational logic based on primitive recursive functions and induction.[30]Key milestones and modern evolution
The 1980s marked a pivotal era for formal methods with the emergence of influential specification and verification techniques. The Z notation, a model-oriented formal specification language based on set theory and first-order logic, was developed by Jean-Raymond Abrial in 1977 at the Oxford University Computing Laboratory and further refined by Oxford researchers through the 1980s.[31] Concurrently, the SPIN model checker, an on-the-fly verification tool for concurrent systems using Promela as its input language, began development in 1980 at Bell Labs and saw its first public release in 1991, enabling efficient detection of liveness and safety properties in distributed software.[32] Another key advancement was Cleanroom software engineering, introduced in the mid-1980s by Harlan Mills and colleagues at IBM, which emphasized mathematical correctness through incremental development, statistical testing, and formal proofs to achieve high-reliability software without debugging.[33] Milner's work also evolved with the introduction of the Calculus of Communicating Systems (CCS) in 1980, complementing CSP for modeling concurrency. In the 1990s and 2000s, formal methods transitioned toward broader industrial adoption, particularly in hardware verification and standardization. IBM extensively applied formal techniques, including theorem proving and model checking, to verify the PowerPC microprocessor family starting in the mid-1990s, with tools like the Microprocessor Test Generation (MPTg) system used across multiple processor designs to ensure functional correctness and reduce verification time.[34] This effort exemplified the shift to formal methods in complex hardware, where traditional simulation proved insufficient for exhaustive coverage. Complementing this, the IEEE Std 1016, originally published in 1987 as a recommended practice for software design descriptions, was revised in 1998 to incorporate formal specification views, facilitating its integration into software engineering processes for critical systems throughout the 2000s. The 2010s witnessed the rise of highly automated tools that enhanced scalability and usability of formal methods. Advances in satisfiability modulo theories (SMT) solvers and bounded model checkers, such as those integrated into tools like Z3 and CBMC, enabled verification of larger software and hardware systems with minimal manual intervention, as demonstrated in industrial applications for embedded systems. By the late 2010s and into the 2020s, formal methods began integrating with artificial intelligence, particularly for verifying neural networks to ensure robustness against adversarial inputs; techniques like abstract interpretation and SMT-based bounds propagation were applied post-2020 to certify properties such as safety in autonomous systems.[35] Government initiatives, including DARPA's Trusted and Assured Microelectronics (TAM) program launched in 2020, further promoted formal methods for safety-critical ML components in hardware-software co-design. Recent trends through 2025 have focused on scalability via machine learning-assisted proofs, with the Lean theorem prover seeing significant enhancements through integration with large language models (LLMs) for automated tactic selection and proof synthesis. For instance, studies have shown LLMs improving proof completion rates in Lean by generating intermediate lemmas, reducing human effort in formalizing complex mathematical and software properties.[36] These developments underscore formal methods' evolution toward hybrid human-AI workflows, enabling verification of AI systems themselves while maintaining rigorous guarantees.Uses
Specification
Formal specification in formal methods involves translating informal natural language requirements into precise mathematical notations to eliminate ambiguity and ensure a clear understanding of system behavior. This process uses formal languages grounded in mathematical logic, such as first-order predicate logic, to express properties and constraints rigorously. For instance, predicate logic allows the definition of system states and operations through predicates that describe relationships between variables, enabling unambiguous representation of requirements that might otherwise be misinterpreted in natural language descriptions.[10] The specification process typically proceeds through stepwise refinement, starting from high-level abstract models and progressively adding details toward concrete implementations. Abstract specifications focus on "what" the system must achieve, often using operational semantics, which describe behavior through step-by-step execution rules on an abstract machine, or denotational semantics, which map program constructs directly to mathematical functions denoting their computational effects. This refinement ensures that each level preserves the properties of the previous one, facilitating a structured development path while maintaining correctness.[37][38] Key concepts in formal specification include invariants, which are conditions that must hold true throughout system execution, and pre- and post-conditions, which specify the state before and after an operation, respectively. A prominent formalism for these is the Hoare triple, denoted as , where is the precondition, is the statement or program segment, and is the postcondition; it asserts that if holds before executing , then will hold afterward, assuming terminates. Invariants and these conditions provide a foundation for reasoning about program correctness without delving into implementation details.[39] One major advantage of formal specification is its ability to detect inconsistencies and errors early in the development lifecycle, often during the specification phase itself, by enabling mathematical analysis of requirements. This early validation reduces the cost of fixes compared to later stages and supports downstream activities like verification, where specifications serve as unambiguous benchmarks for proving implementation fidelity. Additionally, the rigor of formal notations promotes better communication among stakeholders and enhances overall system reliability in critical applications.[10][40]Synthesis
Synthesis in formal methods refers to the automated generation of implementations or designs that provably satisfy given high-level specifications, ensuring correctness by construction. This process typically involves deductive synthesis, where theorem proving is used to derive programs from logical specifications by constructing proofs that guide the implementation, or constructive synthesis, which employs automata-theoretic techniques to build systems from temporal logic formulas. For instance, deductive approaches treat synthesis as a theorem-proving task, transforming specifications into executable code through inference rules and constraint solving.[41][42][43] Key techniques in formal synthesis leverage program synthesis tools grounded in satisfiability modulo theories (SMT) solvers, which search for implementations that meet formal constraints while producing artifacts guaranteed to be correct with respect to the input specification. These methods often integrate refutation-based learning to iteratively refine candidate solutions, enabling the synthesis of complex structures like recursive functions or reactive systems. SMT-based synthesis excels in domains requiring precise handling of data types and arithmetic, as it encodes the synthesis problem as a satisfiability query over theories such as linear integer arithmetic. By focusing on bounded search spaces or templates, these tools generate efficient, verifiable outputs without exhaustive enumeration. Recent advances as of 2024 include AI-assisted synthesis for safety-critical autonomous systems, improving scalability and handling hybrid dynamics.[44][45][46] Representative examples illustrate the practical application of synthesis in formal methods. In hardware design, synthesis from hardware description languages (HDLs) or higher-order logic specifications automates the creation of synchronous circuits, as seen in tools that compile recursive function definitions into clocked hardware modules while preserving behavioral equivalence. For software, Alloy models can drive multi-concept synthesis, where relational specifications are used to generate programs handling multiple interacting concerns, such as data structures with concurrent access. NASA's Prototype Verification System (PVS) supports synthesis through its code generation capabilities, enabling the extraction of verified C code from applicative specifications in safety-critical avionics contexts.[47][48][49] A primary challenge in formal synthesis algorithms is ensuring completeness, meaning the method finds a solution if one exists within the specified language, and termination, guaranteeing the search process halts in finite time. These issues arise due to the undecidability of general synthesis problems, prompting techniques like bounded synthesis or inductive learning to approximate solutions while bounding computational resources. Relative completeness results, where termination implies a valid program if assumptions hold, provide theoretical guarantees but require careful scoping of the search space to avoid non-termination in practice.[50][51][52]Verification
Verification in formal methods involves rigorously proving or disproving that a system implementation satisfies its formal specification, providing mathematical assurance of correctness beyond empirical testing. This process targets exhaustive analysis of the system's behavior to identify any deviations from intended properties, distinguishing it from partial checks like simulation. By establishing formal relations between models, verification ensures that all possible executions align with the specification, mitigating risks in critical systems where failures could have severe consequences.[53] The core goal of verification is to perform exhaustive checking through techniques such as equivalence relations, simulation, or induction, exemplified by bisimulation relations between models. Bisimulation defines a behavioral equivalence where states in two models are indistinguishable if they agree on observable actions and can mutually simulate each other's transitions, enabling reduction of state spaces while preserving key properties for comprehensive analysis. This approach guarantees that the implementation matches the specification across all reachable states, often computed via iterative refinement akin to induction.[12] Verification addresses several types of properties: functional verification ensures behavioral correctness by confirming that the system produces expected outputs for all inputs; safety properties assert that no undesirable "bad" states are ever reached; and liveness properties guarantee that desired "good" states will eventually occur from any execution path. Safety violations are detectable in finite prefixes of execution traces, while liveness requires arguments of progress, such as well-founded orderings, to prevent infinite loops without achievement. Functional correctness typically combines safety (partial correctness) and liveness (termination) to fully validate system behavior. Recent developments as of 2025 include enhanced model checking tools integrated with machine learning for handling large-scale systems in autonomous applications.[54][55] The verification process begins with mapping the implementation model to the specification, often using a shared semantic framework to align their representations, followed by deriving proof obligations as formal assertions to be checked. For instance, in finite-state systems, model checking exhaustively explores the state space to validate these obligations against temporal logic properties. This mapping ensures that implementation details, such as code or hardware descriptions, are refined from or equivalent to the abstract specification, with proof obligations capturing refinement relations or invariant preservation.[53] Key metrics evaluate the effectiveness of verification efforts, including state space coverage, which measures the proportion of reachable states or transitions analyzed to confirm exhaustiveness, and the incidence of false positives from abstraction techniques that may introduce spurious counterexamples due to over-approximation. Coverage is assessed by mutating models and checking if alterations affect property satisfaction, ensuring non-vacuous verification; false positives are mitigated by refining abstractions to balance precision and scalability. These metrics guide the thoroughness of proofs, with high coverage indicating robust assurance against uncovered errors.[56]Techniques
Specification languages
Formal specification languages provide a mathematical foundation for unambiguously describing the behavior, structure, and properties of systems in formal methods. These languages enable precise modeling by defining syntax and semantics that support rigorous analysis, refinement, and verification. They are essential for capturing requirements without implementation ambiguities, facilitating the transition from abstract specifications to concrete designs. Specification languages are broadly categorized into model-oriented and property-oriented approaches. Model-oriented languages focus on constructing explicit mathematical models of the system's state and operations, allowing for detailed simulation and refinement. In contrast, property-oriented languages emphasize axiomatic descriptions of desired behaviors and invariants, often using logical predicates to assert what the system must satisfy without prescribing how. This distinction influences their applicability: model-oriented suits constructive design, while property-oriented excels in abstract validation.[57]Model-Oriented Languages
Model-oriented specification languages represent systems through abstract data types, state spaces, and operation definitions, typically grounded in set theory and predicate logic. Their syntax includes declarations for types, variables, and schemas or functions that define state transitions. Semantics are often denotational, mapping specifications to mathematical structures, though some support operational interpretations for executability. These languages prioritize constructive descriptions, enabling stepwise refinement toward implementations. VDM (Vienna Development Method), originating in the 1970s at IBM's Vienna laboratory, exemplifies this category with its VDM-SL (Specification Language). VDM-SL uses a typed functional notation for defining state invariants and pre/postconditions, such as specifying a stack's operations with explicit preconditions like "the stack is not empty for pop." Its semantics combine denotational models for static aspects with operational traces for dynamic behavior, supporting proof obligations for refinement. Tool support includes the Overture IDE for editing, type-checking, and animation of VDM-SL specifications. Z notation, developed in the late 1970s at Oxford University by Jean-Raymond Abrial and formalized by Mike Spivey, employs schema calculus to encapsulate state and operations. Schemas, such as one for a file system defining known elements and current directory, combine declarations and predicates in a boxed notation for modularity. Z's semantics are model-theoretic, based on Zermelo-Fraenkel set theory with predicate calculus, providing a denotational interpretation of schemas as relations between states. Tools like Community Z Tools and ProofPower offer parsing, type-checking, and theorem proving integration for Z specifications. Alloy, introduced by Daniel Jackson in the early 2000s, extends model-oriented approaches with relational first-order logic for lightweight modeling. Its syntax declares signatures (sets) and fields (relations), as in modeling a file system withsig File { parent: one Dir }. Alloy's semantics are declarative, translated to SAT or Alloy Analyzer's bounded solver for automatic instance finding and counterexample generation. The Alloy Analyzer tool supports visualization, simulation, and checking of models up to configurable scopes, balancing expressiveness with decidable analysis via bounded scopes.
