Promise_theory

Promise theory

Promise theory

Method of analysis for systems of interacting components


Promise Theory is a method of analysis suitable for studying any system of interacting components. In the context of information science, Promise Theory offers a methodology for organising and understanding systems by modelling voluntary cooperation between individual actors or agents, which make public their 'intentions' to one another in the form of promises. Promise Theory is grounded in graph theory and set theory.[1]

Promise graph example
An example Promise Theory diagram illustrating partial ordering of agents by promise.

The goal of Promise Theory is to reveal the behavior of a whole by taking the viewpoint of the parts rather than the whole. In other words, it is a bottom-up, constructionist view of the world. Promise Theory is not a technology or design methodology. It doesn't advocate any position or design principle, except as a method of analysis.[2]

Promise Theory is being used in a variety of disciplines ranging from network (SDN)[3] and computer systems management[4] to organizations[5] and finance.[6]

History

An early form of Promise Theory was proposed by physicist and computer scientist Mark Burgess in 2004,[3] initially in the context of information science, in order to solve observed problems with the use of obligation-based logics in computer management schemes, in particular for policy-based management.[7]

A collaboration between Burgess and Dutch computer scientist Jan Bergstra refined the model of a promise, which included the notion of impositions and the role of trust. The cooperation resulted in several books and many scientific papers covering a range of different applications.[1][6][8][9][10][11][12][13]

In spite of wider applications of Promise Theory, it was originally proposed by Burgess as a way of modelling the computer management software CFEngine and its autonomous behaviour. CFEngine had been under development since 1993[4] and Burgess had found that existing theories based on obligations were unsuitable as "they amounted to wishful thinking".[14] Consequently, CFEngine uses a model of autonomy - as implied by promise theory - both as a way of avoiding distributed inconsistency in policy and as a security principle against external attack. As of January 2023 more than 2700 companies are using CFEngine worldwide.[15]

Outside the configuration management and DevOps disciplines, Promise Theory had a slow start. In the essay Promise You A Rose Garden (2007)[14] Burgess used a more popular, less academic style, but it failed to widen the general visibility of the concept at the time. A few years later, in 2012, things changed when Cisco began using Promise Theory in their growing SDN initiatives, also known as 'ACI', Application Centric Infrastructure.[16] The tech media picked up the usage in 2013[17][3][18][19] which led to a number of applications of Promise Theory in new disciplines in the years following, such as biology,[20] supply chain management,[21] design,[22] business/leadership[23][5] and systems architecture.[24] Tim O'Reilly discusses Promise Theory in his bestseller WTF: What's the Future.[25]

Key ideas

Promise Theory is described as a modeling tool or a method of analysis suitable for studying any system of interacting components. It is not a technology or design methodology and does not advocate any position or design principle, except as a method of analysis.[citation needed]

Agents

Agents in Promise Theory are said to be autonomous, meaning that they are causally independent of one another. This independence implies that they cannot be controlled from without, they originate their own behaviours entirely from within, yet they can rely on one another's services through the making of promises to signal cooperation. Agents are thus self-determined until such a time as they partially or completely give up their independence by promising to accept guidance from other agents.[citation needed]

Agents may be very simple, like a heading in a HTML document, or very complex, like a name server in a network e.g. a DNS server. Regardless of internal complexity, and agent embody mechanisms in order to keep promises. A HTML heading makes promises about its own rendering via CSS statements.[26] A DNS server promises to answer requests for addresses given a name.[14] The first extremely simple, the latter much more sophisticated. These differences in internal process complexity lead to a definition of so-called ‘semantic scaling' of agent complexity.

Intentions and outcomes

Agents in Promise Theory may have `intentions'. An intention may be realized by a behaviour or a target outcome. Intentions are thus made concrete by defining a set of acceptable outcomes associated with each intention. An outcome is most useful when it describes an invariant or mathematical fixed point in some description of states, because this can be both dynamically and semantically stable.

Each intention expresses a quantifiable outcome, which may be described as a state of an agent. Intentions are sometimes described as targets, goals, or desired states. The selection of intentions by an agent is left unexplained so as to avoid questions about free will.[1]

Agents express their intentions to one another by 'promise' or by 'imposition'. This provides a measure by which they can 'assess' whether intentions are fulfilled or not (i.e. whether promises are kept).[citation needed]

Promises

Promises arise when an agent shares one of its intentions with another agent voluntarily, e.g. by publishing its intent. The method of sharing is left to the modeller to explain.

For example, an object, such as a door handle, is an agent that promises to be suitable for opening a door, although it could be used for something else, e.g. for digging a hole in the ground. We cannot assume that agents will accept the promises given in the spirit in which they were intended, because every agent has its own context and capabilities. The promise of 'door handleness' could be expressed by virtue of its physical form or by having a written label attached in some language. An agent that uses this promise can assess whether the agent keeps its promise, or is 'fit for purpose'. Any agent can decide this for itself.

A promise may be used voluntarily by another agent in order to influence its usage of the other agent. Promises facilitate interaction, cooperation, and tend to maximize an intended outcome. Promises are not commands or deterministic controls.[citation needed]

Autonomy

Obligations, rather than promises have been the traditional way of modelling behaviour - in society, in technology and other areas.[27] While still dominant, the obligation based model has known weaknesses, in particular in areas like scalability and predictability, because of its rigidness and lack of dynamism.

Promise Theory's point of departure from obligation logics is the idea that all agents in a system should have autonomy of control—i.e. that they cannot be coerced or forced into a specific behaviour. Obligation theories in computer science often view an obligation as a deterministic command that causes its proposed outcome. In Promise Theory an agent may only make promises about its own behaviour. For autonomous agents it is meaningless to make promises about another's behaviour.

Although this assumption could be interpreted morally or ethically, in Promise Theory this is simply a pragmatic `engineering' principle, which leads to a more complete declaration of the intended roles of the actors or agents in a system: When making assumptions about others' behaviour is disallowed, one is forced to document every promise more completely in order to make predictions which in turn will reveal possible failure modes by which cooperative behaviour could fail.

Command and control systems like those that motivate obligation theories can easily be reproduced by having agents voluntarily promise to follow the instructions of another agent (this is also viewed as a more realistic model of behaviour). Since a promise can always be withdrawn, there is no contradiction between voluntary cooperation and command and control.

In Philosophy and Law a promise is often viewed as something that leads to an obligation. Promise Theory rejects that point of view. Bergstra and Burgess state that the concept of a promise is quite independent of that of obligation.[27]

Economics

Promises can be valuable to the promisee or even to the promiser. They might also lead to costs. There is thus an economic story to tell about promises. The economics of promises naturally motivate `selfish agent' behaviour and Promise Theory can be seen as a motivation for game theoretical decision making, in which multiple promises play the role of strategies in a game.[28]

Promise Theory has also been used to model and build new insights into monetary systems.[6]

Emergent behaviour

In computer science, the Promise theory describes policy governed services, in a framework of completely autonomous agents, which assist one another by voluntary cooperation alone. It is a framework for analyzing realistic models of modern networking, and as a formal model for swarm intelligence.[29]

Promise theory may be viewed as a logical and graph theoretical framework for understanding complex relationships in networks, where many constraints have to be met, which was developed at Oslo University College, by drawing on ideas from several different lines of research conducted there, including policy based management, graph theory, logic and configuration management. It uses a constructivist approach that builds conventional management structures from graphs of interacting, autonomous agents. Promises can be asserted either from an agent to itself or from one agent to another and each promise implies a constraint on the behavior of the promising agent. The atomicity of the promises makes them a tool for finding contradictions and inconsistencies.

Agency as a model of systems in space and time

The promises made by autonomous agents lead to a mutually approved graph structure, which in turn leads to spatial structures in which the agents represent point-like locations. This allows models of smart spaces, i.e. semantically labeled or even functional spaces, such as databases, knowledge maps, warehouses, hotels, etc., to be unified with other more conventional descriptions of space and time. The model of semantic spacetime uses promise theory to discuss these spacetime concepts.

Promises are more mathematically primitive than graph adjacencies, since a link requires the mutual consent of two autonomous agents, thus the concept of a connected space requires more work to build structure. This makes them mathematically interesting as a notion of space, and offers a useful way of modeling physical and virtual information systems.[30]

Promise Theory, Agile Transformation and Social Science

The Open Leadership Network and Open Space Technology organizers Daniel Mezick and Mark Sheffield invited Promise Theory originator Mark Burgess to keynote at the Open Leadership Network's Boston conference in 2019. This led to applying the formal development of Promise Theory to teach agile concepts. Burgess later extended the lecture notes into an online study course,[31] which he claims prompted an even deeper study of the concepts of social systems, including trust and authority.[32][33]


References

  1. Bergstra, Jan A.; Burgess, Mark (2019). Promise Theory: Principles and Applications (Second ed.). XtAxis Press.
  2. Burgess, Mark (2015). Thinking in Promises. O'Reilly.
  3. Burgess, Mark (Summer 1995). "Cfengine: a site configuration engine" (PDF). USENIX Computing Systems. 8 (3). Berkeley, CA, USA: USENIX. Retrieved 2014-08-22.
  4. Bergstra, Jan A.; Burgess, Mark (2019). Money, Ownership and Agency: As an Application of Promise Theory. XtAxis Press. ISBN 978-1696588379.
  5. Burgess, Mark (2005). "An Approach to Understanding Policy Based on Autonomy and Voluntary Cooperation". Ambient Networks. Lecture Notes in Computer Science. Vol. 3775. pp. 97–108. doi:10.1007/11568285_9. ISBN 978-3-540-29388-0.
  6. Promise Theory: Case Study on the 2016 Brexit Vote. ASIN 1974545334.
  7. Promises and Threats by Asymmetric Nuclear-Weapon States. ASIN 1673128211.
  8. Bergstra, J. A.; Burgess, M. (2019). "A Promise Theoretic Account of the Boeing 737 Max MCAS Algorithm Affair". arXiv:2001.01543 [cs.OH].
  9. Bergstra, J. A.; Burgess, M. (2020). "Candidate Software Process Flaws for the Boeing 737 Max MCAS Algorithm and Risks for a Proposed Upgrade". arXiv:2001.05690 [cs.OH].
  10. Sheffield, Mark; Mezick, Daniel (2018). Inviting Leadership. Freestanding Press. p. 82.
  11. O'Reilly, Tim (2017). WTF: What's the Future". p. 118.
  12. Burgess, Mark (October 2014). "Promise theory - what is it?". Linux Journal.
  13. A static theory of promises (Report). 2014. arXiv:0810.3294.
  14. "Voluntary Economic Cooperation in Policy Based Management" (PDF). Archived from the original (PDF) on 6 January 2007. Retrieved 14 January 2022.
  15. M. Burgess, S. Fagernes (2006), Promise theory - a model of autonomous objects for pervasive computing and swarms, Oslo University College, ISBN 0-7695-2622-5

Share this article:

This article uses material from the Wikipedia article Promise_theory, and is written by contributors. Text is available under a CC BY-SA 4.0 International License; additional terms may apply. Images, videos and audio are available under their respective licenses.