The Platform Beta Release 1.0 on the Anvil


Ingine; Inc™, The™, DiracIngine™, MARPLE™ are all Ingine Inc © and Trademark Protected; also The is Patent Pending IP belonging to Ingine; Inc™.

Screen Shot 2016-09-01 at 8.32.18 PM

High Performance Cloud based Cognitive Computing Platform

The below figure depicts the healthcare analytics challenge as the order of complexity is scaled.

1. Introduction Beta Release 1.0

It is our pleasure to introduce startup venture Ingine; Inc that brings to market The™Cognitive Computing Platform for the Healthcare market, delivering Medical Automated Reasoning Programming Language Environment (MARPLE) capability based on the mathematics borrowed from several disciplines and notably from late Prof Paul A M Dirac’s Quantum Mechanics.

The™; is a High Performance Cloud Computing Platformdelivering HealthCare Large-Data Analytics capability derived from an ensemble of bio-statistical computations. The automated bio-statistical reasoning is a combination of “deterministic” and “probabilistic” methods employed against both structured and unstructured large data sets leading into Cognitive Reasoning.

The™; delivers Medical Automated Reasoning based on a Medical Automated Programming Language Environment (MARPLE) capability, so better achieving 2nd order semantic interoperability1 in the Healthcare ecosystem. (Appendix Notes)

The™ is a result of several years of efforts with Dr. Barry Robson; former Chief Scientific Officer, IBM Global Healthcare, Pharmaceutical and Life Science. His research has been in developing quantum math driven exchange and inference language achieving semantic interoperability, while also enabling Clinical Decision Support System, that is inherently Evidence Based Medicine (EBM). The solution, besides enabling EBM, also delivers knowledge graphs for Public Health surveys including those sought by epidemiologists. Based on Dr Robson’s experience in the biopharmaceutical industry and pioneering efforts in bioinformatics, this has the data mining driven potential to advance pathways planning from clinical to pharmacogenomics.

The™; brings the machinery of Quantum Mechanics to Healthcare analytics; delivering a comprehensive data science experience that covers both Patient Health and Population Health (Epidemiology) analytics, driven by a range of bio-statistical methods from descriptive to inferential statistics, leading into evidence driven medical reasoning.

The™; transforms the large clinical data sets generated by interoperability architectures, such as in Health Information Exchange (HIE) into “semantic lake” representing the Health ecosystem that is more amenable to bio-statistical reasoning and knowledge representation. This capability delivers evidence-based knowledge needed for Clinical Decision Support System, better achieving Clinical Efficacy by helping to reduce medical errors.

The™; platform working against large clinical data sets or while residing within the large Patient Health Information Exchange (HIE) works in creating opportunity for Clinical Efficacy, while it also facilitates in the better achievement of “Efficiencies in the Healthcare Management” that Accountable Care Organization (ACO) seeks.

Our endeavors have resulted in the development of revolutionary Data Science to deliver Health Knowledge by Probabilistic Inference. The solution developed addresses critical areas in both scientific and technical, notably the healthcare interoperability challenges of delivering semantically relevant knowledge both at patient health (clinical) and public health level (Accountable Care Organization).

2. WhyThe™?

The basic premise in engineering The™ is in acknowledging the fact that in solving knowledge extraction from the large data sets (both structured and unstructured), one is confronted by very large data sets riddled by high-dimensionality and uncertainty.

Generally in solving insights from the large data sets the order in complexity is scaled as follows:-

A. Insights around :- “what” 

For large data sets, descriptive statistics are adequate to extract a “what” perspective. Descriptive statistics generally delivers statistical summary of the ecosystem and the probabilistic distribution.

B. Univariate Problem :- “what” 

Considering some simplicity in the variables relationships or is cumulative effects between the independent variables (causing) and the dependent variables (outcomes):-

a) Univariate regression (simple independent variables to dependent variables analysis)

b) Correlation Cluster – shows impact of set of variables or segment analysis.


[From above link:- In machine learningcorrelation clustering or cluster editing operates in a scenario where the relationships between the objects are known instead of the actual representations of the objects. For example, given a weighted graph G = (V,E), where the edge weight indicates whether two nodes are similar (positive edge weight) or different (negative edge weight), the task is to find a clustering that either maximizes agreements (sum of positive edge weights within a cluster plus the absolute value of the sum of negative edge weights between clusters) or minimizes disagreements (absolute value of the sum of negative edge weights within a cluster plus the sum of positive edge weights across clusters). Unlike other clustering algorithms this does not require choosing the number of clusters k in advance because the objective, to minimize the sum of weights of the cut edges, is independent of the number of clusters.]

C. Multivariate Analysis (Complexity increases) :- “what”

a) Multiple regression (considering multiple univariate to analyze the effect of the independent variables on the outcomes)

b) Multivariate regression – where multiple causes and multiple outcomes exists

All the above are still discussing the “what” aspect. When the complexity increases the notion of independent and dependent variables become non-deterministic, since it is difficult to establish given the interactions, potentially including cyclic paths of influence in a network of interactions, amongst the variables. A very simple example in just a simple case is that obesity causes diabetes, but the also converse is true, and we may also suspect that obesity causes type 2 diabetes cause obesity… In such situation what is best as “subject” and what is best as “object” becomes difficult to establish. Existing inference network methods typically assume that the world can be represented by a Directional Acyclic Graph, more like a tree, but the real world is more complex than that that: metabolism, neural pathways, road maps, subway maps, concept maps, are not unidirectional, and they are more interactive, with cyclic routes. Furthermore, discovering the “how” aspect becomes important in the diagnosis of the episodes and to establish correct pathways, while also extracting the severe cases (chronic cases which is a multivariate problem). Indeterminism also creates an ontology that can be probabilistic, not crisp.

Most ACO analytics addresses the above based on the PQRS clinical factors, which are all quantitative. Barely useful for advancing the ACO into solving performance driven or value driven outcomes most of which are qualitative.

D. Neural Net :- “what”

The above discussed challenges of analyzing multivariate pushes us into techniques such as Neural Net; which is the next level to Multivariate Regression Statistical Approach…. where multiple regression models are feeding into the next level of clusters, again an array of multiple regression models.

The Neural Net method still remains inadequate in exposing “how” probably the human mind is organized in discerning the health ecosystem for diagnostic purposes, for which “how”, “why”, “when” etc becomes imperative to arrive at accurate diagnosis and target outcomes efficiently. Its learning is “smudged out”. A little more precisely put: it is hard to interrogate a Neural Net because it is far from easy to see what are the weights mixed up in different pooled contributions, or where they come from.

“So we enter Probabilistic Computations which is as such Combinatorial Explosion Problem”.

E. Hyperbolic Dirac Net (Inverse or Dual Bayesian technique): – “how”, “why”, “when” in addition to “what”.

Note:- Beta Release 1.0 only addresses HDN transformation and inference query against the structured data sets and Features A, B and E. However, as a non-packaged solution C and D features can still be explored.

Release 2.0 will deliver full A.I driven reasoning capability MARPLE working against both structured and unstructured data sets. Furthermore, it will be designed to be customized for EBM driven “Point Of Care” and “Care Planning” productized user experience.

The™offers a comprehensive bio-statistical reasoning experience in the application of the data science as discussed above that blends descriptive and inferential statistical studies.

The™; is a High Performance Cloud Computing Platformdelivering HealthCare Large-Data Analytics capability derived from an ensemble of bio-statistical computations. The automated bio-statistical reasoning is a combination of “deterministic” and “probabilistic” methods employed against both structured and unstructured large data sets leading into Cognitive Reasoning.

Given the challenge of analyzing against the large data sets both structured (EHR data) and unstructured data; the emerging Healthcare analytics are around above discussed methods D and E; Ingine Inc is unique in the Hyperbolic Dirac Net proposition.

Q-UEL Toolkit for Medical Decision Making :- Science of Uncertainty and Probabilities

Screen Shot 2016-08-24 at 11.07.49 AM

Quantum Universal Exchange Language

Emergent | Interoperability | Knowledge Mining | Blockchain


  1. It is a toolkit / framework
  2. Is an Algorithmic Language for constructing Complex System
  3. Results into a Inferential Statistical mechanism suitable for a highly complex system – “Hyperbolic Dirac Net”
  4. Involves an approach that is based on the premise that a Highly Complex System driven by the human social structures continuously strives to achieve a higher order in the entropic journey by continuos discerning the knowledge hidden in the system that is in continuum.
  5. A System in Continuum seeking Higher and Higher Order is a Generative System.
  6. A Generative System; Brings System itself as a Method to achieve Transformation. Similar is the case for National Learning Health System.
  7. A Generative System; as such is based on Distributed Autonomous Agents / Organization; achieving Syndication driven by Self Regulation or Swarming behavior.
  8. Essentially Q-UEL as a toolkit / framework algorithmically addresses interoperability, knowledge mining and blockchain; while driving the Healthcare Eco-system into Generative Transformation achieving higher nd higher orders in the National Learning Health System.
  9. It has capabilities to facilitate medical workflow, continuity of care, medical knowledge extraction and representation from vast large sets of structured and unstructured data, automating bio-statistical reasoning leading into large data driven evidence based medicine, that further leads into clinical decision support system including knowledge management and Artificial Intelligence; and public health and epidemiological analysis.


A Large Chaotic System driven by Human Social Structures has two contending ways.

a. Natural Selection – Adaptive – Darwinian – Natural Selection – Survival Of Fittest – Dominance

b. Self Regulation – Generative – Innovation – Diversity – Cambrian Explosion – Unique Peculiarities – Co Existence – Emergent

Accountable Care Organization (ACO) driven by Affordability Care Act transforms the present Healthcare System that is adaptive (competitive) into generative (collaborative / co-ordinated) to achieve inclusive success and partake in the savings achieved. This is a generative systemic response contrasting the functional and competitive response of an adaptive system.

Natural selection seems to have resulted in functional transformation, where adaptive is the mode; does not account for diversity.

Self Regulation – seems like is a systemic outcome due to integrative influence (ecosystem), responding to the system constraints. Accounts for rich diversity.

The observer learns generatively from the system constraints for the type of reflexive response required (Refer – Generative Grammar – Immune System –

From the above observation, should the theory in self regulation seem more correct and that adheres to laws of nature, in which generative learning occurs. Then, the assertion is “method” is offered by the system itself. System’s ontology has an implicate knowledge of the processes required for transformation (David Bohm – Implicate Order)

For very large complex system,

System itself is the method – impetus is the “constraint”.

In the video below, the ability for the cells to creatively create the script is discussed which makes the case for self regulated and generative complex system in addition to complex adaptive system.


Further Notes on Q-UEL / HDN :-

  1. That brings Quantum Mechanics (QM) machinery to Medical Science.
  2. Is derived from Dirac Notation that helped in defining the framework for describing the QM. The resulting framework or language is Q-UEL and it delivers a mechanism for inferential statistics – “Hyperbolic Dirac Net”
  3. Created from System Dynamics and Systems Thinking Perspective.
  4. It is Systemic in approach; where System is itself the Method.
  5. Engages probabilistic ontology and semantics.
  6. Creates a mathematical framework to advance Inferential Statistics to study highly chaotic complex system.
  7. Is an algorithmic approach that creates Semantic Architecture of the problem or phenomena under study.
  8. The algorithmic approach is a blend of linguistics semantics, artificial intelligence and systems theory.
  9. The algorithm creates the Semantic Architecture defined by Probabilistic Ontology :- representing the Ecosystem Knowledge distribution based on Graph Theory

To make a decision in any domain, first of all the knowledge compendium of the domain or the system knowledge is imperative.

System Riddled with Complexity is generally a Multivariate System, as such creating much uncertainty

A highly complex system being non-deterministic, requires probabilistic approaches to discern, study and model the system.

General Characteristics of Complex System Methods

  • Descriptive statistics are employed to study “WHAT” aspects of the System
  • Inferential Statistics are applied to study “HOW”, “WHEN”, “WHY” and “WHERE” probing both spatial and temporal aspects.
  • In a highly complex system; the causality becomes indeterminable; meaning the correlation or relationships between the independent and dependent variables are not obviously established. Also, they seem to interchange the position. This creates dilemma between :- subject vs object, causes vs outcomes.
  • Approaching a highly complex system, since the priori and posterior are not definable; inferential techniques where hypothesis are fixed before the beginning the study of the system become enviable technique.

Review of Inferential Techniques as the Complexity is Scaled

Step 1:- Simple System (turbulence level:-1)

Frequentist :- simplest classical or traditional statistics; employed treating data random with a steady state hypothesis – system is considered not uncertain (simple system). In Frequentist notions of statistics, probability is dealt as classical measures based only on the idea of counting and proportion. This technique is applied to probability to data, where the data sets are rather small.

Increase complexity: Larger data sets, multivariate, hypothesis model is not established, large variety of variables; each can combine (conditional and joint) in many different ways to produce the effect.

Step 2:- Complex System (turbulence level:-2)

Bayesian :- hypothesis is considered probabilistic, while data is held at steady state. In Bayesian notions of statistics, probability is of the hypothesis for a given sets of data that is fixed. That is, hypothesis is random and data is fixed. The knowledge extracted contains the more subjectivist notions of uncertainty, belief, reliability, or confidence often used in automated inference and decision support systems.

Additionally the hypothesis can be explored only in an acyclic fashion creating Directed Acyclic Graphs (DAG)

Increase the throttle on the complexity: Very large data sets, both structured and unstructured,  Hypothesis random, multiple Hypothesis possible, Anomalies can exist, There are hidden conditions, need arises to discover the “probabilistic ontology” as they represent the system and the behavior within.

Step 3: Highly Chaotic Complex System (turbulence level:-3)

Certainly DAG is now inadequate, since we need to check probabilities as correlations and also causations of the variables, and if they conform to a hypothesis producing pattern, meaning some ontology is discovered which describes the peculiar intrinsic behavior among a specific combinations of the variables to represent a hypothesis condition. And, there are many such possibilities within the system, hence  very chaotic and complex system.

Now the System itself seems probabilistic; regardless of the hypothesis and the data. This demands Multi-Lateral Cognitive approach

Telandic …. “Point – equilibrium – steady state – periodic (oscillatory) – quasiperiodic – Chaotic – and telandic (goal seeking behavior) are examples of behavior here placed in order of increasing complexity”

A Highly Complex System, demands a Dragon Slayer – Hyperbolic Dirac Net (HDN) driven Statistics (BI-directional Bayesian) for extracting the Knowledge from a Chaotic Uncertain System.

Part B – Healthcare Interoperability, Standards and Data Science (Resolving the Problem)

Srinidhi Boray | Ingine, Inc |


Introducing, Ingine, Inc. it is a startup in its incipient stages of developing BioIngine platform, which brings advancement in data science around Interoperability. Particularly with healthcare data mining and analytics dealing with medical knowledge extraction. Below are some of the lessons learned discussed while dealing with the healthcare transformation concerns, especially with the ONC’s Interoperability vision.

As an introduction, want to include the following passage from the book

The Engines of Hippocrates: From the Dawn of Medicine to Medical and Pharmaceutical Informatics

By Barry Robson, O. K. Baek


Healthcare administration has often been viewed as one of the most conservative of institutions. This is not simply a matter of the inertia of any complex bureaucratic system. A serious body with an impressive history and profound responsibilities cannot risk unexpected disruptions to public service by changing with every fashionable new convenience, just for the sake of modernity. A strong motivation is needed to change a system on which lives depend and which, for all its faults, is still for the most part an improvement on anything that went before. However, this is also to be balanced against the obligation of healthcare, as an application of science and evolving human wisdom, to make appropriate use of the new findings and technologies available. This is doubly indicated when significant inefficiencies and accidents look as if they can be greatly relieved by upgrading the system. Sooner or later something has to give, and the pressure of many such accumulating factors can sometimes force a relatively entrenched system to change in a sudden way, just as geological pressures can precipitate an earthquake. An Executive Forum on Personalized Medicine organized by the American College of Surgeons in New York City in October 2002 similarly warned of the increasingly overwhelming accumulation of arguments demanding reform of the current healthcare system…if there is to be pain in making changes to an established system, then it makes sense to operate quickly, to incorporate all that needs to be incorporated and not spin out too much the phases of the transitions, and lay a basis for ultimately assimilating less painfully all that scientific vision can now foresee. But scientific vision is of course not known for its lack of imagination and courage, and is typically very far from conservative, still making an element of future shock inevitable in the healthcare industry.

  1. Complicated vs Complexity

A) Generally approaching to characterize a system, there are two views, complicated and complex. Complicated is with problems of system operations and population management, while complex problems are about multi-variability with an individual patient diagnosis.

Below link discusses providing better scenarios regarding complicated vs complexity

Generally, all management concerns around operations, payment models, healthcare ecosystem interactions, etc deal with delivering the systemic efficiencies. These are basically complicated problems residing in the system, which when resolved yield the hidden efficiencies.

All those that affect the delivery of the clinical efficacy have to deal with complex problem. Mostly owing to the high dimensionality (multi-variability) of the longitudinal patient data.

When both, complicated and complex concerns are addressed the Healthcare as an overarching complex system will begin to yield the desired performance driven outcomes.

B) Standards around Interoperability has generally dealt with following three levels of health information technology interoperability:


From the above link:-

1) Foundational; 2) Structural; and 3) Semantic.

1 – “Foundational” interoperability allows data exchange from one information technology system to be received by another and does not require the ability for the receiving information technology system to interpret the data.

2 – “Structural” interoperability is an intermediate level that defines the structure or format of data exchange (i.e., the message format standards) where there is uniform movement of healthcare data from one system to another such that the clinical or operational purpose and meaning of the data is preserved and unaltered. Structural interoperability defines the syntax of the data exchange. It ensures that data exchanges between information technology systems can be interpreted at the data field level.

3 – “Semantic” interoperability provides interoperability at the highest level, which is the ability of two or more systems or elements to exchange information and to use the information that has been exchanged. Semantic interoperability takes advantage of both the structuring of the data exchange and the codification of the data including vocabulary so that the receiving information technology systems can interpret the data. This level of interoperability supports the electronic exchange of patient summary information among caregivers and other authorized parties via potentially disparate electronic health record (EHR) systems and other systems to improve quality, safety, efficiency, and efficacy of healthcare delivery.

The above levels of interoperability only deal with achieving semantic compatibility between systems in the data transacted from the large number of myriad systems (EHRs) while they converge into a heterogeneous architecture (HIE / IoT). This only deals with the complicated concerns within the system. They do not necessarily deal with the extraction and discernment of the knowledge hidden in the complex health ecosystem system. To achieve this for some simplicity sake, let us define need for a second order semantic interoperability that concerns with the data mining approaches required in the representation of the systemic medical knowledge. It is this medical knowledge; implicit, explicit and tacit that all together form evidence based medicine much desired to facilitate any clinical decision support system.

C) In the present efforts around Interoperability, which centers mostly around data standards (HL7v2, HLv3, FHIR, C-CDA, ICD-10, LOINC, SNOMED etc) and clinical quality measures (QRDA); only complicated concerns have been addressed and not necessarily the complex problems. This is the vexation in the quality measures reporting. While this has advanced the adoption of EHR by the hospitals, it is still far from it becoming an effective decision support tool for the physicians

It must be noted that in the MU2 criteria, it is suggested that besides achieving health information exchange pivotal to the creation of Accountable Care Organization (ACO), at the least five-health priority or critical health risk conditions must be addressed employing clinical decision support system. Deservedly created, this point creates a need for addressing clinical efficacy, in addition to achieving best possible system efficiencies leading to systemic performance driven outcomes. This means a much deeper perspective is required to be included in the Interoperability efforts to better drive data science around data mining that can help better engage physicians in the realization of the performance driven outcomes. Rather than allowing physicians to be encumbered by the reimbursement model driven EHRs. Also, although most EHR vendors employ C-CDA to frame the longitudinal patient view, they do not necessarily send all the data to the Health Information Exchange, this results into truncating the full view of the longitudinal patient records to the physicians.

D) Physician, Primary Care, Cost in Rendering and Shortage Physician workforce

When dealing with the primary care, it is desired that today’s physicians who are over-burdened, moving forward works as a team lead engaging variety of healthcare professionals, while also better enabling trained nurse practitioners. Furthermore, also rendering the work in a lesser-cost environment while moving away from higher cost environments such as hospitals and emergency care facilities. This also means moving away from service-based models into performance based payment models becomes imperative.

It must be noted that dealing with the way an organization generally works reassigning responsibilities both horizontally and vertically, has to do only with the complicated concerns of the system, not the complex problem. Again it must be emphasized that data mining related to evidence based medicine, which is in a way knowledge culled from the experiences of the cohorts within the health ecosystem, will play a vital role in improving the much desired clinical efficacy leading ultimately to better health outcomes. This begins to address the complex systemic problems, while also better engaging the physicians who find the mere data entry into the EHR cumbersome and intrusive; and not able to derive any clinical decision support from the integration of the System of systems (SoS).

  1. Correlation vs Causations

A) While we make a case for better enabling evidence based medicine (EBM) driven by data mining as a high priority in the interoperability scheme of things, we also would like to point out the need for creating thorough systematic review aided by automation which is vital to EBM. This also means dealing with Receiver-Operating Characteristic (ROC) Curves

From the above link:-

“”The consensus of expert opinion based upon systematic reviews can either result in a solid and confident unanimous opinion, a reliable opinion with serious minority objections, a genuine controversy with no objective resolution, or simply the conclusion that we currently lack sufficient evidence and do not know the answer.””

Also, another reference to:-

Reflections on the Nature and Future of Systematic Review in Healthcare. By:- Dr. Barry Robson

In the recent times Bayesian statistics has emerged as a gold standard to developing curated EBM ( and; in this context we would like to draw attention that while correlation is important as discussed in the above linked article, which is developed from the consensus of the cohorts in the medical community, it is also important to ascertain the causation. This demands need for a holistic Bayesian statistics as proposed in the new algorithms, including those built on proven ideas in physics advancing the scope of the Bayesian Statistics as developed by Dr. Barry Robson. The approach and its impact on the Healthcare Interoperability and analytics are discussed in the link provided below.

From the above link: –


We extend Q-UEL, our universal exchange language for interoperability and

inference in healthcare and biomedicine, to the more traditional fields of public health surveys. These are the type associated with screening, epidemiological and cross-sectional studies, and cohort studies in some cases similar to clinical trials. “”There is the challenge that there is some degree of split between frequentist notions of probability as (a) classical measures based only on the idea of counting and proportion and on classical biostatistics as used in the above conservative disciplines, and (b) more subjectivist notions of uncertainty, belief, reliability, or confidence often used in automated inference and decision support systems. Samples in the above kind of public health survey are typically small compared with our earlier “Big Data” mining efforts. An issue addressed here is how much impact on decisions should sparse data have. “””””

B) Biostatistics, Algebra, Healthcare Analytics and Cognitive Computing

Another interesting aspect that emerges is the need for biostatistics and such many doctors with MD qualification are getting additionally qualified in Public Health Management, which also deals with Biostatistics. Dealing with population health one hand and clinical efficacy on the other, Interoperability via biostatistics has to deliver both views macro wrt systemic outcomes and at the micro level clinical efficacies. Developing such capabilities means much grander vision for Interoperability, as discussed in the OSEHRA, VA sponsored Open Source Efforts in making VistA available to the world market at a fraction cost. More discussion on the OSEHRA forum in the below link.

From the above link:-

“”””Tom Munnecke – The Original Architect of VistA – This move to a higher level of abstraction is a bit like thinking of things in terms of algebra, instead of arithmetic. Algebra gives us computational abilities far beyond what we can do with arithmetic. Yet, those who are entrenched in grinding through arithmetic problems have a disdain for the abstract facilities of algebra.””””

Interesting point to note in the discussions on the above link, is that a case is being made for the role of data science (previously called Knowledge Engineering during last three decades) driving better new algorithms, including those built on proven ideas in physics in the Healthcare Interoperability. This helps in advancing the next generations of the EHR capabilities, eventually emerging as a medical science driven cognitive computing platform. The recommendation is in the employ of advances in the data science in moving the needle from developing a deterministic or a predicated System of systems (SoS) based on schemas such as FHIM (, that proves design laborious and is outmoded, to harnessing the data locked in the heterogeneous system by the employ of advanced Bayesian statistics, new algorithms, including those built on proven ideas in physics and especially exploitation of the algebra. This approach delivered on a BigData architecture as a Cognitive Computing Platform with schema-less approaches has a huge benefit in terms of cost, business capability and time to market, delivering medical reasoning from the healthcare ecosystem as realized by the interoperability architectures.

Part A – Healthcare Interoperability Measures:- Cartesian Dilemma (Diagnosis)

Those in blue in the below content are reproduced from the referenced links.Slide06

Definition of Cartesian Dilemma; per Alexander Christopher

(what eyes sees and the mind sees are two different things)

Cartesian Dilemma

From above link

“””””Alexander has been inexorably led to the revolutionary necessity of revising our basic picture of the universe to include a conception of the personal nature of order and our belonging to the world in which the wholeness of space and the extent to which it is alive is perceived as rooted in the plenum behind the visible universe, “the luminous ground” that holds us all. This form of extended objective truth will ultimately resolve our Cartesian dilemma by teaching us a new view of order and a new cosmology in which objective reality “out there” and a personal reality “in here” are thoroughly connected and the bifurcation of nature healed.””””””

“”To Rene Descartes the “Method” (1638) was a convenient mental trick but its success has left us with a mindset that conceives of the universe as a machine without any intrinsic value: the realms of human experience and of feeling are simply absent from the Cartesian world. Whilst inspiring generations of architects and many others from all walks of life concerned with the fate of the earth, Alexander’s ultimately life changing work has understandably provoked powerful opposition from those invested within the establishment of the old paradigm. Social disorder, mental illness, ecological degradation, these and many other problems are due to a misunderstanding of the structure of matter and the nature of the universe and, until quite recently, there has been no coherent way of explaining the order that we respond to and love in nature.””


Affordability Care Act and HITECH Act lead into EHR Incentive Program. Based on the EHR Incentive Program CMS has already payed out 24+ Billions of dollars to Eligible Participants. Has it or will it drive the envisioned Healthcare Interoperability still remains a big question. Specifically will it be possible to mine the millions of records and discover opportunity for improvement? Without emphasis on clinical decision support will it be possible to achieve efficacy in the healthcare delivery, while also advancing the opportunities for “pay for performance” outcomes?

To advance EHR adoption in the Healthcare Ecosystem CMS proposed formation of Accountable Care Organization

From the above link

“”The Pioneer ACO Model is designed for health care organizations and providers that are already experienced in coordinating care for patients across care settings. It will allow these provider groups to move more rapidly from a shared savings payment model to a population-based payment model on a track consistent with, but separate from, the Medicare Shared Services Program. And it is designed to work in coordination with private payers by aligning provider incentives, which will improve quality and health outcomes for patients across the ACO, and achieve cost savings for Medicare, employers and patients.””

Importantly CMS proposed roadmap for EHR Adoption based on Meaningful Use (MU) 3 Stages, in the hope of advancing interoperability in the healthcare ecosystem ultimately achieving performance driven model, where the payment models shifts from “pay for service” towards “pay for performance”. Looking at the Healthcare ecosystem, one must take note that achieving efficiency is in the healthcare management; while achieving efficacy is in the healthcare delivery.

You will see in the end of the discussion that somehow efforts of the EHR Incentive Program lays more emphasis on the helathcare efficiency without paying required attention to clinical efficacy. This leads to the systemic entropic discontinuity that can be described by the Boltzmann constant.

This results into missed Line of Sight, where the established “objective”s at the IT / EHR level do not deliver all the required the “business capabilities” or the output and hence the desired “transformative outcomes” are not realized.

From the above link:-

“”In statistical mechanicsBoltzmann’s equation is a probability equation relating the entropy S of an ideal gas ( or consider healthcare ecosystem) to the quantity W, which is the number of microstates corresponding to a given macrostate.”””

Following are the EHR Adoption Meaningful Use Stages:-

MU Stage 1 :- Achieves electronic capture of the patient data (Data Capture and Sharing)

MU Stage 2 :- Achieves Health Information Exchanges (Advances co-ordinated clinical processes)

MU Stage 3:- Target Improved Outcomes ( achieved by moving the payment model from pay for service to pay for performance)

The eligible participants, physicians, hospitals and the ACOs have to demonstrate that they have met the MU criteria in stages. To demonstrate that they have met the requirements, first of all it is required to demonstrate that the data being captured adhere to a prescribed format. This is ascertained by MU attestation.

Additionally, the eligible participants are required to submit quality measures reports to CMS

From the above link

“””” Quality Measures and Performance Standards

Quality data reporting and collection support quality measurement, an important part of the Shared Savings Program. Before an ACO can share in any savings generated, it must demonstrate that it met the quality performance standard for that year. There are also interactions between ACO quality reporting and other CMS initiatives, particularly the Physician Quality Reporting System (PQRS) and meaningful use. The sections below provide resources related to the program’s 33 quality measures, which span four quality domains: Patient / Caregiver Experience, Care Coordination / Patient Safety, Preventive Health, and At-Risk Population. Of the 33 measures, 7 measures of patient / caregiver experience are collected via the CAHPS survey, 3 are calculated via claims, 1 is calculated from Medicare and Medicaid Electronic Health Record (EHR) Incentive Program data, and 22 are collected via the ACO Group Practice Reporting Option (GPRO) Web Interface.”””” Instruments/QualityMeasures/index.htm/lredirect=/QUALITYMEASURES/

National Quality Forum (NQF) endorsed for CMS reports are :

  • The Hospital Inpatient Quality Reporting (IQR) Program,
  • The Hospital Outpatient Quality Reporting (OQR) Program,
  • The Physician Quality Reporting System (PQRS), and
  • Others as directed by CMS, such as long-term care settings and ambulatory care settings

The CMS quality reporting is based on the schematic derived from HL7, termed QRDA

Click to access qrda_ep_hqr_guide_2015.pdf

From the above link

Overview of QRDA

“””The Health Level Seven International (HL7) QRDA is a standard document format for the exchange of electronic clinical quality measure (eCQM) data. QRDA reports contain data extracted from electronic health records (EHRs) and other information technology systems. QRDA reports are used for the exchange of eCQM data between systems for a variety of quality measurement and reporting initiatives, such as the Centers for Medicare & Medicaid Services (CMS) EHR Incentive Program: Meaningful Use Stage 2 (MU2).1

The Office of the National Coordinator for Health Information Technology (ONC) adopted QRDA as the standard to support both QRDA Category I (individual patient) and QRDA Category III (aggregate) data submission approaches for MU2 through final rulemaking in September 2012.2 CMS and ONC subsequently released an interim final rule in December 2012 that replaced the QRDA Category III standard adopted in the September 2012 final rule with an updated version of the standard.3 QRDA Category I and III implementation guides (IGs) are Draft Standards for Trial Use (DSTUs). DSTUs are issued at a point in the standards development life cycle when many, but not all, of the guiding requirements have been clarified. A DSTU is tested and then taken back through the HL7 ballot process to be formalized into an American National Standards Institute (ANSI)-accredited normative standard.

QRDA is a subset of CDA HL7 Standard; QRDA is a constraint on the HL7 Clinical Document Architecture (CDA), a document markup standard that specifies the structure and semantics of clinical documents for the purpose of exchange.4 To streamline implementations, QRDA makes use of CDA templates, which are business rules for representing clinical data consistently. Many QRDA templates are reused from the HL7 Consolidated CDA (C-CDA) standard5, which contains a library of commonly used templates that have been harmonized for MU2. Templates defined in the QRDA Category I and III IGs enable consistent representations of quality reporting data to streamline implementations and promote interoperability.”””

On the contrary we have Office Of National Coordinator (ONC) stipulate and regulate standards to achieve Healthcare Interoperability

ONC Roadmap Vision in the below link

From above link:-

Sadly, although Evidence based is discussed, data mining and concerns around algorithm development is missing.


Connecting Health and Care for the Nation: A Shared Nationwide Interoperability Roadmap version 1.0 (Roadmap) [PDF – 3.7 MB] supports the vision that ONC outlined in Connecting Health and Care for the Nation: A 10 Year Vision to Achieve An Interoperable Health IT Infrastructure [PDF – 607 KB]. The Roadmap, shaped by stakeholder input, lays out a clear path to catalyze the collaboration of stakeholders who are going to build and use the health IT infrastructure. The collaborative efforts of stakeholders is crucial to achieving the vision of a learning health system where individuals are at the center of their care; providers have a seamless ability to securely access and use health information from different sources; an individual’s health information is not limited to what is stored in electronic health records (EHRs), but includes information from many different sources and portrays a longitudinal picture of their health, not just episodes of care; and where public health agencies and researchers can rapidly learn, develop, and deliver cutting edge treatments.

“””””””” roadmap-update/

There is no doubt that ONC aspires to achieve true Healthcare Interoperability, by bringing more clarity to the Health Information Exchange (HIE) as discussed in the below link.

Interoperability vs Health Information Exchange: Setting the Record Straight

ONC under its purview has Office of Standards and Technology, which drives the Interoperability Standards; and it acknowledges that there are numerous challenges in realizing the ONC roadmap; as discussed in the below link

Interoperability Standards – Shades of Gray

Also ONC specifies roadmap in achieving MU stages for physicians, hospitals and ACOs ( HIE)

Specifically for the Semantic Interoperability it recommends Consolidated – Clinical Document Architecture ( C-CDA).

CDA helps in representing a comprehensive view of the patient; complete birth-to-death view – Longitudinal Record.

Also ONC Interoperability Specification Address the Following three levels (Not adequate to achieve EBM driven CDSS):-

There are three levels of health information technology interoperability:  1) Foundational; 2) Structural; and 3) Semantic.

1 – “Foundational” interoperability allows data exchange from one information technology system to be received by another and does not require the ability for the receiving information technology system to interpret the data.

2 – “Structural” interoperability is an intermediate level that defines the structure or format of data exchange (i.e., the message format standards) where there is uniform movement of healthcare data from one system to another such that the clinical or operational purpose and meaning of the data is preserved and unaltered. Structural interoperability defines the syntax of the data exchange. It ensures that data exchanges between information technology systems can be interpreted at the data field level.

3 – “Semantic” interoperability provides interoperability at the highest level, which is the ability of two or more systems or elements to exchange information and to use the information that has been exchanged. Semantic interoperability takes advantage of both the structuring of the data exchange and the codification of the data including vocabulary so that the receiving information technology systems can interpret the data. This level of interoperability supports the electronic exchange of patient summary information among caregivers and other authorized parties via potentially disparate electronic health record (EHR) systems and other systems to improve quality, safety, efficiency, and efficacy of healthcare delivery.

Desired or Recommended 2nd Order Semantic Interoperability

Probabilistic Ontology Driven Knowledge Engineering


Chronically ill patients are complex health care cases that require the coordinated interaction of multiple professionals. A correct intervention of these sort of patients entails the accurate analysis of the conditions of each concrete patient and the adaptation of evidence-based standard intervention plans to these conditions. There are some other clinical circumstances such as wrong diagnoses, unobserved comorbidities, missing information, unobserved related diseases or prevention, whose detection depends on the capacities of deduction of the professionals involved.

< diagnosis > < procedures > < outcomes > [triple store]


From the above points it must be noted that QRDA and C-CDA achieves different things. Unfortunately, against MU attestation and quality reports that are filed by the eligible participants (physicians, hospitals and ACOs) based on QRDA (especially PQRA), CMS runs the EHR incentives program. Whereas, in the MU2 stage ( as per ONC), it is also required by the participants to demonstrate that they have achieved interoperability within ACO, while implementing HIE, this requires C-CDA. This stage must demonstrate that coordinated clinical processes have been achieved.

Also, clinical decision support system (CDSS) has been established addressing at least 5 critical or clinical priority areas.  Unfortunately this particular capability does not seems to be addressed adequately by the ACOs; who only pursue to demonstrate quality measures have been achieved which necessarily does not mean clinical efficacy have been addressed. 

It seems an important architectural problem has been glossed over by the policy designers, who proposed quality measures model with the motivation for capturing the metrics that eventually demonstrate “pay for performance”; and somehow assumed that the proposed metrics based on QRDA also demonstrate that the clinical efficacies have been achieved. This leads into systemic entropic discontinuity, where the efforts at macro states that represents healthcare management leading into healthcare efficiency  is not necessarily a cumulative realization for the efforts at the micro states which represents gaining clinical efficacy. This entropic discountuinity between the macro state and the micro states is measured by Boltzmann Constant.

Link to more discussion on micro states and macro states within a complex system. Basically discusses for a given complex system, and for all the efforts towards the input; the entropy arrested and created loss, so the output is a actually created incurring loss. This means the systemic efficiency incurred losses and did not realize all the benefits arising out of the clinical efficacy. This is a model problem which inaccurately represents the “phenomenon of interest”.

To achieve Clinical Decision Support System capability which rather plays a very important role in enhancing clinical efficacy, developing data mining driven Evidence Based Medicine capability is imperative. This capability does not seem as being achieved because most HIE / ACO is being developed around QRDA; although discussed in the ONC Connecting Health and Care for the Nation: A Shared Nationwide Interoperability Roadmap version 1.0 (Roadmap) [PDF – 3.7 MB]; unless data mining related algorithmic challenges are addressed which means standards beyond mere capture of the required data fields, interoperability efforts will be in vain.

Role of EBM in achieving CDSS discussed on following sites

CMS Site

NIH Site

As such it must be noted clinical errors is one among the highest risk becoming the No 3 Killer in the US.

From above link

“””It’s a chilling reality – one often overlooked in annual mortality statistics: Preventable medical errors persist as the No. 3 killer in the U.S. – third only to heart disease and cancer – claiming the lives of some 400,000 people each year. At a Senate hearing Thursday, patient safety officials put their best ideas forward on how to solve the crisis, with IT often at the center of discussions. “””


Bioingine (; a Cognitive Computing Platform transforms the patient information (millions of records) created by the HIE into Ecosystem Knowledge Landscape that is inherently evidence based, allowing for study of the Tacit Knowledge, as discovered from the millions of patient records (large data sets) by mining and knowledge inference in an automated way. This is achieved employing AI, Machine Learning and such techniques. Thereby, creating Clinical Decision Support System.

Applying Quantum Theory for Deep Healthcare Analytics (Semantic Algebra) – Dr. Barry Robson (RQSA)

Quantum driven Cognitive Computing

RQSA Theory – Develops algorithm for The QEXL Approach; overcoming limitations in gold standard Bayesian Network; while allowing for creation of generative models. Bayesian as such is an adaptive technique.

(March 8 2013, Version March 10 2013)


1.1. Purpose and Background. The following document describes principle features of a mathematical system of practical importance in probabilistic ontology and semantics, and their applications inference and automated reasoning. Whilst applications are wide, healthcare may be the most pressing need [1].  The focus here is on aspects of practical importance in (a) decision support systems as Expert Systems (ES) derived from probabilistic rules formulated by experts, (b) decision support systems based on automated unsupervised data mining of structured data (DM), and (c) the Semantic Web (SW) and mining of unstructured data, here, essentially text analytics (TA).  The primary areas of interest to the author is in Evidence Based Medicine (EBM), Comparative Effectiveness Research (CER), epidemiology, and Clinical Decision support Systems (CDSS), and bioinformatics, which provide useful reference points and test beds for more general application.

1.2. The Need for a New Approach.  In general we are beset by a plethora of approaches for reasoning probabilistically when many probabilistic rules are available, and the same uncertainty about best practice poses problems for the SW to go probabilistic [2].  The easiest relations between things to handle probabilistically, i.e. to extract, quantify, and reason with, are “are associated with”, as in “A is associated with B and C and D” [2-6], and it has found many biomedical applications [7-14]. From many such and self probabilities or information, we can express “A if B and C and D” are readily derived. These are of course not the only relationships between things that we use when communicating information by natural language, but the latter are noteworthy in forming the basis of the Bayesian network or Bayes Net (BN) [15], which is a gold standard for probabilistic inference. It is theoretically well founded. However, the reason that we see a plethora of further approaches is that BN’s adherence to a strict set of axioms to avoid apparent difficulties makes it very restricted in application. A BN (one that is truly a BN by original definition confines) confines itself to use of conditional probabilities that reflect the above “if” relationship, to multiplicative operations implying just logical AND, it considers only one direction of conditionality when in general we are interested in inference about etiology or causes as well as outcomes, and not least it confines itself to acyclic directed graphs when networks of knowledge representation are in practice rich in cyclic paths. That last is because the ideal full treatment is a fully connected graph, not neglecting any relationship, which is necessarily dominated by cycles.  Note that neglect of relationships by a BN is equivalent to saying that they are they are there with probability 1, and whilst to be fair to BNs it is true that this implies absence of information I = −log(P), it remains an extreme assumption that  is evidently not justified when we consider the other conditional probability terms and in the context of all available data, often called the problem of coherence.

1.3. Advantages of the Present Approach. The advantages of the present approach can be referenced with respect to BN as the gold standard that does not  traditionally provide the following.

(1) Bidirectional inference, i.e. etiologies as well as outcomes.

(2) Intrinsic treatment of coherence as Bayes Rule.

(3) Cyclic paths are allowed, fall out naturally as part of the theory, and do not require iteration.

(4) Not confined to AND logic

(5) Not confined to conditional probabilities. Relators and operators may be used symbolically albeit with probabilities, or as matrices or algorithms.

(6) Probability distributions represented by vectors.

(7) Metarules with binding variables such to generate new rules and evolve the old, evolving the old. Metarules are also used to define words from simpler vocabularies.

(8) Handling of negation and, when there are double negation etc., conversion to canonical forms.

(9) Reconciliation into one rule of rules that overlap in information content or are semantically equivalent, including reconciliation of their probabilities.

1.4. The Pursuit of Best Practice for the Theoretical Basis. Quantum mechanics (QM) claims to be a system of best practice for representing observations and for inference from them on all scales, although the notorious predictions by QM in the narrow band of scale of everyday human experience  has  discouraged investigation of applicability. This is due to the perception of QM as wave mechanics, even though we are not for everyday purposes usually interested in inference about waves. For that reason, the method is based more specifically on the larger QM of Paul A. M. Dirac [16-17] who established the theoretical basis of particle physics.  Penrose [18] provides an excellent primer.  That there is sufficient breadth to Dirac’s perception of QM to encompass semantics is indicated by Dirac’s Nobel Prize Banquet Speech in 1933, “The methods of theoretical physics should be applicable to all those branches of thought in which the essential features are expressible with numbers”. Dirac was  certainly modestly referring to his extensions to physics as further extensible to  human language and thought, because  in interviews he explicitly excluded poetry and  (more controversially) economics as subjective.

The author has published extensively in areas of some relevance, but the bibliography [19-31] refers to publications since 2005 which have some relevance to an idea first broached in Ref. [31]. The present report collates the observations that have survived as useful, adjusts some nomenclature as well as perceptions, and adds integrating material.  “Best Practice” would be presumptive for these or indeed anyone’s publications  (though “pursuit of Best Practice” validly reflects the intent). Indeed, these publications show various degrees of development from some naïve initial observations, and represent a learning curve, and so are presented in reverse chronological order. This is because whilst the author has some formal training as doctorates in, essentially, biophysics and theoretical and computational chemistry, the quantum mechanics of molecules then touched (and still does) only upon aspects related to pre-Dirac quantum mechanics. That pre-Dirac era is essentially that of Schrödinger that  describes quantum mechanics as wave mechanics. It is based on the imaginary number i, the number such that ii = −1. The problematic wave nature arises because exponentials of i-complex values are periodic or wave functions: eiq =  cos(q) + i sin(q). However, Dirac rediscovered another imaginary number, although it was first noted by Cockle in 1848 and relates to a broader Clifford calculus originating roughly around that time.  It has very different consequences, as follows.



2.1. The Hyperbolic Imaginary Number. The “new” imaginary number is here represented as h such that hh = +1.  Dirac developed what is now known as the Clifford-Dirac algebra. It, and h, arose in consideration of the origins of mass, introduced general relativity into quantum mechanical systems, and founded the current “standard model” of particle physics. Actually, the Clifford Dirac algebra and particle physics has several distinct imaginary numbers (that can all be represented as matrices). However they fall into the two general classes, of character i in that their squares are −1, and of character h in that their squares are +1. Otherwise, two different flavors of imaginary number  are anticommutative, meaning that if the order is changed in which they form a product, the sign of the product changes. This includes hi = −ih.  In the core theory presented here, different flavors of imaginary number do not meet up in the equations, and the focus is primarily upon h in isolation, meaning that focus is on purely hyperbolic complex or h-complex algebra (although real numbers are of course present, and that algebra sometimes delivers purely real-valued results, because hh = +1, and h−h = 0). In physics, those of h character include Dirac’s linear operator s, g0 (or gtime), and g5. Significantly for what follows,  they particularly appear  in physics as the equivalent to ½ (1+h) and ½ (1−h) multiplied by real or usually i-complex expressions, which are called spinors. As dual spinors in which these two spinors are involved, they relate to key expressions in quantum field theory, as well as the in theory discussed here.

2.2. The h-complex Hyperbolic Function. Unlike in i-complex algebra where eiq =  cos(q) + isin(q), we are not usually concerned with making inference about waves, which may be important in physics and chemistry but not much in everyday life. As it happens, h is frequently called the hyperbolic number, and that is so because ehq  is a hyperbolic, not a trigonometric and so periodic function (see next the Section after next for an interesting example and its physical consequences). We can express the h-complex hyperbolic function in the author’s notation (which will be valuable later below).

ehq  =  i*eq + ie+q   =  cosh(q) + h sinh(q)                                                                                (1)

where the notation means

i =  ½ (1+h)

i*   =  ½ (1−h)                                                                                                                                   (2)

For completeness and to demonstrate consistency with quantum mechanics, as well as for a few practical applications for present purposes (use of waves and wavelets), note that a somewhat more general description of quantum mechanics is given by

e+hiq  =  ehiq   =   i*eiq + ie+iq  =   cos(q) + hi sin(q)                                                                (3)

The full treatment following Dirac would be to resolve i into three kinds, one for each dimension of space, implying a Clifford-Dirac product of four imaginary numbers that leads to i =  ½ (1+g5) and i* =  ½ (1-g5), to same effect as g5 is a flavor of h:  i.e. g5 g5  = +1. Although algebraically it does not of course matter, the author writes the i* term as the “lead term” i.e. as the focus of attention, for several reasons, one of which will become apparent later.

2.3. Complex Conjugation. The asterisk * used above and in other contexts below means forming the complex conjugate, i.e. changing the sign of the imaginary part, equivalent to replacing all +h by –h and vice versa. It is applied much more generally in the theory than just to rationalize i* and i*.  Usually defined for i, we need to assume that it is extensible to h. However, in cases less relevant here where we need to think in terms of both imaginary numbers, note that we need to think of applying complex conjugation as  (hi)* = ih = –hi, not h*i* = hi.

2.4. Physical and Statistical Interpretations.  As indicated above, the relation to classical as opposed to wave behavior arises because ehq  can be expressed in terms of hyperbolic functions. That means it can be expressed in terms of Gaussian functions (“normal distributions”) and their reciprocals by choice of variables in q. An example from physics is the choice of distance x = xt – x0 from starting point x0 in quantum mechanical expressions where a particle moves with momentum p=mx/t, giving q = -2pmx2/ht, with mass m, time t and Planck’s constant. Incidentally, Planck’s constant h is not of course be confused with the hyperbolic number h, and so when appearing elsewhere in discussions of physics the latter is usually written by the author in italic bold, as h, to make clear that distinction. The normalization procedure (Dirac recipe) is described later below, but for present purposes it suffices that the exponential of -2pmx2/ht is proportional to the probability of the absolute value |x|, and more generally, and as a practical application for the present context, it represents a normal distribution q = -½x2/s2 where starting point x0 becomes a mean value and √(ht/4pm) becomes the standard deviation s. The use of h does to some extent dictate a model of physical observation. To show consistency with quantum mechanics and interestingly consistency with the collapse of the wave function (e.g. Penrose reduction) interpretation, the above essentially suggests a kind of diffusion model, in which in isolation an i-complex description as a wave is the lowest energy state, but an observation of the particle as a particle and not a wave, or analogous physical perturbations, swings the description to a now lower energy h-complex one by “forcing the particle description”. This establishes a new x0 around which the particle collapses, although there is a accuracy of measurement conveyed by a standard deviation s  in general, and at very least, for the most accurate measurements that are in principle possible, we have s = √(ht/4pm). Following the perturbation, the i-complex description is restored as the lowest energy state[1]. This relates to the physicists’ spread of the wave function, but it would be billions of years to see the phenomena for an object on the scale of a household object (by which time other entropic considerations would have had their effect). It is interesting to note that the collapse of the wave function seen this way does not necessarily imply a discontinuous ih jump (or hii jump in the broader description of Eqn. 3) but a progressive rotation in time t or to an extent governed by the energy of the perturbation that may be interpretable as a field, of which demanding to measure it as a particle rather than wave is just a particularly strong case. The practical application here is that one can consider wave packets or wavelets that are progressively localized wave descriptions, and a Gaussian in the limit of being maximally localized. The applications outside of physics appear, however, to be in specialist areas such as probabilistic treatment of image analysis, and distributions generally are better described in terms of h-complex vectors, described later.

In consequence, we are considering the primary applications below as relating to the case when the parameters in q, and notably time, are fixed so that the exponentials merely relate to the single probability value such as P(A) of a state, event, observation, measurement or description A. P(A) that can empirically replace the concept of the exponential (as the statistical weight) and any normalizing factor for it.   That said, the exponential form will make appearance in which the physicists’ q  identifies with Fano’s  mutual information  I(A; B) between A and B, as described below, though we will also think of eI(A;B) as association constant K(A; B). The logarithm of the wave function y proportional to eq is seen as information that is somehow encoded. Note that eiI for any information I is a periodic function that bounds the different information and resultant probability values that one may have to the interval 0..2p. In contrast, ehI does not. h may be interpreted as adding the capacity for additional information that localizes the wave function as a Gaussian due to observations made, with a precision due to variance(i.e. square of standard deviation s2)  as a counterpart of the physicists’ ht/4pm, which relates to the physicists’ notion of  q as the action written in units of Plank’s constant h. For consistency with the physicist’s interpretation as wave function collapse due to loss and movement of information from the system, the treatment below should therefore be seen, reasonably enough,  as the gain of information to the observer. In practice the information comes from sampling and counting of an everyday system as a population, or from our belief in the result that we would obtain by doing so.

2.5. Eigenvalues of h. Unlike the case of i that has imaginary eigenvalues +i and –i, we can also (as Dirac noted) replace h by its eigenvalues that are real, either +1, or −1. This is equivalent to treating i  and  i* as linear operators with eigenvalues 0 and 1, but specifically meaning that we can set i =  1 and i*  =  0  to get one solution, and i =  0 and i* =  1  to get the other, giving two plausible physically real eigensolutions, or two sets of i-complex ones when expressions are also i-complex. Basic quantum mechanical texts gloss over this, jumping straight to eiq  and e+iq as the intuitive solutions and solving each separately,  but Dirac made it clear by stating that a wave function is always decomposable into two parts, one a product with  i and one a product with  i* (although he didn’t use that notation). In physics, they typically relate to solutions in matters of direction in time or chirality (handedness) and more generally to direction in conditionality. That is meant in the same sense that conditional probability P(A|B) = PA, B)/P(B) is of  reverse conditionality to P(B|A) = P(A, B)/P(A). In other words, the two eigensolutions do not imply indeterminacy in the sense that multiple eigenvalues would be possible interpretations.  Rather, they simply relate to two directions of inference in the network and two directions corresponding of effect of the terms in it. However, we cannot compute P(A|B) from P(B|A) or vice versa by taking the adjoint  † as the transpose and/or complex conjugate of either one of them because a classical probability is a scalar value, and has no imaginary part. In other words, P(A|B) is purely a symbolic adjoint of P(B|A). However, the effect of our i and i* operators is to render real values h-complex, and it should be held in mind that since (i)* = i* and (i*)* = i*, then (i*P(A|B) + iP(B|A))* = (i*P(B|A) + iP(A|B))*.  Then given the (0,1) eigenvalues of these operators, we have P(A|B) and P(B|A) as the two eigensolutions.

2.6. Iota Algebra. Alternatively to thinking in terms of of Eqn. 3, one can for present purposes think of pre-Dirac quantum mechanics in which h replaces i, which in physics is called the Lorentz rotation, and it is arguably a generalization of the Wick rotation in which time t is replaced by imaginary time it to render quantum mechanical expressions classical. The resulting purely h-complex algebra takes some practice and excessive familiarity with i-complex algebra can sometime be more a hindrance than a help, because one can jump to conclusions that do not hold when h replaces i, and conversely miss important algebraic opportunities that h provides. Fortunately, h-complex algebra can be rendered in a form making manipulation much easier than in i-complex algebra. The above spinor forms, more generally quantities of form i*x and iywhere  x and y are not h-complex,  are so-called by analogy with Dirac’s treatment, and can be considered as comprising spinor operators i*and i with a very convenient algebra of their own. We can usually avoid discussion any of h by using them, i.e. by using i algebra or iota algebra alone. Its simple properties include the idempotent property ii = i and i*i* = i* which for example means that ei = i  and  ei* = i*, and similarly all powers and logarithms leave i or i* changed, but not of course the non-h-complex terms they multiply.  They also include the annihilation property  ii* =  i*i = 0 that by annihilating terms in greatly facilitates multiplying h-complex expressions, and the normalization property  i+i*  =  i*+i  = 1, with the important effect that if a dual spinor form ix  +i*y where x and y are real values, then ix +i*y = x + y.  Note also that (ix +i*y)* = (i*y +i*x), important as the general statement that was implied in stating (i*P(A|B) + iP(B|A))* = (i*P(B|A) + iP(A|B))*.   Eqns. 1 and 3 cover trigonometry and hyperbolic functions. That covers almost all the new algebra that we need here, but for completeness, because the Riemann zeta function z(s=1, n)  = 1 + 2-s + 3-s +…+ n-s,  is used in data mining and treatment of finite data to estimate information and probability values, it should be noted that z(x+hy, n) = i* z(x−y, n)  + i z(x+y, n).

2.7. Dirac Notation. Let  q be a function of A and B where are A and B are observed states or events that can take on particular values, being prepared that way, or by observation. More precisely, using the author’s notation, we should write A:=a and B:=b for these states or events as measurements. Here A and B are metadata or data descriptors such as momentum in physics or systolic blood pressure in medical life, and the specific values they have are a and b respectively, constituting the orthodata or specific manifestation of A and B. However, we can take A and B etc. as implying A:=a, B:=b etc. for brevity. We have, for example, in Dirac’s bracket or braket notation,  where we can speak of a bra part <A| and a ket part |B>,

<A|B>  =  ke+hiq  =  kehiq   =   i*keiq + ike+iq  =   k(cos(q) + hi sin(q))                                  (4)

However, we shall assume that <A|B> is purely hyperbolic, i.e. there is a Lorentz rotation which here implies i → 1, and although this is not usually the case in quantum mechanical texts, a purely hyperbolic <A|B>  is a valid solution. Above, k is a real valued constant dependent on the nature and scale of the system, and importantly it must be such that k relates to P(A) and P(B) meaning P(A:=a) and P(B:=b) so that the probability  of a measurement value is purely one of chance without a prior observation of one of A and B, or in quantum mechanical language, without preparing a value of A or B.  For example, in quantum mechanics texts considering a particle on a circular orbit of length L, we see k=1/L2.


2.8. Dirac Recipe for Observable Probabilities. The fact that we set a set a prior value of A and B, and then measure B or A as conditional upon it, means that that we think in terms of  calculating  P(A|B) and P(B|A), in quantum mechanics a process that algebraically implies first a ket |B> or bra |A> normalization as the preparation of B and A.  We can write the bra normalized <A|B> as `<A|B> and the ket normalized<A|B>, whatever that might mean algebraically at this stage. In fact, in the earlier example q = 2pmx2/ht, we are conceptually obliged to apply ket normalization since q = -2pmx2t/ht becomes q = +2pmx2/ht in the term ike+iq  which can exceed an upper valid probability of 1 if  t  and m are, as in our everyday world,  positive.  Following Dirac’s recipe for observable probabilities we apply after ket normalization  P(A|B) =  <A|B> (<A|B>)*, sometimes written as the square of the absolute magnitude |<A|B>|2 according to the Born rule but implying that <A|B> is ket normalized according to the Dirac recipe. In fact, the interpretation as  |<A|B>|2  is peculiar to i-complex algebra and is not presumed here, and the more fundamental interpretation is that observation implies a projection operator P = |A><B|  acting on vectors <A| or |B> (a vector interpretation is preserved in the current theory – see later below) such that

<A| P |B> = <A|  |B><A| | B> = <A|B><B|A> = <A|B><A|B>*                                          (5)

We cannot yet express <A|B> as a function of P(A|B) and P(B|A), but the solution must satisfy

<A|B> = i *<A|B> +  i<A|B>  = i *<A|B> +  i<A|B>  = i *<A|B> +  i (<A|B>)*

= <A|B> <A|B>                                                                  (6)

P(A|B)  =<A|B> (<A|B>)*                                                                                                   (7)

P(B|A)  = <A|B> (<A|B>)*  = (<A|B>*) ((<A|B>*))*                                                      (8)

We can see the requirements for <A|B> emerging from the above by inspection, but when we apply it to Eqns. 1,3, and 4, we find that P(A|B) = P(B|A), which is the special case P(A) = P(B).

2.9. Conjugate and Non-Conjugate Variables. To move towards the required theory, we note that  Eqns. 1,3, and 4 have a conjugate symmetry not suitable for our general and more classical purposes. That is, they are composed as i*x + iy.  such that xy = 1 for all values of x and y. It arises because A:=a is a simple  function f of B:=b and vice versa, so that the probabilities are predetermined as P(A:=a) = P(f(B:=b)) = P(f(A:=a)) = P(B). That is not generally true in quantum mechanics either, but relates to the important special case of conjugate variables such as momentum and position, or energy and time, and generally where the action A = (A:=a)(B:=b) such that  q = 2pA/h. Classically, we can also be measuring values which are also such conjugate variables, like pressure P and volume V in the gas law PV = RT where R is a constant and when the absolute temperature T is constant, or current I and resistance R in the electrical engineering equation V= IR for constant voltage V. However, such cases in inference are rare, and in practice the extent to which P(A:=a) ≈ P(f(B:=b)) is more interesting as deducible from inference than as input, and moreover as part of a more general description of a relationship between A and B, i.eas the association constant

K(A; B) = P(A, B) / P(A)P(B) =  eI(A; B)                                                                                (9)

where I(A; B) is Fano’s mutual information, between A and B, noting K(A; B) = K(B; A) and I(A; B) = I(B; A)..  We will generally only require for our applications that measurements and observations are such that 0 ≤ xy ≤ 1, and more specifically that 0 ≤ x ≤ 1 and 0 ≤ y ≤ 1, because we will relate them directly to empirical probabilities from data mining or human assignment.  We can think of conjugate symmetry xy < 1 as the mother or prototype form, and we might say in physics that the dual spinor is a system in an asymmetric field that breaks the conjugate symmetry. Whilst this is most generally is of no relevance here, there is arguably one exception:  an observation or measurement that implies the above described  normalization as part of the Dirac recipe is an interaction analogous to an asymmetric field,  and  it sets one of x or y to 1.

2.10. The Braket as a Simple Linear Function of Empirical Probabilities.  To all the above consideration there is only one and simple interpretation:

<A|B>  = i*P(A|B) +  i P(B|A) = [i *P(A) +  iP(B)] K(A; B) = [i *P(A) +  iP(B)] eI(A; B)

½ h[P(A|B) + P(B|A) + ½ h [P(B|A) – P(A|B)]                                     (10)

which can be shown to satisfy Eqns. 6-9.  Several things may be noted here and in Sections immediately following. First, we can see form the last line, analogous to the quantum mechanical Hermitian commutator form, that it is not entirely true that it is the only solution; we can replace +h by –h and satisfy most of the above discussion, with one exception, that in the case of conjugate variables it was argued that we must normalize the ket, not the bra, and that can be shown to imply Eqn. 10 as written. Second, specifically in the h-complex algebra, following the Dirac recipe does not suggest that we consider the square roots of probabilities in such expressions. In particular, our bra and ket normalizations become

<A|B>  = i*P(A|B) +  i                                                                              (11)

<B|A>  = i* +  i P(B|A)                                                                              (12)

Multiplying these by their own complex conjugates delivers P(A|B) and P(B|A) respectively. Computationally, is equivalent to using the following where Re and Im are the real and imaginary parts.

P(A|B) = Re<A|B> – Im<A|B>                                                                                              (13)

P(B|A) = Re<A|B> + Im<A|B>                                                                                             (14)

2.11. Classical Probabilistic Behavior. Above it was noted that h-complex functions exhibits classical local distribution functions.. The further point about Eqn. 10 is that it also yields classical probabilistic behavior. Notably for the chain rule P(A|C) ≈  P(A|B)P(B|C) and so P(C|A) ≈ P(C|A)  that assume A and C independence, which is often physically the case, we obtain

<A|C> ≈ <A|B><B|C> = i*P(A|B)P(B|C) +  iP(C|B)P(B|A)

= [i*P(A) +  iP(B)] K(A; B) K(B; C)                                          (15)

and similarly for <A|B><B|C><C|D> and so on.

2.12. Observation Brakets. There is an important special case of Eqns. 11 and 12 when A and B are statically independent, i.e. K(A; B) = 1, and when we are at least 100% sure that we are performing a preparation or measurement, which is here represented by ? where P(?) = 1. It gives us two terms which are the counterparts of prior self-probabilities in an inference network.

<B|?>  = i * P(B|?)  +  i   =   i* P(B) +  i                                                         (16)

<?|A>  = i* +  i P(A|?)  = i* +  i P(B)                                                             (17)

Note that the following t can be readily shown with a little standard algebra and recalling i *+ i = 1 so  that  i*x + ix = x ,

<?|A><A|B><B|C><C|D<D|E><E|?>   = P(A, B)P(B, C)P(C,D)P(D|E) / P(B)P(C)P(D)   (18)

which is an estimate of  <A|E> = P(A|E), a joint probability that estimates the joint probability P(A, B, C, D, E).

2.13. Dirac Nets and Coherence.  A network built from brakets (or bra-operator-kets – see later below) may be dubbed a Dirac Net. It is common as in Bayes Nets to consider a joint probability, although we can always make it conditional on say X or (X,Y) by dividing by P(X) and P(X, Y), which makes most sense of course when X and Y are states and events in the network. We can make a joint probability first by providing a full provision of observational brakets as all terminal nodes of the network, with a caveat as shortly to be mentioned.  An advantage of  Dirac Nets is that they can be used to ensure that the network is coherent with itself and all available data , meaning that all P(A|B)P(B) = P(B|A)P(A) and so on, which is Bayes Theorem. Ironically, Bayes Nets as usually defined as acyclic directed graphs do not consider this, since they only see one direction, P(A|B)P(B). If there is not such balance overall, or if it is conditional, and in a   non-trivial way (e.g. missing probabilities are not 1), a Dirac Net value will have an imaginary part, positive or negative. In conversion of a Bayes Net to a Dirac Net, we will encounter branch points such as for example P(A| B, C) P(B|D)P(C|E). They allow that B and C are not independent in P(A| B, C)  in one direction, but that they are independent in  P(B|D)P(C|E)  in the other direction. This will typically show up as a complex and not purely real value for a network. Consideration show that we need to correct this (arguable) mistreatment by a Bayes Net by multiplying by

<? ; B, C> =  i* + iK(B; C) = <? | B, C> / <? | B > <? | C>                                                             (19)

Note that the construction <A|B><B|C>  <B|D>  etc., where we have a branch with two As, or indeed  <C|A>  <A|D> as a branch in the other direction, is valid. We need to correct accordingly in the first case by i* + i K(A; A) and in the second case by i*K(A; A) + i, where K(A; A) = P(A, A) / P(A)P(A). That is, by 1/P(A) if they are absolutely indistinguishable so that P(A, A) = P(A). But that is not necessarily the case if A is Bernoulli countable state, when P(A, A) = P(A)P(A).in that case no such correction is required. The general case of degrees of distinguishability  is considered immediately after the following Section.

2.14. Cyclic Paths.  Oddly, traditional Bayes Nets deny the possibility of cyclic paths, yet in a Dirac Net using observational brakets creates cyclic paths in order to get the answer we require.  A full and proper use of observational brakets giving a purely real value indicates a joint probability, not a conditional one, but it applies whatever additional states, say F  replace ?. Indeed it applies in replacement by two different states, say F and G, providing P(F) = P(G). It is an important feature of cyclic paths in an inference network formed by these methods that they are purely real, and appear to pose no special problems when a system is described where probabilities are in steady state, or sampled on a timescale much shorter than that over which it evolves. Such considerations show that <A|B><A|B>* = <A|B><B|A>* that yields observable probabilities  is a simple case of such as cyclic path.

2.15. Distinguishability of States and Events. Actually, the simplest cyclic path is <A|A>. In quantum mechanics , texts often state that <A|A> = 1, but that is not in general true, and is just one possibility for the distinguishability of A and B on a continuum for <A|B> as a real line defined by P(A|B) = P(B|A), i.e.  P(A) = P(B).  <A|B> = 1 is the case when A and B are absolutely indistinguishable, so that P(A) = P(B) = 1, because P(A, A) =A. If they are distinguishable by recurrence so that the As can be counted, as when counting males in a population, then recurrences are independent (Bernoulli sampling) and <A|A> = P(A), because P(A,A) = P(A)P(A) Note here that <A, A, A | A> =  P(A)3  and so on have meaning as concurrence of As.  If A and B are totally indistinguishable, they are mutually exclusive, giving the orthogonal case, because P(A,B) = 0.  Of course, many interesting cases do not satisfy P(A|B) = P(B|A), and we need to move from a real line to a plane, which can be described by a complex value. The valid region for probabilities in such a plane is in our case  the  h-complex iota space. It is contained by the vectors connecting the values 0  →  i  →  1 → i* 0. With that, we are now ready to proceed to the ontological interpretation implying the verb to be, and on to other verbs and relationships in general, which will require a h-complex vector and matrix interpretation analogous to that in quantum mechanics.



3.1 Introduction and Overview of Agenda. Conditionality discussed in Section 1 is an example case of a relationship, but has several interpretations.  Quantum mechanics does not well differentiate <B|A> from , since conditionality is typically seen as inevitably due to information relayed through cause and effect. The author uses “ontology” as closer to its older meaning of use of categorical relationships, i.e. the use of, in English, the verb “to be” and certain other related verbs as the relational operator or relator. More precisely, the term ontology is used here for the different interpretations of the relationship between nouns or noun phrases A and B in P(A|B) and <A|B>, that would reflect the way we sample to get counts that lead to the conditional probabilities. They can be distinguished by convenient readable choice of relator word or phrase in the form <A| relator |B> such as ‘if’. Whereas it is tempting, and indeed desirable to use symbols for ontological relators in particular, in applications natural language is used for readability.

Important examples follow. Here we shall not be too fussy about plurality of nouns and corresponding verb persons at present, although it can have effect in assigning probabilities. Rather, the more general existential notions of “some” and of the universal notions of “all” are the focus here(see later below).  The asterisk that implies complex conjugation can for the moment be considered as used symbolically, to indicate and active-passive inversion of the relator on which it acts.

if* =  implies (general conditionality)

when* = then (coincidence in time)

are* = include (categorical, set theoretic)

causes* = is caused by (causality)

“Semantics” is larger: it also includes other verbs such as verbs of action, where we have no choice but to write <A| relator |B> because there is way to construe that meaning by <A|B>.  Other verbs primarily differ in the probabilities they convey, so the verb “to be” is regarded as the mother form from which they are derived. Meaning, in contrast, comes (a) from the h-complex probabilistic knowledge network of h-complex terms that specifies meaning as the network context, and (b) from definitions that are not fundamentally different in action from specifications of syllogisms and other logical laws, that evolve the network to generate new probabilistic statements.

In both ontological and full semantic treatment, we can think of the nodes such as A as states having self, marginal, or  prior probabilities P(A), but they can also be viewed as parameters that set P(A|B) and P(B|A), as can be seen in Eqn. 10.  Often that is the same thing, but we may not know these probabilities associated with the states as nodes, and subjectively at least, it is easier to assign conditional probabilities P(A|B) and P(B|A), from which P(A) and P(B) follow, given KA; B) = eI(A; B), the value of which can be changed to describe the probabilistic properties of  different and diverse relationships.  More correctly node probabilities are to be seen as h-complex values vectors of state  |A> where we will equally well need <A|, but we can easily get <A| from |A>* as discussed below. The minimal perspective is that A and B are vectors of one lement, and more precisely the observational brakets discussed above, < ?|A>  =  i* + iP(A) = <A|?>* = (i*P(A) + iP(A))*. However they can be full vectors. Relators are operators that act on <A| or |B> first, to same effect, establishing the probability of the specified relationship. The development below is an agenda essentially follows symbolic manipulation  → semiquantitative manipulation → symbolic projection with quantification → sufficient vectors→ distribution vectors. We can work at any stage on this continuum between symbolic and a full vector-matrix treatment.

3.2 Hermitian Operators as Relationship Operators.  In general, any statements <A| relator |B> represent relationships in the network called probabilistic rules, or simply rules, that the applications import as XML-like tags. We can see a knowledge network as consisting of nodes or vertices A, B, C,  that represent nouns and noun phrases as analogous to the physicists’’ states, or measurements concerning the actual values of states, and vertices between them as the relationships. The two pieces of relationship information in each direction of each edge are

<A| relator |B> =                                                                                         (20)


<B| relator |A> = <A| relator* |B>                                                                                          (21)

These two equations each individually represent the active-passive inverse form of the statement with no change in meaning, as in and <‘type 2 diabetes’ | is caused by | obesity>, where “is cause by” ≡ causes*. The two equations are connected by the complex conjugate of the whole rule:

*  =                                                                                        (22)

So we can thing of a directed edge as associated with as a single complex value, encoding both directions of the relationship. The above define ontological and semantic relationships as Hermitian operators, an important class of QM operators related to data from observations and measurements. Were they not Hermitian, then * = *, as in *  =  <‘type 2 diabetes’ | is caused by | obesity>, which is true as active-passive inversion, an example of semantic equivalence, but it now loses the ability to represent two different directions of distinct effect, and in terms of the meaning, the graph is no longer a directed graph.

In the following few Sections, we would appear to be making use of the above symmetry rules in a way that is, if not exactly simply symbolic, nonetheless nominally quantitative. That this is not necessarily the case is discussed much later below. We shall focus first largely on the general semantic implications of the categorical case that is, after all extensible to verbs of action: = The two reasons why this not a good idea ouside of formulating the case with non-categorical verbs is that (a)  it multiplies the number of nodes represented in states by having a variety of properties associated with each same noun, and (b) we will typically want to inference by referring to the object noun as a state, here cat, rather than to a quality of the subject noun, here dogs as cat-chasers.

3.3. Existential and Universal Quantification. In the ontological interpretation that is specifically categorical, <A|B> is translated as follows.

<A|B> =  = =   = <A| are B>*                    (23)

One consequence is that we can think of  <A|B> in a very simple way.

  =  i P(“A are B”) +  i*P(“B are A”)

=   ½ [(“A are B”) +  P(“B are A”)] + ½ h[[(“A are B”) −  P(“B are A”)]           (24)

Note that we have switched i and  i * for  a more readable styles, as it is nice that the first term iP(“A are B”)  images .  It is important to note that the existential notions of  “some” and the universal notions of “all” are not required in these expressions, although they could be. Rather, they would relate to specific choices  the values of P(“A are B”) and P(“B are A”) dictate that.  When they are used,  they follow a QM rule that is here applied to quantifiers such as ‘the’, ‘a’, ‘one’, ‘two’ ‘many’: when such an entity is moved outside the bra, its complex conjugate is taken (however, when moved outside the ket, it is unchanged.

< quantifier A | relator |B>  = < A | quantifier*  relator |B>                                                   (25)

< A | relator | relator |B>  = < A |  relator quantifier*  relator |B>                                          (26)

So,  for example, = . So armed, we can express the extent of existential quantification as

= = Re<A| are |B> = ½ [(“A are B”) +  P(“B are A”)]                                                                                                                                                       (27)

= Im <A| are |B> = ½  h[(“A are B”) P(“B are A”)],

Im <A| are |B>  >   0                                                               (28)

= Im <A| are |B> = ½  h[(“A are B”) P(“B are A”)],

Im <A| are |B>  <   0                                                               (29)

The above “greater than” or “less than” is for ease of interpretation, but it really represents a continuum, not a discontinuity. To put it another way,  Im <A| are |B> reflects the universal case on a scale −1 to +1  mapping from  “all A are B” to “all B are A”, and if is approximately zero , we can simply say that “some A are B” and no more. It is clear logical why the existential case subtends the universal: if all A are B, or all B are A, it necessarily follows that some A are B (and some B are A).

3.4. Trivially and Non-trivially Hermitian Relationships. More generally, relators are non-trivially Hermitian,  because   is not the same as in meaning or in probability. Or rather, they have that capacity, because a trivially Hermitian operator obeys the rules in Section 2.1. In addition, however, the latter obeys = as in < Jack | marries | Jill>, and note that whilst we can write “gets married by” = married*, here married = married*. Note that English often carries different meanings in verbs as if they were overloaded operators.  Saying “the priest marries Jack and Jill” and even “the priest marries Jill” are not taken by the human reader as the same kind of meaning of “marries”, but we would need rules from the context to say that in the second case, considering whether in that religion priests can get married. The priest performs the ceremony of marriage, or “causes to be married”. Text analytics would need to perform context dependent task to distinguish the meaning, and we might say that a second operator holds with marries(2)* = ‘got married by’(2), but marries ≠ married*, so part of the distinct definition is that the verb is in the second meaning is  not trivially Hermitian.  We could write < The priest | causes | marriage> being careful that there is not another marriage performed by the priest in the same relevancy set of rules, and we note that causes and of  are non-trivially Hermitian. For ‘of’, however, a categorical interpretation requires caution. In principle, we could say that aces of spades are the the same thing as the spades of aces, since the set of spades includes an ace, and vice versa. However ‘of’ carries linguistically an non-trivial Hermitian sense, of “owned by” or that in that B is a larger set of things than A. These difficulties vanish, to some extent, in the present system. Rather, <A| relator |B> quantifies the trivial or non-trivial nature of the spefiic relation in the context of a statement. It carries asymmetry as potentially different probabilities. P(Jack marries Jill”) = P(“Jill marries Jack”) but so <A| relator |B> is then purely real, but P(“The priest marries Jill”) > P(“Jill marries the priest”) even if neither can, a prior, be said to be said to probability 0 or 1.

3.5. Para-ontological Relationships. Note that

<A| are equivalent to |B> = <A|B> =<B|A> = = 1, if P(A|B) = P(B|A)             (30)

In some sense, there is even a mother form of <A|B>, which is when A and B are seen together  at random. Following Section 2.15, if there is no such association, such that A and B are independent, then <A|B> = P(A) P(B), and mutual information I(A; B) = 0 (i.e. K(A; B) = 1). Eqn. 10 then becomes, say,

= <A|B>  = <B|A> = i*P(A)  +  i P(B)

=  ( i*P(A)  +  i)  ( i*  +  i P(B))    =   <A|?><?|B>                              (31)

This lies in the scope of QM, but just means that we have the special case P(A) =P(A|B) and P(B|A) = P(A). By the Dirac recipe

<A|B> (<A|B>)*  =   ( i*P(A)  +  i)  ( i*P(A)  +  i) * = <A|?><?|A>

= ( i*P(A)  +  iP(A)) = P(A)                                                                         (32)

But we are not confined to conjugate variables, and can bra-normalize.

<A|B> (<A|B>)*  =   ( i* +  iP(B)) ( i*+  iP(B)) *  <B|?><?|B> = P(B)                                     (33)

In general <A|?><?|A>  is not a relation but stands for the self probability P(A) of a node A in the network. There in this way no information describing the relationship, so this can be computed de novo as required from the self-probabilities P(A) and P(B) of nodes A and B. We do not need a rule (however, recall that its omission in a purely multiplicative network of which we wish to express the join probability implies that it is there with probability 1, not P(A)P(B)).

3.6. Negation. It must always be the case that <A|?><?|A>  > 0 although to be pedantic we might say “providing A exists”. But it does not follow that <A|?><?|B>  > 0. Note, then, the case when A and B are so distinguishable that they are mutually exclusive.

=  <A|B> = 0     =>    = 1                                                        (34)

It is the case of orthogonal vectors <A| and |B> discussed later below. This is actually an important case, because it is quite plausible to build a network in which nodes A, B, C, have distinct and non-overlapping meanings, or that some nodes in a network do. Cats and dogs can be nodes, and in the ontological interpretation = , evidently the value is zero. But it does not mean that, for example,   is zero.

At first glance the rule, that when we move a quantifier outside of the bra we take the complex conjugate, at least symbolically, need not be applied here. That is because at first glance  it seems that the words ‘no’, ‘not’, ‘none’ and ‘non-‘ do not obviously change the meaning when so moved, as if trivially Hermitian, e.g. not = not*. The situation is more subtle.

< no A | are |B> = < A | none are |B> = <A| are not |B> = <A| are |not B> =

= = = 1              (35)

The last equality   = , which is not completely obvious to initial contemplation, is the logical law of the contrapositive. It includes, for example, “mammals are cats” ≡ “non-cats are non-mammals”, which takes a momen’s thought.   But even more subtly, it holds quantitatively  only under certainty, otherwise, using our original example, and can have different associated probabilities. Conversely we can see that does not have the same value as in any event, because while the latter is absolutely true as read, the former is only very occasionally true, in the sense that non-birds can be reptiles or fish, or trees etc. For these reasons we apply the out-of-bra and out-of-ket rules, and distinguish between

= <A| are |not B> =

=  = <A| are not* |B>                                                      (36)

It may well be argued that these are not the same thing anyway, even to causal inspection, but then consistent with that, we are saying that ‘not*’ is different to ‘not’. The real point is it conveniently places the two distinct forms of negation within the relator phrase, i.e. as a property of the relationship.

3.7. Subjective Quantification as Semiquantitative Quantification.  In several Sections following, how statements in a network interact depends on assigning probabilities that are valid combinations in that context, even if not necessarily true as reasonable estimates. If data is structured, even containing relationships, we can establish detailed probability values at least for that set of data. But reasonable estimates are of course desirable in every case, including  textual and anecdotal statements about relationships, and need not be confined to the idea of an authoritative statement having probability 1 merely because the author stated it.  This is harder, but probabilities are nonetheless constrained or guided in many examples considered so far. We could frequently at least say that forward and reverse values are equal, or one specified one is a lot larger than the other, but the other is definitely not zero. For example P(“cats are mammals”) = 1, while the reverse probability P(“mammals are cats”) is problematic. Consistent with the forward direction, it is reasonable that such probabilities relate to the size of sets they describe. For example, given that what you are sampling is a mammal, what is the chance of it being a cat, or what proportion of individual mammalian animals are also cats?  Without any further conditions, this would require knowledge of the number of mammals and number of cats (number of individuals, not species). It is the kind of calculation that some people, and especially demographers, like to do as an exercise in indirect estimation. For example, in a detailed analysis Legay estimated 400 million cats in the world (although others estimate more than 500 million).  Estimates of the number of mammalian species is relevant to estimating the number of individual mammals, and is  around 6000 From that one guesses that the number of mammals, if cats were representative, is 400 x 6000 x 1000000 = 2.4 x 1012.  Importantly also, if the 6000 mammalian species were equally populated, we could then say that P(cats | mammals) = P(“mammals are cats”) ≈ 1/6000 ≈  0. 00017. That gives some idea of the magnitudes if figures encountered,  but assuming equal sized populations a bad assumption. Typically such distributions follow Zipf’s law that predicts that out of a population M of N elements, the probability P(e(k) | M)  of elements of rank k is

P(e(k) | M)  =  (z(s, k)  ̶   z(s, k ̶ 1)) / z(s, N)                                                                         (37)

where we have written it maximally in terms of Riemann’s zeta function [3] defined as  z(s, n)  =  1 + 2 ̶  s + 3 ̶  s + …+ n ̶  s ).  The zeta function has more general significance as the amount of information obtained by counting things. Although it does not immediately solve the P(cats | mammals)  estimation problem, we note that P(cats | mammals) also estimated in terms of zeta functions is in our example case

P(cats | mammals)  =  ez(s=1, n[cats, mammals])  ̶  z(s=1, n[mammals]) =  ez(s=1, n[cats])  ̶  z(s=1, n[mammals])         (38)

Here n[A] in general means “number of A”, i.e. the counted number, or observed frequency,  of A. Note n[cats, mammals] = n[cats], since all cats are mammals. Putting the above together (along with, strictly speaking, considerations of the next Section) we have for 6000 mammalian species, and with cats ranked k[cats] as the k[cats] most populous species,

P(cats | mammals) = P(“mammals are cats”)

= ez(s=1, n[cats, mammals])  ̶  z(s=1, n[mammals])

= ez(s=1, n[cats])  ̶  z(s=1, n[mammals])

= (z(s, k[cats])  ̶   z(s, k[cats] ̶ 1) / z(s, N=6000)

=  k[cats] ̶ s / z(s, N=6000)

logeP(cats | mammals) =  z(s=1, n[mammals])  ̶   z(s=1, n[cats])

=  s logek  +  loge z(s, N=6000)

For the simplest case of s = 1             and large amounts of data,

P(cats | mammals) = P(“mammals are cats”)

= ez(s=1, n[cats])  ̶  z(s=1, n[mammals])

≈   n[cats]) / n[mammals]

= (z(s, k[cats])  ̶   z(s, k[cats] ̶ 1) / z(s, N=6000)

=  1 /   ( k[cats] ( loge(6000) + 0.5772))

z(s=1, n[mammals])  ̶   z(s=1, n[cats])

=  logek  +  loge (loge z(s, N=6000) + 0.5772))

≈    logek  +  2

where 0.5772… = g,  the Euler-Mascheroni constant relating logarithms and zeta functions That loge ( loge z(s=1, N=6000) + 0.5772)) ≈ 2 is a reasonable approximation for a broad number of  estimates of any number of taxonomic groups considered of which the group of specific interest, here domestic cats, is one: for N=100,  it is more precisely 1.5, for N=1000,000, it is more precisely 2.6. We can note now that 1 ≤ k ≤ N, since it is the kth of the N groups. Hence logek cannot itself exceed the value of  approximately 2 and so, again approximately, it  lies in the range 0…2.  For the 10th in the ranked list it would be about 0.8, for the  100th about 1.5, for the 1000th about 1.9, and for  the 3000th (the median of the ranked list), it would be about 2.1. As a ball park estimate of what we expect for  z(s=1, n[mammals])  ̶   z(s=1, n[cats]), we have z(s=1, n[mammals])  ̶   z(s=1, n[cats]) ≈ 4. In consequence,

P(cats | mammals) = P(“mammals are cats”) = ez(s=1, n[cats])  ̶  z(s=1, n[mammals])  ≈  0.018.

However, we assumed s=1, and s can be critical in the Zip’s distribution. For large s and large N, z(s, N) → 1

P(cats | mammals) = P(“mammals are cats”)

= ez(s=1, n[cats])  ̶  z(s=1, n[mammals])

= (z(s, k[cats])  ̶   z(s, k[cats] ̶ 1) / z(s, N=6000)

=  k[cats] ̶ s + 1


z(s=1, n[mammals])  ̶   z(s=1, n[cats])

=  logek  +  loge (loge z(s, N=6000) + 0.5772))

≈   2s, s >> 1

and so

P(cats | mammals) = P(“mammals are cats”)  ≈  e ̶ 2s, s >> 1, and as reasoned earlier, P(cats | mammals) = P(“mammals are cats”)  →  ~ 0.018 when s → 1. Note that if we simply set s=1 in e ̶ 2s, then e ̶ 2  ≈ 0.14. The above 0.018 is two orders of magnitude higher than the value of 0.00017 reasoned  even earlier above if all species in the Mammalia are equally densely populated, but that kind of qualitative disagreement is what may be expected by the more realistic distribution of mammalian species that we see, and by Zipf’s law. It suggests that s must be significantly greater than 1. More precisely, we have   ̶ loge0. 00017 = 8.68 ≈ 2s, and so s ≈ 4.

3.9. Objective Treatment in Terms of Zeta Functions. The bases of the follow considerations are found in Refs. [3-6] and references therein. It  is useful for calculations to note that we can rewrite Eqn. 10 in zeta function terms, with N as total data. Retaining expected frequencies, this is as follows.

<A|B>   = [ i*P(A)  +  iP(B)]eI(A; B)  

               =  [ i* ez(s=1, n[A])  ̶  z(s=1, N)  +  i ez(s=1,n[B])  ̶  z(s=1, N)]  ez(s=1,n[A, B])  ̶  e[A, B])                

               =  [ i* ez(s=1, n[A])i ez(s=1,n[B])] ez(s=1,n[A, B])]   ̶  z(s=1, e[A, B])  ̶  z(s=1, N)                                   (39)

Here e[ ] is an expected frequency, calculated on the classical, e.g.chi-square test basis: e[A, B] = n[A]n[B]/N = n[A]P(B) = P(A)n[B]. In the cat and mammal example, the above becomes

i* ez(s=1, n[mammals])   i ez(s=1,n[cats])]ez(s=1, n[cats, mammals]])   ̶  z(s=1,e[ cats, mammals]) e  ̶  z(s=1, n[animals])      

=  i* ez(s=1, n[mammals])   iez(s=1,n[cats])]e  ̶ z(s=1,n[mammals])  ez(s=1, n[animals])  e  ̶  z(s=1, n[animals])              

=  i* ez(s=1, n[mammals])   iez(s=1,n[cats])]e  ̶ z(s=1,n[mammals])

=   i*   i ez(s=1,n[mammals])  ̶  z(s=1, n[cats])                                                                                   (40)

The above estimate of P(cats | mammals) reflects the choice of  z(s=1, n[cats, mammals])  ̶  z(s=1, n[mammals])  as the expected information E( I(mammals | cats) | D[cats, mammals]) given data D.  Incidentally, note that n[cats, mammals] means n[cats & mammals], and the converse probability P(mammals | cats) =1 reflects the fact that we know that all cats are mammals, so we could replace n[cats, mammals]) by n[cats].

For what follows,  note that we can get rid of the sometimes problematic total amount of data N by avoiding expected frequencies e[   ].

<A|B>   = [ i*P(A|B)  +  iP(B|A)]

               =  [ i* ez(s=1, n[A, B])  ̶  z(s=1, n[B])  +  i ez(s=1,n[A,B])  ̶  z(s=1, n[A]            )

               =  [ i* ez(s=1, n[B])i ez(s=1,n[A])] ez(s=1,n[A, B])]                                                                            (41)

3.10. Subjective Quantification. The two preceding Sections essentially addressed subjective and objective data, in that order. We can represent either, or combine both, as follows

Importantly, we can include expected frequencies of another kind [1,5,6], namely b[  ] that are based on subjective prior belief about the values.

<A|B>   = [ i*P(A|B)  +  iP(B|A)]

                  =  [ i* ez(s=1, n[B])+b[B])i ez(s=1,n[A])+b[A])] ez(s=1,n[A, B]+b[A, B])

≈  [ i* Be(A|B)ez(s=1, n[A])iBe(B|A) ez(s=1,n[B])] ez(s=1,n[A, B])]                                   (42)

Here Be( ) are the pior, probability-like, degrees of belief, in the values, Be(A|B) = ez(s=1, b[A|B]-b[A] and Be(B|A) = ez(s=1, b[A|B]-b[A]. If we have zero objective (frequentist) information from counting, of course <A|B>   = i* Be(A|B)  +  iBe(B|A). The use of Be(  ) rapidly becomes a good approximation as the b[ ] increase; if any of their their values approach  then the full zeta  function representation should be used.

3.11. Chain of Effect.  Further development benefits from the above quantitative and semi-quantitative ideas and the constraints imposed on the nature of the probabilities when used collectively in inference. Notably, when consider AB |B> CD |D> have to either assume that some kind of linker such as <B|C> orBC |C> has a value of approximately 1, or provide a linker. A note should be made here on why we care. After all, the example   addresses little more than (a) establishing a web of relationships that help define nouns and noun-phrases, (b) can occasionally (albeit rarely) be important  inference in deducing that dogs chase some mice-chasers, and (c) for <mice|?> establishing a collective truth of the statements. But there can be an important meaning when there is an implicit or explicit chain of effect.   Categorical and causal relationships are a clear case of such. However, catching, transporting, selling, and eating fish can have important meaning as a chain of effect for an epidemiologist when one of several possible lakes are contaminated and the outcome is an incidence of food poisoning, and when they wish to establish most probable origins as well as predict outcomes. Indeed, every action exerted or something in the vicinity leaves some sort of trace on where and what it acts upon (e.g. culprit DNA), this being a principle of forensic science. Some are, however, more interesting than others. A large knowledge network may contain interesting and uninteresting cases from a probabilistic inference perspective, but even the uninteresting cases can affect the probabilities finally deduced from a joint probability that addresses a topic of interest. Those rules that are directly relevant constitute the relevancy set.

3.12. Symbolic Projection. There is an intermediate possibility between the scalar treatment so far, and the vector-matrix approach to follow. It is useful when probabilities are to be assigned by a human expert. For a more general relationship such as a verb of action, it is convenient to think of a symbolic projection of value from the operator into the bra and ket. It is an intermediate step towards a less extensively symbolic treatment.

<A| relator |B> =  i* P(A:=relator,  B:=relator*)   +  iP(B:=relator, A:=relator*)

=   i* <A:=relator | B:=relator*> +  i<B:=relator | A:=relator*>                                (41)

By thinking of the relator as “causes”, which is a kind of  prototype action verb, we have a reference point allowing us to say the following under the causal interpretation of conditional probabilities as a joint probability.

<A| causes |B> =  i* P(A:=causes,  B:=causes*)   +  iP(B:=causes, A:=causes*)

= <B|A> =  i* P(B|A)   +  iP(A|B)


That is, P(A|B) = P(B:=causes, A:=causes*). Consistent with this, <?|A><A| relatorAB |B><B| relatorBC |C><C|?> requires linkers such as i*P(B:=relator*) and iP(A:=relator*)  to fill in gaps to complete the chain rule such as P(A|C)  ≈P(A|B)P(B|C). Including links and associating then with appropriate brakets in a network intended to calculate a joint probability, it can readily be shown to be equivalent to using the following for each braket. It gives us Eqn. 41. <?|A><A| relatorAB |B><B| relatorBC |C><C|?>  does not, as a consequence,  contain redundant terms, and although Eqn. 35 is not conditional in its probabilities, it remains asymmetric and typically has an h-complex value. So in asking a human about the probability that A causes B and the probability for the converse, the question in this case is as to with what probability doogs  doing chasing occurs with cats being chased, and the probability that cats doing chasing occurs with dogs being chased.

There is an algebra that links the metadata operator ‘:=’ to conditionality, e.g. P(mammals:=cats) = P(mammals | cats). It suggests P(A|B) = P(A:=causes,  B:=causes*)  = P(A , B  | causes, causes*), but that is not correct because it loses the information as to which of A and B are the cause. It exemplifies the limitations of an approach that is still essentially in terms of conditional probabilities and <A|B>, when it is QM’s <A| operator |ket> that is needed.

3.13. Sufficient Vectors. The following has the limitation that it is only really suitable for (a) ontological relations (i.e. that are some kind of interpretation of <A|B>, not necessarily categorical), and (b) trivially Hermitian relators in general. Actually, the overall non-trivial Hermitian effect can be represented, but it is dependent on choosing particular different values for the probabilities P(A), P(B) etc. implied by the nodes. That may conflict with values that we want the nodes to have when engaged in other interactions. That <A| and |B> are vectors of at least two elements is not strictly true in the Dirac notation. The significance of the notation <A|  and |A> is that  one is the transpose of the other, but we add the requirement that one is also the complex conjugate of the other, and so the general requirement is,

<A| = |A>*                                                                                                                              (24)

The above double * and T consideration follows from the use of complex algebra (i-complex and h-complex) to represent directions of effect, as in <A|B> = <B|A>* and = *. But if x is a scalar real number, it is unchanged by the transpose when seen as a vector or matrix of one element, and unchanged by complex conjugation because its imaginary part is zero, and the Dirac notation as implying the above transformations means that

<x| =  x* = x,    |x> = x,                                                                                                            (25)

On the other hand, it also follows that if we address a complex scalar value say x+hy where y is also a scalar real value,

< x+hy | =  (x+hy)* = xhy,    | x+hy>  = x+hy,                                                                       (26)

Consistent with that, but not requiring that the vectors are of one element, we can think of the observation brakets as vectors (not necessarily of one element), and a projection matrix over the whole space as an identity matrix|?><!–?| = I, then times eI(A; B) To see the relation with QM for operator, we can write

<a|  relator=”” |b=””>  =  <a|  <b=””>|?><!–?| eI(A; B)   |B>  =  <a|  <b=””>I eI(A; B)   |B>                                         (27)

It is sufficient to think of vectors of one element  <a| ==”” <a|<b=””>?> = i *P(A) + i, and  |B> = <?|B> = i * + iP(B), because the above equation holds true for I = 1.  Recall however that we can write <a|b> = [i *P(A) + iP(B)] eI(A; B)   = <a|?> eI(A; B)  <?|B>, so we can write

<a|  relator=”” |b=””>  =  <a|  e<sup=””>I(A; B)   |B>  =  <a|?>  eI(A; B)   <?|B>  = <a|b>                             (28)

What this curious equation means is that <a|?>  eI(A; B)   <?|B>  is the case <a|?><?|B>  for I(A; B) =0, and A and B are independent. In this case, the relator here inserts mutual information on A and B as independent. There is just a single eigenvalue, though we could consider that an relator implies a distinct eigenvalue from the space of all possible relators. In that sense we can represent the probabilistic effect of any relator I eI(A; B)   just by inserting into <a|b> an exponential of a new I(A; B). We can accept any I(A; B) because it has the property of being independent of P(A) and P(B) respectively, since it abstracts them as in I(A; B) = lnP(A, B) – ln(A) – lnP(B). I(A; B) varies much less than P(A) and P(B) when used as a metric that is measured from  on population and applied to another, including the population of a single patient for which we wish to perform inference as to best action. However, the limitations mentioned, at the beginning of this Section, hold.

3.14. Operators Acting on Orthogonal Sufficient Vectors. The following works for orthogonal vectors, i.e. when <a|b> = 0. The extent to which that is not actually a restriction is discussed below. Recall that [a, b; c, d] [p, q]T = [ap+bq, cp+dq]T is the product of a matrix with a column vector, and [p, q] [a, b; c, d] = [pa+qc, pb+qd] is the product of a row vector with a matrix. Let us first consider the matrix

Q(A, B) = [0,  i eI(A;B) ; i*eI(A;B),0]  =  [0,  ii*,0]  eI(A;B)                                                    (29)

withectors <a| and=”” |b=””> are of consistent form satisfying <a| ==”” |a=””>*.

|B>  = [i*P(B),  i)]T

<b|  = =”” [<b=””>iP(B),  i*]

<a|  ==”” [<b=””>iP(A),  i*]                                                                                                                    (30)

Q(A,B) |B>  =  [0,  i eI(A;B) ; i*eI(A;B),0]  [i*P(B),  i)] =   [ie I(A;B), i*P(B) e I(A;B)]T                        (31)

<a| Q(A,B) |B>  =    [iP(A), i*] [iP(e I(A;B), i*P(B)e I(A;B)] T = iP(A)e I(A;B) + i*P(B)e I(A;B)

=i*P(A|B) +iP(B|A)                                                                                                 (32)

This is not a restriction if we ensure that all relationships are of character. In other words, even for the basic ontological cases we use [0,  ii*,0] eI(A;B)  as a joining operator, and one that stands for  (using our English type notation of earlier) ‘if’, ‘whe’n,’ is caused by’, and ‘include’. If we want to have another relation we use a new I(A; B) to represent its different probabilities. This attempt has a serious problem, however, in the complex conjugates of the matrices the idempotent multiplications ii= i  and i*i* = i*are replaced by the annihilations are  ii* = i*i = 0, and so the net result is <a| Q(A,B) |B>  =   0. We would be confined to replacing, for example   by <b| include=”” |a=””>, because = <a| are=”” |b=””>  = 0.   Constructions like   by <b| include=”” |a=””> are semantically equivalent in the categorical interpretation, but irksome and restricting. However, it has its uses, notably as the effective matrix with time dependent modifiers. For example,

<a|  now=”” destroys=”” |b=””>  = 1

= 0

This is beyond scope of the present introduction. For present purposes, it is good that we can prevent annihilation by using the matrix [0, i*+i eI(A;B) ; i*eI(A;B)+i,0]  that leaves an idempotent multiplication. Using vectors

|B> = [i*,  i P(B)]T

<a| ==”” [<b=””>i, i* P(A)]                                                                                                                      (33)

we have

R(A,B) |B> = [0, i*+i eI(A;B) ; i*eI(A;B)+i,0]  [i*,  i P(B)]T =[iP(B)e I(A;B), i*e I(A;B)]T                         (34)

<a| R(A,B) |B>  =  [i, i* P(A)] [iP(B)e I(A;B), i*e I(A;B)] T = iP(B)e I(A;B) + i*P(A)e I(A;B)

=i*P(A|B) +iP(B|A)                                                                                                 (36)

We can again change I(A; B) to represent the probabilities involved in a new relationship.

3.15. Distribution Vectors. In general, the appearance of an operator implies in QM that it is a matrix and that we see <a| and=”” |b=””> as vectors between which the operator sits, and it can act on either <a| or=”” |b=””> first to same effect.  Following QM exactly, we have

<a| = =”” [<a|y<sub=””>0>, <a|y1>, <a|y2>, ….]                                                                                  (32)

|B> =  [<y0|B>, <y1|B>, <y2|BA>, ….]T                                                                               (33)

which obey the general rules described in the next section with T here indication transposition to a column vector, and for which vector multiplication satisfies the QM law of composition of probability amplitudes (and exemplifies a kind of inference network).

<a|b> = Si = 0,1,2,3,…n<a|yi><yi||B>                                                                                       (34)

The problem is that  in QM,  y is the universal wave function (universal quantum state) to which the probabilities of all other states A, B, C, etc can be referred with high precision. To do that it represents a information repository with a trans-astronomic, indeed by definition, cosmological, number of bits 0,1,0,0,1,1,1,…In practice, QM practitioners  focus not on the universe but a specific subsystem of interest.  In our case, we need to choose an ubiquitous if not universal state that is of interest. The obvious one for A as, say, obesity, or diabetes type 2, or systolic blood pressure  140 mmHg, or systolic blood pressure greater than  140 mmHg, is that y  is age, and the indices of  y 0, y1, y2 are its value (how many years old).  Whatever the base chosen (here age), vectors represent probability distributions when expressed as above, and so represent distribution vectors.

Use of matrices to act on such distributions is rare in our current applications and beyond scope of tis introduction. Note, however, that we can construct projection operators generally as Pi = 0,1,2,3,…n|yi><yi||. We can sum over many P to obtain a new P  if their component ket and bra are not orthogonal, and the result is the identity operator I if we sum over the  the whole larger space, not just n-dimensional, meaning we consider  more elements than just the n elements in the above vectors. We can multiply a projection operator by the analogue of the exponential of mutual information to obtain an operator (which is not a projection operator, since the effect of including this exponential is that result squared does not return itself, but itself times the square of  this exponential).


  • Robson, B., and Baek, O.K. (2009) “The Engines of Hippocrates: From the Dawn of Medicine to Medical and Pharmaceutical Informatics” B. Robson and Ok Beak (2009), John Wiley & Sons
  2. Robson , B. (2003)  “clinical and Pharmacogenomic Data Mining. 1. the generalized theory of expected information and application to the development of tools” J. Proteome Res. (Am. Chem. Soc.) 283-301, 2
  3. Robson, B.,  and Mushlin, R.  (2004) “clinical and Pharmacogenomic Data Mining.. 2. A Simple Method for the Combination of Information from Associations and Multivariances to Facilitate Analysis, Decision and Design in Clinical Research and Practice. J. Proteome Res. (Am. Chem. Soc.) 3(4); 697-711
  4. Robson, B (2005) . Clinical and Pharmacogenomic Data Mining: 3. Zeta Theory As a General Tactic for Clinical Bioinformatics. J. Proteome Res. (Am. Chem. Soc.) 4(2); 445-455
  5. Robson, B. (2008) Clinical and Pharmacogenomic Data Mining: 4. The FANO Program and Command Set as an Example of Tools for Biomedical Discovery and Evidence Based Medicine” J. Proteome Res., 7 (9), pp 3922–3947
  6. Mullins, I. M., Siadaty, M. S., Lyman, J., Scully, K., Garrett, C. T., Miller, W. G., Robson, B., Apte, C., Weiss, S., Rigoutsos, Platt, D., Cohen, S., Knaus, W. A. (2006) “Data mining and clinical data repositories: Insights from a 667,000 patient data set” Computers in Biology and Medicine, Dec;36(12):1351-77
  7. Robson, B., Li, J., Dettinger, R., Peters, A., and Boyer, S.K. (2011), Drug discovery using very large numbers of patents. General strategy with extensive use of match and edit operations. Journal of Computer-Aided Molecular Design 25(5): 427-441 (2011)
  8. Svinte, M., Robson, B., and Hehenberger, H.(2007)  “Biomarkers in Drug Dvelopment and Patient Care” Burrill 2007 Person. Med. Report. Vol. 6,  3114 – 3126. 8
  9. Robson, B. and McBurney, R. (2013) “The Role of Information, Bioinformatics, and Genomics” pp 77-94 in  Drug Discovery and Development. Technology in Transition, Churcill Livingstone, Elsevier
  10. Robson, B (2013) “Rethinking Global Interoperability in Healthcare. Reflections and Experiments of an e-Epidemiologist from Clinical Record to Smart Medical Semantic Web” Johns Hopkins Grand Rounds  Lectures 80245ac77f9d4fe0a2a2 bbf300caa8be1d
  11. Robson, B. (2013)“Towards New Tools for Pharmacoepidemiology”, Advances in Pharmacoepidemiology and Drug Safety, 1:6,, in press.
  12. Robson, B. and McBurney, R.  (2012) “The Role of Information, Bioinformatics and Genomics”, pp77-94 In Drug Discovery and Development: Technology In Transition, Second Edition, Ed.  Hill, R.G., Rang, P. Eds. Elsevier Press.
  13. Robson, B., Li, J., Dettinger, R., Peters, A., and Boyer, S.K. (2011), “Drug Discovery Using Very Large Numbers Of Patents. General Strategy With Extensive Use Of Match And Edit Operations”, Journal of Computer-Aided Molecular Design 25(5): 427-441 (2011)
  16. Dirac,. P. M. (1930) “The Principles of Quantum Mechanics”, Oxford University Press.
  17. Penrose, R.  (2004) “The Road to Reality: A Complete Guide to the Laws of the Universe”, Vintage Press
  18. Robson, B. “Schrödinger’s Better Patients”, Lecture and Synopsis  University of North Carolina, (4/16/2012)
  19. Robson, B. (2012) “Towards Automated Reasoning for Drug Discovery and Pharmaceutical Business Intelligence”,  Pharmaceutical Technology and Drug Research, 2012 1: 3 ( 27 March 2012 )
  20. Robson, B.,  Balis, UGC , and Caruso, T.P. (2012), “Considerations for a Universal Exchange Language for HealthcareIEEE Healthcom ’11 Conference Proceedings, June 13-15, 2011, Columbia, MO pp 173-176
  21. Robson, B (2009),  “Towards Intelligent Internet-Roaming Agents for Mining and Inference from Medical Data”, Studies in Health Technology and Informatics,
    Vol. 149  pp 157-177

    • Robson,B.  (2009),  “Links Between Quantum Physics and Thought  (for Medical A.I. Decision Support Systems) ”, Studies in Health Technology and Informatics, Vol. 149, pp 236-248
    • Robson. B (2009) “Artificial Intelligence for Medical Decisions”  14th Future of Health Technology Congress,  MIT , September 29-30 2010 and Robson. B (2009) “Using Deep Models of Medicine and Common Sense to Answer ad hoc Clinical Queries” 14th Future of Health Technology Congress,  MIT , September 28-29, 2009.
    • Robson B., and Baek OK (2009), “The Engines of Hippocrates: From the Dawn of Medicine to Medical and Pharmaceutical Infromatics.” BOOK,  600 pages, pub. Wiley.
    • Robson B., and  Vaithilingam A.   (2009) “Drug Gold and Data Dragons. Myths and Realities of Data Mining in the Pharmaceutical Industry” in Pharmaceutical Data Mining: Approaches and Applications for Drug Discovery,  Ed. Konstantin V. Balakin, Pub. Wiley
    • Robson B.  (2008)  “Clinical and Pharmacogenomic Data Mining: 4. The FANO Program and Command Set as an Example of Tools for Biomedical Discovery and Evidence Based Medicine” J. Proteome Res. (Am. Chem. Soc.). ” J. Proteome Res., 7 (9), pp 3922–3947
    • Robson B., and Vaithilingam, A. (2008)  “Protein Folding Revisited” in Molecular Biology and Translational Science, Vol. 84 Ed. Kristi A.S. Gomez., Elsevier Inc.
    • Robson B. and Vaithilingam, A. (2007)  A. “New Tools for Epidemiology, Data Mining, and Evidence Based Medicine”. Poster,  10th World Congress in Obstetrics & Gynecology, Grand Cayman, 2007.
    • Robson, B. (2007)  “Data Mining and Inference Systems for Physician Decision Support in Personalized Medicine”. Lecture and Circulated Report at the 1st Annual Total Cancer Care Summit, Bahamas 2007
    • Robson B.  (2007) “The New Physician as Unwitting Quantum Mechanic: Is Adapting Dirac’s Inference System Best Practice for Personalized Medicine, Genomics and Proteomics?”  J Proteome Res. (Am. Chem. Soc.), Vol. 6, No. 8: pp 3114 – 3126

    [1] It should be declared that a Gaussian function can also be reached within an i-complex description provided that we see a particle as a harmonic oscillator in the ground state, for which the wave function solution is well known to be a Gaussian function. The particle is then seen as oscillating around the fixed point x0. In that sense, h-complex quantum mechanical descriptions relate to ground state harmonic oscillations in the i-complex description,  but do not explain why a wave and not particle is a lower energy state in the absence of perturbation (unless energy is withdrawn from the system by perturbation). However, the rest mass as a kind of ground state of motion in the Dirac equation for mass predicts the mass as due to an oscillation. Here the rest mass is computed via igtime ∂y/∂t where gtime is a hyperbolic number, suggesting the more general hi description as starting point.

Quantum Theory driven (QEXL Approach) Cognitive Computing Architecture resolving Healthcare Interoperability (BigData – HIE/ ACO )

[healthcare cognitive computing platform]

Conquering Uncertainties Creating Infinite Possibilities

(Possible application :- Achieving Algorithm Driven ACO)



The QEXL Approach is a Systems Thinking driven technique that has been designed with the intension of developing “Go To Market” solutions for Healthcare Big Data applications requiring integration between Payor, Provider, Health Management (Hospitals), Pharma etc; where the systemic complexities tethering on the “edge of chaos” pose enormous challenges in achieving interoperability owing to existence of plethora of healthcare system integration standards and management of the unstructured data in addition to structured data ingested from diverse sources. Additionally, The QEXL Approach targets for the creation of Tacit  Knowledge Sets by inductive techniques and probabilistic inference from the diverse sets of data characterized by volume, velocity and variability. In fact, The QEXL Approach facilitates algorithmic driven Proactive Public Health Management, while rendering business models achieving Accountable Care Organization most effective.

The QEXL Approach is an integrative multivariate declarative cognitive architecture proposition to develop Probabilistic Ontology driven Big Data applications creating interoperability among Healthcare systems. Where, it is imperative to develop architecture that enable systemic capabilities such as Evidence Based Medicine, Pharmacognomics, biologics etc; while also creating  opportunities for studies such as Complex Adaptive System (CAS). Such approach is vital to develop ecosystem as an response to mitigate the Healthcare systemic complexities. Especially CAS studies makes it possible to integrate both macro aspects (such as epidemiology) related to Efficient Heathcare Management Outcomes ; and micro aspects (such as  Evidence Based Medicine and Pharmacogenomics that helps achieve medicine personalization) achieving Efficacy in the Healthcare delivery, to help achieve systemic integrity. In The QEXL Approach QEXL stands for “Quantum Exchange Language”, and Q-UEL is the initial proposed language. The QEXL Consortium embraces Quantal Semantics, Inc; (NC) and Ingine, Inc; (VA), and collaborates with The Dirac Foundation (UK), which has access to Professor Paul Dirac’s unpublished papers. The original consortium grew as a convergence of responses to four stimuli:

  1. The “re-emerging” interest in Artificial Intelligence (AI) as “computational thinking”, e.g. under the American Recovery Act;
  2. The President’s Council of Advisors on Science and Technology December 2010 call for an “XML-like” “Universal Exchange Language” (UEL) for healthcare;
  3. A desire to respond to the emerging Third World Wide Web (Semantic Web) by an initiative based on generalized probability theory  – the Thinking Web; and
  4. In the early courses of these  efforts, a greater understanding  of what Paul Dirac meant in his  Nobel Prize dinner speech where he stated that quantum mechanics should be applicable to all aspects of human thought.

The QEXL Approach

The QEXL Approach is developed based on considerable experiences in Expert Systems, linguistic theory, neurocognitive science, quantum mechanics, mathematical and physics-based approaches in Enterprise Architecture, Internet Topology, Filtering Theory, Semantic Web, Knowledge Lifecycle Management, and principles of Cloud Organization and Integration. The idea for well-formed probabilistic programming reasoning language is simple.  Importantly, also, the more essential features of it for reasoning and prediction are correspondingly simple such that the programmers are not necessarily humans, but structured and unstructured (text-analytic) “data mining” software robots. We have constructed a research prototype Inference Engine (IE) network (and more generally a program) that “simply” represents a basic Dirac notation and algebra compiler, with the caveat that it extends to Clifford-Dirac algebra; notably a Lorentz rotation of the imaginary number i (such that ii = -1) to the hyperbolic imaginary number h (such that hh = +1) corresponding to Dirac’s s, and gtime or g5) is applied.

[Outside the work of Dr. Barry Robson, this approach has not been tried in the inference and AI fields, with one highly suggestive exception: since the late 1990s it has occasionally been used in the neural network field by T. Nitta and others to solve the XOR problem in a single “neuron” and to reduce the number of “neurons” generally. Also suggestively, in particle physics it may be seen as a generalization of the Wick rotation time i x time used by Richard Feynman and others to render wave mechanics classical.  It retains the mathematical machinery and philosophy of Schrödinger’s wave mechanics but, instead of probability amplitudes as wave amplitudes, it yields classical but complex probability amplitudes encoding two directions of effect: “A acts on B, and B differently on A”. It maps to natural language where words relate to various types of real and imaginary scalar, vector, and matrix quantities. Dirac’s becomes the XML-like semantic triple . ]  

The QEXL Approach involves following  interdependent components.

  • Q-UEL (Probabilistic Inference + Phenomenon Of Interest): Addresses global issues that potentially pervade all human endeavors, and hence universal interoperability is of key importance
  •  (Inference Engine + Semantic Inferencing): Project addressing universal meaning underlying diverse natural languages on the Internet, and the use of that in knowledge representation
  • Inference Engine + Decentralized Infra: A link infrastructure for intra- and inter-cloud interoperability and integration in a coherent high level “metaware” environment. This component can also be explored to be replaced with simpler industry ready solutions such as MarkLogic® Enterprise NoSQL Database on Hadoop Distributed File System.

In an endeavor of this kind the partitions-of-work are inevitably artificial; it is important that this does not impede the integrity of optimal solutions.  The most important aspect in The QEXL Approach is, in essence where architecturally Probabilistic Inference (PI) and Data Architecture for the Inference Engine (IE)  is designed to be cooperative; software robots are created while PI and IE interact; and the inference knowledge gained by the PI and IE provide rules for solvers (robots) to self compile and conduct queries etc. This is therefore the grandeur of the scheme: This approach will have facilitated programming by nice compilers so that writing the inference network is easy, but it is not required to write the inference net as input code to compile, with the exception of reusable metarules as Dirac expressions with variables to process other rules by categorical and higher order logic. The robots are designed and programmed to do the remaining coding required to perform as solvers. So the notion of a compiler disappears under the hood. The robots are provided with well-formed instructions as well formed queries. Once inferences are formed, different “what – if” questions can be asked. Given that probability or that being the case, what is the chance of… and so on. It is as if having acquired knowledge, Phenomenon Of Interest (POI) is in a better state to explore what it means. Hyperbolic Dirac Networks (HDNs) are inference networks capable of overcoming the limitations imposed by Bayesian Nets (and statistics) and creating generative models richly expressing the “Phenomenon Of Interest” (POI) by the action of expressions containing binding variables. This may be thought of as an Expert System but analogous to Prolog data and Prolog programs that act upon the data, albeit here a “probabilistic Prolog”. Upfront should be stated the advantages over Bayes Nets as a commonly used inference method, but rather than compete with such methods the approach may be regarded as extending them. Indeed a Bayes Net as a static directed acyclic conditional probability graph is a subset of the Dirac Net as a static or dynamic general bidirectional graph with generalized logic and relationship operators, i.e. empowered by the mathematical machinery of Dirac’s quantum mechanics.

 The QEXL Approach Theory :- Robson Quantitative Semantics Algebra (RQSA)

Developed by Dr. Barry Robson

Theory :- The QEXL Approach based on Robson Quantitative Semantics Algebra – RQSA (Link to development of algorithm – overcoming limitations of Gold Stand Bayesian Network – to solve uncertainty while developing probabilistic ontology)

Impact Of The QEXL Approach

Impact of The QEXL Approach creating Probabilistic Ontology based on Clifford-Dirac algebra has immense opportunity in advancing the architecture to tackle large looming problems involving System of Systems; in which vast uncertain information emerge. Generally, as such systems are designed and developed employing Cartesian methods; such systems do not offer viable opportunity to deal with vast uncertain information when ridden with complexity. Especially when the context complexity poses multiple need for ontologies, and such a system inherently defies Cartesian methods. The QEXL Approach develops into an ecosystem response while it overcomes the Cartesian dilemma (link to another example for Cartesian Dilemma) and allows for generative models to emerge richly expressing the POI. The models generatively develops such that the POI behavior abstracted sufficiently lend for the IE and the Solvers to varieties of studies based on evidence and also allows for developing systemic studies pertaining to Complex Adaptive System and Complex Generative Systems afflicted by multiple cognitive challenges. Particularly, The QEXL Approach has potential to address complex challenges such as evidence based medicine (EBM); a mission that DoD’s Military Health System envisions while it modernizes its Electronics Health Record System – Veterans Health Information Systems and Technology Architecture (VistA). Vast potential also exists in addressing Veteran Administration’s (VA) Million Veteran Program (MVP); an effort by VA to consolidate genetic, military exposure, health, and lifestyle information together in one single database. By identifying gene-health connections, the program could consequentially advance disease screening, diagnosis, and prognosis and point the way toward more effective, personalized therapies.

Although The QEXL Approach is currently targeted to the healthcare and pharmaceutical domains where recognition of uncertainty is vital in observations, measurements and predictions, and probabilities underlying a variety of medical metrics, the scope of application is much more general. The QEXL Approach is to create a generic multivariate architecture for complex system characterized by Probabilistic Ontology that employing generative order will model “POI” facilitating creation of “communities of interest” by self-regulation in diverse domains of interest, requiring integrative of disciplines to create complex studies. The metaphor of “Cambrian Explosion” may aptly represent the enormity of the immense possibilities in advancing studies that tackle large systemic concerns riddled with uncertain information and random events that The QEXL Approach can stimulate.


The inference engine can be conceptualized into solutions such as MarkLogic NoSQL + Hadoop (HDFS).

It is interesting to note that in the genesis of evolving various NoSQL solutions based on Hadoop few insights have emerged related to need for designing the components recognizing their cooperative existence.

The Goal of The QEXL Approach: Is all about Contextualization 

The goal employing The QEXL Approach is to enable the realization of cognitive multivariate architecture for Probabilistic Ontology, advancing the Probabilistic Ontology based architecture for context specific application; such as Healthcare. Specifically, The QEXL Approach will develop PI  that helps in the creation of generative models that depicts the systemic behavior of the POI riddled with vast uncertain information. Generally, uncertainty in the vast information is introduced by the System of Systems complexity that is required to resolve multiples of ontologies, standards etc., these further introduce cognitive challenges. The further goal of The QEXL Approach is to overcome such challenges, by addressing interoperability at all levels, including the ability to communicate data and knowledge in a way that recognizes uncertainty in the world, so that automated PI and decision-making is possible. The aim is semiotic portability, i.e. the management of signs and symbols that deals especially with their function and interactions in both artificially constructed and natural languages. Existing systems for managing semantics and language are mostly systems of symbolic, not quantitative manipulation, with the primary exception of BayesOWL.  RQSA, or Robson Quantitative Semantic Algebra by its author Dr. Barry Robson, to distinguish it from other analogous systems, underlies Q-UEL. It is the development of (a) details of particular aspect of Dirac’s notation and algebra that is found to be of practical importance in generalizing and correctly normalizing Bayes Nets according to Bayes Theorem (i.e. controlling coherence, which ironically Bayes Nets usually neglect, as they are unidirectional), (b) merged with the treatment of probabilities and information based on finite data using the Riemann Zeta function that he has employed for many years in bioinformatics and data mining (, and (c) the extension to more flavors of hyperbolic imaginary number to encode intrinsic “dimensions of meaning” under a revised Rojet’s thesaurus system.

The Layers of the Architecture Created by The QEXL Approach

The QEXL Layered View

The QEXL Layered View

Layer 1- Contextualization: Planning, Designing driven by Theories 

A.    Probabilistic Ontology creating Inferencing leading into Evidence Based Medicine

i.     Aspects addressed by Q-UEL Tags and Kodaxil Inferencing

  1. Autonomy / Solidarity
  2. Inferencing (Kodaxil and Q – UEL)
  3. MetaData
  4. Security / Privacy
  5. Consented vs Un-consented Data
  6. Creating Incidence Rule (predicated – Q-UEL and Kodaxil)

ii.     Kodaxil:-  Enforcing Semantics across data sources (global text and data interoperability) – universal meaning underlying diverse natural languages on the Internet

iii.     Fluxology:- Logical Meta Data Cloud (A link infrastructure for intra- and inter-cloud interoperability and integration in a international setting)

  1. Adaptive
  2. Emergent Data Usage Patterns (networks of networks – enables by Probabilistic Ontology rules)
  3. Modeless Emergent Hierarchies
  4. Federation and Democratization Rule for Data (contract, trust, certificates, quality)

B.    Development of Probabilistic Model Representing Universal Abstraction of Phenomenon Of Interest

C.   Targeting Architecture to Application

  • Evidence Based Medicine
  • Genomics
  • Systemic Healthcare Studies
  • etc

Layer 2 – A: Operational Architecture (Logical )

A.    Reference Architecture

  1. Business Con Ops (Use cases)
  2. Conceptual Target Solution Architecture

Layer 2 – B: Data Management – Data Ingestion and Processing 

  1.  The processing of entries in the source data into form suitable for data mining
  2. The data mining of that processed data to obtain summary rules
  3. The capture of the appropriate released summary rules for inference

B.    Data Storage and Retrieval, Transactions

  1. Secure Storage and Retrieval
  2. Enable Secure Transactions
  3. Secure Data Exchange among several stake-holders and data owners

C.    Data Lifecycle, Data Organization Rules, Data Traceability to the Events, 

  1. Security and privacy by encryption and disaggregation of the EHR in a manner that is balanced against authorized access for extraction of global clinical and biomedical knowledge.
  2. Mechanisms for fine-grained consent permitting sharing and data mining.
  3. Mechanisms for secure alerting of patient or physician by backtrack when an authorized researcher or specialist notes that a patient is at risk.
  4. Structure and format that allows all meaningful use cases to be applied in reasonable time, including large-scale data mining.
  5. Assemblies across sources and data users forming contextual work patterns
  6. Hardened Security Framework

D.    Large EHR repository scaling

E.    Data Mining Rules

F.     Extracting and creating Incidence Rules

G.    Experimenting, observing and creating Semantic Inferences

H.    Visualization 

 The below two layers can be implemented in varieties of BigData platforms such as Hortonworks, Pivotal, Altiscale

Layer 3 – Application Layer (Schema-less for structured and unstructured Knowledge Repository – KRS)

Layer 4 – Infrastructure Architecture (Physical) (Hadoop and MapReduce for Large Data File-management and Processing; and Distributed / Concurrent Computations)

Generative Transformation :- System is the Method

Simple View – Cognition

Multi-Lateral View – Cognition

System is the Method:-

Application :- Accountable Care Organization (ACO) – Complex Adaptive / Generative System – CAS Modeling

ACO objective is to achieve Systemic Healthcare Effectiveness through High Quality Rendering at Least Possible Cost, by co-ordinated efforts, while the engaging to achieve a share in  the savings. This is completely a different Systemic behavior, such as seen in swarming of birds.

Cognition – result of social observation
Where each social unit employs relevant symbols to capture the knowledge and each of these knowledge could be a different level of abstraction. Enterprise or System Architecture as a architecture is sum of several architecture abstractions (various social observations), where architecture is considered as set of decisions. Architecture is established to navigate the system complexities. System architecture is described by sets of abstractions and system occupies different orders to mitigate complexities owing to entropy in a very complex system (messy).

The above challenge of observation (discernment) by different stakeholders can be brought into an assimilation, realized from the integrative influence, where different disciplines intermingle and create an ecosystem influence. Such findings are being studied to understand the rich diversity that emerged during cambrian explosion. Example the the geological influence on the sedimentation and its impact on the living organisms developing functional capabilities deriving from certain calcite properties as in bones, teeth etc

Combining discussion around Context and Integrative Social Inquiry

A. Observer (unconditioned and defeating imposition) merely observing system and asking only one question “why” – in context

(incorporating Bohm – Science, Order and Creativity – significance of social abilities in languaging, sensing, cognition, assimilation etc playing a role in formation of situational experience / knowledge).

B. Within system complexity becoming more complex – higher order of complexity. System “constraints” are experienced.

Change which is a constant by itself, has ‘conflict’ traveling along side constantly. This is a paradoxical phenomenon. Conflict is because of the inherent ‘constraint’ that is pushing the system to seek transformation. The constraint emerges due to the conflict in the resources, for which all living organisms are fighting for its own sustenance. Seeking to resolve the conflict, in order to overcome the constraint, the system transforms and it undergoes entropy occupying different increasing orders, as it mitigates levels of complexities in a complex system.

C. Observer (set of observers) record the “order” of the system as a result of integrative influences. Observers have to acquire ways to achieve equilibrium in the new order of complexity, by overcoming system constraints.

D. There are two contending ways the society can respond to achieve harmony within complex system.

From – De’Arcy Wentworth Thomson (DWT) – Growth and Form

and Form and Transformation – Gerry Webster and Brian Goodwin; this book makes a case for generative development in biology

a. Natural Selection – Adaptive – Darwinian – Natural Selection – Survival Of Fittest – Dominance

b. Self Regulation – Generative – Innovation – Diversity – Cambrian Explosion – Unique Peculiarities – Co Existence – Emergent

Accountable Care Organization (ACO) driven by Affordability Care Act transforms the present Healthcare System that is adaptive (competitive) into generative (collaborative / co-ordinated) to achieve inclusive success and partake in the savings achieved. This is a generative systemic response contrasting the functional and competitive response of an adaptive system.

Natural selection seems to have resulted in functional transformation, where adaptive is the mode; does not account for diversity.

Self Regulation – seems like is a systemic outcome due to integrative influence (ecosystem), responding to the system constraints. Accounts for rich diversity.

The observer learns generatively from the system constraints for the type of reflexive response required (Refer – Generative Grammar – Immune System –

From the above observation, should the theory in self regulation seem more correct and that adheres to laws of nature, in which generative learning occurs. Then, the assertion is “method” is offered by the system itself. System’s ontology has an implicate knowledge of the processes required for transformation (David Bohm – Implicate Order)

For very large complex system,

System itself is the method – impetus is the “constraint”.

In the video below, the ability for the cells to creatively create the script is discussed which makes the case for self regulated and generative complex system in addition to complex adaptive system.