PCT Ontology Modelling
This document is about Permissive Commons Technology - Ontology Modelling. The high-level introductory description of what ontologies and ontology modelling is about, has been moved to the page about UnderstandingOntologies.
In-order for CognitiveAI-WIP systems to meaningfully support the notion of 'common sense', semantic web related ontology technology is sought to be advanced. In-turn, the objective is to seek to address a number of underlying issues that are considered to be problems with it.
QuickLinks
- OntologyModelling - general introductory notes
- PCTWebizenUseOfOntology - info about the applied use of ontology
- VocabularyModelling - general notes about language modelling
- EnglishLanguageModelling - specifically about english (NLP) ontology modelling
- BiosphereOntologies - (tbd better: needs top-level ontology)
- SocioSphereOntologies - (tbd better: needs top-level & BiosphereOntology)
Webizen Diagram: A Basic UseCase Illustration
The above diagram was produced as part of the process of seeking to get some sort of grasp on how to address the GuardianshipRelations problems.
Further details about the applied usecases are described: PCTWebizenUseOfOntology
Part of the purpose of the illustration, is to denote the difference between a persons 'private and personal' inforg (seeAlso: The Semantic Inforg & The “Human Centric Web” — Reality Check, Tech.) and the resources that are electronically supplied as a consequence of social involvement with others.
This is intended to be empowered via ValuesCredentials which in-turn require core ontologies; that are sought to be delivered using the PCT ecosystems.
Thereafter; the effect of these components, are intended to support the private operation of 'Webizen' (Agents).
NOTE: the end-concept of a 'webizen', is that it is a distinguishable class of robot, that is owned by an entity who is responsible for how it is made to operate; and that, all human beings using them, must first have their own. This in-turn acts to support, HumanCentricDigitalIdentity, HumanCentricAI via tooling, that will include a HumanCentricOntology structure - that can be further improved upon overtime
The objective for now, is to define the requirements to establish a functional Webizen ecosystem that is then able to be used to further progress the works, upon a common set of principals and technology that supports those socio-economic and technological foundations
Technical Requirements
The solution that can work, shouldn't consume more than 128MB or at most 1GB of RAM, depending on how the graph model may be processing - complex graphs; which should require upto 2GB RAM to operate in a basic way. The use of GPU Processing may be employed, but shouldn't be required for whatever the basic solution is. Some GPUs do not have tensor cores whilst others may not support cuda or similar. Conversely, in future implementations - the use of Neuromorphic Processors is being considered, and perhaps there are significantly beneficial applications for the use of these sorts of hardware components to expressly support the way these fundamental processes are processed.
The complexity of the model will in-turn have various repercussions on performance and resource requirements. The language model should also be designed to support voice interaction (ie: VoiceToText, and TextToVoice); which in-turn means it should support phonetic analysis.
The language model should be able to work offline; the solution should not require continual communications with a public (cloud) API.
Perhaps moreover - the difference between 'NLP' and what this process is seeking to achieve, is that the software is not seeking to create an archive of the knowledge of human kind that is available to an AI agent to turn into one massive AI 'language model'; rather, what i'm seeking to achieve is the development of a very well defined vocabulary model, that can support the development of database structures, AI and therein - ontologies & interfaces.
In-order to scope the broad notional concept; there's a question about whether to slim it down or first work to define a 'gold plated' model ie: all the qualities that one might wish for if computing resources for all users wasn't a problem... and thereafter, seek to slim it down as required?
Problem Statement
HTTP(s) Ontologies are published online using a HTTP server and ICANN DNS.
PROBLEM 1: Often these resources disappear and may only later be found by searching for a record of it in archive.org or on github.
PROBLEM 2: The documents (and embedded semantics) may change without notice; and without a known mechanism to distinguish between the intended semantic meanings when defining the document vs. the future definitions that are engendered as a consequence of the content of an ontological dependency changing.
PROBLEM 3: The level of know-how relating to the production of ontologies can often lead to unintended meanings being semantically defined.
PROBLEM 4: Tools are most-often defined to be built around OWL which has a top-level concept of 'thing' and not everything relating to our conscious experience of life, is about things; nor are all entities things.
PROBLEM 5: In-order to decentralise and empower human kind, there must be a way to decentralise AI. There are various complex reasons, implications and consequences; and whilst these sorts of ideologies may not be universally sought, for those who do want to ensure a means for tooling to support the needs of themselves and others as human beings in a way that requires Decentralised (Human Centric) AI - decentralising ontologies (and other assets) in a manner that can (with the use of cryptography alongside other patterns) provide a fit-for-purpose capacities to participatorily create, make use of and rely upon electronic resources - is essential for all persons who exist in socio-economic structures that are managed using electronic systems.
As far as i am aware, there are not presently well-defined or well-known solutions that are 'fit for purpose' in relation to the requirements of the webizen project; The technical challenges include,
Defining a way to create ontologies (documents) that are referencable; and, a. 'tamper evident' b. temporally distinguishable. c. Able to be targeted as a specific version or group; with logic support for different usecases.
Defining cryptography methods 'ecosystem' that support the various usecases. (permissions, provonance, etc.)
Support inferencing methods for 'common sense' related functions; including, a means to process and formulate appropriate responses if the derivative of a functional request renders a result that is 'unclear' (ie: not binary - Yes/No)
Social Factors;
a. Provide explainability of decision making processes; including, via the use of permission systems that may distinguish the rights of access to resources based upon external factors (ie: a victim / perpetrator making claims, vs. the capacities of a judge/jury in a court of law)
b. Support for Children (ie: Content classifications)
c. Support for storing and maintaining access to the information created by persons who have become deceased.
d. Self-Archiving related methodologies; ideally, 'useful content' should never dissapear entirely; unless it becomes the case that no human being on the planet (voluntarily) wants to store it; or that it relates to some other complex use-case involving dignity (even then, perhaps the information should be stored somewhere as evidence, idk. perhaps laws?)
e. Support for basic economic considerations to be built into the protocols: This is not about the commercial abuse of a payments system, rather, it is about seeking to ensure that there is a caculation done on the biosphere and socio-sphere implications, as to provide protocol level evidence of such things as energy use and a means to abolish the support that is presently wide-spread witth respect to issues such as digital slavery.
Design Implications
Somehow there needs to be;
- A foundational ontology that defines requirements for a data-structure.
- A methodology that is able to be deployed over multiple Non-HTTP(s)Protocols
Broader Considerations
DNS: it is entirely possible that part of the solution may be to advance solutions that provide an update to DNS. If this is the case, the solution would need to work Side-By-Side with existing DNS.
SocioEconomics: There are material costs and gains associated with every human being, irraspective of age or competency. Existing systems have defines means to extract economic value in relation to any human entity; but they have not well defined how to ensure every human entity is also entitled to be a beneficiary; and indeed also, how that might work.
The GuardianshipRelations document provides some initial texts in relation to this topic.
PCT Ontology Modelling Requirements
TheValuesProject seeks to employ various pre-existing well-known instruments; such as those that have been created via the United Nations and various other internationally well regarded groups and institutions, that are not presently available in an SemanticWeb based format. Some information about existing SemWebOntologies is provided via that link (although, not exhaustively).
In-order to support the SafetyProtocols and related ecosystems, some core work on the ontology methods is required; which will in-turn lead to the development of an ecosystem method for SocioSphereOntologies and BiosphereOntologies to compliment existing systems that are well defined for owl:thing
related requirements.
Additionally,
- well known ontologies such as FOAF are now fairly outdated;
- Most reasoning / ontologicial systems are very much centered aruond the use of OWL; which has a top-level predicate defining anything that employs it as a sub-classes of
owl:thing
- Ontologies are currently designed to refer to HTTP-URIs. By employing Non-HTTP(s)Protocols a method to define how to support various POCAppRequirements is thereby invoked.
- A foundational structure needs to be defined to thereafter built upon; and some consideration is being made about what the most appropriate approach may well be.
Thereafter, once these considerations are better understood; then, there is an ability to create the ontological structures of those ontology documents, including references to other .wellknown
ontologies and related intended inferencing models, can be better modelled.
PCT Ontology Modelling Method
Ground-Up ontological modelling ideas....
Atomic Ontologies
The concept of Atomic Ontologies - is about the idea of forming an ontological approach that describes one term; rather than an entire syntax. thereafter, a group of these atomic ontological resources could be used to define more comprehensive ontological structures.
Foundational Requirements: RDF Dictionaries?
In-order to address the semantic ontology problems, perhaps the best way is to define a new approach to forming the foundational ontology via a different sort of modelling technique.
An idea of how this may be achieved; is via the use of a dictionary as a sourcefile; for Generating a semantic dictionary using AI (NN/DL/ML); which may then act as a datasource for the creation of an app that provides support for modelling an ontological structure based upon those fundamental predicates.
This initial process, would likely not include an array of terms that relate to brands and various areas of science, yet, the same process could be undertaken with other dictionaries.
Thereby providing a means to create a mapping technique between the historical / traditional foundational predicates; and in-turn, the ability to generate more meaningful relationships with the meaning of other terms; and thereby provide an underlying context for producing ontological (knowledge) modelling.
Whilst i'm not sure if this theory has merit; i'm working to figure out how it might work now...
Part of the consideration is; whether and/or how these sorts of requirements might fit into an ecosystem that is designed to employ CognitiveAI-WIP related applied computer science theories and principles.
NowTherefore; a few questions with unknown answers.
Forming Trustworthy Ontological Structures
The existing Sematnic Web tooling, is built upon HTTP(s) Identifiers and as noted otherwise, there are an array of qualities that are considered to be issues.
Given the permissive commons ecosystem works provide a foundational opportunity to redefine how these sorts of things are implemented; in a new and different sort of way that may extend beyond merely ensuring support via protocols other than HTTP(s) and related version control and permissions related ecosystem characteristics.
I am now therefore working through some ideas that will remain largely undefined until the though processes are somewhat advanced (as a consequence of having done the work).
Discovery work
The content below is the discovery work that is being defined as it is written / produced, that is seeking to process a series of complex problems without a known answer or defined approach; as to catalogue related considerations in-order to create one (an approach)
How to best construct an artificial learning model that provides a coheasive data-set as to support a systematic method for foundational principals that can then in-turn be employed for producing SocioSphereOntologies, BiosphereOntologies and most importantly a HumanCentricOntology.
The first objective would be to describe a method that can be used to create a resource (software) that is able to be used with the english language.
(it goes without saying, that other languages should be supported - but i'm focusing firstly, upon the language i mostly use and best know)
QUESTION: is it better to start with the use of english or the use of mathematics; or is it best to ensure the use of both, in-order to create a bridge between human use of linquistics and the computational use of human language, in a way that's in-turn going to be used computationally.
English Language Concepts
A few core concepts;
- Sentence - a set of words that is complete in itself, typically containing a subject and predicate, conveying a statement, question, exclamation, or command, and consisting of a main clause and sometimes one or more subordinate clauses.
- word - a single distinct meaningful element of speech or writing, used with others (or sometimes alone) to form a sentence and typically shown with a space on either side when written or printed.
- morphemes - a meaningful morphological unit of a language that cannot be further divided (e.g. in, come, -ing, forming incoming ).
- Grammar - the whole system and structure of a language or of languages in general, usually taken as consisting of syntax and morphology (including inflections) and sometimes also phonology and semantics.
- meaning - what is meant by a word, text, concept, or action.
- known - recognized, familiar, or within the scope of knowledge.
- explained - make (an idea or situation) clear to someone by describing it in more detail or revealing relevant facts.
- antecedent - a thing that existed before or logically precedes another.
- consequent - following as a result or effect.
- Translation - the process of translating words or text from one language into another.
- sense - a faculty by which the body perceives an external stimulus; one of the faculties of sight, smell, hearing, taste, and touch.
- Common Sense - good sense and sound judgement in practical matters.
- Definition - a statement of the exact meaning of a word, especially in a dictionary.
NOTE: above definitions are generally provided in relation to the use of the term as a noun.
1. NOUN - A noun is the name of a person, place, thing, or idea. 2. PRONOUN - A pronoun is a word used in place of a noun. 3. VERB - A verb expresses action or being. 4. ADJECTIVE - An adjective modifies or describes a noun or pronoun. 5. ADVERB - An adverb modifies or describes a verb, an adjective, or another adverb. 6. PREPOSITION - A preposition is a word placed before a noun or pronoun to form a phrase modifying another word in the sentence. 7. CONJUNCTION - A conjunction joins words, phrases, or clauses. 8. INTERJECTION - An interjection is a word used to express emotion.
One of the Antecedent to words is letters of an alphabet, which may in-turn be expressed in an array of computational codes - here's a tool that illustrates this.
Computational Language Concepts.
Whilst seeking to make the note, i also understand the fundamental nature of it; apologies,
Computers fundamentally 'speak' binary. Various abstrations are thereby created, which are fundamentally based upon mathematics.
There are various 'command' structures that relate to different computational models; relating both, to hardware and software based instruction-sets. The way through which this is performed has a significant impact on the amount of power (energy) and time (amount of processing required) to achieve some sort of derivative function. Higher level languages or abstractions (from a computing point of view) are able to engender results that other lower-level methodologies are able to achieve with less time (computational horsepower) / energy.
Computers are not manufactured to comprehend natural (human) language; although, it is increasingly 'done' via increasingly sophistocated systems, etc.
Consideration:
If a possible approach to generate a foundational ontology relating to the use of the english language is able to be produced (and consequentially via use / contributions, refined); is it better to start with the predicates relating to computational language concepts, or human (natural) language concepts? or, is a AI model defined to process a model that involves both?
Whilst the diagram is most certainly flawed / incorrect (requires more work) the general gist of the concept that i'm working through is difficult to conceptually construct; so i hope this helps (even if / as it has errors) - noting in particular, the importance of Centricity factors...
Therein - the general idea or concept is; to process the english language via a AI Process (meaning DL, ML, NN, CNN, etc.) in a manner that can thereafter be used to define a tool (Software Program) for producing RDF based ontologically defined assets / code (documents).
The intended useful purpose of this program is to foundationally support the development of PCT Ontologies, which are in-turn a critical requirement for the production of PCT more broadly; including but not limited to, furnishing the means for others to form alternative and/or different foundational ontological schemas and related processes for logic mapping.
Processing Dimentionalities
- Time
- Energy
Both time and energy have complex attribution frameworks associated with the useful application of those concepts to different constituent 'ecosystems' processes.
- Correctness and implication of incorrectness
Methods to consider uncertainty.
- Centricity
Employing an Agent Centric Approach which employs the use of cryptography; whilst ensuring the operational use of it, is HumanCentric.