Documenti di Didattica
Documenti di Professioni
Documenti di Cultura
These practical problems are also partially a consequent of 1. 2. Some specific institutional behaviors related to traditional program management. Confusion caused by long term, and heavily funded Artificial Intelligence marketing activities
As a general proposition, through out the federal government, quality metrics are not guiding management decisions supporting: 1) Quick transitions from database centered information technology to XML based Semantic Web technology. 2) Transitions from XML repositories to ontology mediated Total Information Awareness, with Informational Transparency, in Secure Channels.
?
DHS Ontology
1) 2) 3) 4) World-wide Trade Data Investigation Targeting Risks, Threats and Vulnerabilities Policy Enforcement Emerging Semantic Web Standards
is not complete
Measurement is part of the semantic extraction task, and is accomplished with a known set of techniques Latent semantic technologies Some sort of n-gram measurement with encoding into hash tables or internal ontology representation (CCM and NdCore, perhaps AeroText and Converas process ontology (?), Orbs, Hilbert encoding, CoreTalk/Cubicon.
One model for semantic extraction explicitly focuses on the first two aspects of the AIPM; e.g. instrumentation/measurement and data-encoding/interpretation
Actionable Intelligence Process Model has an action-perception event cycle. Stratified ontology supports the use of this cycle to produce knowledge of attack and anticipatory mechanisms based on the measurement of substructural categorical invariance.
Work flow and process ontology is available as a basis for encoding knowledge of anticipatory response mechanisms.
Categorical Invariance is measured, using Orbs (Ontology referential bases) for example or CCM (Contiguous Connection Model) encoding, and organized as a resource for RDF triples using some lite OWL OIL.
Benchmark
Transition
Community
Community
Community
Community Community
Localized Ontologies
Community
Real time acquisition of new concepts, and modifications to existing concepts are made via an piece of software called Ontology Use Agent Ontology Use Agents
Community
Localized Ontologies
First principles
First, and before all else, an computer based ontology is a
{ concepts }
In the natural, physical, sciences an ontology is the causes of those things that one can interact with in Nature. Physical science informs us that a formative process is involved in the expression of natural ontology in natural settings.
The set of our personal and private concepts is often thought to be the causes of how we reason as humans. This metaphor is operational in many peoples understanding about the nature and usefulness of machine encoded ontology. But this metaphor can also mislead us!!!!
Extensive literatures indicate that the Artificial Intelligence (AI) mythology has lead many to believe that the reasoning of an ontology might be the same as the reasoning of a human in all cases.
This inference is not proper because the truthful-ness of this inference has not been demonstrated by natural science, and perhaps cannot be demonstrated no matter what the levels of government funding for AI. AI is discounted in Tim Berners Lees notion of the Semantic Web.
The point being made here is that the notion of inference is very different depending if one is talking about the human side or the machine side of the Semantic Web.
One consequence of acknowledging this difference is to elevate the work of the authors of the OASIS standards, in particular Topic Maps. In Topic Maps we have an open world assumption and very little emphasis on computational inference. Human knowledge is represented in a shallow form, and visualization is used to manage this representation. Computation with topic maps AND OWL ontologies work together with XML repositories.
First principles
{ concepts }
Let is use only set theory to consider Tim Berners Lees notion of Semantic Web.
Subsetting function
B is a subset of C
First principles
{ concepts } At this point we have various possible consequences.
1) The small(er) B ontology might simply be viewed by a human and actions taken outside of the information system Subsetting function
b)
B is a subset of C
First principles
Situational Ontology
Software
The ontology sub-setting function has pulled part, but not all, of the background into a situational focus. This first principle is consistent with perceptual physics and thus is informed by natural science.
(The following slides are from OntologyStreams Ontology Presentation VII General Background)
Presentation Contents
Functional specs Ontology Use Start-up Use Model Model: Steady State Ontology System Components: Framework for Query Entity Data Access: Steady State Ontology System
Functional specs
Functional specs:
1. Human-centric: must be human (individual) centric in design and function 2. Support data retrieval: must act as a data retrieval mechanism 3. Event structure measurement: must assist in the definition of data acquisition requirements on an on-going basis 4. Interactive: must support multiple interacting ontologies
5. Real Time: must aid in real time problem solving and in the long term management of specific sets of concepts
Note: Ontology mediated knowledge systems have operational properties that are quite different from traditional relational database information systems. These five functional specs have been reviewed by a small community of professional ontologists, as has been deemed correct for knowledge systems.
Transaction process
Entity updates Inferences about
Knowledge Base
Query entities
Transaction process
Entity updates inferences
Knowledge Base
Query entities
Reasoner
Instance data
Transaction process
Entity updates inferences
Instance data
Instance data may be remote or local. Local data is on the same network as the knowledge base.
Transaction processes
updates
Query Manager
inferences
Inference Manager
Data
Reasoner
Query entities
Data Access Mgr
Knowledge Base
Ontology Mgr
Instance data
Pipes and Threads
Data Object
{ < subject, verb, predicate > } Data and the data is either XML or a data structure such as one would have as a C construct. The Data Access Manager must manage the mapping between local data stores (sometimes having millions of elements) and the set of concepts. The remote data may have many persistence forms, and will be accessed via a data object.
Data Access Mgr
Knowledge Base
** We use RDF and OWL as a standard to create minimal and well knowledge inference capabilities.
Transaction processes
updates
Query Manager
inferences
Inference Manager
Data
Reasoner
Query entities
Data Access Mgr
Knowledge Base
Ontology Mgr
A Query Entity is itself a type of light ontology. It develops knowledge about the user(s) and about the query process.
Transaction processes
Query Manager
Query entities
Real time analysis is supported through the development and use of query entities. These entities have regular structure and are managed within a Framework.
Objective: We convert a stream of event measurements into an transaction ontology, and create auxiliary processes that will use a general risk ontology, an ontology about process optimization, and other utility ontologies.
Later the number of fields in any "findings data flow" may increase or decrease without us caring at all.
Knowledge Base
Query Manager
Each event will have a weakly w structured (free form text) and structured s component. So we use the notation
e = w/s or e(i) = w(i)/s(i) .
Instance data
{ s(i) }
Notation e(i) = w(i)/s(i) An event is measured by filling in slots in a data entry form, and by typing in natural language into comments fields in these entry forms.
Semantic extraction is performed using one of several tools, or tools in combination with each other Discrete analysis
Observation: Given real data, one can categorize the set of events due to the nature of the information filled in.
Discrete analysis is mostly the manual development of ontology through the study of natural categories in how the data is understood by humans.
{ w(i) }
Semantic extraction
{ w(i) }
Each association is made as exploratory activities with specific goals. Each s(i) is a record from a single table. Suppose there are 120 columns. Each column has values, sometimes empty. Fix the counter at *. Let s(*, j) , j = 1, . . . , 120 be the columns. We can call these columns also using the term slot.
Regularity in data flow is caused by the events occurring in the external world. Thus the instances of specific data in data records provide to the knowledge system a measurement of the events in the external world.
{ s(i) }
Discrete analysis
Now for each s(*, j) list the values that are observed to be in that column. These values are the possible fillers for the associated slot.
XML and related standards community open and protected standards (CoreTalk, Rosetta Net, ebXML) .NET component management J2EE frameworks Spring
The role of community 1) A community of practice provide a reification process that is Human centric (geographical-community / functional-community) 2) Each community may have locally instantiated OWL ontology with topic map visualization.
a) Consistently and completeness is managed locally as a property of the human individuals, acting within a community, and a locally persisted history of informational transactions with his/her agent b) Individual agents can query for and acquire XML based Web Services, negotiate with other agents and create reconciliation processes involving designated agencies.
Establishing coherence in natural knowledge representation 1) Coherence is a physical phenomenon seen in lasers
a) b) c) Brain function depends critically on electromagnetic coherence Incoherence, e.g. non-rationality, and in-completeness are two separate contrasting issues to the issue of coherence Mathematics, and therefore computer science and logic, have completeness and consistency issues that are well established and well studied
3) Human usability often is a function of a proper balance between logic and agility
2) Observation: Human cognitive capabilities have far more agility than current generation decision support systems 3) The topic map standard (2001, Newcomb and Briezski ) was specifically developed to address the non-agility of Semantic Web standards based on OWL and RDF. (Ontopia, Steve Pepper) 4) Combining XML repositories, OWL, distributed agent architectures and Topic Maps is expressed as Stratified Ontology Management
Detection of novelty
Scenario: an targeting and search analyst at the Port of Seattle is only partially aware of why she feels uncomfortable about some characteristic of a shipment from Sweden. The feeling is expressed in a hand written finding and fed into a document management repository for findings. A targeting and search analyst at the Port of LA expresses a fact about a similar shipment without knowing of her colleagues sense of discomfort.
1) Conceptual roll-up techniques are used on a minute by minute basis to create a viewable topic map over occurrences of concepts expressed in findings. 2) Link analysis connects an alert about uncertainty in the Seattle finding to the fact from LA to produce new information related to a known vulnerability and attack pattern. 3) New knowledge forms are propagated into OWL instantiated ontology and rules and viewed using Topic Maps.
Agent architecture
Scenario: Human analysts provide situational awareness via tacit knowledge, personally agent mediated interactions with agent structures, and human to human communications. A model of threats and vulnerabilities has evolved but does not account for a specific new strategy being developed by a smart smuggler. The smuggler games the current practices in order to bring illegal elements into the United States
1) 2) The model of threats and vulnerabilities expresses as a reification process from various techniques and encoded OWL/Protg ontology with rules Global Ontology: The model is maintained via near real time agency processes under the observation, active review, of knowledge engineers and program managers working with knowledge of policy and event structure Local Ontology: Information is propagated to individual analysts via alerts and ontology management services controlled by the localized agent (of the person)
3)
New (1/30/2005) tutorial on automated extraction of ontology from free form text:
http://www.bcngroup.org/beadgames/anticipatoryWeb/23.htm