Semantic spectrum

The semantic spectrum, sometimes referred to as the ontology spectrum, the smart data continuum, or semantic precision, is a series of increasingly precise or rather semantically expressive definitions for data elements in knowledge representations, especially for machine use.

At the low end of the spectrum is a simple binding of a single word or phrase and its definition. At the high end is a full ontology that specifies relationships between data elements using precise URIs for relationships and properties.

With increased specificity comes increased precision and the ability to use tools to automatically integrate systems, but also increased cost to build and maintain a metadata registry.

Some steps in the semantic spectrum include the following:
 * 1) Glossary : A simple list of terms and their definitions. A glossary focuses on creating a complete list of the terminology of domain-specific terms and acronyms. It is useful for creating clear and unambiguous definitions for terms, and because it can be created with simple word processing tools, few technical tools are necessary.
 * 2)  Controlled vocabulary : A simple list of terms, definitions and naming conventions. A controlled vocabulary frequently has some type of oversight process associated with adding or removing data element definitions to ensure consistency. Terms are often defined in relationship to each other.
 * 3)  Data dictionary : Terms, definitions, naming conventions and one or more representations of the data elements in a computer system. Data dictionaries often define data types, validation checks such as enumerated values and the formal definitions of each of the enumerated values.
 * 4)  Data model : Terms, definitions, naming conventions, representations and one or more representations of the data elements as well as the beginning of specification of the relationships between data elements including abstractions and containers.
 * 5)  Taxonomy : A complete data model in an inheritance hierarchy where all data elements inherit their behaviors from a single "super data element". The difference between a data model and a formal taxonomy is the arrangement of data elements into a formal tree structure where each element in the tree is a formally defined concept with associated properties.
 * 6)  Ontology : A complete, machine-readable specification of a conceptualization using URIs (and then IRIs) for all data elements, properties and relationship types. The W3C standard language for representing ontologies is the Web Ontology Language (OWL). Ontologies frequently contain formal business rules formed in discrete logic statements that relate data elements to each another.

Typical questions for determining semantic precision
The following is a list of questions that may arise in determining semantic precision.


 * Correctness: : How can correct syntax and semantics be enforced? Are tools (such as XML Schema) readily available to validate syntax of data exchanges?
 * Adequacy/Expressiveness/Scope: : Does the system represent everything that is of practical use for the purpose? Is an emphasis being placed on data that is externalized (exposed or transferred between systems)?
 * Efficiency: : How efficiently can the representation be searched/queried and possibly reasoned on?
 * Complexity: : How steep is the learning curve for defining new concepts, querying for them or constraining them? Are there appropriate tools for simplifying typical workflows? (See also: ontology editor)
 * Translatability: : Can the representation easily be transformed (e.g. by Vocabulary-based transformation) into an equivalent representation so that semantic equivalence is ensured?

Determining location on the semantic spectrum
Many organizations today are building a metadata registry to store their data definitions and to perform metadata publishing. The question of where they are on the semantic spectrum frequently arises. To determine where your systems are, some of the following questions are frequently useful.


 * 1) Is there a centralized glossary of terms for the subject matter?
 * 2) Does the glossary of terms include precise definitions for each terms?
 * 3) Is there a central repository to store data elements that includes data types information?
 * 4) Is there an approval process associated with the creation and changes to data elements?
 * 5) Are coded data elements fully enumerated?  Does each enumeration have a full definition?
 * 6) Is there a process in place to remove duplicate or redundant data elements from the metadata registry?
 * 7) Is there one or more classification schemes used to classify data elements?
 * 8) Are document exchanges and web services created using the data elements?
 * 9) Can the central metadata registry be used as part of a Model-driven architecture?
 * 10) Are there staff members trained to extract data elements that can be reused in metadata structures?

Strategic nature of semantics
Today, much of the World Wide Web is stored as Hypertext Markup Language. Search engines are severely hampered by their inability to understand the meaning of published web pages. These limitations have led to the advent of the Semantic web movement.

In the past, many organizations that created custom database application used isolated teams of developers that did not formally publish their data definitions. These teams frequently used internal data definitions that were incompatible with other computer systems. This made Enterprise Application Integration and Data warehousing extremely difficult and costly. Many organizations today require that teams consult a centralized data registry before new applications are created.

The job title of an individual that is responsible for coordinating an organization's data is a Data architect.

History
The first reference to this term was at the 1999 AAAI Ontologies Panel. The panel was organized by Chris Welty, who at the prodding of Fritz Lehmann and in collaboration with the panelists (Fritz, Mike Uschold, Mike Gruninger, and Deborah McGuinness) came up with a "spectrum" of kinds of information systems that were, at the time, referred to as ontologies. The "ontology spectrum" picture appeared in print in the introduction to Formal Ontology and Information Systems: Proceedings of the 2001 Conference. The ontology spectrum was also featured in a talk at the Semantics for the Web meeting in 2000 at Dagstuhl by Deborah McGuinness. McGuinness produced a paper describing the points on that spectrum that appeared in the book that emerged (much later) from that workshop called "Spinning the Semantic Web." Later, Leo Obrst extended the spectrum into two dimensions (which technically is not really a spectrum anymore) and added a lot more detail, which was included in his book, The Semantic Web: A Guide to the Future of XML, Web Services, and Knowledge Management.

The concept of the Semantic precision in business systems was popularized by Dave McComb in his book Semantics in Business Systems: The Savvy Managers Guide published in 2003 where he frequently uses the term Semantic Precision.

This discussion centered around a 10 level partition that included the following levels (listed in the order of increasing semantic precision):
 * 1) Simple Catalog of Data Elements
 * 2) Glossary of Terms and Definitions
 * 3) Thesauri, Narrow Terms, Relationships
 * 4) Informal "Is-a" relationships
 * 5) Formal "Is-a" relationships
 * 6) Formal instances
 * 7) Frames (properties)
 * 8) Value Restrictions
 * 9) Disjointness, Inverse, Part-of
 * 10) General Logical Constraints

Note that there was formerly a special emphasis on the adding of formal is-a relationships to the spectrum which has been dropped.

The company Cerebra has also popularized this concept by describing the data formats that exist within an enterprise in their ability to store semantically precise metadata. Their list includes:
 * 1) HTML
 * 2) PDF
 * 3) Word Processing documents
 * 4) Microsoft Excel
 * 5) Relational databases
 * 6) XML
 * 7) XML Schema
 * 8) Taxonomies
 * 9) Ontologies

What these concepts share in common is the ability to store information with increasing precision to facilitate intelligent agents.