Misplaced Pages

Semantic spectrum

Article snapshot taken from Wikipedia with creative commons attribution-sharealike license. Give it a read and then ask your questions in the chat. We can research this topic together.
This article has multiple issues. Please help improve it or discuss these issues on the talk page. (Learn how and when to remove these messages)
This article relies excessively on references to primary sources. Please improve this article by adding secondary or tertiary sources.
Find sources: "Semantic spectrum" – news · newspapers · books · scholar · JSTOR (February 2007) (Learn how and when to remove this message)
This article is in list format but may read better as prose. You can help by converting this article, if appropriate. Editing help is available. (March 2021)
This article is written like a personal reflection, personal essay, or argumentative essay that states a Misplaced Pages editor's personal feelings or presents an original argument about a topic. Please help improve it by rewriting it in an encyclopedic style. (March 2021) (Learn how and when to remove this message)
(Learn how and when to remove this message)

The semantic spectrum, sometimes referred to as the ontology spectrum, the smart data continuum, or semantic precision, is a series of increasingly precise or rather semantically expressive definitions for data elements in knowledge representations, especially for machine use.

At the low end of the spectrum is a simple binding of a single word or phrase and its definition. At the high end is a full ontology that specifies relationships between data elements using precise URIs for relationships and properties.

With increased specificity comes increased precision and the ability to use tools to automatically integrate systems, but also increased cost to build and maintain a metadata registry.

Some steps in the semantic spectrum include the following:

  1. Glossary: A simple list of terms and their definitions. A glossary focuses on creating a complete list of the terminology of domain-specific terms and acronyms. It is useful for creating clear and unambiguous definitions for terms, and because it can be created with simple word processing tools, few technical tools are necessary.
  2. Controlled vocabulary: A simple list of terms, definitions and naming conventions. A controlled vocabulary frequently has some type of oversight process associated with adding or removing data element definitions to ensure consistency. Terms are often defined in relationship to each other.
  3. Data dictionary: Terms, definitions, naming conventions and one or more representations of the data elements in a computer system. Data dictionaries often define data types, validation checks such as enumerated values and the formal definitions of each of the enumerated values.
  4. Data model: Terms, definitions, naming conventions, representations and one or more representations of the data elements as well as the beginning of specification of the relationships between data elements including abstractions and containers.
  5. Taxonomy: A complete data model in an inheritance hierarchy where all data elements inherit their behaviors from a single "super data element". The difference between a data model and a formal taxonomy is the arrangement of data elements into a formal tree structure where each element in the tree is a formally defined concept with associated properties.
  6. Ontology: A complete, machine-readable specification of a conceptualization using URIs (and then IRIs) for all data elements, properties and relationship types. The W3C standard language for representing ontologies is the Web Ontology Language (OWL). Ontologies frequently contain formal business rules formed in discrete logic statements that relate data elements to each another.

Typical questions for determining semantic precision

This section does not cite any sources. Please help improve this section by adding citations to reliable sources. Unsourced material may be challenged and removed. (March 2021) (Learn how and when to remove this message)

The following is a list of questions that may arise in determining semantic precision.

Correctness:
How can correct syntax and semantics be enforced? Are tools (such as XML Schema) readily available to validate syntax of data exchanges?
Adequacy/Expressiveness/Scope:
Does the system represent everything that is of practical use for the purpose? Is an emphasis being placed on data that is externalized (exposed or transferred between systems)?
Efficiency:
How efficiently can the representation be searched/queried and possibly reasoned on?
Complexity:
How steep is the learning curve for defining new concepts, querying for them or constraining them? Are there appropriate tools for simplifying typical workflows? (See also: ontology editor)
Translatability:
Can the representation easily be transformed (e.g. by Vocabulary-based transformation) into an equivalent representation so that semantic equivalence is ensured?

Determining location on the semantic spectrum

This section does not cite any sources. Please help improve this section by adding citations to reliable sources. Unsourced material may be challenged and removed. (March 2021) (Learn how and when to remove this message)

Many organizations today are building a metadata registry to store their data definitions and to perform metadata publishing. The question of where they are on the semantic spectrum frequently arises. To determine where your systems are, some of the following questions are frequently useful.

  1. Is there a centralized glossary of terms for the subject matter?
  2. Does the glossary of terms include precise definitions for each terms?
  3. Is there a central repository to store data elements that includes data types information?
  4. Is there an approval process associated with the creation and changes to data elements?
  5. Are coded data elements fully enumerated? Does each enumeration have a full definition?
  6. Is there a process in place to remove duplicate or redundant data elements from the metadata registry?
  7. Is there one or more classification schemes used to classify data elements?
  8. Are document exchanges and web services created using the data elements?
  9. Can the central metadata registry be used as part of a Model-driven architecture?
  10. Are there staff members trained to extract data elements that can be reused in metadata structures?

Strategic nature of semantics

This section does not cite any sources. Please help improve this section by adding citations to reliable sources. Unsourced material may be challenged and removed. (March 2021) (Learn how and when to remove this message)

Today, much of the World Wide Web is stored as Hypertext Markup Language. Search engines are severely hampered by their inability to understand the meaning of published web pages. These limitations have led to the advent of the Semantic web movement.

In the past, many organizations that created custom database application used isolated teams of developers that did not formally publish their data definitions. These teams frequently used internal data definitions that were incompatible with other computer systems. This made Enterprise Application Integration and Data warehousing extremely difficult and costly. Many organizations today require that teams consult a centralized data registry before new applications are created.

The job title of an individual that is responsible for coordinating an organization's data is a Data architect.

History

The first reference to this term was at the 1999 AAAI Ontologies Panel. The panel was organized by Chris Welty, who at the prodding of Fritz Lehmann and in collaboration with the panelists (Fritz, Mike Uschold, Mike Gruninger, and Deborah McGuinness) came up with a "spectrum" of kinds of information systems that were, at the time, referred to as ontologies. The "ontology spectrum" picture appeared in print in the introduction to Formal Ontology and Information Systems: Proceedings of the 2001 Conference. The ontology spectrum was also featured in a talk at the Semantics for the Web meeting in 2000 at Dagstuhl by Deborah McGuinness. McGuinness produced a paper describing the points on that spectrum that appeared in the book that emerged (much later) from that workshop called "Spinning the Semantic Web." Later, Leo Obrst extended the spectrum into two dimensions (which technically is not really a spectrum anymore) and added a lot more detail, which was included in his book, The Semantic Web: A Guide to the Future of XML, Web Services, and Knowledge Management.

The concept of the Semantic precision in business systems was popularized by Dave McComb in his book Semantics in Business Systems: The Savvy Managers Guide published in 2003 where he frequently uses the term Semantic Precision.

This discussion centered around a 10 level partition that included the following levels (listed in the order of increasing semantic precision):

  1. Simple Catalog of Data Elements
  2. Glossary of Terms and Definitions
  3. Thesauri, Narrow Terms, Relationships
  4. Informal "Is-a" relationships
  5. Formal "Is-a" relationships
  6. Formal instances
  7. Frames (properties)
  8. Value Restrictions
  9. Disjointness, Inverse, Part-of
  10. General Logical Constraints

Note that there was formerly a special emphasis on the adding of formal is-a relationships to the spectrum which has been dropped.

The company Cerebra has also popularized this concept by describing the data formats that exist within an enterprise in their ability to store semantically precise metadata. Their list includes:

  1. HTML
  2. PDF
  3. Word Processing documents
  4. Microsoft Excel
  5. Relational databases
  6. XML
  7. XML Schema
  8. Taxonomies
  9. Ontologies

What these concepts share in common is the ability to store information with increasing precision to facilitate intelligent agents.

See also

References

  1. Lassila, Tim Berners-Lee, James Hendler and Ora (2001-05-01). "The Semantic Web". Scientific American. Retrieved 2024-05-05.{{cite web}}: CS1 maint: multiple names: authors list (link)
Categories: