Knowledge Science

Knowledge Science

Knowledge Science is the discipline of understanding the mechanics through which humans and software-based machines "know," "learn," "change," and "adapt" their own behaviors. Throughout recorded history, knowledge has been made explicit through symbols, text and graphics on media such as clay, stone, papyrus, paper and most recently, as digitally stored representations. The digital effort began in the early 1970s when knowledge science was recognized as a vigorous field of study beginning with the development of natural language learning programs funded by the National Science Foundation (NSF). Now, knowledge science experts are engaged in a debate between:
* Meaning as represented by language-based propositions that adhere to universal truth-conditions, and
* The quantum relativist view, that meaning exists as a condition under which it can be verified and certified as acceptable without regard to universal truth-conditions.Knowledge science and knowledge representations encompass ontology, epistemology, and axiology considerations. This articles presents an overview of this new field of study, relying on historical data to provide insight and understanding, while it addresses the two schools of knowledge science.

Early Beginnings

The challenge of digitally representing knowledge (meaning) within the machine environment, as opposed to simply using language to describe who, what, when, where and how much information, started in 1969/70 under the sponsorship of the National Science Foundation (NSF). Three initial projects were founded for the study of "Natural Language Processing." These projects included the University of California, Irvine Physics Computer Development Project, headed by Alfred Bork and Research Assistant, Richard L. Ballard; the Mitre Ticit Project conducted at the University of Texas, later moved to the University of Utah; and the Plato Project, at the University of Illinois, Champaign. Over 140 natural language dialog programs were created between 1970 and 1978. UCI's California Physics Computer Development Project conducted approximately 55 educational programs and spearheaded development throughout the UC system. Initial projects were conducted on Teletype Type33, paper tape punch machine that operated at a 110 baud rate. In 1976, NSF cited Richard L. Ballard, then co-director of the Physics Computer Development Project at Irvine, for the "first application of artificial intelligence to conceptual learning" (Expert Systems)."

In 1984, Doug Lenat started the Cyc Project with the objective of codify, in machine-usable form, millions of pieces of knowledge that comprise human common sense. CycL presented a proprietary knowledge representation schema that utilized first-order relationships. Cyc [cite web
last = Lenat
first = Douglas
authorlink = Douglas Lenat
title = Hal's Legacy: 2001's Computer as Dream and Reality. From 2001 to 2001: Common Sense and the Mind of HAL
publisher = Cycorp, Inc.
url = http://www.cyc.com/cyc/technology/halslegacy.html
accessdate = 2006-09-26
] , the fruits of which were spun out of MCC into Cycorp in 1994. In 1986, he estimated the effort to complete Cyc would be 250,000 rules and 350 man-years of effort [cite book
last = The Editors of Time-Life Books
title = Understanding Computers: Artificial Intelligence
publisher = Time-Life Books
date = 1986
location = Amsterdam
pages = p84
id = ISBN 0-7054-0915-5
] . As of 2006, Lenat continues his work on Cyc at Cycorp.

This effort in 1989, permitted the Swiss High-Energy Particle Physics lab (SGML), Conseil Européenne pour la Recherche Nucleaire (CERN), to develop an instance of the Cyc project for their needs. This set into motion the work of Tim Berners-Lee, of CERN at the time, who conceived and developed anchor tags, also called hyperlinks, to link text and documents.

Thereafter in August 1991, Tim Berners-Lee released to the world in the Usenet forum alt.hypertext. This development ultimately enabled the decentralized publishing and exchange of information through the use of HTML (HTML) and its accompanying transmission protocol (HTTP).

In the same year Richard L. Ballard compiled a new technology called Mark 2, a system that utilized a single relational database table with four fields designed to store unique identifier codes that were pre-defined, stored, then later looked-up to populate the value-model/object/attribute/value table fields. Ballard called his value-based knowledge system "theory-based semantics," signaling a divide between language-based and theory-based knowledge representation systems. Mark 2 was deployed through "statements of work" with projects for the NASA, DoD, U.S. Navy and U.S. Air Force among others.

An Emerging Knowledge Science

In October 1994, the World Wide Web Consortium (W3C) was founded at the Massachusetts Institute of Technology, Laboratory for Computer Science (MIT/LCS) in collaboration with CERN, where the World Wide Web originated. This effort received support from the Defense Advanced Research Projects Agency (DARPA) and the European Commission. This organization would serve as a focus point for standardizing mechanisms of information exchange over the Internet. Thereafter, in 1995, Richard L. Ballard released Mark 2, v2.0 that utilized the earlier theory-based semantic representation technology, but included an ad hoc, 2nd order "if this, then that" knowledge representation. Mark 2 v2.0 was deployed through "statements of work" with the DoD, U.S. Navy, NASA, DEA, FAA, Office of the White House and private enterprise.

Then, in 1995, DUBLIN CORE National Center for Supercomputing Applications and the OCLC (OCLC) held a joint workshop to discuss metadata semantics in Dublin, Ohio. At this event, called simply the "OCLC/NCSA Metadata Workshop", more than 50 people discussed how a core set of semantics for Web-based resources would be extremely useful for categorizing the Web for easier search and retrieval. The participants dubbed the result "Dublin Core metadata" based on the location of the workshop. Shortly thereafter, in 1995, HTML (HTML), inspired by SGML and incorporating hyperlinks, was first standardized. This language allowed the transmission and viewing of web pages and lead to an explosion in the popularity of the Internet.

Advancing previous efforts, in 1998, XML (XML) was introduced, createdby a W3C Working Group under the leadership of
Jon Bosak. Its goal was to combine the approachable simplicity of HTML with the extensibility of SGML while avoiding the shortcomings of each. Its popularity sparked widespread interest in tagging text in an effort to advance the goal of achieving machine-based knowledge representation.

At this same time, conducting parallel research and advanced engineering on the Mark 2 product, Richard L. Ballard completed the last of 50 projects called the Avionics Prototype Tool. That project demonstrated that theory-based knowledge representation would support virtual integation of disparate database systems by modeling theory, the conditional reasoning power of the human brain, to understand and make-sense of the data content stored within database systems. This project, in turn, demonstrated that pre-defined schemas, language-based conventions, and a reliance on self-consistent logic were the cause of conventional system complexity and interoperability. In response to these findings, Ballard redirected his work to the development of a revolutionary technology called Mark 3, theory-based semantics. Mark 3 would demonstrate that knowledge could be explicitly represented within the machine environment, and that software-based machines could reason with this content to answer value-based questions the way people do.

Language-based Knowledge Representation

Language-based knowledge representation was advanced in 1999 when Resource Description Framework (RDF), an XML-based extension of an earlier 1996 PICS technology, was deployed to enhance content description. RDF drew upon submissions by Microsoft and Netscape and upon the Dublin Core/Warwick Framework. RDF is used primarily to organize and express document properties. The specific needs of different resource types, such as authorization structures or versioning, necessitated a schema language similar to XML DTD, called the RDF Schema (RDF-S) language.

On January 12, 1999, Executive Order 13111 called SCORM (Sharable Content Object Reference Model), was signed, tasking the United States Department of Defense to take the lead in working with other federal agencies and the private sector to develop common specifications and standards for technology-based learning. SCORM was developed as a way to integrate and connect the work of these organizations in support of the DoD's Advanced Distributed Learning (ADL) Initiative.

In 2001, a new markup language was specified called DAML+OIL DARPA Agent Markup Language. This new specification was based on RDF, XML and SGML. The Ontology Interface Layer, or Ontology Interchange Language (OIL) component. This new specification provided a means for description logic to be integrated into the programming language for the purpose of extracting meaning from the ontological framework.

Within two years, Web Ontology Language (OWL) emerged for defining and instantiating Web ontologies. OWL was designed for use by applications that need to process the content of information instead of just presenting information to humans. It facilitates greater machine interpretability of Web content than that supported by XML, RDF, and RDF Schema (RDF-S) by providing additional vocabulary along with a formal semantics. OWL has since evolved into OWL-S and other formats.

Currently languaged-based researchers are working on a numerical reprentation using integers to define a language-independent structured knowledge exchange.

Theory-based Knowledge Representation

Building on previous research and development, Richard L. Ballard continued the development and articulation of a Physical Theory of Knowledge and Computation delivered as a keynote presentation at the September 8-12, 2003 Seybold conference entitled: "The Future of Publishing Technology - Part 1". Within this presentation, Ballard defined the bit capacity of human thought, how theory and information produce situation awareness, and how software-based machines can faithfully represent every form of knowledge and reason with that knowledge the way people do.

References

Alfred M. Bork, Physics Computer Development Project (PCDP), Progress Report. February 19, 1971, non-journal, University of California, Irvine. (Contains a list of computer dialogues developed with a short description of each. The types of dialogs developed are: (1) development of an interactive proof, (2) assistance in problem solving, (3) diagnosing and filling in limitations in the student's mathematical background, (4)simulations, and (5) quizzes containing minimm performance standards.

Alfred Bork and Richard Ballard, The Physics Computer Development Project, Journal of College Science Teaching, Vol. II, No. 4, April 1973.

Alfred Bork, Current Status of the Physics Computer Development Project. January 3, 1975, Non-Journal ("With support from the National Science Foundation and the University of California, the Physics Computer Development Projects have produced in the last six years computer based material in a wide variety of modes. Among the major products are science learning dialogs, graphic additions to APL (A Programming Language), the underlying software, and the authoring system.)


Wikimedia Foundation. 2010.

Игры ⚽ Нужен реферат?

Look at other dictionaries:

  • Science and the Church — • Dicsusses the relationship between the two subjects Catholic Encyclopedia. Kevin Knight. 2006. Science and the Church     Science and the Church      …   Catholic encyclopedia

  • Science, technology, society and environment education — Science, technology, society and environment (STSE) education, originates from the science technology and society (STS) movement in science education. This is an outlook onyou are my amazment science education that emphasizes the teaching of… …   Wikipedia

  • knowledge — knowledge, science, learning, erudition, scholarship, information, lore are comparable when they mean what is known or can be known, usually by an individual but sometimes by human beings in general. Knowledge applies not only to a body of facts… …   New Dictionary of Synonyms

  • Science Olympiad — is a primarily American elementary, middle school, or high school team competition that requires knowledge of various science topics and engineering ability. Over 14,000 teams from 47 U.S. states and Canada compete each year.HistoryScience… …   Wikipedia

  • Science, Technology, & International Affairs — (STIA) is the study of how science and technology affect diplomacy and international relations, especially in the areas of security, environment, energy, health, business, and development.As a university subject STIA is found primarily at the… …   Wikipedia

  • Knowledge-based systems — According to the Free On line Dictionary of Computing (FOLDOC), a knowledge based system is a program for extending and/or querying a knowledge base.The [http://www.computeruser.com/resources/dictionary/ Computer User High Tech Dictionary]… …   Wikipedia

  • SCIENCE —    in English science has come to be associated with the physical sciences such as physics and chemistry but the MEANING of the term is much wider, namely the systematic classification of knowledge. Science as knowledge needs to be distinguished… …   Concise dictionary of Religion

  • Science — Sci ence, n. [F., fr. L. scientia, fr. sciens, entis, p. pr. of scire to know. Cf. {Conscience}, {Conscious}, {Nice}.] 1. Knowledge; knowledge of principles and causes; ascertained truth of facts. [1913 Webster] If we conceive God s sight or… …   The Collaborative International Dictionary of English

  • Science and British philosophy: Boyle and Newton — G.A.J.Rogers INTRODUCTION Achievements in the natural sciences in the period from Nicholas Copernicus (1473– 1543) to the death of Isaac Newton (1642–1727) changed our whole understanding of the nature of the universe and of the ways in which we… …   History of philosophy

  • Science in the Middle Ages — Science, and particularly geometry and astronomy, was linked directly to the divine for most medieval scholars. Since God created the universe after geometric and harmonic principles, to seek these principles was therefore to seek and worship God …   Wikipedia

Share the article and excerpts

Direct link
Do a right-click on the link above
and select “Copy Link”