The Semantic Web

Reference

Summary

Berners-Lee, Hendler, and Lassila sketch an extension of the World Wide Web in which content carries machine-processable meaning, enabling software agents to perform non-trivial tasks on behalf of users. The article opens with the now-famous Pete-and-Lucy vignette: Lucy’s agent coordinates medical appointments, cross-references insurance providers, negotiates schedules with Pete’s agent, and replans when a preferred plan is rejected. The authors stress that this does not require HAL-level AI; it requires that web pages embed structured semantics that off-the-shelf tools can author and that agents can consume.

The vision rests on three interlocking commitments. First, meaning is expressed through structured data formats — XML for syntax, RDF for triples (subject–predicate–object with URIs), and ontologies that give classes, properties, and inference rules a shared interpretation. Second, the Web’s decentralisation is preserved: unlike traditional knowledge-representation systems that require a single central vocabulary, the Semantic Web accepts paradoxes and unanswerable questions as the price of scale, and lets ontologies be published, linked, and partially reused. Third, agents are the consumers: they roam from page to page, chain inferences across ontologies, evaluate trust via digital signatures and proofs, and negotiate on behalf of human users. The article walks through examples of rule-based inference, ontology alignment (e.g., reconciling “zip code” across sites), and the architecture of trust (“Oh, yeah?” buttons) that lets an agent verify claims by following proofs.

Throughout, the tone is programmatic. The authors concede that knowledge representation has existed since long before the Web and is “in a state comparable to that of hypertext before the advent of the Web” — useful demos but no global connective tissue. The Semantic Web’s contribution is not new KR theory but linking KR into a single global system, exactly as hypertext was the connective tissue for documents. The payoff is machine-to-machine data integration, delegated tasks, and a qualitative shift in what individual users and organisations can accomplish through agents.

Key Ideas

  • Machine-processable meaning on the Web: structured data with URI-identified terms, not just rendered documents.
  • Layered stack: URIs → XML → RDF (triples) → Ontologies (OWL-like) → Logic/Rules → Proof → Trust.
  • Decentralised knowledge representation: no central ontology; partial agreement, mapping between vocabularies.
  • Agents as consumers: the reason the Semantic Web matters is that software agents can act on the structured content, including negotiating, scheduling, verifying.
  • Ontologies provide shared classes, properties, and inference rules; critical for cross-site data integration.
  • Trust via proofs and digital signatures: an agent should be able to demand the derivation of a claim and check it.
  • Pete-and-Lucy scenario as the canonical multi-agent, multi-ontology, real-world task.
  • Hypertext analogy: KR is where hypertext was pre-Web — good ideas, no global connective layer; the Semantic Web is that layer.

Connections

Conceptual Contribution

The paper’s achievement is not a new formalism but a unifying architectural stance: meaning on the Web is cheap only if it is decentralised, layered, and consumed by agents. By pairing ontology-based data with autonomous task-performing agents, Berners-Lee, Hendler, and Lassila fuse two traditions — KR and MAS — that had been circling each other since the 1970s. For agent communication this is the moment when shared ontologies stop being a lab concern and become a web-scale precondition for agent interoperability; every subsequent discussion of common ontologies, ACL content languages, and cross-platform agent protocols inherits the Semantic Web’s layered stack and its decentralised-by-design philosophy.

Tags

#semantic-web #ontologies #knowledge-representation #agents #web-architecture #RDF #OWL #Berners-Lee #Hendler #Lassila #foundational

Backlinks