The Semantic Web
Reference
- Berners-Lee, Tim; Hendler, James; & Lassila, Ora (2001). “The Semantic Web.” Scientific American 284(5): 34–43, May 17 2001.
- PDF: jmvidal.cse.sc.edu/library/berners-lee01a.pdf
- Also: scientificamerican.com/article/the-semantic-web
- Lassila’s site: lassila.org/publications/2001/SciAm.html
- Local:
semanticweb2001.pdf
Summary
Berners-Lee, Hendler, and Lassila sketch an extension of the World Wide Web in which content carries machine-processable meaning, enabling software agents to perform non-trivial tasks on behalf of users. The article opens with the now-famous Pete-and-Lucy vignette: Lucy’s agent coordinates medical appointments, cross-references insurance providers, negotiates schedules with Pete’s agent, and replans when a preferred plan is rejected. The authors stress that this does not require HAL-level AI; it requires that web pages embed structured semantics that off-the-shelf tools can author and that agents can consume.
The vision rests on three interlocking commitments. First, meaning is expressed through structured data formats — XML for syntax, RDF for triples (subject–predicate–object with URIs), and ontologies that give classes, properties, and inference rules a shared interpretation. Second, the Web’s decentralisation is preserved: unlike traditional knowledge-representation systems that require a single central vocabulary, the Semantic Web accepts paradoxes and unanswerable questions as the price of scale, and lets ontologies be published, linked, and partially reused. Third, agents are the consumers: they roam from page to page, chain inferences across ontologies, evaluate trust via digital signatures and proofs, and negotiate on behalf of human users. The article walks through examples of rule-based inference, ontology alignment (e.g., reconciling “zip code” across sites), and the architecture of trust (“Oh, yeah?” buttons) that lets an agent verify claims by following proofs.
Throughout, the tone is programmatic. The authors concede that knowledge representation has existed since long before the Web and is “in a state comparable to that of hypertext before the advent of the Web” — useful demos but no global connective tissue. The Semantic Web’s contribution is not new KR theory but linking KR into a single global system, exactly as hypertext was the connective tissue for documents. The payoff is machine-to-machine data integration, delegated tasks, and a qualitative shift in what individual users and organisations can accomplish through agents.
Key Ideas
- Machine-processable meaning on the Web: structured data with URI-identified terms, not just rendered documents.
- Layered stack: URIs → XML → RDF (triples) → Ontologies (OWL-like) → Logic/Rules → Proof → Trust.
- Decentralised knowledge representation: no central ontology; partial agreement, mapping between vocabularies.
- Agents as consumers: the reason the Semantic Web matters is that software agents can act on the structured content, including negotiating, scheduling, verifying.
- Ontologies provide shared classes, properties, and inference rules; critical for cross-site data integration.
- Trust via proofs and digital signatures: an agent should be able to demand the derivation of a claim and check it.
- Pete-and-Lucy scenario as the canonical multi-agent, multi-ontology, real-world task.
- Hypertext analogy: KR is where hypertext was pre-Web — good ideas, no global connective layer; the Semantic Web is that layer.
Connections
- Semantic Web — the cluster topic this paper names.
- Ontologies — ontologies are the load-bearing layer of the Semantic Web stack.
- Handbook On Ontologies — systematic treatment of the ontology engineering the paper demands.
- Ontolingua Portable Ontology Specifications — prior art on shareable ontology specifications that feeds into the Semantic Web plan.
- Toward Principles for the Design of Ontologies Used for Knowledge Sharing — Gruber’s principles underlie the Semantic Web’s ontology layer.
- A Common Ontology Of ACLs, ACL Design Principles — the ACL literature provides the agent-to-agent communication layer that Pete’s and Lucy’s agents would need.
- Multi-Agent Systems, Intelligent Agents Theory and Practice, On Agent-Based Software Engineering — the agent side of the vision; Jennings’s “knowledge-level interaction” is the social counterpart to the Semantic Web’s data-level interoperability.
- A Framework for Representing Knowledge — frames, slots, defaults prefigure RDF/OWL class-property structure.
- Knowledge Representation — the historical discipline whose ideas the Semantic Web proposes to “link into a single global system”.
- Abstract Agent Interfaces — interface-level agreement aligning with ontology-level agreement.
Conceptual Contribution
The paper’s achievement is not a new formalism but a unifying architectural stance: meaning on the Web is cheap only if it is decentralised, layered, and consumed by agents. By pairing ontology-based data with autonomous task-performing agents, Berners-Lee, Hendler, and Lassila fuse two traditions — KR and MAS — that had been circling each other since the 1970s. For agent communication this is the moment when shared ontologies stop being a lab concern and become a web-scale precondition for agent interoperability; every subsequent discussion of common ontologies, ACL content languages, and cross-platform agent protocols inherits the Semantic Web’s layered stack and its decentralised-by-design philosophy.
Tags
#semantic-web #ontologies #knowledge-representation #agents #web-architecture #RDF #OWL #Berners-Lee #Hendler #Lassila #foundational