Get AI summaries of any video or article — Sign up free
Semantic MediaWiki as Knowledge Graph Interface thumbnail

Semantic MediaWiki as Knowledge Graph Interface

Semantic MediaWiki·
5 min read

Based on Semantic MediaWiki's video on YouTube. If you like this content, support the original creators by watching, liking and subscribing to their content.

TL;DR

Semantic MediaWiki turns a MediaWiki instance into a knowledge management system by adding typed semantic annotations to wiki pages and enabling structured querying.

Briefing

Semantic MediaWiki positions a familiar collaborative wiki interface as a practical knowledge-graph front end—letting teams store structured entities, query them, and publish results as tables, maps, timelines, and machine-readable exports. The core pitch is that it turns a MediaWiki instance (the engine behind Wikipedia) into a flexible knowledge management system for custom domains, without forcing users to abandon the page-based workflow they already know.

At the center is an open-source extension stack that adds semantic structure to wiki pages. After installing Semantic MediaWiki, users keep the benefits of MediaWiki—versioning for every edit, collaborative editing, namespaces/categories for weak structuring, and a strong API/community—then gain the ability to attach typed properties and relationships to pages. Data can be entered directly through wiki markup, generated via templates, or collected through form-like extensions such as Page Forms. The system also supports semantic web standards, including RDF output and “triple star” support, making it easier to align a wiki-based knowledge base with knowledge-graph tooling.

The talk frames Semantic MediaWiki as a knowledge graph tool using a comprehensive knowledge-graph definition: concepts map to classes/categories, properties map to property pages (including data types), relationships are expressed through annotations in wiki text or templates, and entity descriptions live on wiki pages. It also supports schema and metadata (for example, last editor), taxonomies via categories and subcategories, and links to external identifiers and vocabularies. That structure enables FAIR-aligned practices—such as reusing controlled vocabularies and exporting data—while keeping the “instance data” anchored in the wiki’s page model.

A concrete example shows how a page about Vienna can carry typed attributes like population and coordinates, plus an external identifier (e.g., a Wikidata ID). Categories place the page into a taxonomy, while templates and forms streamline data entry. Internal querying then powers dynamic views: an Ask query can retrieve all pages in a category with a given property value (such as entities named after Sigmund Freud) and render results as tables or maps. Semantic MediaWiki offers multiple result formats, including JSON, RDF, BibTeX, and vCard, plus visualization-oriented outputs like timelines, charts, and tag clouds.

For teams that want a traditional triple store or RDF workflow, the interface can extend beyond the wiki. Semantic MediaWiki stores extra semantic data in SQL tables (often on MariaDB, SQLite, or PostgreSQL), and can optionally integrate search back ends like ElasticSearch. For SPARQL/RDF graph usage, it can connect to existing triple stores via connectors, or export RDF dumps for ingestion into external knowledge graphs. The talk also notes extensions for SPARQL querying and RDF endpoint options, while cautioning that some approaches may be less maintained.

Finally, knowledge-graph maintenance is treated as a first-class concern. The system supports “semantic gardening” through curation workflows and user rights, enabling property health checks (outdated or missing annotations, uniqueness issues, redirects/equality), and lightweight inferencing through subproperties, subcategories, and redirect-based equality. Compared with Wikibase (the technology behind Wikidata), Semantic MediaWiki is described as more flexible in its data model and more wiki-native for editing and internal querying, while both typically rely on external triple stores for SPARQL. The overall message: Semantic MediaWiki can serve as a knowledge-graph interface that keeps governance, entry, querying, and publishing inside a collaborative wiki environment—while still offering paths to RDF and triple-store ecosystems.

Cornell Notes

Semantic MediaWiki adds semantic structure to a MediaWiki site, turning wiki pages into typed entities and relationships that behave like a knowledge graph. It supports knowledge-graph concepts such as classes/categories, properties with data types, and relationships expressed through wiki annotations and templates, with RDF-oriented outputs for machine use. Internal Ask queries can retrieve and display results as tables, maps, timelines, and exports like JSON, RDF, BibTeX, and vCard. For teams needing SPARQL and triple stores, Semantic MediaWiki can connect to external RDF stores or export RDF dumps for ingestion. This combination matters because it preserves collaborative, versioned editing while enabling structured querying and knowledge-graph interoperability.

How does Semantic MediaWiki turn ordinary wiki pages into knowledge-graph elements?

It layers semantic extensions on top of MediaWiki so that each page can act as an “instance” with typed properties and relationships. Categories and subcategories provide taxonomy structure, property pages define properties (often with data types), and relationships are expressed via semantic annotations in wiki text or through templates. Entity descriptions remain on wiki pages, while schema-like information and metadata (such as last editor) can also be represented. The result is a flexible data model where the wiki page is the core container for graph nodes and their attributes.

What makes internal querying useful for knowledge-graph interfaces?

Semantic MediaWiki includes an internal query language (Ask queries) that can select pages based on category membership and property constraints, then render results in multiple formats. For example, a query can retrieve all pages in a category that have a property value matching a specific entity (like pages with a “named after” property pointing to Sigmund Freud) and display the results as a table. Changing the output format can switch the same data into map views using coordinates or into other visualization/export formats.

How does the system support data entry and governance without abandoning the wiki workflow?

Data entry can be done directly with semantic annotations, but templates and Page Forms-style extensions enable structured form-like input so contributors don’t need to write annotation markup. Governance is handled through “semantic gardening”: user rights and curation roles can manage property information and run checks for issues such as outdated or missing annotations, property uniqueness problems, and redirect/equality handling. This keeps the knowledge graph maintainable as it grows.

What interoperability options exist for RDF and SPARQL workflows?

Semantic MediaWiki stores semantic data in SQL tables (commonly MariaDB, SQLite, or PostgreSQL) and can optionally integrate ElasticSearch for improved search. For RDF/triple-store workflows, it can export RDF dumps for ingestion into external knowledge graphs, or connect directly to existing triple stores via configured connectors. SPARQL can be supported through triple-store connections, and there are extensions that provide SPARQL querying or RDF endpoint capabilities (with some noted as potentially less maintained).

How does Semantic MediaWiki compare with Wikibase in terms of modeling and editing?

Semantic MediaWiki is described as having a more flexible data model: properties can be predefined via property pages but can also be declared on the fly during annotation. Wikibase follows the Wikidata-style model with fixed property patterns (e.g., Q and P identifiers). Editing differs too: Semantic MediaWiki relies on MediaWiki-style editing plus optional Page Forms extensions, while Wikibase has a different approach to forms and typically depends on external SPARQL/triple-store infrastructure for querying.

Review Questions

  1. What mapping would you use to represent a domain’s entities, properties, and relationships inside Semantic MediaWiki (including where categories and property pages fit)?
  2. How would you design an Ask query and choose an output format (table vs map vs JSON/RDF) to support a specific knowledge-graph use case?
  3. What maintenance tasks fall under “semantic gardening,” and how do user rights and redirect/equality features help keep the graph consistent?

Key Points

  1. 1

    Semantic MediaWiki turns a MediaWiki instance into a knowledge management system by adding typed semantic annotations to wiki pages and enabling structured querying.

  2. 2

    Categories/subcategories provide taxonomy structure, while property pages define properties and data types used to annotate entities.

  3. 3

    Ask queries can retrieve entities based on category and property constraints and render results as tables, maps, timelines, and exports like JSON, RDF, BibTeX, and vCard.

  4. 4

    Templates and Page Forms-style extensions support form-based data entry, reducing the need for contributors to write annotation markup manually.

  5. 5

    Interoperability with RDF/SPARQL is supported via RDF dumps for ingestion and via configurable connectors to external triple stores.

  6. 6

    Knowledge-graph maintenance is supported through “semantic gardening,” including curation roles, property health checks, and redirect-based equality/inferencing features.

  7. 7

    Compared with Wikibase, Semantic MediaWiki emphasizes wiki-native flexibility in modeling and internal querying, while SPARQL typically still relies on external triple-store infrastructure.

Highlights

Semantic MediaWiki anchors knowledge-graph “instances” directly in wiki pages, preserving collaborative editing and versioning while adding typed semantics.
The same underlying data can be queried with Ask and displayed in multiple ways—tables, maps, timelines, and machine exports like RDF and JSON.
Knowledge-graph upkeep is treated as operational work: semantic gardening includes property health checks, uniqueness validation, and redirect-based equality.
For triple-store users, Semantic MediaWiki can either export RDF dumps or connect to external RDF stores so SPARQL workflows remain possible.

Topics

  • Semantic MediaWiki
  • Knowledge Graph Interface
  • RDF Export
  • SPARQL Connectors
  • Semantic Gardening

Mentioned