Skip to content

Use Cases

QLever comes with a number of ready-to-use Qleverfiles for a variety of datasets. To get one of these Qleverfiles, run qlever setup-config <name>; for example qlever setup-config wikidata. Using that Qleverfile, you can then download the data, load it into QLever, and start a QLever server using simple qlever commands as described in the quickstart.

The following table provides an overview of the available Qleverfiles and the associated system requirements. The numbers measured on a machine with an AMD Ryzen 9 9950X processor (16 cores), 128 GB of RAM, and sufficient disk space. The disk consumption from QLever (excluding the dataset itself) during index building is slightly larger than the final index size (twice as much is always sufficient). No more RAM is needed to queries than for building the index. The RAM consumption at query time can be limited in the settings.

Dataset Endpoint Number of triples Index size RAM (Indexing) Indexing time Description
dblp Official (QLever) ~1.5 B ~25 GB ~20 GB ~30 m DBLP is a computer science bibliography. This dataset contains the DBLP dataset with citations from OpenCitations.
dbpedia Official, QLever Demo ~845 M - - ~20 m The DBpedia dataset contains information extracted from the structured Wikipedia elements like the infoboxes.
default N/A N/A N/A N/A A minimal QLeverfile that can be used to build a custom configuration for your dataset.
dnb Official (QLever) ~240 M - ~5 GB ~5 m The German National Library (DNB) is the central archival library for Germany. The dataset contains information on the publications in the DNB's inventory.
fbeasy QLever Demo ~362 M - ~10 GB ~10 m FBEasy is a simplified and cleaned version of the freebase dataset.
freebase QLever Demo ~3.1 B - ~20 GB ~4 m Freebase was a collaborative knowledge base, which became read-only in 2015, and was then superseded by Wikidata
imdb QLever Demo ~382 M - ~5 GB ~5 m A dataset containing basic movie and ratings information from the film and series database IMDb. The dataset is available at https://datasets.imdbws.com/.
ohm-planet QLever Demo ~5.8 B ~60 GB ~10 GB ~1 h The OpenHistoricalMap is a OpenStreeMap like dataset of the world throughout history. The data modelling is similar to the osm-planet dataset. osm2rdf is required to build the index.
olympics QLever Demo ~1.8 M ~20 MB ~1 GB ~10 s A demo dataset with historic data on the olympic winter and summer games from Athens 1896 to Rio 2016. More information can be found in the repository of this dataset.
orkg QLever Demo ~3 M - - N/A The Open Research Knowledge Graph (ORKG) describes research papers mainly from biology and computer science.
osm-country N/A N/A N/A N/A A template for running an engine with OpenStreetMap data for a single country. The data model is identical to the osm-planet dataset. osm2rdf is required to build the index.Adjust CONTINENT and COUNTRY to your needs. All available continents and countries can be found at Geofabrik.
osm-planet QLever Demo ~204 B ~1.5 TB ~60 GB ~20 h OpenStreetMap is an open-source map. This is the complete OpenStreetMap planet dataset including nodes, ways and relations. The geometries can be queried using GeoSPARQL. Uses an RDF dump of the dataset from https://osm2rdf.cs.uni-freiburg.de which was converted using osm2rdf.
osm-planet-from-pbf QLever Demo ~204 B ~2.5 TB ~60 GB ~40 h The osm-planet dataset with the difference that the data ist generated directly from the latest dump which requires osm2rdf.
pubchem QLever Demo ~26 B ~350 GB ~20 GB ~6 h PubChem is a database of chemical data.
scientists QLever Demo ~370 M - - ~20 s Scientists and associated triples and text (used for QLever-internal testing)
uniprot QLever Demo ~240 B ~ 3 TB ~60 GB ~40 h UniProt is a database of proteins.
vvz This dataset is used internally and not publicly available. The dataset contains information on the teaching (lectures, lecturers, rooms, examination regulations) at the faculty of engineering at the University Freiburg. In the future this dataset will power the faculties course catalog.
wikidata QLever Demo ~ 21 B ~500 GB ~20 GB ~5 h Wikidata is a collaborative knowledge base (think Wikipedia for structured data) under the umbrella of the Wikimedia fundation.
wikipathways - - - ~20 s Wikipathways is a database of biological pathways. The latest snapshot will be used by default. Set RELEASE to a snapshot from https://data.wikipathways.org/ to used a different snapshot.
yago-4 QLever Demo ~2.5 B - ~10 GB ~4 h Yago-4 is built on Wikidata but simplifies the data. The changes to Wikidata include huma-readable identifiers and different classes and properties.