There are different ways to install Wikibase. The most common are:

When you install Wikibase and your Wikibase Installation is public, please register you installation in the Wikibase Registry

Install Wikibase using Git and Composer

This screencast shows how to install Wikibase using Git and Composer:


Prerequisites to install a Wikibase instance:

  • A web server with PHP and a database server. The detailed requirements are listed on
  • Git version control system
  • Composer dependency manager

Get MediaWiki and extension files

  1. Run git clone --branch REL1_32 mediawiki in your console or command line. This will clone the MediaWiki version 1.32 into a subdirectory “?mediawiki” of your current directory. (Be sure to checkout the current MediaWiki version before running the command and replace REL1_32 with the most current MediaWiki version.)
  2. Enter the subdirectory (cd mediawiki) and run composer install --no-dev, which installs all dependencies required by MediaWiki. MediaWiki extension may be installed using composer install <package>:<version> as done in the following. Extensions are listed on and packages of the extensions are also distributed through the Composer packages repository for being able to easily install them using Composer.
  3. To install the Vector skin, which is the default of Wikipedia, run composer install mediawiki/vector-skin:dev-REL1_32. This will download the Vector skin files and all dependencies.

    The specified version (dev-REL1_32) must match the version of MediaWiki (with all MediaWiki extensions featuring a dev- prefix). Otherwise, compatibility issues may occur.

  4. To install the Universal Language Selector that allows specifying labels and descriptions in a vast range of languages, run composer install mediawiki/universal-language-selector:dev-REL1_32.
  5. To install Wikibase, run composer install wikibase/wikibase:dev-REL1_32.

Set up MediaWiki and extension

Now it’s time to set up MediaWiki using the web interface.

  1. In your browser, navigate to the path MediaWiki was installed into as served by the web server, e.g.: http://localhost/mediawiki. You are prompted with a hint that let’s you start the setup process. One of the more relevant parts of the setup is to specify the database MediaWiki should install it’s database tables to.
  2. After completing the setup, you are prompted to download the file LocalSettings.php and copy that into the installation directory of MediaWiki. Having done so, open LocalSettings.php with a text editor, because the extensions installed above still need to be activated.
  3. While the Vector skin is activated automatically, the Universal Language Selector needs to be activated manually by adding the line wfLoadExtension( 'UniversalLanguageSelector' ); at the bottom of LocalSettings.php.
  4. To activate the Wikibase Repo, add the following lines at the bottom of LocalSettings.php:

    $wgEnableWikibaseRepo = true;
    $wgEnableWikibaseClient = false;
    require_once "$IP/extensions/Wikibase/repo/Wikibase.php";
    require_once "$IP/extensions/Wikibase/repo/ExampleSettings.php";

    For setting up a Wikibase Client or a Repo/Client, see instructions on

  5. Wikibase requires its own tables in the database to store its information. In order to create these tables required by Wikibase switch back to your console or command line and run php maintenance/update.php in your MediaWiki installation directory.

Verify Wikibase installation

Wikibase ist now installed, set up, and ready to use. For verifying your installation, access you Wiki in the browser and click Special pages in the Tools section of the menu. At the bottom of the page is a section titled Wikibase. Clicking Create a new item let’s you create you first Wikibase item.

Install Wikibase with Docker

This Video will show installing the Wikibase Docker image:


All information about the Docker installation can be found in the Wikibase Docker image repository.

Installing a stand-alone Wikibase instance and SPARQL Query Service with Docker

The Wikibase Docker image repository contains an example docker compose file that can be used with the images also contained in this repo. Following these instructions, you’ll be able to install a stand-alone instance of Wikibase, as well as fully functional SPARQL endpoint and query service, complete with a data visualisation frontend and query helper (similar to the Wikidata Query Service).

Individual documentation for each image used in this example can be found in the Wikibase Docker image repository Readme file.

WARNING: Currently this example requires 3GB+ of memory. If you are running Docker for Windows or Mac and have containers run within a VM (the default) make sure you set your VM memory allocation to 4GB. If you don’t do this, the query service may fail to run.

Installing Docker and the Wikibase images

You’ll need to install Docker, Docker Compose and clone images from this repository. For a single script to run that should work on a Wikimedia Toolforge (wmflabs) VM running Debian Jessie with 4GB memory, take a look at in this repository.

1) Install Docker (community edition) & Docker compose

Note: Docker Compose is included in the Docker community edition, so most likely you won’t need to install it separately.

2) Copy the docker-compose.yml file

This file is only meant as an example, your modification might be needed.


If you want to develop the Dockerfiles then you should clone this repo and instead use the docker-compose-build.yml file.

3) Setting up and starting the containers

Note: In linux you must either be in the docker group or use sudo!

Pulling / updating the images
docker-compose pull
Starting the containers
docker-compose up

To set up quickstatements follow the instructions in the quickstatements README

Stopping the containers
docker-compose stop
Removing the containers
docker-compose down

This will keep all data stored by MySQL, Mediawiki and the Query Service in Docker volumes.

Removing both the containers and data

WARNING: this will remove ALL of the data you had added to Wikibase and the Query Service.

docker-compose down --volumes

Backing up data from Docker volumes

All data for Wikibase and the Query Service is stored in Docker volumes. You can create compressed copies of these volumes to use as backups or to hand off to other users.

Volume backups will only work if you use the same image when restoring / using the backup data. If you are backing up your mysql data you may also want to just take an SQL dump.

You can see all Docker volumes created by using the following command:

docker volume ls | grep wikibase-docker

You can grab a zip of each volumes by doing the following:

docker run -v wikibase-docker_mediawiki-mysql-data:/volume -v /tmp/wikibase-data:/backup --rm loomchild/volume-backup backup mediawiki-mysql-data

You, or someone else, can then restore the volume by doing the following:

docker run -v wikibase-docker_mediawiki-mysql-data:/volume -v /tmp/wikibase-data:/backup --rm loomchild/volume-backup restore mediawiki-mysql-data

Backing up data using mysqldump

If using volume-backup for the database does not work because of InnoDB tables (check here), you can achieve data backup using mysqldump.

Backing up with mysqldump

docker exec wikibase-docker_mysql_1 mysqldump -u wikiuser -psqlpass my_wiki > backup.sql

Restoring from mysqldump backup file

docker exec wikibase-docker_mysql_1 mysql -u wikiuser -psqlpass my_wiki < backup.sql

Accessing your Wikibase instance and the Query Service UI

Access the following hosts:

Creating content and querying it

You can start creating items and properties to start populating your Wikibase instance via these special pages

Once you’ve created your first items and statements, you’ll be able to query them and visualize the results through the Query Service UI.

Exporting data as a JSON or RDF dump

Get a JSON dump from wikibase:

docker-compose exec wikibase php ./extensions/Wikibase/repo/maintenance/dumpJson.php

Get an RDF dump from Wikibase:

docker-compose exec wikibase php ./extensions/Wikibase/repo/maintenance/dumpRdf.php


Using Open Stack to run a custom Wikibase instance

This Video will show you, how to do run a custom Wikibase with OpenStack.


To get all information please read Using Open Stack to run custom Wikibase.

Getting Started

  1. Create an account, if you don’t have one already
  2. Add billing details
  3. Click the button to setup horizon
  4. Navigate to “Project”

    Compute Instances Launch Instance

  5. Name: wikibase1, Any availability zone, count 1
  6. Source, latest ubuntu (currently 18.04)
  7. Flavour: c1.large (medium should also work)
  8. Key pair, create one and download the PEM :)

    Launch Instance

Add a public IP

  1. From the instances page, click on the actions menu on the right hand side and click “Associate floating IP”
  2. Click the “+” button to add a new IP
  3. Click “Allocate IP && Associate”

Add firewall rules

  1. Navigate to Project » Compute » Access & Security » Security Groups
  2. Click on “Manage Rules” for the default group
  3. Allow Ingress on port 22 (for ssh) for your IP (or all IPs, depending on how secure you want to be)
  4. Allow Ingress on the ports for Wikibase and Query Service UI access (defaults are at 8181 and 8282) from all IPs

Login to the instance

ssh ubuntu@ # where the IP is the public IP you assigned

Install deps

sudo apt-get update
  1. Install Docker, follow the instructions on
  2. Install Docker Compose, follow the instructions on

Get your Docker Compose

In the future there will be a lovely UI that you can download a docker-compose.yml. Currently you have to look at the example and work from that replacing various default values.

  • All values that should be replaced are included next to “CONFIG” comments in the yaml
  • wget # Make your modifications, or just download your modified file
sudo docker-compose up -d #Wait for a few seconds
sudo docker-compose ps #check the status

Finally, there will be an empty Wikibase ready for use. When the Wikibase Cloud screen is visible, it is ready to receive input. If the instance runs on the same infrastructure as Wikidata, linked data can be added using the same data model allowing for better integration between Wikidata and the freshly installed Wikibase instance. We are now able to populate this Wikibase instance running on the Fuga Cloud server by using the same Python framework used to add data to Wikidata.