Data Libraries


  • How do data libraries work?

  • What are they good for?

  • How can I use them?

  • How can I setup permissions for them?

  • Setup a data library

  • Manage permissions

  • Import data from disk

Time estimation: 30 minutes
Supporting Materials
Last modification: Jul 9, 2021
License: Tutorial Content is licensed under Creative Commons Attribution 4.0 International License The GTN Framework is MIT

Data libraries are a great way to provide structured repositories of data to everyone on your server, or to a subset of your users. Datasets in libraries do not count against user quotas, so they are commonly used to provide the results of a sequencing run or similar project to a group of users on the servers.

screenshot of data libraries


  1. Setup
  2. Importing Data
    1. from History
    2. from User Directory
    3. from import Directory (Admins only)
  3. Automatically Populating a Data Library
  4. GTN Data
  5. Using Data from Libraries


Before we can import local data, we need to configure Galaxy to permit this. Additionally we will setup an example data library which we can use for demonstrative purposes.

hands_on Hands-on: Setting up Data Libraries

  1. We will add a pre-task to clone a data repository into your machine. We will use this as the source for a library dataset.

    --- a/galaxy.yml
    +++ b/galaxy.yml
    @@ -8,6 +8,9 @@
         - name: Install Dependencies
             name: ['acl', 'bzip2', 'git', 'make', 'python3-psycopg2', 'tar', 'virtualenv']
    +    - git:
    +        repo: ''
    +        dest: /libraries/
         - galaxyproject.postgresql
         - role: natefoo.postgresql_objects
  2. Edit the file group_vars/galaxyservers.yml and set the following variables:

    --- a/group_vars/galaxyservers.yml
    +++ b/group_vars/galaxyservers.yml
    @@ -29,6 +29,8 @@ miniconda_manage_dependencies: false
    +    library_import_dir: /libraries/admin
    +    user_library_import_dir: /libraries/user
         tool_data_table_config_path: /cvmfs/,/cvmfs/
         dependency_resolvers_config_file: "{{ galaxy_config_dir }}/dependency_resolvers_conf.xml"
         containers_resolvers_config_file: "{{ galaxy_config_dir }}/container_resolvers_conf.xml"
  3. Run the playbook:

    code-in Input: Bash

    ansible-playbook galaxy.yml

Importing Data

There are multiple options for importing data from your server, we’ll go through all of your choices below. But first, let’s take a quick look at the example library structure we’ve provided.

code-in Input: Bash

tree /libraries

code-out Output: Bash

├── admin
│   └── admin-wildtype.fna
├── example-library.yaml
└── user
    │   └── user-wildtype.fna
        └── user-wildtype.fna

4 directories, 5 files

We have a directory named admin, which will be available to all admin users (we set library_import_dir: /libraries/admin earlier.)

Additionally we have a user directory, below the user directory are more directories with the user’s email as they directory key. Data can be placed in here, and it will become accessible to those users (we set user_library_import_dir: /libraries/user for this.)

An add datasets dropdown menu in galaxy showing the options from history, from user directory, and under admins only, from import directory

from History

This is by far the easiest and most convenient option for small datasets, or datasets that are just already in a history

A select box is shown listing files in the user's history

You can easily select multiple files and quickly import them.

from User Directory

If user directories are configured, as we did at the beginning of this tutorial, then users will be able to import any files under their personal directory. This can be used for a wide variety of setups, e.g. providing the output of sequencing machines to users. This can point to the same directory structure that’s used by the FTP service, if you want your users to be able to import files directly from FTP.

Import popup with a list of files with one file, user-wildtype.fna, and buttons for configuring import behaviour.

This will enable the option for everyone, any unprivileged user with a folder in the user_library_import_dir directory and permissions on a data library can import things from their import directory.

tip Tip: Do you not see a directory here?

We pre-created a directory for someone registered as, but if you are logged in with an account registered with a different email, you’ll see nothing.

You can fix this by going into /libraries/user and ln -s whatever-email-you-used Then it should appear in Galaxy.

from import Directory (Admins only)

Similarly to the user import directories, there is the option to have an admin only import area. If one admin is responsible for creating the data libraries and importing data this can be a good option.

Same as the previous image, import popup listing options and one file, admin-wildtype.fna

An important feature of data libraries importing is the ability to “link files instead of copying”. If you know your data will not be deleted (e.g. sequencing data sitting on an NFS), then you can use this option to further decrease the amount of storage your Galaxy server requires. The datasets can be used as if they were normal datasets, imported by users and analysed, but not imported into the Galaxy data storage area.

Automatically Populating a Data Library

If your data is accessible via URL, you can write a yaml file to import and setup the data library automatically. We’ve included this file in the example data libraries repository we cloned at the beginning:

code-in Input: Bash

cat /libraries/example-library.yaml

code-out Output: Bash

  type: library
  name: Mouse sequencing project
  description: some data
  synopsis: samples collected from somewhere
- url:
  src: url
  ext: fasta
- name: A directory
  description: Exome sequencing means that all protein-coding genes in a genome are
  - url:
    src: url
    ext: fastqsanger
  - url:
    src: url
    ext: fastqsanger

Let’s try setting that up in our Galaxy!

Tip: Getting your API key

Galaxy admin accounts are specified as a comma-separated email list in the admin_users directive of galaxy.yml . If you have set up your Galaxy server using the Galaxy Installation with Ansible tutorial, this is set to

  1. In your browser, open your Galaxy homepage
  2. Log in, or register a new account, if it’s the first time you’re logging in
  3. Go to User -> Preferences in the top menu bar, then click on Manage API key
  4. If there is no current API key available, click on Create a new key to generate it
  5. Copy your API key to somewhere convenient, you will need it throughout this tutorial

hands_on hands-on: importing data library yaml.

  1. re-activate the virtualenv you created for the ephemeris tool management tutorial.

    code-in input: bash

    . ~/ephemeris_venv/bin/activate

    tip missing?

    then you might need to re-run the steps:

    virtualenv -p python3 ~/ephemeris_venv
    . ~/ephemeris_venv/bin/activate
    pip install ephemeris
  2. we’ll use the setup-data-libraries command to install the data in this yaml file into a library in our galaxy.

    code-in input: bash

    setup-data-libraries -g https://your-galaxy -a <api-key> --training -i /libraries/example-library.yaml --legacy

    code-out output

    library name: mouse sequencing project
  3. this command is safe to re-run. for example if you update the yaml, it will simply report that the library exists before ensuring all files exist in their appropriate location:

    library name: mouse sequencing project
    library already exists! id: f597429621d6eb2b

<img src=”../../images/data/imported.png” alt=”Screenshot of data libraries, we’re in a library folder named “Mouse sequencing project” and a directory and single file are shown. The file has an ugly URL as its name” loading=”lazy”>

That’s it! You should be able to see your newly created data library in your Galaxy.

Got lost along the way?

If you missed any steps, you can compare against the reference files, or see what changed since the previous tutorial.

Note that we’ve used some special flags here, --training and --legacy. Training sets some defaults that make sense for the GTN (mostly around library descriptions / etc.)


This enables the use of legacy APIs even for newer Galaxies that should have a batch upload API enabled. Unfortunately the new batch upload API is not able to update existing data libraries in place and will always create new libraries with the same name. So currently --legacy is quite useful for maintaining a YAML file, and running setup-data-libraries regularly whenever that file has updates.

But this comes with the downside that the entire URL is used for the filename.

GTN Data

This process scales quite well. Galaxy Europe, in their mission to support all of the GTN materials on their server setup a shared-data repository, a single giant YAML file with all of the data from the GTN tutorials. This was then expanded to other UseGalaxy.* servers to provide synced training libraries for all of the major servers.

Do you want the training data synced on your server? If so join the shared data repository! If you provide us with a non-admin API key and a data library to upload data into, we can sync this automatically.

Using Data from Libraries

Users can now conveniently use datasets in libraries when they are running analyses, without having to import them first.

The tool form provides a button on the right of every dataset selector which allows users to “Browse datasets”

<img src=”../../images/data/select-from-lib0.png” alt=”tool form with the folder icon to the right of the dataset selector shown, the button with tooltip “Browse datasets” is highlighted” loading=”lazy”>

Users can then choose from datasets in their history, or browse through the data libraries

<img src=”../../images/data/select-from-lib1.png” alt=”Popup with “Data Libraries”, datasets from histories, and an upload button” loading=”lazy”>

Here users can see every file accessible to them in the data library to begin analysing.

Same popup as previous image, but the file listing now shows the contents of the data library we created during the hands-on step with Mouse data

Key points

  • Data libraries are a great way to share data with groups of users

Frequently Asked Questions

Have questions about this tutorial? Check out the FAQ page for the Galaxy Server administration topic to see if your question is listed there. If not, please ask your question on the GTN Gitter Channel or the Galaxy Help Forum


Did you use this material as an instructor? Feel free to give us feedback on how it went.

Click here to load Google feedback frame

Citing this Tutorial

  1. Helena Rasche, Saskia Hiltemann, 2021 Data Libraries (Galaxy Training Materials). Online; accessed TODAY
  2. Batut et al., 2018 Community-Driven Data Analysis Training for Biology Cell Systems 10.1016/j.cels.2018.05.012

details BibTeX

author = "Helena Rasche and Saskia Hiltemann",
title = "Data Libraries (Galaxy Training Materials)",
year = "2021",
month = "07",
day = "09"
url = "\url{}",
note = "[Online; accessed TODAY]"
    doi = {10.1016/j.cels.2018.05.012},
    url = {},
    year = 2018,
    month = {jun},
    publisher = {Elsevier {BV}},
    volume = {6},
    number = {6},
    pages = {752--758.e1},
    author = {B{\'{e}}r{\'{e}}nice Batut and Saskia Hiltemann and Andrea Bagnacani and Dannon Baker and Vivek Bhardwaj and Clemens Blank and Anthony Bretaudeau and Loraine Brillet-Gu{\'{e}}guen and Martin {\v{C}}ech and John Chilton and Dave Clements and Olivia Doppelt-Azeroual and Anika Erxleben and Mallory Ann Freeberg and Simon Gladman and Youri Hoogstrate and Hans-Rudolf Hotz and Torsten Houwaart and Pratik Jagtap and Delphine Larivi{\`{e}}re and Gildas Le Corguill{\'{e}} and Thomas Manke and Fabien Mareuil and Fidel Ram{\'{\i}}rez and Devon Ryan and Florian Christoph Sigloch and Nicola Soranzo and Joachim Wolff and Pavankumar Videm and Markus Wolfien and Aisanjiang Wubuli and Dilmurat Yusuf and James Taylor and Rolf Backofen and Anton Nekrutenko and Björn Grüning},
    title = {Community-Driven Data Analysis Training for Biology},
    journal = {Cell Systems}

Congratulations on successfully completing this tutorial!