De novo transcriptome assembly, annotation, and differential expression analysis

Overview
Creative Commons License: CC-BY Questions:
  • Which biological questions are addressed by the tutorial?

  • Which bioinformatics techniques are important to know for this type of data?

Objectives:
  • The learning objectives are the goals of the tutorial

  • They will be informed by your audience and will communicate to them and to yourself what you should focus on during the course

  • They are single sentences describing what a learner should be able to do once they have completed the tutorial

  • You can use Bloom’s Taxonomy to write effective learning objectives

Requirements:
Time estimation: 3 hours
Supporting Materials:
Published: Feb 13, 2020
Last modification: Nov 9, 2023
License: Tutorial Content is licensed under Creative Commons Attribution 4.0 International License. The GTN Framework is licensed under MIT
purl PURL: https://gxy.io/GTN:T00290
rating Rating: 2.0 (2 recent ratings, 7 all time)
version Revision: 8

As a result of the development of novel sequencing technologies, the years between 2008 and 2012 saw a large drop in the cost of sequencing. Per megabase and genome, the cost dropped to 1/100,000th and 1/10,000th of the price, respectively. Prior to this, only transcriptomes of organisms that were of broad interest and utility to scientific research were sequenced; however, these developed in 2010s high-throughput sequencing (also called next-generation sequencing) technologies are both cost- and labor- effective, and the range of organisms studied via these methods is expanding.

Examining non-model organisms can provide novel insights into the mechanisms underlying the “diversity of fascinating morphological innovations” that have enabled the abundance of life on planet Earth. In animals and plants, the “innovations” that cannot be examined in common model organisms include mimicry, mutualism, parasitism, and asexual reproduction. De novo transcriptome assembly is often the preferred method to studying non-model organisms, since it is cheaper and easier than building a genome, and reference-based methods are not possible without an existing genome. The transcriptomes of these organisms can thus reveal novel proteins and their isoforms that are implicated in such unique biological phenomena.

(source)

Agenda

In this tutorial, we will cover:

  1. Read cleaning (20 minutes)
    1. Get data
    2. Quality control
    3. Read cleaning with Trimmomatic
    4. Quality control after cleaning
  2. Assembly (120 minutes - computing)
    1. Assembly with Trinity
  3. Assembly assessment / cleaning
    1. Checking of the assembly statistics
    2. Remapping on the raw transcriptome
    3. Merge the mapping tables and compute normalizations
    4. Compute contig Ex90N50 statistic and Ex90 transcript count
    5. Transcriptome annotation completeness
    6. Filter low expression transcripts
    7. Checking of the assembly statistics after cleaning
  4. Annotation
    1. Generate gene to transcript map
    2. Peptide prediction
    3. Similarity search
    4. Find signal peptides
    5. Find transmembrane domains
    6. Search again profile database
    7. Transcriptome annotation using Trinotate
  5. Differential Expression (DE) Analysis
    1. Remapping on the filtered transcriptome using
    2. Merge the mapping tables and compute a TMM normalization
    3. RNASeq samples quality check
    4. Differential expression analysis
    5. Extract and cluster differentially expressed transcripts
    6. Partition genes into expression clusters
  6. Conclusion

Read cleaning (20 minutes)

Known sequencing biases:

  • Unknown nucleotides (Ns)
  • Bad quality nucleotides
  • Hexamers biases (Illumina. Now corrected ?)

Why do we need to correct those?

  • To remove a lot of sequencing errors (detrimental to the vast majority of assemblers)
  • Because most de-bruijn graph based assemblers can’t handle unknown nucleotides

Get data

Hands-on: Data upload
  1. Create a new history for this tutorial

    To create a new history simply click the new-history icon at the top of the history panel:

    UI for creating new history

  2. Import the 12 fq.gz into a List of Pairs collection named fastq_raw
    • Option 1: from a shared data library (ask your instructor)
    • Option 2: from Zenodo using the URLs given below

      DOI.

    https://zenodo.org/record/3541678/files/A1_left.fq.gz
    https://zenodo.org/record/3541678/files/A1_right.fq.gz
    https://zenodo.org/record/3541678/files/A2_left.fq.gz
    https://zenodo.org/record/3541678/files/A2_right.fq.gz
    https://zenodo.org/record/3541678/files/A3_left.fq.gz
    https://zenodo.org/record/3541678/files/A3_right.fq.gz
    https://zenodo.org/record/3541678/files/B1_left.fq.gz
    https://zenodo.org/record/3541678/files/B1_right.fq.gz
    https://zenodo.org/record/3541678/files/B2_left.fq.gz
    https://zenodo.org/record/3541678/files/B2_right.fq.gz
    https://zenodo.org/record/3541678/files/B3_left.fq.gz
    https://zenodo.org/record/3541678/files/B3_right.fq.gz
    
    • Copy the link location
    • Click galaxy-upload Upload Data at the top of the tool panel

    • Click on Collection on the top

    • Click on Collection Type and select List of Pairs

    • Select galaxy-wf-edit Paste/Fetch Data
    • Paste the link(s) into the text field

    • Press Start

    • Click on Build when available

    • Enter a name for the collection

      • fastq_raw
    • Click on Create list (and wait a bit)

    As an alternative to uploading the data from a URL or your computer, the files may also have been made available from a shared data library:

    1. Go into Data (top panel) then Data libraries
    2. Navigate to the correct folder as indicated by your instructor.
      • On most Galaxies tutorial data will be provided in a folder named GTN - Material –> Topic Name -> Tutorial Name.
    3. Select the desired files
    4. Click on Add to History galaxy-dropdown near the top and select as Datasets from the dropdown menu
    5. In the pop-up window, choose

      • “Select history”: the history you want to import the data to (or create a new one)
    6. Click on Import

  3. Rename the datasets
  4. Check that the datatype

    • Click on the galaxy-pencil pencil icon for the dataset to edit its attributes
    • In the central panel, click galaxy-chart-select-data Datatypes tab on the top
    • In the galaxy-chart-select-data Assign Datatype, select datatypes from “New type” dropdown
      • Tip: you can start typing the datatype into the field to filter the dropdown menu
    • Click the Save button

  5. Add to each database a tag corresponding to …

    Datasets can be tagged. This simplifies the tracking of datasets across the Galaxy interface. Tags can contain any combination of letters or numbers but cannot contain spaces.

    To tag a dataset:

    1. Click on the dataset to expand it
    2. Click on Add Tags galaxy-tags
    3. Add tag text. Tags starting with # will be automatically propagated to the outputs of tools using this dataset (see below).
    4. Press Enter
    5. Check that the tag appears below the dataset name

    Tags beginning with # are special!

    They are called Name tags. The unique feature of these tags is that they propagate: if a dataset is labelled with a name tag, all derivatives (children) of this dataset will automatically inherit this tag (see below). The figure below explains why this is so useful. Consider the following analysis (numbers in parenthesis correspond to dataset numbers in the figure below):

    1. a set of forward and reverse reads (datasets 1 and 2) is mapped against a reference using Bowtie2 generating dataset 3;
    2. dataset 3 is used to calculate read coverage using BedTools Genome Coverage separately for + and - strands. This generates two datasets (4 and 5 for plus and minus, respectively);
    3. datasets 4 and 5 are used as inputs to Macs2 broadCall datasets generating datasets 6 and 8;
    4. datasets 6 and 8 are intersected with coordinates of genes (dataset 9) using BedTools Intersect generating datasets 10 and 11.

    A history without name tags versus history with name tags

    Now consider that this analysis is done without name tags. This is shown on the left side of the figure. It is hard to trace which datasets contain “plus” data versus “minus” data. For example, does dataset 10 contain “plus” data or “minus” data? Probably “minus” but are you sure? In the case of a small history like the one shown here, it is possible to trace this manually but as the size of a history grows it will become very challenging.

    The right side of the figure shows exactly the same analysis, but using name tags. When the analysis was conducted datasets 4 and 5 were tagged with #plus and #minus, respectively. When they were used as inputs to Macs2 resulting datasets 6 and 8 automatically inherited them and so on… As a result it is straightforward to trace both branches (plus and minus) of this analysis.

    More information is in a dedicated #nametag tutorial.

Quality control

Hands-on: Task description
  1. FastQC tool with the following parameters:
    • “Short read data from your current history”: fastq_raw (collection)

Read cleaning with Trimmomatic

Hands-on: Task description
  1. Trimmomatic tool with the following parameters:
    • “Single-end or paired-end reads?”: Paired-end (as collection)
    • “Select FASTQ dataset collection with R1/R2 pair”: fastq_raw
    • “Perform initial ILLUMINACLIP step?”: Yes
    • “Adapter sequences to use”: TruSeq3 (additional seqs) (paired-ended, for MiSeq and HiSeq)
    • In “Trimmomatic Operation”:
      • param-repeat “Insert Trimmomatic Operation”
        • “Select Trimmomatic operation to perform”: Cut bases off end of a read, if below a threshold quality (TRAILING)
      • param-repeat “Insert Trimmomatic Operation”
        • “Select Trimmomatic operation to perform”: Cut bases off start of a read, if below a threshold quality (LEADING)
      • param-repeat “Insert Trimmomatic Operation”
        • “Select Trimmomatic operation to perform”: Sliding window trimming (SLIDINGWINDOW)
      • param-repeat “Insert Trimmomatic Operation”
        • “Select Trimmomatic operation to perform”: Drop reads with average quality lower than a specific level (AVGQUAL)
          • “Minimum length of reads to be kept”: 25
      • param-repeat “Insert Trimmomatic Operation”
        • “Select Trimmomatic operation to perform”: Drop reads below a specified length (MINLEN)
          • “Minimum length of reads to be kept”: 50
    • “Output trimmomatic log messages?”: Yes
  2. Rename the Dataset Collection
    • Trimmomatic on collection XX: paired -> fastqc_cleaned
    Comment

    You can check the Trimmomatic log files to get the number of read before and after the cleaning

    Input Read Pairs: 10000
    Both Surviving: 8804 (88.04%)
    Forward Only Surviving: 491 (4.91%)
    Reverse Only Surviving: 456 (4.56%) Dropped: 249 (2.49%)
    
    1. Click on the collection
    2. Click on the name of the collection at the top
    3. Change the name
    4. Press Enter

Quality control after cleaning

Hands-on: Task description
  1. FastQC tool with the following parameters:
    • “Short read data from your current history”: fastqc_cleaned (collection)

Assembly (120 minutes - computing)

Assembly with Trinity

Hands-on: Task description
  1. Trinity tool with the following parameters:
    • “Are you pooling sequence datasets?”: Yes
      • “Paired or Single-end data?”: Paired-end collection
        • “Strand specific data”: No
    • “Run in silico normalization of reads”: No
    • In “Additional Options”:
      • “Use the genome guided mode?”: No
  2. Rename the Trinity output
    • Trinity on data 52, data 51, and others: Assembled Transcripts -> transcriptome_raw.fasta
    • Click on the galaxy-pencil pencil icon for the dataset to edit its attributes
    • In the central panel, change the Name field
    • Click the Save button

Assembly assessment / cleaning

Checking of the assembly statistics

Hands-on: Task description
  1. Trinity Statistics tool with the following parameters:
    • “Trinity assembly”: transcriptome_raw.fasta
    Comment

    This step, even with this toy dataset, will take around 2 hours

Remapping on the raw transcriptome

Hands-on: Task description
  1. Align reads and estimate abundance tool with the following parameters:
    • “Transcripts”: transcriptome_raw.fasta
    • “Paired or Single-end data?”: Paired
      • “Left/Forward strand reads” -> Multiple datasets
        • Click on the Folder button at the right
          • Type to Search: left
          • Select the 6 Trimmomatic on ..._left.fq.gz
      • “Right/Reverse strand reads” -> Multiple datasets
        • Click on the Folder button at the right
          • Type to Search: right
          • Select the 6 Trimmomatic on ..._left.fq.gz
      • “Strand specific data”: Yes
    • “Abundance estimation method”: Salmon
    • In “Additional Options”:
      • “Trinity assembly?”: Yes
  2. Rename the 6 * isoforms counts :(
    • Check in the information panel (i icon) the lineage of your file (ex: A1_left.fq.gz … )
    • Rename the datasets: A1_raw, A2_raw, A3_raw, B1_raw, B2_raw, B3_raw.
    Comment

    If you check at the Standard Error messages of your outputs. You can get the Mapping rate

    1. Click on one dataset
    2. Click on the little i icon
    3. Click on Tool Standard Error: stderr
      [2019-11-14 15:44:21.500] [jointLog] [info] Mapping rate = 44.4358%
      
    Comment

    At this stage, you can now delete some useless datasets

    • Trimmomatic on collection XX: unpaired
    • Align reads and estimate abundance on *: genes counts Note that the dataset are just hidden. You can delete them permanently and make some room in the history options (the little wheel icon)

Merge the mapping tables and compute normalizations

Hands-on: Task description
  1. Build expression matrix tool with the following parameters:
    • “Abundance estimates”: A1_raw, A2_raw, A3_raw, B1_raw, B2_raw, B3_raw
    • “Abundance estimation method”: Salmon
Question

What are the three tables?

  1. estimated RNA-Seq fragment isoform counts (raw counts)`
  2. matrix of isoform TPM expression values (not cross-sample normalized)
  3. matrix of TMM-normalized expression values

Compute contig Ex90N50 statistic and Ex90 transcript count

Hands-on: Task description
  1. Compute contig Ex90N50 statistic and Ex90 transcript count tool with the following parameters:
    • “Expression matrix”: Build expression matrix: matrix of TMM-normalized expression values
    • “Transcripts”: transcriptome_raw.fasta
  2. Click on the visulization icon on the dataset Compute contig Ex90N50 statistic and Ex90 transcript count: ExN50 statistics
    1. Scatterplot - Creates a 2D-scatterplot from tabular datapoints
    2. “X Column”: select the Columns 1
    3. “Y Column”: select the Columns 2

What we get

ExN50_plot_toy_dataset.

What we should get with a real dataset

ExN50_plot. (source)

Transcriptome annotation completeness

Hands-on: Task description
  1. Busco tool with the following parameters:
    • “Sequence to analyse”: transcriptome_raw.fasta
    • “Mode”: transcriptome
    • “Lineage”: eukaryota_odb9

RNASeq samples quality check Graphs.

Filter low expression transcripts

Hands-on: Task description
  1. Filter low expression transcripts tool with the following parameters:
    • “Trinity assembly”: transcriptome_raw.fasta
    • “Expression matrix”: Build expression matrix: matrix of isoform TPM expression values (not cross-sample normalized)
    • “Minimum expression level required across any sample”: 1.0
    • “Isoform filtering method”: Keep all isoforms above a minimum percent of dominant expression
      • “Minimum percent of dominant isoform expression”: 1
    Comment

    If you check at the Standard Error messages of your outputs. You can get the Retained rate

    1. Click on one dataset
    2. Click on the little i icon
    3. Click on Tool Standard Error: stderr
       Retained 2096 / 2102 = 99.71% of total transcripts.
      
  2. Rename the output
    • Filter low expression transcripts on data 42 and data 14: filtered low expression transcripts -> transcriptome_filtered.fasta

Checking of the assembly statistics after cleaning

Hands-on: Task description
  1. Trinity Statistics tool with the following parameters:
    • “Trinity assembly”: transcriptome_filtered.fasta

Annotation

Generate gene to transcript map

Hands-on: Task description
  1. Generate gene to transcript map tool with the following parameters:
    • “Trinity assembly”: transcriptome_filtered.fasta

Peptide prediction

Hands-on: Task description
  1. TransDecoder tool with the following parameters:
    • “Transcripts”: transcriptome_filtered.fasta
    • In “Training Options”:
      • “Select the training method”: Train with the top longest ORFs
Hands-on: Task description
  1. Diamond tool with the following parameters:
    • “What do you want to align?”: Align amino acid query sequences (blastp)
    • “Input query file in FASTA or FASTQ format”: TransDecoder on data XXX: pep
    • “Select a reference database”: Uniprot Swissprot
    • “Format of output file”: BLAST Tabular
    • In “Method to restrict the number of hits?”: Maximum number of target sequences
      • “The maximum number of target sequence per query to report alignments for”: 1
  2. Rename the Diamond output
    • Diamond on data XXX -> Diamond (blastp)
  3. Diamond tool with the following parameters:
    • “What do you want to align?”: Align DNA query sequences (blastx)
    • “Input query file in FASTA or FASTQ format”: transcriptome_filtered.fasta
    • “Select a reference database”: Uniprot Swissprot
    • “Format of output file”: BLAST Tabular
    • In “Method to restrict the number of hits?”: Maximum number of target sequences
      • “The maximum number of target sequence per query to report alignments for”: 1
  4. Rename the Diamond output
    • Diamond on data XXX -> Diamond (blastx)
    Comment

    Note that you can both use Diamond tool or the NCBI BLAST+ blastp tool and NCBI BLAST+ blast tool

Find signal peptides

Hands-on: Task description
  1. SignalP 3.0 tool with the following parameters:
    • “Fasta file of protein sequences”: TransDecoder on data XXX: pep

Find transmembrane domains

Hands-on: Task description
  1. TMHMM 2.0 tool with the following parameters:
    • “FASTA file of protein sequences”: TransDecoder on data XXX: pep

Search again profile database

Hands-on: Task description
  1. hmmscan tool with the following parameters:
    • “Sequence file”: TransDecoder on data XXX: pep

Transcriptome annotation using Trinotate

Hands-on: Task description
  1. Trinotate tool with the following parameters:
    • “Transcripts”: transcriptome_filtered.fasta
    • “Peptides”: TransDecoder on data XXX: pep
    • “Genes to transcripts map”: Generate gene to transcript map on data XXX: Genes to transcripts map
    • “BLASTP: Peptides vs Uniprot.SwissProt”: Diamond (blastp)
    • “BLASTX: Transcripts vs Uniprot.SwissProt”: Diamond (blastx)
    • “HMMER hmmscan: Peptides vs PFAM”: Table of per-domain hits from HMM matches of TransDecoder on data XXX: pep against the profile database
    • “TMHMM on Peptides”: TMHMM results
    • “SignalP on Peptides”: SignalP euk results
    • “Let Galaxy downloading the Trinotate Pre-generated Resource SQLite database”: Yes

Differential Expression (DE) Analysis

Remapping on the filtered transcriptome using

Hands-on: Task description
  1. Align reads and estimate abundance tool with the following parameters:
    • “Transcripts”: transcriptome_filtered.fasta
    • “Paired or Single-end data?”: Paired
      • “Left/Forward strand reads” -> Multiple datasets
        • Click on the Folder button at the right
          • Type to Search: left
          • Select the 6 Trimmomatic on ..._left.fq.gz
      • “Right/Reverse strand reads” -> Multiple datasets
        • Click on the Folder button at the right
          • Type to Search: right
          • Select the 6 Trimmomatic on ..._left.fq.gz
      • “Strand specific data”: Yes
    • “Abundance estimation method”: Salmon
    • In “Additional Options”:
      • “Trinity assembly?”: Yes
  2. Rename the 6 * isoforms counts :(
    • Check in the information panel (i icon) the lineage of your file (ex: A1_left.fq.gz … )
    • Rename the datasets: A1, A2, A3, B1, B2, B3.
    Comment

    If you check at the Standard Error messages of your outputs. You can get the Mapping rate

    1. Click on one dataset
    2. Click on the little i icon
    3. Click on Tool Standard Error: stderr
      [2019-11-14 15:44:21.500] [jointLog] [info] Mapping rate = 44.4358%
      
    Comment

    At this stage, you can now delete some useless datasets

    • Align reads and estimate abundance on *: genes counts Note that the dataset are just hidden. You can delete them permanently and make some room in the history options (the little wheel icon)

Merge the mapping tables and compute a TMM normalization

Hands-on: Task description
  1. Build expression matrix tool with the following parameters:
    • “Abundance estimates”: A1, A2, A3, B1, B2, B3
    • “Abundance estimation method”: Salmon
  2. Describe samples and replicates tool with the following parameters:
    • “Samples”
      • “1: Samples”:
        • “Full sample name”: A1
        • “Condition”: A
      • “2: Samples”:
        • “Full sample name”: A2
        • “Condition”: A
      • …:
      • “6: Samples”:
        • “Full sample name”: B3
        • “Condition”: B

RNASeq samples quality check

Hands-on: Task description
  1. RNASeq samples quality check tool with the following parameters:
    • “Expression matrix”: Build expression matrix: estimated RNA-Seq fragment isoform counts (raw counts)
    • “Samples description”: Describe samples

Differential expression analysis

Hands-on: Task description
  1. Differential expression analysis tool with the following parameters:
    • “Expression matrix”: Build expression matrix: estimated RNA-Seq fragment isoform counts (raw counts)
    • “Sample description”: Describe samples (the last one)
    • “Differential analysis method”: DESeq2

Extract and cluster differentially expressed transcripts

Hands-on: Task description
  1. Extract and cluster differentially expressed transcripts tool with the following parameters:
    • In “Additional Options”:
      • “Expression matrix”: Build expression matrix: estimated RNA-Seq fragment isoform counts (raw counts)
      • “Sample description”: Describe samples
      • “Differential expression results”: Differential expression results on data XXX and data XXX
      • “p-value cutoff for FDR”: 1
      • “Run GO enrichment analysis”: No
    Comment

    “p-value cutoff for FDR”: 1 Don’t do this at home! It’s because we have a Toy Dataset. The cutoff should be around 0.001

Partition genes into expression clusters

Hands-on: Task description
  1. Partition genes into expression clusters tool with the following parameters:
    • “RData file”: Extract and cluster differentially expressed transcripts: RData file
    • “Method for partitioning genes into clusters”: Cut tree based on x percent of max(height) of tree

Conclusion

Sum up the tutorial and the key takeaways here. We encourage adding an overview image of the pipeline used.