From fde2f4dadb9b24638ca7b99358b78f95ca1b9bae Mon Sep 17 00:00:00 2001
From: Caughlin Bohn <caughlin.bohn@huskers.unl.edu>
Date: Mon, 13 Jul 2020 14:30:33 -0500
Subject: [PATCH] /bin/sh to /bin/bash

---
 .../blast_with_allinea_performance_reports.md             | 2 +-
 .../lammps_with_allinea_performance_reports.md            | 2 +-
 .../ray_with_allinea_performance_reports.md               | 2 +-
 .../alignment_tools/blast/create_local_blast_database.md  | 2 +-
 .../alignment_tools/blast/running_blast_alignment.md      | 4 ++--
 .../bioinformatics_tools/alignment_tools/blat.md          | 2 +-
 .../bioinformatics_tools/alignment_tools/bowtie.md        | 2 +-
 .../bioinformatics_tools/alignment_tools/bowtie2.md       | 2 +-
 .../alignment_tools/bwa/running_bwa_commands.md           | 2 +-
 .../bioinformatics_tools/alignment_tools/clustal_omega.md | 2 +-
 .../alignment_tools/tophat_tophat2.md                     | 2 +-
 .../app_specific/bioinformatics_tools/biodata_module.md   | 4 ++--
 .../bamtools/running_bamtools_commands.md                 | 2 +-
 .../samtools/running_samtools_commands.md                 | 2 +-
 .../data_manipulation_tools/sratoolkit.md                 | 2 +-
 .../bioinformatics_tools/de_novo_assembly_tools/oases.md  | 2 +-
 .../bioinformatics_tools/de_novo_assembly_tools/ray.md    | 2 +-
 .../de_novo_assembly_tools/soapdenovo2.md                 | 2 +-
 .../trinity/running_trinity_in_multiple_steps.md          | 8 ++++----
 .../velvet/running_velvet_with_paired_end_data.md         | 4 ++--
 .../running_velvet_with_single_end_and_paired_end_data.md | 4 ++--
 .../velvet/running_velvet_with_single_end_data.md         | 4 ++--
 .../bioinformatics_tools/pre_processing_tools/cutadapt.md | 2 +-
 .../bioinformatics_tools/pre_processing_tools/prinseq.md  | 4 ++--
 .../bioinformatics_tools/pre_processing_tools/scythe.md   | 2 +-
 .../bioinformatics_tools/pre_processing_tools/sickle.md   | 4 ++--
 .../pre_processing_tools/tagcleaner.md                    | 2 +-
 .../app_specific/bioinformatics_tools/qiime.md            | 2 +-
 .../reference_based_assembly_tools/cufflinks.md           | 2 +-
 .../removing_detecting_redundant_sequences/cap3.md        | 2 +-
 .../removing_detecting_redundant_sequences/cd_hit.md      | 2 +-
 content/applications/app_specific/dmtcp_checkpointing.md  | 4 ++--
 content/applications/app_specific/fortran_c_on_hcc.md     | 4 ++--
 content/applications/app_specific/mpi_jobs_on_hcc.md      | 4 ++--
 .../applications/app_specific/running_gaussian_at_hcc.md  | 6 +++---
 .../running_ocean_land_atmosphere_model_olam.md           | 4 ++--
 content/applications/user_software/using_singularity.md   | 4 ++--
 content/submitting_jobs/job_dependencies.md               | 8 ++++----
 38 files changed, 58 insertions(+), 58 deletions(-)

diff --git a/content/applications/app_specific/allinea_profiling_and_debugging/allinea_performance_reports/blast_with_allinea_performance_reports.md b/content/applications/app_specific/allinea_profiling_and_debugging/allinea_performance_reports/blast_with_allinea_performance_reports.md
index ec4291ed..d0a891d8 100644
--- a/content/applications/app_specific/allinea_profiling_and_debugging/allinea_performance_reports/blast_with_allinea_performance_reports.md
+++ b/content/applications/app_specific/allinea_profiling_and_debugging/allinea_performance_reports/blast_with_allinea_performance_reports.md
@@ -9,7 +9,7 @@ with Allinea Performance Reports (`perf-report`) on Crane is shown below:
 
 {{% panel theme="info" header="blastn_perf_report.submit" %}}
 {{< highlight batch >}}
-#!/bin/sh
+#!/bin/bash
 #SBATCH --job-name=BlastN
 #SBATCH --nodes=1
 #SBATCH --ntasks=16
diff --git a/content/applications/app_specific/allinea_profiling_and_debugging/allinea_performance_reports/lammps_with_allinea_performance_reports.md b/content/applications/app_specific/allinea_profiling_and_debugging/allinea_performance_reports/lammps_with_allinea_performance_reports.md
index cdb77498..eae0db2a 100644
--- a/content/applications/app_specific/allinea_profiling_and_debugging/allinea_performance_reports/lammps_with_allinea_performance_reports.md
+++ b/content/applications/app_specific/allinea_profiling_and_debugging/allinea_performance_reports/lammps_with_allinea_performance_reports.md
@@ -9,7 +9,7 @@ below:
 
 {{% panel theme="info" header="lammps_perf_report.submit" %}}
 {{< highlight batch >}}
-#!/bin/sh
+#!/bin/bash
 #SBATCH --job-name=LAMMPS
 #SBATCH --ntasks=64
 #SBATCH --time=12:00:00
diff --git a/content/applications/app_specific/allinea_profiling_and_debugging/allinea_performance_reports/ray_with_allinea_performance_reports.md b/content/applications/app_specific/allinea_profiling_and_debugging/allinea_performance_reports/ray_with_allinea_performance_reports.md
index 28b25aa5..1a21e017 100644
--- a/content/applications/app_specific/allinea_profiling_and_debugging/allinea_performance_reports/ray_with_allinea_performance_reports.md
+++ b/content/applications/app_specific/allinea_profiling_and_debugging/allinea_performance_reports/ray_with_allinea_performance_reports.md
@@ -8,7 +8,7 @@ with Allinea PerformanceReports (`perf-report`) is shown below:
 
 {{% panel theme="info" header="ray_perf_report.submit" %}}
 {{< highlight batch >}}
-#!/bin/sh
+#!/bin/bash
 #SBATCH --job-name=Ray
 #SBATCH --ntasks-per-node=16
 #SBATCH --time=10:00:00
diff --git a/content/applications/app_specific/bioinformatics_tools/alignment_tools/blast/create_local_blast_database.md b/content/applications/app_specific/bioinformatics_tools/alignment_tools/blast/create_local_blast_database.md
index 4caf0d48..d01dfee4 100644
--- a/content/applications/app_specific/bioinformatics_tools/alignment_tools/blast/create_local_blast_database.md
+++ b/content/applications/app_specific/bioinformatics_tools/alignment_tools/blast/create_local_blast_database.md
@@ -15,7 +15,7 @@ where **input_reads.fasta** is the input file containing all sequences that need
 Simple example of how **makeblastdb** can be run on Crane using SLURM script and nucleotide database is shown below:
 {{% panel header="`blast_db.submit`"%}}
 {{< highlight bash >}}
-#!/bin/sh
+#!/bin/bash
 #SBATCH --job-name=Blast_DB
 #SBATCH --nodes=1
 #SBATCH --ntasks-per-node=1
diff --git a/content/applications/app_specific/bioinformatics_tools/alignment_tools/blast/running_blast_alignment.md b/content/applications/app_specific/bioinformatics_tools/alignment_tools/blast/running_blast_alignment.md
index e4de1150..6ed8935c 100644
--- a/content/applications/app_specific/bioinformatics_tools/alignment_tools/blast/running_blast_alignment.md
+++ b/content/applications/app_specific/bioinformatics_tools/alignment_tools/blast/running_blast_alignment.md
@@ -52,7 +52,7 @@ Basic SLURM example of nucleotide BLAST run against the non-redundant **nt** BL
 
 {{% panel header="`blastn_alignment.submit`"%}}
 {{< highlight bash >}}
-#!/bin/sh
+#!/bin/bash
 #SBATCH --job-name=BlastN
 #SBATCH --nodes=1
 #SBATCH --ntasks-per-node=8
@@ -92,7 +92,7 @@ Basic SLURM example of protein BLAST run against the non-redundant **nr **BLAS
 
 {{% panel header="`blastx_alignment.submit`"%}}
 {{< highlight bash >}}
-#!/bin/sh
+#!/bin/bash
 #SBATCH --job-name=BlastX
 #SBATCH --nodes=1
 #SBATCH --ntasks-per-node=8
diff --git a/content/applications/app_specific/bioinformatics_tools/alignment_tools/blat.md b/content/applications/app_specific/bioinformatics_tools/alignment_tools/blat.md
index 9113bca2..6d41e9c9 100644
--- a/content/applications/app_specific/bioinformatics_tools/alignment_tools/blat.md
+++ b/content/applications/app_specific/bioinformatics_tools/alignment_tools/blat.md
@@ -24,7 +24,7 @@ $ blat
 Running BLAT on Crane with query file `input_reads.fasta` and database `db.fa` is shown below:
 {{% panel header="`blat_alignment.submit`"%}}
 {{< highlight bash >}}
-#!/bin/sh
+#!/bin/bash
 #SBATCH --job-name=Blat
 #SBATCH --nodes=1
 #SBATCH --ntasks-per-node=1
diff --git a/content/applications/app_specific/bioinformatics_tools/alignment_tools/bowtie.md b/content/applications/app_specific/bioinformatics_tools/alignment_tools/bowtie.md
index 41aaf03d..ae79a451 100644
--- a/content/applications/app_specific/bioinformatics_tools/alignment_tools/bowtie.md
+++ b/content/applications/app_specific/bioinformatics_tools/alignment_tools/bowtie.md
@@ -28,7 +28,7 @@ Bowtie supports both single-end (`input_reads.[fasta|fastq]`) and paired-end (`
 An example of how to run Bowtie alignment on Crane with single-end fastq file and `8 CPUs` is shown below:
 {{% panel header="`bowtie_alignment.submit`"%}}
 {{< highlight bash >}}
-#!/bin/sh
+#!/bin/bash
 #SBATCH --job-name=Bowtie
 #SBATCH --nodes=1
 #SBATCH --ntasks-per-node=8
diff --git a/content/applications/app_specific/bioinformatics_tools/alignment_tools/bowtie2.md b/content/applications/app_specific/bioinformatics_tools/alignment_tools/bowtie2.md
index 8d091433..f6054dbf 100644
--- a/content/applications/app_specific/bioinformatics_tools/alignment_tools/bowtie2.md
+++ b/content/applications/app_specific/bioinformatics_tools/alignment_tools/bowtie2.md
@@ -34,7 +34,7 @@ where **index_prefix** is the generated index using the **bowtie2-build** co
 An example of how to run Bowtie2 local alignment on Crane with paired-end fasta files and `8 CPUs` is shown below:
 {{% panel header="`bowtie2_alignment.submit`"%}}
 {{< highlight bash >}}
-#!/bin/sh
+#!/bin/bash
 #SBATCH --job-name=Bowtie2
 #SBATCH --nodes=1
 #SBATCH --ntasks-per-node=8
diff --git a/content/applications/app_specific/bioinformatics_tools/alignment_tools/bwa/running_bwa_commands.md b/content/applications/app_specific/bioinformatics_tools/alignment_tools/bwa/running_bwa_commands.md
index 02ad6683..7e96edbc 100644
--- a/content/applications/app_specific/bioinformatics_tools/alignment_tools/bwa/running_bwa_commands.md
+++ b/content/applications/app_specific/bioinformatics_tools/alignment_tools/bwa/running_bwa_commands.md
@@ -25,7 +25,7 @@ where **index_prefix** is the index for the reference genome generated from **bw
 Simple SLURM script for running **bwa mem** on Crane with paired-end fastq input data, `index_prefix` as reference genome index, SAM output file and `8 CPUs` is shown below:
 {{% panel header="`bwa_mem.submit`"%}}
 {{< highlight bash >}}
-#!/bin/sh
+#!/bin/bash
 #SBATCH --job-name=Bwa_Mem
 #SBATCH --nodes=1
 #SBATCH --ntasks-per-node=8
diff --git a/content/applications/app_specific/bioinformatics_tools/alignment_tools/clustal_omega.md b/content/applications/app_specific/bioinformatics_tools/alignment_tools/clustal_omega.md
index 6aeebfae..95b625b2 100644
--- a/content/applications/app_specific/bioinformatics_tools/alignment_tools/clustal_omega.md
+++ b/content/applications/app_specific/bioinformatics_tools/alignment_tools/clustal_omega.md
@@ -33,7 +33,7 @@ $ clustalo -h
 Running Clustal Omega on Crane with input file `input_reads.fasta` with `8 threads` and `10GB memory` is shown below:
 {{% panel header="`clustal_omega.submit`"%}}
 {{< highlight bash >}}
-#!/bin/sh
+#!/bin/bash
 #SBATCH --job-name=Clustal_Omega
 #SBATCH --nodes=1
 #SBATCH --ntasks-per-node=8
diff --git a/content/applications/app_specific/bioinformatics_tools/alignment_tools/tophat_tophat2.md b/content/applications/app_specific/bioinformatics_tools/alignment_tools/tophat_tophat2.md
index b8047568..c0590f6a 100644
--- a/content/applications/app_specific/bioinformatics_tools/alignment_tools/tophat_tophat2.md
+++ b/content/applications/app_specific/bioinformatics_tools/alignment_tools/tophat_tophat2.md
@@ -30,7 +30,7 @@ Prior running TopHat/TopHat2, an index from the reference genome should be built
 An example of how to run TopHat2 on Crane with paired-end fastq files `input_reads_pair_1.fastq` and `input_reads_pair_2.fastq`, reference index `index_prefix` and `8 CPUs` is shown below:
 {{% panel header="`tophat2_alignment.submit`"%}}
 {{< highlight bash >}}
-#!/bin/sh
+#!/bin/bash
 #SBATCH --job-name=Tophat2
 #SBATCH --nodes=1
 #SBATCH --ntasks-per-node=8
diff --git a/content/applications/app_specific/bioinformatics_tools/biodata_module.md b/content/applications/app_specific/bioinformatics_tools/biodata_module.md
index c4b7c0a3..97750365 100644
--- a/content/applications/app_specific/bioinformatics_tools/biodata_module.md
+++ b/content/applications/app_specific/bioinformatics_tools/biodata_module.md
@@ -43,7 +43,7 @@ $ ls $BLAST
 An example of how to run Bowtie2 local alignment on Crane utilizing the default Horse, *Equus caballus* index (*BOWTIE2\_HORSE*) with paired-end fasta files and 8 CPUs is shown below:
 {{% panel header="`bowtie2_alignment.submit`"%}}
 {{< highlight bash >}}
-#!/bin/sh
+#!/bin/bash
 #SBATCH --job-name=Bowtie2
 #SBATCH --nodes=1
 #SBATCH --ntasks-per-node=8
@@ -64,7 +64,7 @@ bowtie2 -x $BOWTIE2_HORSE -f -1 input_reads_pair_1.fasta -2 input_reads_pair_2.f
 An example of BLAST run against the non-redundant nucleotide database available on Crane is provided below:
 {{% panel header="`blastn_alignment.submit`"%}}
 {{< highlight bash >}}
-#!/bin/sh
+#!/bin/bash
 #SBATCH --job-name=BlastN
 #SBATCH --nodes=1
 #SBATCH --ntasks-per-node=8
diff --git a/content/applications/app_specific/bioinformatics_tools/data_manipulation_tools/bamtools/running_bamtools_commands.md b/content/applications/app_specific/bioinformatics_tools/data_manipulation_tools/bamtools/running_bamtools_commands.md
index 07a02752..8cc866a1 100644
--- a/content/applications/app_specific/bioinformatics_tools/data_manipulation_tools/bamtools/running_bamtools_commands.md
+++ b/content/applications/app_specific/bioinformatics_tools/data_manipulation_tools/bamtools/running_bamtools_commands.md
@@ -19,7 +19,7 @@ where the option **-format** specifies the type of the output file, **input_a
 Running BamTools **convert** on Crane with input file `input_alignments.bam` and output file `output_reads.fastq` is shown below:
 {{% panel header="`bamtools_convert.submit`"%}}
 {{< highlight bash >}}
-#!/bin/sh
+#!/bin/bash
 #SBATCH --job-name=BamTools_Convert
 #SBATCH --nodes=1
 #SBATCH --ntasks-per-node=1
diff --git a/content/applications/app_specific/bioinformatics_tools/data_manipulation_tools/samtools/running_samtools_commands.md b/content/applications/app_specific/bioinformatics_tools/data_manipulation_tools/samtools/running_samtools_commands.md
index 0a9787f7..c5e5c2f5 100644
--- a/content/applications/app_specific/bioinformatics_tools/data_manipulation_tools/samtools/running_samtools_commands.md
+++ b/content/applications/app_specific/bioinformatics_tools/data_manipulation_tools/samtools/running_samtools_commands.md
@@ -17,7 +17,7 @@ where **input_alignments.[bam|sam]** is the input file with the alignments in BA
 Running **samtools view** on Crane with `8 CPUs`, input file `input_alignments.sam` with available header (**-S**), output in BAM format (**-b**) and output file `output_alignments.bam` is shown below:
 {{% panel header="`samtools_view.submit`"%}}
 {{< highlight bash >}}
-#!/bin/sh
+#!/bin/bash
 #SBATCH --job-name=SAMtools_View
 #SBATCH --nodes=1
 #SBATCH --ntasks-per-node=8
diff --git a/content/applications/app_specific/bioinformatics_tools/data_manipulation_tools/sratoolkit.md b/content/applications/app_specific/bioinformatics_tools/data_manipulation_tools/sratoolkit.md
index d2704b9f..250963ff 100644
--- a/content/applications/app_specific/bioinformatics_tools/data_manipulation_tools/sratoolkit.md
+++ b/content/applications/app_specific/bioinformatics_tools/data_manipulation_tools/sratoolkit.md
@@ -21,7 +21,7 @@ $ fastq-dump [options] input_reads.sra
 An example of running **fastq-dump** on Crane to convert SRA file containing paired-end reads is:
 {{% panel header="`sratoolkit.submit`"%}}
 {{< highlight bash >}}
-#!/bin/sh
+#!/bin/bash
 #SBATCH --job-name=SRAtoolkit
 #SBATCH --nodes=1
 #SBATCH --ntasks-per-node=1
diff --git a/content/applications/app_specific/bioinformatics_tools/de_novo_assembly_tools/oases.md b/content/applications/app_specific/bioinformatics_tools/de_novo_assembly_tools/oases.md
index 82278e88..14458383 100644
--- a/content/applications/app_specific/bioinformatics_tools/de_novo_assembly_tools/oases.md
+++ b/content/applications/app_specific/bioinformatics_tools/de_novo_assembly_tools/oases.md
@@ -28,7 +28,7 @@ Oases has a lot of parameters that can be found in its [manual](https://www.ebi.
 A simple SLURM script to run Oases on the Velvet output stored in `output_directory/` with minimum transcript length of `200` is shown below:
 {{% panel header="`oases.submit`"%}}
 {{< highlight bash >}}
-#!/bin/sh
+#!/bin/bash
 #SBATCH --job-name=Velvet_Oases
 #SBATCH --nodes=1
 #SBATCH --ntasks-per-node=1
diff --git a/content/applications/app_specific/bioinformatics_tools/de_novo_assembly_tools/ray.md b/content/applications/app_specific/bioinformatics_tools/de_novo_assembly_tools/ray.md
index 7e57db1f..99d4f63b 100644
--- a/content/applications/app_specific/bioinformatics_tools/de_novo_assembly_tools/ray.md
+++ b/content/applications/app_specific/bioinformatics_tools/de_novo_assembly_tools/ray.md
@@ -41,7 +41,7 @@ Ray supports odd values for k-mer equal to or greater than 21 (`-k <kmer_value>`
 Simple SLURM script for running Ray with both paired-end and single-end data with `k-mer=31`, `8 CPUs` and `4 GB RAM per CPU` is shown below:
 {{% panel header="`ray.submit`"%}}
 {{< highlight bash >}}
-#!/bin/sh
+#!/bin/bash
 #SBATCH --job-name=Ray
 #SBATCH --ntasks=8
 #SBATCH --time=168:00:00
diff --git a/content/applications/app_specific/bioinformatics_tools/de_novo_assembly_tools/soapdenovo2.md b/content/applications/app_specific/bioinformatics_tools/de_novo_assembly_tools/soapdenovo2.md
index d4248506..eae39ef8 100644
--- a/content/applications/app_specific/bioinformatics_tools/de_novo_assembly_tools/soapdenovo2.md
+++ b/content/applications/app_specific/bioinformatics_tools/de_novo_assembly_tools/soapdenovo2.md
@@ -97,7 +97,7 @@ After creating the configuration file **configFile**, the next step is to run th
 Simple SLURM script for running SOAPdenovo2 with `k-mer=31`, `8 CPUSs` and `50GB of RAM` is shown below:
 {{% panel header="`soapdenovo2.submit`"%}}
 {{< highlight bash >}}
-#!/bin/sh
+#!/bin/bash
 #SBATCH --job-name=SOAPdenovo2
 #SBATCH --nodes=1
 #SBATCH --ntasks-per-node=8
diff --git a/content/applications/app_specific/bioinformatics_tools/de_novo_assembly_tools/trinity/running_trinity_in_multiple_steps.md b/content/applications/app_specific/bioinformatics_tools/de_novo_assembly_tools/trinity/running_trinity_in_multiple_steps.md
index 086a9ed4..22fa8383 100644
--- a/content/applications/app_specific/bioinformatics_tools/de_novo_assembly_tools/trinity/running_trinity_in_multiple_steps.md
+++ b/content/applications/app_specific/bioinformatics_tools/de_novo_assembly_tools/trinity/running_trinity_in_multiple_steps.md
@@ -10,7 +10,7 @@ weight = "10"
 The first step of running Trinity is to run Trinity with the option **--no_run_inchworm**:
 {{% panel header="`trinity_step1.submit`"%}}
 {{< highlight bash >}}
-#!/bin/sh
+#!/bin/bash
 #SBATCH --job-name=Trinity_Step1
 #SBATCH --nodes=1
 #SBATCH --ntasks-per-node=8
@@ -29,7 +29,7 @@ Trinity --seqType fq --max_memory 100G --left input_reads_pair_1.fastq --right i
 The second step of running Trinity is to run Trinity with the option **--no_run_chrysalis**:
 {{% panel header="`trinity_step2.submit`"%}}
 {{< highlight bash >}}
-#!/bin/sh
+#!/bin/bash
 #SBATCH --job-name=Trinity_Step2
 #SBATCH --nodes=1
 #SBATCH --ntasks-per-node=8
@@ -48,7 +48,7 @@ Trinity --seqType fq --max_memory 100G --left input_reads_pair_1.fastq --right i
 The third step of running Trinity is to run Trinity with the option **--no_distributed_trinity_exec**:
 {{% panel header="`trinity_step3.submit`"%}}
 {{< highlight bash >}}
-#!/bin/sh
+#!/bin/bash
 #SBATCH --job-name=Trinity_Step3
 #SBATCH --nodes=1
 #SBATCH --ntasks-per-node=8
@@ -67,7 +67,7 @@ Trinity --seqType fq --max_memory 100G --left input_reads_pair_1.fastq --right i
 The fourth step of running Trinity is to run Trinity without any additional option:
 {{% panel header="`trinity_step4.submit`"%}}
 {{< highlight bash >}}
-#!/bin/sh
+#!/bin/bash
 #SBATCH --job-name=Trinity_Step4
 #SBATCH --nodes=1
 #SBATCH --ntasks-per-node=8
diff --git a/content/applications/app_specific/bioinformatics_tools/de_novo_assembly_tools/velvet/running_velvet_with_paired_end_data.md b/content/applications/app_specific/bioinformatics_tools/de_novo_assembly_tools/velvet/running_velvet_with_paired_end_data.md
index 2f52cc87..6e6dc24c 100644
--- a/content/applications/app_specific/bioinformatics_tools/de_novo_assembly_tools/velvet/running_velvet_with_paired_end_data.md
+++ b/content/applications/app_specific/bioinformatics_tools/de_novo_assembly_tools/velvet/running_velvet_with_paired_end_data.md
@@ -10,7 +10,7 @@ weight = "10"
 The first step of running Velvet is to run **velveth**:
 {{% panel header="`velveth.submit`"%}}
 {{< highlight bash >}}
-#!/bin/sh
+#!/bin/bash
 #SBATCH --job-name=Velvet_Velveth
 #SBATCH --nodes=1
 #SBATCH --ntasks-per-node=8
@@ -30,7 +30,7 @@ velveth output_directory/ 43 -fastq -longPaired -separate input_reads_pair_1.fas
 After running **velveth**, the next step is to run **velvetg** on the `output_directory/` and files generated from **velveth**:
 {{% panel header="`velvetg.submit`"%}}
 {{< highlight bash >}}
-#!/bin/sh
+#!/bin/bash
 #SBATCH --job-name=Velvet_Velvetg
 #SBATCH --nodes=1
 #SBATCH --ntasks-per-node=8
diff --git a/content/applications/app_specific/bioinformatics_tools/de_novo_assembly_tools/velvet/running_velvet_with_single_end_and_paired_end_data.md b/content/applications/app_specific/bioinformatics_tools/de_novo_assembly_tools/velvet/running_velvet_with_single_end_and_paired_end_data.md
index fa355b11..187e020b 100644
--- a/content/applications/app_specific/bioinformatics_tools/de_novo_assembly_tools/velvet/running_velvet_with_single_end_and_paired_end_data.md
+++ b/content/applications/app_specific/bioinformatics_tools/de_novo_assembly_tools/velvet/running_velvet_with_single_end_and_paired_end_data.md
@@ -10,7 +10,7 @@ weight = "10"
 The first step of running Velvet is to run **velveth**:
 {{% panel header="`velveth.submit`"%}}
 {{< highlight bash >}}
-#!/bin/sh
+#!/bin/bash
 #SBATCH --job-name=Velvet_Velveth
 #SBATCH --nodes=1
 #SBATCH --ntasks-per-node=8
@@ -30,7 +30,7 @@ velveth output_directory/ 51 -fasta -short input_reads.fasta -fasta -shortPaired
 After running **velveth**, the next step is to run **velvetg** on the `output_directory/` and files generated from **velveth**:
 {{% panel header="`velvetg.submit`"%}}
 {{< highlight bash >}}
-#!/bin/sh
+#!/bin/bash
 #SBATCH --job-name=Velvet_Velvetg
 #SBATCH --nodes=1
 #SBATCH --ntasks-per-node=8
diff --git a/content/applications/app_specific/bioinformatics_tools/de_novo_assembly_tools/velvet/running_velvet_with_single_end_data.md b/content/applications/app_specific/bioinformatics_tools/de_novo_assembly_tools/velvet/running_velvet_with_single_end_data.md
index 82efa872..4528bc6c 100644
--- a/content/applications/app_specific/bioinformatics_tools/de_novo_assembly_tools/velvet/running_velvet_with_single_end_data.md
+++ b/content/applications/app_specific/bioinformatics_tools/de_novo_assembly_tools/velvet/running_velvet_with_single_end_data.md
@@ -10,7 +10,7 @@ weight = "10"
 The first step of running Velvet is to run **velveth**:
 {{% panel header="`velveth.submit`"%}}
 {{< highlight bash >}}
-#!/bin/sh
+#!/bin/bash
 #SBATCH --job-name=Velvet_Velveth
 #SBATCH --nodes=1
 #SBATCH --ntasks-per-node=8
@@ -30,7 +30,7 @@ velveth output_directory/ 31 -fasta -short input_reads.fasta
 After running **velveth**, the next step is to run **velvetg** on the `output_directory/` and files generated from **velveth**:
 {{% panel header="`velvetg.submit`"%}}
 {{< highlight bash >}}
-#!/bin/sh
+#!/bin/bash
 #SBATCH --job-name=Velvet_Velvetg
 #SBATCH --nodes=1
 #SBATCH --ntasks-per-node=8
diff --git a/content/applications/app_specific/bioinformatics_tools/pre_processing_tools/cutadapt.md b/content/applications/app_specific/bioinformatics_tools/pre_processing_tools/cutadapt.md
index 6f8770f0..d0e82b28 100644
--- a/content/applications/app_specific/bioinformatics_tools/pre_processing_tools/cutadapt.md
+++ b/content/applications/app_specific/bioinformatics_tools/pre_processing_tools/cutadapt.md
@@ -25,7 +25,7 @@ $ cutadapt --help
 Simple Cutadapt script that trims the adapter sequences **AGGCACACAGGG** and **TGAGACACGCA** from the 3' end and **AACCGGTT** from the 5' end of single-end fasta input file is shown below:
 {{% panel header="`cutadapt.submit`"%}}
 {{< highlight bash >}}
-#!/bin/sh
+#!/bin/bash
 #SBATCH --job-name=Cutadapt
 #SBATCH --nodes=1
 #SBATCH --ntasks-per-node=1
diff --git a/content/applications/app_specific/bioinformatics_tools/pre_processing_tools/prinseq.md b/content/applications/app_specific/bioinformatics_tools/pre_processing_tools/prinseq.md
index 09d2fd44..ea0ad94e 100644
--- a/content/applications/app_specific/bioinformatics_tools/pre_processing_tools/prinseq.md
+++ b/content/applications/app_specific/bioinformatics_tools/pre_processing_tools/prinseq.md
@@ -27,7 +27,7 @@ The output format (`-out_format`) can be **1** (fasta only), **2** (fasta and qu
 Simple PRINSEQ SLURM script for single-end fasta data and fasta output format is shown below:
 {{% panel header="`prinseq_single_end.submit`"%}}
 {{< highlight bash >}}
-#!/bin/sh
+#!/bin/bash
 #SBATCH --job-name=PRINSEQ
 #SBATCH --nodes=1
 #SBATCH --ntasks-per-node=1
@@ -59,7 +59,7 @@ The output format (`-out_format`) can be **1** (fasta only), **2** (fasta and qu
 Simple PRINSEQ SLURM script for paired-end fastq data and fastq output format is shown below:
 {{% panel header="`prinseq_paired_end.submit`"%}}
 {{< highlight bash >}}
-#!/bin/sh
+#!/bin/bash
 #SBATCH --job-name=PRINSEQ
 #SBATCH --nodes=1
 #SBATCH --ntasks-per-node=1
diff --git a/content/applications/app_specific/bioinformatics_tools/pre_processing_tools/scythe.md b/content/applications/app_specific/bioinformatics_tools/pre_processing_tools/scythe.md
index e907b013..f19e2bb0 100644
--- a/content/applications/app_specific/bioinformatics_tools/pre_processing_tools/scythe.md
+++ b/content/applications/app_specific/bioinformatics_tools/pre_processing_tools/scythe.md
@@ -25,7 +25,7 @@ $ scythe --help
 Simple Scythe script that uses the `illumina_adapters.fa` file and `input_reads.fastq` is shown below:
 {{% panel header="`scythe.submit`"%}}
 {{< highlight bash >}}
-#!/bin/sh
+#!/bin/bash
 #SBATCH --job-name=Scythe
 #SBATCH --nodes=1
 #SBATCH --ntasks-per-node=1
diff --git a/content/applications/app_specific/bioinformatics_tools/pre_processing_tools/sickle.md b/content/applications/app_specific/bioinformatics_tools/pre_processing_tools/sickle.md
index 7550b450..53726405 100644
--- a/content/applications/app_specific/bioinformatics_tools/pre_processing_tools/sickle.md
+++ b/content/applications/app_specific/bioinformatics_tools/pre_processing_tools/sickle.md
@@ -28,7 +28,7 @@ where **input_reads.fastq** is the input file of sequencing data in fastq form
 Simple SLURM Sickle script for Illumina single-end reads input file `input_reads.fastq` is shown below:
 {{% panel header="`sickle_single.submit`"%}}
 {{< highlight bash >}}
-#!/bin/sh
+#!/bin/bash
 #SBATCH --job-name=Sickle
 #SBATCH --nodes=1
 #SBATCH --ntasks-per-node=1
@@ -56,7 +56,7 @@ where **input_reads_pair_1.fastq** and **input_reads_pair_2.fastq** are the in
 Simple SLURM Sickle script for Sanger paired-end reads input files `input_reads_pair_1.fastq` and `input_reads_pair_2.fastq` is shown below:
 {{% panel header="`sickle_paired.submit`"%}}
 {{< highlight bash >}}
-#!/bin/sh
+#!/bin/bash
 #SBATCH --job-name=Sickle
 #SBATCH --nodes=1
 #SBATCH --ntasks-per-node=1
diff --git a/content/applications/app_specific/bioinformatics_tools/pre_processing_tools/tagcleaner.md b/content/applications/app_specific/bioinformatics_tools/pre_processing_tools/tagcleaner.md
index eeb9a893..6619c5a7 100644
--- a/content/applications/app_specific/bioinformatics_tools/pre_processing_tools/tagcleaner.md
+++ b/content/applications/app_specific/bioinformatics_tools/pre_processing_tools/tagcleaner.md
@@ -25,7 +25,7 @@ $ tagcleaner.pl --help
 Simple TagCleaner script for removing known 3' and 5' tag sequences (`NNNCCAAACACACCCAACACA` and `TGTGTTGGGTGTGTTTGGNNN` respectively) is shown below:
 {{% panel header="`tagcleaner.submit`"%}}
 {{< highlight bash >}}
-#!/bin/sh
+#!/bin/bash
 #SBATCH --job-name=TagCleaner
 #SBATCH --nodes=1
 #SBATCH --ntasks-per-node=1
diff --git a/content/applications/app_specific/bioinformatics_tools/qiime.md b/content/applications/app_specific/bioinformatics_tools/qiime.md
index dfec13f0..8d3e2c54 100644
--- a/content/applications/app_specific/bioinformatics_tools/qiime.md
+++ b/content/applications/app_specific/bioinformatics_tools/qiime.md
@@ -35,7 +35,7 @@ Sample QIIME submit script to run **pick_open_reference_otus.py** is:
 
 {{% panel header="`qiime.submit`"%}}
 {{< highlight bash >}}
-#!/bin/sh
+#!/bin/bash
 #SBATCH --job-name=QIIME
 #SBATCH --nodes=1
 #SBATCH --ntasks-per-node=8
diff --git a/content/applications/app_specific/bioinformatics_tools/reference_based_assembly_tools/cufflinks.md b/content/applications/app_specific/bioinformatics_tools/reference_based_assembly_tools/cufflinks.md
index 711471e1..b48c94f6 100644
--- a/content/applications/app_specific/bioinformatics_tools/reference_based_assembly_tools/cufflinks.md
+++ b/content/applications/app_specific/bioinformatics_tools/reference_based_assembly_tools/cufflinks.md
@@ -22,7 +22,7 @@ $ cufflinks -h
 An example of how to run Cufflinks on Crane with alignment file in SAM format, output directory `cufflinks_output` and 8 CPUs is shown below:
 {{% panel header="`cufflinks.submit`"%}}
 {{< highlight bash >}}
-#!/bin/sh
+#!/bin/bash
 #SBATCH --job-name=Cufflinks
 #SBATCH --nodes=1
 #SBATCH --ntasks-per-node=8
diff --git a/content/applications/app_specific/bioinformatics_tools/removing_detecting_redundant_sequences/cap3.md b/content/applications/app_specific/bioinformatics_tools/removing_detecting_redundant_sequences/cap3.md
index c3f17b64..0bc7928e 100644
--- a/content/applications/app_specific/bioinformatics_tools/removing_detecting_redundant_sequences/cap3.md
+++ b/content/applications/app_specific/bioinformatics_tools/removing_detecting_redundant_sequences/cap3.md
@@ -21,7 +21,7 @@ An example of how to run basic CAP3 SLURM script on Crane is shown
 below:
 {{% panel header="`cap3.submit`"%}}
 {{< highlight bash >}}
-#!/bin/sh
+#!/bin/bash
 #SBATCH --job-name=CAP3
 #SBATCH --nodes=1
 #SBATCH --ntasks-per-node=1
diff --git a/content/applications/app_specific/bioinformatics_tools/removing_detecting_redundant_sequences/cd_hit.md b/content/applications/app_specific/bioinformatics_tools/removing_detecting_redundant_sequences/cd_hit.md
index 4fa711fc..c905ce22 100644
--- a/content/applications/app_specific/bioinformatics_tools/removing_detecting_redundant_sequences/cd_hit.md
+++ b/content/applications/app_specific/bioinformatics_tools/removing_detecting_redundant_sequences/cd_hit.md
@@ -35,7 +35,7 @@ CD-HIT is multi-threaded program, and therefore, using multiple threads is recom
 Simple SLURM CD-HIT script for Crane with 8 CPUs is given in addition:
 {{% panel header="`cd-hit.submit`"%}}
 {{< highlight bash >}}
-#!/bin/sh
+#!/bin/bash
 #SBATCH --job-name=CD-HIT
 #SBATCH --nodes=1
 #SBATCH --ntasks-per-node=8
diff --git a/content/applications/app_specific/dmtcp_checkpointing.md b/content/applications/app_specific/dmtcp_checkpointing.md
index 9de1c055..19786d37 100644
--- a/content/applications/app_specific/dmtcp_checkpointing.md
+++ b/content/applications/app_specific/dmtcp_checkpointing.md
@@ -66,7 +66,7 @@ on crane is shown below:
 
 {{% panel theme="info" header="dmtcp_blastx.submit" %}}
 {{< highlight batch >}}
-#!/bin/sh
+#!/bin/bash
 #SBATCH --job-name=BlastX
 #SBATCH --nodes=1
 #SBATCH --ntasks=8
@@ -98,7 +98,7 @@ following submit file:
 
 {{% panel theme="info" header="dmtcp_restart_blastx.submit" %}}
 {{< highlight batch >}}
-#!/bin/sh
+#!/bin/bash
 #SBATCH --job-name=BlastX
 #SBATCH --nodes=1
 #SBATCH --ntasks=8
diff --git a/content/applications/app_specific/fortran_c_on_hcc.md b/content/applications/app_specific/fortran_c_on_hcc.md
index 877bd5d0..9a2c002f 100644
--- a/content/applications/app_specific/fortran_c_on_hcc.md
+++ b/content/applications/app_specific/fortran_c_on_hcc.md
@@ -123,7 +123,7 @@ line.
 
 {{% panel header="`submit_f.serial`"%}}
 {{< highlight bash >}}
-#!/bin/sh
+#!/bin/bash
 #SBATCH --mem-per-cpu=1024
 #SBATCH --time=00:01:00
 #SBATCH --job-name=Fortran
@@ -137,7 +137,7 @@ module load compiler/gcc/4.9
 
 {{% panel header="`submit_c.serial`"%}}
 {{< highlight bash >}}
-#!/bin/sh
+#!/bin/bash
 #SBATCH --mem-per-cpu=1024
 #SBATCH --time=00:01:00
 #SBATCH --job-name=C
diff --git a/content/applications/app_specific/mpi_jobs_on_hcc.md b/content/applications/app_specific/mpi_jobs_on_hcc.md
index 18201221..5d3792e5 100644
--- a/content/applications/app_specific/mpi_jobs_on_hcc.md
+++ b/content/applications/app_specific/mpi_jobs_on_hcc.md
@@ -212,7 +212,7 @@ main program name.
 
 {{% panel header="`submit_f.mpi`"%}}
 {{< highlight bash >}}
-#!/bin/sh
+#!/bin/bash
 #SBATCH --ntasks=5
 #SBATCH --mem-per-cpu=1024
 #SBATCH --time=00:01:00
@@ -226,7 +226,7 @@ mpirun ./demo_f_mpi.x
 
 {{% panel header="`submit_c.mpi`"%}}
 {{< highlight bash >}}
-#!/bin/sh
+#!/bin/bash
 #SBATCH --ntasks=5
 #SBATCH --mem-per-cpu=1024
 #SBATCH --time=00:01:00
diff --git a/content/applications/app_specific/running_gaussian_at_hcc.md b/content/applications/app_specific/running_gaussian_at_hcc.md
index 45aa43e6..a903fe15 100644
--- a/content/applications/app_specific/running_gaussian_at_hcc.md
+++ b/content/applications/app_specific/running_gaussian_at_hcc.md
@@ -96,7 +96,7 @@ Content of Gaussian SLURM submission file `run-g09-general.slurm`:
 
 {{% panel theme="info" header="run-g09-general.slurm" %}}
 {{< highlight batch >}}
-#!/bin/sh
+#!/bin/bash
 #SBATCH -J g09
 #SBATCH --nodes=1 --ntasks-per-node=4
 #SBATCH --mem-per-cpu=2000
@@ -164,7 +164,7 @@ Submit your initial **g09** job with the following SLURM submission file:
 
 {{% panel theme="info" header="Submit with dmtcp" %}}
 {{< highlight batch >}}
-#!/bin/sh
+#!/bin/bash
 #SBATCH -J g09-dmtcp
 #SBATCH --nodes=1 --ntasks-per-node=16
 #SBATCH --mem-per-cpu=4000
@@ -214,7 +214,7 @@ resume your interrupted job:
 
 {{% panel theme="info" header="Resume with dmtcp" %}}
 {{< highlight batch >}}
-#!/bin/sh
+#!/bin/bash
 #SBATCH -J g09-restart
 #SBATCH --nodes=1 --ntasks-per-node=16
 #SBATCH --mem-per-cpu=4000
diff --git a/content/applications/app_specific/running_ocean_land_atmosphere_model_olam.md b/content/applications/app_specific/running_ocean_land_atmosphere_model_olam.md
index 2efd86d9..67dee23d 100644
--- a/content/applications/app_specific/running_ocean_land_atmosphere_model_olam.md
+++ b/content/applications/app_specific/running_ocean_land_atmosphere_model_olam.md
@@ -206,7 +206,7 @@ USE_HDF5=1
 
 {{% panel theme="info" header="Sample submit script for PGI compiler" %}}
 {{< highlight batch >}}
-#!/bin/sh
+#!/bin/bash
 #SBATCH --ntasks=8                                          # 8 cores
 #SBATCH --mem-per-cpu=1024                                  # Minimum memory required per CPU (in megabytes)
 #SBATCH --time=03:15:00                                     # Run time in hh:mm:ss
@@ -223,7 +223,7 @@ mpirun /path/to/olam-4.2c-mpi
 
 {{% panel theme="info" header="Sample submit script for Intel compiler" %}}
 {{< highlight batch >}}
-#!/bin/sh
+#!/bin/bash
 #SBATCH --ntasks=8                                          # 8 cores
 #SBATCH --mem-per-cpu=1024                                  # Minimum memory required per CPU (in megabytes)
 #SBATCH --time=03:15:00                                     # Run time in hh:mm:ss
diff --git a/content/applications/user_software/using_singularity.md b/content/applications/user_software/using_singularity.md
index 801f3d62..c5521bf2 100644
--- a/content/applications/user_software/using_singularity.md
+++ b/content/applications/user_software/using_singularity.md
@@ -78,7 +78,7 @@ Using Singularity in a SLURM job is similar to how you would use any other softw
 
 {{% panel theme="info" header="Example Singularity SLURM script" %}}
 {{< highlight bash >}}
-#!/bin/sh
+#!/bin/bash
 #SBATCH --time=03:15:00          # Run time in hh:mm:ss
 #SBATCH --mem-per-cpu=4096       # Maximum memory required per CPU (in megabytes)
 #SBATCH --job-name=singularity-test
@@ -201,7 +201,7 @@ For example,
 
 {{% panel theme="info" header="Example SLURM script" %}}
 {{< highlight bash >}}
-#!/bin/sh
+#!/bin/bash
 #SBATCH --time=03:15:00          # Run time in hh:mm:ss
 #SBATCH --mem-per-cpu=4096       # Maximum memory required per CPU (in megabytes)
 #SBATCH --job-name=singularity-test
diff --git a/content/submitting_jobs/job_dependencies.md b/content/submitting_jobs/job_dependencies.md
index 36aa971e..b0dd6b7b 100644
--- a/content/submitting_jobs/job_dependencies.md
+++ b/content/submitting_jobs/job_dependencies.md
@@ -34,7 +34,7 @@ The SLURM submit files for each step are below.
 
 {{%expand "JobA.submit" %}}
 {{< highlight batch >}}
-#!/bin/sh
+#!/bin/bash
 #SBATCH --job-name=JobA
 #SBATCH --time=00:05:00
 #SBATCH --ntasks=1
@@ -49,7 +49,7 @@ sleep 120
 
 {{%expand "JobB.submit" %}}
 {{< highlight batch >}}
-#!/bin/sh
+#!/bin/bash
 #SBATCH --job-name=JobB
 #SBATCH --time=00:05:00
 #SBATCH --ntasks=1
@@ -66,7 +66,7 @@ sleep 120
 
 {{%expand "JobC.submit" %}}
 {{< highlight batch >}}
-#!/bin/sh
+#!/bin/bash
 #SBATCH --job-name=JobC
 #SBATCH --time=00:05:00
 #SBATCH --ntasks=1
@@ -83,7 +83,7 @@ sleep 120
 
 {{%expand "JobC.submit" %}}
 {{< highlight batch >}}
-#!/bin/sh
+#!/bin/bash
 #SBATCH --job-name=JobD
 #SBATCH --time=00:05:00
 #SBATCH --ntasks=1
-- 
GitLab