Skip to content
Snippets Groups Projects
Commit 2c1c6c6d authored by Tim O'Donnell's avatar Tim O'Donnell
Browse files

update

parent b9b80604
No related merge requests found
...@@ -42,23 +42,36 @@ python write_proteome_peptides.py \ ...@@ -42,23 +42,36 @@ python write_proteome_peptides.py \
--chromosome 1 \ --chromosome 1 \
--out proteome_peptides.chr1.csv --out proteome_peptides.chr1.csv
for kind in with_mass_spec no_mass_spec #for kind in with_mass_spec no_mass_spec
do #do
python run_mhcflurry.py \ # python run_mhcflurry.py \
proteome_peptides.chr1.csv \ # proteome_peptides.chr1.csv \
--chunk-size 100000 \ # --chunk-size 100000 \
--batch-size 65536 \ # --batch-size 65536 \
--models-dir "$(mhcflurry-downloads path models_class1_pan)/models.$kind" \ # --models-dir "$(mhcflurry-downloads path models_class1_pan)/models.$kind" \
--allele $(cat alleles.txt) \ # --allele $(cat alleles.txt) \
--out "predictions/chr1.mhcflurry.$kind" \ # --out "predictions/chr1.mhcflurry.$kind" \
--verbosity 1 \ # --verbosity 1 \
--worker-log-dir "$SCRATCH_DIR/$DOWNLOAD_NAME" \ # --worker-log-dir "$SCRATCH_DIR/$DOWNLOAD_NAME" \
--cluster-parallelism \ # --cluster-parallelism \
--cluster-max-retries 15 \ # --cluster-max-retries 15 \
--cluster-submit-command bsub \ # --cluster-submit-command bsub \
--cluster-results-workdir ~/mhcflurry-scratch \ # --cluster-results-workdir ~/mhcflurry-scratch \
--cluster-script-prefix-path $SCRIPT_DIR/cluster_submit_script_header.mssm_hpc.lsf # --cluster-script-prefix-path $SCRIPT_DIR/cluster_submit_script_header.mssm_hpc.lsf
done #done
python run_thirdparty_predictors.py \
proteome_peptides.chr1.csv \
--predictor netmhcpan4 \
--chunk-size 100000 \
--allele $(cat alleles.txt) \
--out "predictions/chr1.netmhcpan4" \
--worker-log-dir "$SCRATCH_DIR/$DOWNLOAD_NAME" \
--cluster-parallelism \
--cluster-max-retries 15 \
--cluster-submit-command bsub \
--cluster-results-workdir ~/mhcflurry-scratch \
--cluster-script-prefix-path $SCRIPT_DIR/cluster_submit_script_header.mssm_hpc.lsf
# Now all peptides # Now all peptides
python write_proteome_peptides.py \ python write_proteome_peptides.py \
...@@ -66,23 +79,38 @@ python write_proteome_peptides.py \ ...@@ -66,23 +79,38 @@ python write_proteome_peptides.py \
"${REFERENCES_DIR}/uniprot_proteins.csv.bz2" \ "${REFERENCES_DIR}/uniprot_proteins.csv.bz2" \
--out proteome_peptides.all.csv --out proteome_peptides.all.csv
for kind in with_mass_spec no_mass_spec #for kind in with_mass_spec no_mass_spec
do #do
python run_mhcflurry.py \ # python run_mhcflurry.py \
proteome_peptides.all.csv \ # proteome_peptides.all.csv \
--chunk-size 500000 \ # --chunk-size 500000 \
--batch-size 65536 \ # --batch-size 65536 \
--models-dir "$(mhcflurry-downloads path models_class1_pan)/models.$kind" \ # --models-dir "$(mhcflurry-downloads path models_class1_pan)/models.$kind" \
--allele $(cat alleles.txt) \ # --allele $(cat alleles.txt) \
--out "predictions/all.mhcflurry.$kind" \ # --out "predictions/all.mhcflurry.$kind" \
--verbosity 1 \ # --verbosity 1 \
--worker-log-dir "$SCRATCH_DIR/$DOWNLOAD_NAME" \ # --worker-log-dir "$SCRATCH_DIR/$DOWNLOAD_NAME" \
--cluster-parallelism \ # --cluster-parallelism \
--cluster-max-retries 15 \ # --cluster-max-retries 15 \
--cluster-submit-command bsub \ # --cluster-submit-command bsub \
--cluster-results-workdir ~/mhcflurry-scratch \ # --cluster-results-workdir ~/mhcflurry-scratch \
--cluster-script-prefix-path $SCRIPT_DIR/cluster_submit_script_header.mssm_hpc.lsf # --cluster-script-prefix-path $SCRIPT_DIR/cluster_submit_script_header.mssm_hpc.lsf
done #done
python run_thirdparty_predictors.py \
proteome_peptides.all.csv \
--predictor netmhcpan4 \
--chunk-size 100000 \
--allele $(cat alleles.txt) \
--out "predictions/all.netmhcpan4" \
--worker-log-dir "$SCRATCH_DIR/$DOWNLOAD_NAME" \
--cluster-parallelism \
--cluster-max-retries 15 \
--cluster-submit-command bsub \
--cluster-results-workdir ~/mhcflurry-scratch \
--cluster-script-prefix-path $SCRIPT_DIR/cluster_submit_script_header.mssm_hpc.lsf
bzip2 proteome_peptides.chr1.csv bzip2 proteome_peptides.chr1.csv
......
0% Loading or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment