Quick Start¶
Examples below use
-profile docker,gpu. Swapgpu→cpuif you don't have an NVIDIA GPU; foundation/structure models will still run on CPU, just slower (~30–60 s per inference call vs sub-second on GPU). seq2ribo and Orthrus are GPU-only and auto-skip undercpu.
How it works¶
RNAZoo is opt-in — only models you provide input for will run. No --ignore_* flags needed.
# Only RNA-FM runs (everything else is skipped automatically)
nextflow run . -profile docker,gpu --rnafm_input my_sequences.fa
Run a single model¶
RNA embeddings (RNA-FM)¶
Translation efficiency (RiboNN)¶
RNA secondary structure (RNAformer)¶
RNA 3D structure (RhoFold)¶
RNA modifications (MultiRM)¶
Run multiple models in parallel¶
Provide inputs for multiple models and they run simultaneously:
nextflow run . -profile docker,cpu \
--rnafm_input sequences.fa \
--rnaformer_input sequences.fa \
--spotrna_input sequences.fa \
--multirm_input sequences.fa
Custom output directories¶
By default, each model writes to results/<model_name>/. Override per-model or globally:
# Custom per-model output
nextflow run . -profile docker,cpu \
--rnafm_input sequences.fa \
--rnafm_outdir my_project/embeddings
# Custom global output root
nextflow run . -profile docker,cpu \
--rnafm_input sequences.fa \
--outdir my_project/results
Use a YAML params file¶
For complex runs, create a YAML file instead of passing many CLI flags. A template is provided at conf/params_template.yml:
cp conf/params_template.yml my_params.yml
# Edit my_params.yml — set _input for models you want to run
nextflow run . -profile docker,cpu -params-file my_params.yml
Example my_params.yml:
outdir: my_results
rnafm_input: data/rna_sequences.fa
rnafm_per_token: true
rnaformer_input: data/rna_sequences.fa
rnaformer_cycling: 6
multirm_input: data/rna_sequences.fa
multirm_alpha: 0.01
Fine-tuning¶
Some models support fine-tuning on your own data. The workflow is: fine-tune to produce a checkpoint, then use that checkpoint for prediction.
RiboNN — transfer learning for a new cell type¶
# Fine-tune on your TE data
nextflow run . -profile docker,cpu \
--ribonn_finetune_input my_te_data.tsv \
--ribonn_finetune_target TE_MyCondition
UTR-LM — fine-tune on custom expression data¶
# Step 1: Fine-tune
nextflow run . -profile docker,cpu \
--utrlm_finetune_input my_utr_data.csv \
--utrlm_finetune_label my_mrl \
--utrlm_finetune_task mrl
# Step 2: Predict with fine-tuned model
nextflow run . -profile docker,cpu \
--utrlm_input new_sequences.fa \
--utrlm_checkpoint results/utrlm_finetune/best_model.pt \
--utrlm_task mrl
See RiboNN and UTR-LM model pages for full details.
Run with Docker directly (no Nextflow)¶
Every model can also be run standalone with Docker:
# RNA-FM embeddings
docker run --rm \
-v /path/to/input.fa:/data/input.fa \
-v /path/to/output:/out \
ghcr.io/ericmalekos/rnazoo-rnafm:latest \
rnafm_predict.py -i /data/input.fa -o /out
See each model's page for its specific Docker command.
Output structure¶
Results are organized by model name: