Running Flair with many samples #396
Labels
Documentation
Update ReadTheDocs manual
duplicate
This issue or pull request already exists
enhancement
New feature or request
mod; collapse
I have read the paper (https://doi.org/10.1038/s41467-020-15171-6)
and the manual (https://flair.readthedocs.io/en/latest/) and I still have a question about
Dear developers for Flair,
Thank you very much for providing this tool. I have a question concerning how to run flair collapse.
I have 167 samples (nanopore; transcriptome; human), each with a fastq file of 20-40GB (unzipped). After running flair correct, concatenating the corrected bed files and splitting according to chromosomes, the resulting bed files ranged from 240 MB to 24 GB. What would be the best way to run collapse in this case? Do I really need to feed all the 167 fastq files? How much resource (core, memory, time) should be expected?
For the record, I tried running collapse using the following command. But it stopped after 30 hours, exceeding the memory limit. I allocated 16 cores and 200 GB memory. I fed all the 167 fastq files.
flair collapse -t 16 -o $out -g $ref --gtf $gtf -q $bed -r $fas --stringent --check_splice --generate_map --annotation_reliant generate
Thank you for your help.
Best regards,
Shenglin
The text was updated successfully, but these errors were encountered: