Releases: HKU-BAL/Clair3
Releases · HKU-BAL/Clair3
v0.1-r9
v0.1-r8
v0.1-r7
- Increased
var_pct_full
in ONT mode from 0.3 to 0.7. Indel F1-score increased ~0.2%, but took ~30 minutes longer to finish calling a ~50x ONT dataset. - Expand fall through to next most likely variant if network prediction has insufficient read coverage (#53 commit 09a7d18, contributor @ftostevin-ont), accuracy improved on complex Indels.
- Streamized pileup and full-alignment training workflows. Reduce diskspace demand in model training (#55 commit 09a7d18, contributor @ftostevin-ont).
- Added
mini_epochs
option in Train.py, performance slightly improved in training a model for ONT Q20 data using mini-epochs(#60, contributor @ftostevin-ont). - Massively reduced disk space demand when outputting GVCF. Now compressing GVCF intermediate files with lz4, five times smaller with little speed penalty.
- Added
--remove_intermediate_dir
to remove intermediate files as soon as no longer needed (#48). - Renamed ONT pre-trained models with Medaka's naming convention.
- Fixed training data spilling over to validation data (#57).
v0.1-r6
v0.1-r5
- Modified data generator in model training to avoid memory exhaustion and unexpected segmentation fault by Tensorflow (contributor @ftostevin-ont ).
- Simplified dockerfile workflow to reuse container caching (contributor @amblina).
- Fixed ALT output for reference calls (contributor @wdecoster).
- Fixed a bug in multi-allelic AF computation (AF of [ACGT]Del variants was wrong before r5).
- Added AD tag to the GVCF output.
- Added the
--call_snp_only
option to only call SNP only (#40). - Added pileup and full-alignment output validity check to avoid workflow crashing (#32, #38).
v0.1-r4
- Install via bioconda.
- Added an ONT Guppy2 model to the images (ont_guppy2). Click here for more benchmarking results. The results show you have to use the Guppy2 model for Guppy2 or earlier data. 3. Added google colab notebooks for quick demo. 4. Fixed a bug then there are too few variant candidates (#28).
v0.1-r3
- Added ulimit -u (max user processes) check (lowers the
THREADS
if the resource is insufficient) and automatic retries on failed jobs (#20, #23, #24). - Added support for user-defined pileup model prefix
--pileup_model_prefix
and full-alignment model prefix--fa_model_prefix
. - Added contig validity check. Contigs specified in BED but not in BAM won't be used.
- Added an ONT Guppy5 model (ont_guppy5).
v0.1-r2
v0.1-r1
v0.1
Initial release