Metagenomics: rationale behind Q trimming
1
0
Entering edit mode
5.7 years ago
f.a.galkin ▴ 40

I am wondering on how I should trim my reads for a metagenomics study. Surely, I don't need extra stringent conditions as Q=30. But how low can I get to keep my reads numerous and long? I have considered checking average quality first and then trimming, say, at half the average quality. But then I'd have different trimming settings for different runs.

What are your ideas on balancing sensitivity/specificity in this case?

illumina bbduk quality trimming metagenomics • 1.5k views
ADD COMMENT
2
Entering edit mode
5.7 years ago

I am wondering on how I should trim my reads for a metagenomics study.

Don't trim your data :) As long as you don't have a significant drop-off of the quality values you would throw away informations that could be useful.

If your data is paired-end you can use bbmerge to correct quality values in regions where paires overlap.

fin swimmer

ADD COMMENT
0
Entering edit mode

That makes sense I guess, I'll just set Q=4 to make sure nts are more likely to be right than wrong. I guess there's no way of finding out if it works other than testing it

ADD REPLY
0
Entering edit mode

Mapper/Aligner and VariantCaller have there own logic how to handle bad quality values. Just let them do their job.

In most cases you will have bad quality values at the end of your reads. If the Mapper/Aligner isn't satisfied with it, it will clipp those bases by it's own.

ADD REPLY

Login before adding your answer.

Traffic: 2054 users visited in the last hour
Help About
FAQ
Access RSS
API
Stats

Use of this site constitutes acceptance of our User Agreement and Privacy Policy.

Powered by the version 2.3.6