-

nihao guest [ sign in / register ]
2025-12-5 16:05:11


Rongye Ye, Lun Li, Shuhui Song. [preprint]Influ-BERT: An Interpretable Model for Enhancing Low-Frequency Influenza A virus Subtype Recognition. https://doi.org/10.1101/2025.07.31.667841
submited by kickingbird at Aug, 3, 2025 8:2 AM from https://doi.org/10.1101/2025.07.31.667841

Influenza A Virus (IAV) poses a continuous threat to global public health due to its wide host adaptability, high-frequency antigenic variation, and potential for cross-species transmission. Accurate recognition of IAV subtypes is cru-cial for the early pandemic warning. Here, we propose Influ-BERT, a domain-adaptive pretraining model based on the transformer architecture. Optimized from DNABERT-2, Influ-BERT constructed a dedicated corpus of approximately 900,000 influenza genome sequences, developed a custom Byte Pair Encoding (BPE) tokenizer, and employ a two-stage training strategy involving domain-adaptive pretraining followed by task-specific fine-tuning. This approach significantly enhanced recognition performance for low-frequency subtypes. Experimental results demonstrate that Influ-BERT outper-forms traditional machine learning methods and general genomic language models (DNABERT-2, MegaDNA) in subtype recognition, achieving a substantial improvement in F1-score, particularly for subtypes H5N8, H5N1, H7N9, H9N2. Furthermore, sliding window perturbation analysis revealed the model´s specific focus on key regions of the IAV genome, providing interpretable evidence supporting the observed performance gains.

See Also:

Latest articles in those days:

[Go Top]    [Close Window]

Related Pages:
Learn about the flu news, articles, events and more
Subscribe to the weekly F.I.C newsletter!


  

Site map  |   Contact us  |  Term of use  |  FAQs |  粤ICP备10094839号-1
Copyright ©www.flu.org.cn. 2004-2025. All Rights Reserved. Powered by FIC 4.0.1
  Email:webmaster@flu.org.cn