PDF (469.2 KB)
Collect
Submit Manuscript
Show Outline
Figures (3)

Tables (3)
TableA 1
Table 1
Table 2
Open Access

Three-Branch BERT-Based Text Classification Network for Gastroscopy Diagnosis Text

Zhichao Wang1,2Xiangwei Zheng1,2()Jinsong Zhang1,2Mingzhe Zhang1,2()
School of Information Science and Engineering, Shandong Normal University, Jinan 250358, China
State Key Laboratory of High-end Server & Storage Technology, Jinan 250013, China
Show Author Information

Abstract

During a hospital visit, a significant volume of Gastroscopy Diagnostic Text (GDT) data are produced, representing the unstructured gastric medical records of patients undergoing gastroscopy. As such, GDTs play a crucial role in evaluating the patient’s health, shaping treatment plans, and scheduling follow-up visits. However, given the free-text nature of GDTs, which lack a formal structure, physicians often find it challenging to extract meaningful insights from them. Furthermore, while deep learning has made significant strides in the medical domain, to our knowledge, there are not any readily available text-based pre-trained models tailored for GDT classification and analysis. To address this gap, we introduce a Bidirectional Encoder Representations from Transformers (BERT) based three-branch classification network tailored for GDTs. We leverage the robust representation capabilities of the BERT pre-trained model to deeply encode the texts. A unique three-branch decoder structure is employed to pinpoint lesion sites and determine cancer stages. Experimental outcomes validate the efficacy of our approach in GDT classification, with a precision of 0.993 and a recall of 0.784 in the early cancer category. In pinpointing cancer lesion sites, the weighted F1 score achieved was 0.849.

References

[1]

Y. Cui, G. Cheng, G. Tian, S. He, and Y. Yan, Secular trends in the mortality of gastrointestinal cancers across China, Japan, the US, and India: An age-period-cohort, joinpoint analyses, and Holt forecasts, Front. Public Health, vol. 10, p. 925011, 2022.

[2]

C. Herrera-Pariente, S. Montori, J. Llach, A. Bofill, E. Albeniz, and L. Moreira, Biomarkers for gastric cancer screening and early diagnosis, Biomedicines, vol. 9, no. 10, p. 1448, 2021.

[3]
A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, Ł. Kaiser, and I. Polosukhin, Attention is all you need, in Proc. 31st Int. Conf. Neural Information Processing Systems, Long Beach, CA, USA, 2017, pp. 6000–6010.
[4]
J. Devlin, M. W. Chang, K. Lee, and K. Toutanova, BERT: Pre-training of deep bidirectional transformers for language understanding, arXiv preprint arXiv: 1810.04805, 2019.
[5]
W. Zaremba, I. Sutskever, and O. Vinyals, Recurrent neural network regularization, arXiv preprint arXiv: 1409.2329, 2014.
[6]
A. Graves, Long short-term memory, in Supervised Sequence Labelling with Recurrent Neural Networks, A. Graves, ed. Berlin, Germany: Springer, 2012, pp. 37–45.
[7]

A. Onan and M. A. Toçoğlu, A term weighted neural language model and stacked bidirectional LSTM based framework for sarcasm identification, IEEE Access, vol. 9, pp. 7701–7722, 2021.

[8]
W. Zhou, J. Du, and X. Ren, Improving BERT fine-tuning with embedding normalization, arXiv preprint arXiv: 1911.03918, 2019.
[9]

L. Jiang, C. Li, S. Wang, and L. Zhang, Deep feature weighting for naive Bayes and its application to text classification, Eng. Appl. Artif. Intell., vol. 52, pp. 26–39, 2016.

[10]
T. Joachims, Making large scale SVM learning practical, https://api.semanticscholar.org/CorpusID:61116019, 1998.
[11]

J. R. Quinlan, Induction of decision trees, Mach. Learn., vol. 1, no. 1, pp. 81–106, 1986.

[12]

L. Breiman, Random forests, Mach. Learn., vol. 45, no. 1, pp. 5–32, 2001.

[13]
Y. Yang and X. Liu, A re-examination of text categorization methods, in Proc. 22nd Annual Int. ACM SIGIR Conf. Research and Development in Information Retrieval, Berkeley, CA, USA, 1999, pp. 42–49.
[14]

S. Minaee, N. Kalchbrenner, E. Cambria, N. Nikzad, M. Chenaghlu, and J. Gao, Deep learning: Based text classification, ACM Comput. Surv., vol. 54, no. 3, pp. 1–40, 2022.

[15]

M. Hughes, I. Li, S. Kotoulas, and T. Suzumura, Medical text classification using convolutional neural networks, Stud. Health Technol. Inform., vol. 235, pp. 246–250, 2017.

[16]

A. Onan, Bidirectional convolutional recurrent neural network architecture with group-wise enhancement mechanism for text sentiment classification, J. King Saud Univ. Comput. Inf. Sci., vol. 34, no. 5, pp. 2098–2117, 2022.

[17]

A. Onan, S. Korukoğlu, and H. Bulut, Ensemble of keyword extraction methods and classifiers in text classification, Expert Syst. Appl., vol. 57, pp. 232–247, 2016.

[18]
Y. Liu, M. Ott, N. Goyal, J. Du, M. Joshi, D. Chen, O. Levy, M. Lewis, L. Zettlemoyer, and V. Stoyanov, RoBERTa: A robustly optimized BERT pretraining approach, arXiv preprint arXiv: 1907.11692, 2019.
[19]

A. Onan, Hierarchical graph-based text classification framework with contextual node embedding and BERT-based dynamic fusion, J. King Saud Univ. Comput. Inf. Sci., vol. 35, no. 7, p. 101610, 2023.

[20]
A. Radford, K. Narasimhan, T. Salimans, and I. Sutskever, Improving language understanding by generative pre-training, https://www.cs.ubc.ca/~amuham01/LING530/papers/radford2018improving.pdf, 2018.
[21]
T. B. Brown, B. Mann, N. Ryder, M. Subbiah, J. Kaplan, P. Dhariwal, A. Neelakantan, P. Shyam, G. Sastry, A. Askell, et al., Language models are few-shot learners, in Proc. 34th Int. Conf. Neural Information Processing Systems, Vancouver, Canada, 2020, pp. 1877–1901.
[22]

S. Ding, S. Hu, X. Li, Y. Zhang, and D. D. Wu, Leveraging multimodal semantic fusion for gastric cancer screening via hierarchical attention mechanism, IEEE Trans. Syst. Man Cybern. Syst., vol. 52, no. 7, pp. 4286–4299, 2022.

[23]
Y. Liu, Fine-tune BERT for extractive summarization, arXiv preprint arXiv: 1903.10318, 2019.
[24]
N. Ketkar, Introduction to keras, in Deep Learning with Python, N. Ketkar, ed. Berkeley, CA, USA: Apress, 2017, pp. 97–111.
International Journal of Crowd Science
Pages 56-63
Cite this article:
Wang Z, Zheng X, Zhang J, et al. Three-Branch BERT-Based Text Classification Network for Gastroscopy Diagnosis Text. International Journal of Crowd Science, 2024, 8(1): 56-63. https://doi.org/10.26599/IJCS.2023.9100031
Metrics & Citations  
Article History
Copyright
Rights and Permissions
Return