Evaluating learned feature aggregators for writer retrieval

Mattick A, Mayr M, Seuret M, Kordon F, Wu F, Christlein V (2024)


Publication Type: Journal article, Original article

Publication year: 2024

Journal

Book Volume: 27

Pages Range: 265-274

Journal Issue: 3

DOI: 10.1007/s10032-024-00482-x

Abstract

Transformers have emerged as the leading methods in natural language processing, computer vision, and multi-modal applications due to their ability to capture complex relationships and dependencies in data. In this study, we explore the potential of transformers as feature aggregators in the context of patch-based writer retrieval, with the objective of improving the quality of writer retrieval by effectively summarizing the relevant features from image patches. Our investigation underscores the complexity of leveraging transformers as feature aggregators in patch-based writer retrieval. While we have experimented with various model configurations, augmentations, and learning objectives, the performance of transformers in this task has room for improvement. This observation highlights the challenges in this domain and emphasizes the need for further research to enhance their effectiveness. By shedding light on the limitations of transformers in this context, our study contributes to the growing body of knowledge in the field of writer retrieval and provides valuable insights for future research and development in this area.

Authors with CRIS profile

How to cite

APA:

Mattick, A., Mayr, M., Seuret, M., Kordon, F., Wu, F., & Christlein, V. (2024). Evaluating learned feature aggregators for writer retrieval. International Journal on Document Analysis and Recognition, 27(3), 265-274. https://doi.org/10.1007/s10032-024-00482-x

MLA:

Mattick, Alexander, et al. "Evaluating learned feature aggregators for writer retrieval." International Journal on Document Analysis and Recognition 27.3 (2024): 265-274.

BibTeX: Download