Title: Scaling transformer neural networks to gigapixel images
Project duration: 5 years
Research Area: Applied AI and Big Data
Our project focuses on developing transformer neural networks for gigapixel image classification, tackling the challenge of processing large-scale pathology images using advanced AI techniques. Pathology images are gigapixel-sized, making them unsuitable for standard neural network architectures. Recently, transformer neural networks have achieved state-of-the-art performance in image processing. Therefore, we aim to develop new transformer-based architectures for gigapixel image classification and compare them with existing methods. We face technical challenges in designing an architecture that can handle images of varying sizes and can be trained efficiently on large datasets. Moreover, the architecture needs to process gigapixel images directly without tiling or using a sliding window. Our goal is to create an end-to-end solution that processes whole slide images and outputs classifications. Additionally, we will explore the explainability of transformer models and compare them to other architectures.
Currently, workflows in computational pathology lack end-to-end solutions: Large histopathology slide images must be tiled and analyzed in patches. This process causes loss of information. Moreover, the lack of global context in individual patches limits classification performance. In addition, optimizing the tiling strategy can be challenging.
In this project, we will develop and adapt transformer neural networks for gigapixel images to address limitations in computational pathology. The advantage of this approach is that it processes the whole slide image at once, providing direct classification without the need for tiling.
Overall, our novel analysis method for gigapixel images will enable the analysis and classification of large microscopic images as a whole. Additionally, our transformer neural network can be applied to analyze different cancer types and relevant biomarkers.