You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
<p>Patch extraction from gigapixel whole-slide images, typically guided by tissue detection methods, forms the backbone of computational pathology workflows,
38
-
and remains a major computational bottleneck. Here we present AtlasPatch, an efficient and scalable slide preprocessing framework designed to enable accurate tissue
39
-
detection and high-throughput patch extraction with minimal computational overhead. To ensure robust and generalizable slide processing, we curated and
40
-
semi-manually annotated a diverse dataset of approximately 35,000 whole-slide image thumbnails spanning multiple cohorts, tissue types, and organ systems. Using this
41
-
corpus, we implement selective finetuning of the normalization layers of the Segment-Anything2 model for efficient thumbnail-level segmentation. This strategy ensures
42
-
fast processing while achieving segmentation accuracy comparable to—or exceeding—that of existing methods. From the thumbnail masks, tissue coordinates are efficiently
43
-
extrapolated to full-resolution slides for precise patch extraction. The entire pipeline is parallelized for both CPU and GPU execution to maximize throughput. We assess
44
-
AtlasPatch across segmentation accuracy, computational complexity, and downstream multiple-instance learning performance, showing consistent predictive power with
45
-
state-of-the-art methods while operating at a fraction of their computational cost.</p>
46
-
47
-
<imgsrc="images/Fig1a.png" alt=""/>
42
+
<h2class="abstract-title">Abstract</h2>
43
+
<p>Whole-slide image (WSI) preprocessing, typically comprising tissue detection followed by patch extraction, is foundational to AI-driven
44
+
and image-based computational pathology workflows. This remains a major computational bottleneck as existing tools either rely on inacurate
45
+
heuristic thresholding for tissue detection, or adopt AI-based approaches trained on limited-diversity data that operate at the patch level,
46
+
incurring substantial computational complexity. We present AtlasPatch, an efficient and scalable slide preprocessing framework for accurate
47
+
tissue detection and high-throughput patch extraction with minimal computational overhead. AtlasPatch’s tissue detection module is trained
48
+
on a heterogeneous and semi-manually annotated dataset of $\sim$35,000 WSI thumbnails, using efficient fine-tuning of the Segment Anything2
49
+
model. The tool extrapolates tissue masks from thumbnails to full-resolution slides to extract patch coordinates at user-specified magnifications,
50
+
with options to stream patches directly into commonly used image encoders for embedding generation or export patch images for storage, all efficiently
51
+
parallelized across CPUs and GPUs to maximize throughput. We assess AtlasPatch across segmentation accuracy, computational complexity, and downstream
52
+
multiple-instance learning, matching state-of-the-art performance while operating at a fraction of their computational cost.</p>
0 commit comments