Masked Autoencoders for Microscopy are Scalable Learners of Cellular Biology
Featurizing microscopy images for use in biological research remains a significant challenge, especially for large-scale experiments spanning millions of images. This work explores the scaling properties of weakly supervised classifiers and self-supervised masked autoencoders (MAEs) when training wi...
Gespeichert in:
Hauptverfasser: | , , , , , , , , , , , , , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | |
container_volume | |
creator | Kraus, Oren Kenyon-Dean, Kian Saberian, Saber Fallah, Maryam McLean, Peter Leung, Jess Sharma, Vasudev Khan, Ayla Balakrishnan, Jia Celik, Safiye Beaini, Dominique Sypetkowski, Maciej Cheng, Chi Vicky Morse, Kristen Makes, Maureen Mabey, Ben Earnshaw, Berton |
description | Featurizing microscopy images for use in biological research remains a
significant challenge, especially for large-scale experiments spanning millions
of images. This work explores the scaling properties of weakly supervised
classifiers and self-supervised masked autoencoders (MAEs) when training with
increasingly larger model backbones and microscopy datasets. Our results show
that ViT-based MAEs outperform weakly supervised classifiers on a variety of
tasks, achieving as much as a 11.5% relative improvement when recalling known
biological relationships curated from public databases. Additionally, we
develop a new channel-agnostic MAE architecture (CA-MAE) that allows for
inputting images of different numbers and orders of channels at inference time.
We demonstrate that CA-MAEs effectively generalize by inferring and evaluating
on a microscopy image dataset (JUMP-CP) generated under different experimental
conditions with a different channel structure than our pretraining data
(RPI-93M). Our findings motivate continued research into scaling
self-supervised learning on microscopy data in order to create powerful
foundation models of cellular biology that have the potential to catalyze
advancements in drug discovery and beyond. |
doi_str_mv | 10.48550/arxiv.2404.10242 |
format | Article |
fullrecord | <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2404_10242</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2404_10242</sourcerecordid><originalsourceid>FETCH-LOGICAL-a672-cb4368b3d6b025e9be02a5b020d8835556e7ba0be3b85e7e4c22eea835110b333</originalsourceid><addsrcrecordid>eNotj8tuwjAQRb3poqL9gK7wDySd-JGYJY36koJYwD6acSYowmDklKr5-wLt6l7pHl3pCPFUQG6ctfCM6Wf4zpUBkxegjLoXzQrHPXdyef6KfPSx4zTKPia5GnyKo4-nSWJiufEYkALLhjEdr1DsZc0hnAMm-TLEEHfTg7jrMYz8-J8zsX173dYfWbN-_6yXTYZlpTJPRpeOdFcSKMsLYlBoLx0657S1tuSKEIg1OcsVG68UM16mogDSWs_E_O_2ptOe0nDANLVXrfampX8B3K9H1g</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Masked Autoencoders for Microscopy are Scalable Learners of Cellular Biology</title><source>arXiv.org</source><creator>Kraus, Oren ; Kenyon-Dean, Kian ; Saberian, Saber ; Fallah, Maryam ; McLean, Peter ; Leung, Jess ; Sharma, Vasudev ; Khan, Ayla ; Balakrishnan, Jia ; Celik, Safiye ; Beaini, Dominique ; Sypetkowski, Maciej ; Cheng, Chi Vicky ; Morse, Kristen ; Makes, Maureen ; Mabey, Ben ; Earnshaw, Berton</creator><creatorcontrib>Kraus, Oren ; Kenyon-Dean, Kian ; Saberian, Saber ; Fallah, Maryam ; McLean, Peter ; Leung, Jess ; Sharma, Vasudev ; Khan, Ayla ; Balakrishnan, Jia ; Celik, Safiye ; Beaini, Dominique ; Sypetkowski, Maciej ; Cheng, Chi Vicky ; Morse, Kristen ; Makes, Maureen ; Mabey, Ben ; Earnshaw, Berton</creatorcontrib><description>Featurizing microscopy images for use in biological research remains a
significant challenge, especially for large-scale experiments spanning millions
of images. This work explores the scaling properties of weakly supervised
classifiers and self-supervised masked autoencoders (MAEs) when training with
increasingly larger model backbones and microscopy datasets. Our results show
that ViT-based MAEs outperform weakly supervised classifiers on a variety of
tasks, achieving as much as a 11.5% relative improvement when recalling known
biological relationships curated from public databases. Additionally, we
develop a new channel-agnostic MAE architecture (CA-MAE) that allows for
inputting images of different numbers and orders of channels at inference time.
We demonstrate that CA-MAEs effectively generalize by inferring and evaluating
on a microscopy image dataset (JUMP-CP) generated under different experimental
conditions with a different channel structure than our pretraining data
(RPI-93M). Our findings motivate continued research into scaling
self-supervised learning on microscopy data in order to create powerful
foundation models of cellular biology that have the potential to catalyze
advancements in drug discovery and beyond.</description><identifier>DOI: 10.48550/arxiv.2404.10242</identifier><language>eng</language><subject>Computer Science - Artificial Intelligence ; Computer Science - Computer Vision and Pattern Recognition ; Computer Science - Learning</subject><creationdate>2024-04</creationdate><rights>http://creativecommons.org/licenses/by-nc-sa/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,777,882</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2404.10242$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2404.10242$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Kraus, Oren</creatorcontrib><creatorcontrib>Kenyon-Dean, Kian</creatorcontrib><creatorcontrib>Saberian, Saber</creatorcontrib><creatorcontrib>Fallah, Maryam</creatorcontrib><creatorcontrib>McLean, Peter</creatorcontrib><creatorcontrib>Leung, Jess</creatorcontrib><creatorcontrib>Sharma, Vasudev</creatorcontrib><creatorcontrib>Khan, Ayla</creatorcontrib><creatorcontrib>Balakrishnan, Jia</creatorcontrib><creatorcontrib>Celik, Safiye</creatorcontrib><creatorcontrib>Beaini, Dominique</creatorcontrib><creatorcontrib>Sypetkowski, Maciej</creatorcontrib><creatorcontrib>Cheng, Chi Vicky</creatorcontrib><creatorcontrib>Morse, Kristen</creatorcontrib><creatorcontrib>Makes, Maureen</creatorcontrib><creatorcontrib>Mabey, Ben</creatorcontrib><creatorcontrib>Earnshaw, Berton</creatorcontrib><title>Masked Autoencoders for Microscopy are Scalable Learners of Cellular Biology</title><description>Featurizing microscopy images for use in biological research remains a
significant challenge, especially for large-scale experiments spanning millions
of images. This work explores the scaling properties of weakly supervised
classifiers and self-supervised masked autoencoders (MAEs) when training with
increasingly larger model backbones and microscopy datasets. Our results show
that ViT-based MAEs outperform weakly supervised classifiers on a variety of
tasks, achieving as much as a 11.5% relative improvement when recalling known
biological relationships curated from public databases. Additionally, we
develop a new channel-agnostic MAE architecture (CA-MAE) that allows for
inputting images of different numbers and orders of channels at inference time.
We demonstrate that CA-MAEs effectively generalize by inferring and evaluating
on a microscopy image dataset (JUMP-CP) generated under different experimental
conditions with a different channel structure than our pretraining data
(RPI-93M). Our findings motivate continued research into scaling
self-supervised learning on microscopy data in order to create powerful
foundation models of cellular biology that have the potential to catalyze
advancements in drug discovery and beyond.</description><subject>Computer Science - Artificial Intelligence</subject><subject>Computer Science - Computer Vision and Pattern Recognition</subject><subject>Computer Science - Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotj8tuwjAQRb3poqL9gK7wDySd-JGYJY36koJYwD6acSYowmDklKr5-wLt6l7pHl3pCPFUQG6ctfCM6Wf4zpUBkxegjLoXzQrHPXdyef6KfPSx4zTKPia5GnyKo4-nSWJiufEYkALLhjEdr1DsZc0hnAMm-TLEEHfTg7jrMYz8-J8zsX173dYfWbN-_6yXTYZlpTJPRpeOdFcSKMsLYlBoLx0657S1tuSKEIg1OcsVG68UM16mogDSWs_E_O_2ptOe0nDANLVXrfampX8B3K9H1g</recordid><startdate>20240415</startdate><enddate>20240415</enddate><creator>Kraus, Oren</creator><creator>Kenyon-Dean, Kian</creator><creator>Saberian, Saber</creator><creator>Fallah, Maryam</creator><creator>McLean, Peter</creator><creator>Leung, Jess</creator><creator>Sharma, Vasudev</creator><creator>Khan, Ayla</creator><creator>Balakrishnan, Jia</creator><creator>Celik, Safiye</creator><creator>Beaini, Dominique</creator><creator>Sypetkowski, Maciej</creator><creator>Cheng, Chi Vicky</creator><creator>Morse, Kristen</creator><creator>Makes, Maureen</creator><creator>Mabey, Ben</creator><creator>Earnshaw, Berton</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20240415</creationdate><title>Masked Autoencoders for Microscopy are Scalable Learners of Cellular Biology</title><author>Kraus, Oren ; Kenyon-Dean, Kian ; Saberian, Saber ; Fallah, Maryam ; McLean, Peter ; Leung, Jess ; Sharma, Vasudev ; Khan, Ayla ; Balakrishnan, Jia ; Celik, Safiye ; Beaini, Dominique ; Sypetkowski, Maciej ; Cheng, Chi Vicky ; Morse, Kristen ; Makes, Maureen ; Mabey, Ben ; Earnshaw, Berton</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a672-cb4368b3d6b025e9be02a5b020d8835556e7ba0be3b85e7e4c22eea835110b333</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Computer Science - Artificial Intelligence</topic><topic>Computer Science - Computer Vision and Pattern Recognition</topic><topic>Computer Science - Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Kraus, Oren</creatorcontrib><creatorcontrib>Kenyon-Dean, Kian</creatorcontrib><creatorcontrib>Saberian, Saber</creatorcontrib><creatorcontrib>Fallah, Maryam</creatorcontrib><creatorcontrib>McLean, Peter</creatorcontrib><creatorcontrib>Leung, Jess</creatorcontrib><creatorcontrib>Sharma, Vasudev</creatorcontrib><creatorcontrib>Khan, Ayla</creatorcontrib><creatorcontrib>Balakrishnan, Jia</creatorcontrib><creatorcontrib>Celik, Safiye</creatorcontrib><creatorcontrib>Beaini, Dominique</creatorcontrib><creatorcontrib>Sypetkowski, Maciej</creatorcontrib><creatorcontrib>Cheng, Chi Vicky</creatorcontrib><creatorcontrib>Morse, Kristen</creatorcontrib><creatorcontrib>Makes, Maureen</creatorcontrib><creatorcontrib>Mabey, Ben</creatorcontrib><creatorcontrib>Earnshaw, Berton</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Kraus, Oren</au><au>Kenyon-Dean, Kian</au><au>Saberian, Saber</au><au>Fallah, Maryam</au><au>McLean, Peter</au><au>Leung, Jess</au><au>Sharma, Vasudev</au><au>Khan, Ayla</au><au>Balakrishnan, Jia</au><au>Celik, Safiye</au><au>Beaini, Dominique</au><au>Sypetkowski, Maciej</au><au>Cheng, Chi Vicky</au><au>Morse, Kristen</au><au>Makes, Maureen</au><au>Mabey, Ben</au><au>Earnshaw, Berton</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Masked Autoencoders for Microscopy are Scalable Learners of Cellular Biology</atitle><date>2024-04-15</date><risdate>2024</risdate><abstract>Featurizing microscopy images for use in biological research remains a
significant challenge, especially for large-scale experiments spanning millions
of images. This work explores the scaling properties of weakly supervised
classifiers and self-supervised masked autoencoders (MAEs) when training with
increasingly larger model backbones and microscopy datasets. Our results show
that ViT-based MAEs outperform weakly supervised classifiers on a variety of
tasks, achieving as much as a 11.5% relative improvement when recalling known
biological relationships curated from public databases. Additionally, we
develop a new channel-agnostic MAE architecture (CA-MAE) that allows for
inputting images of different numbers and orders of channels at inference time.
We demonstrate that CA-MAEs effectively generalize by inferring and evaluating
on a microscopy image dataset (JUMP-CP) generated under different experimental
conditions with a different channel structure than our pretraining data
(RPI-93M). Our findings motivate continued research into scaling
self-supervised learning on microscopy data in order to create powerful
foundation models of cellular biology that have the potential to catalyze
advancements in drug discovery and beyond.</abstract><doi>10.48550/arxiv.2404.10242</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | DOI: 10.48550/arxiv.2404.10242 |
ispartof | |
issn | |
language | eng |
recordid | cdi_arxiv_primary_2404_10242 |
source | arXiv.org |
subjects | Computer Science - Artificial Intelligence Computer Science - Computer Vision and Pattern Recognition Computer Science - Learning |
title | Masked Autoencoders for Microscopy are Scalable Learners of Cellular Biology |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-19T14%3A20%3A53IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Masked%20Autoencoders%20for%20Microscopy%20are%20Scalable%20Learners%20of%20Cellular%20Biology&rft.au=Kraus,%20Oren&rft.date=2024-04-15&rft_id=info:doi/10.48550/arxiv.2404.10242&rft_dat=%3Carxiv_GOX%3E2404_10242%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |